Module:excerpt/testcases
![]() | This is the test cases page for the module Module:Excerpt. Results of the test cases. |
Errors
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
No page given |
No page given |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
No page given |
No page given |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
No page given |
No page given |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Title «2 > 1» is not valid |
Title «2 > 1» is not valid |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Page 'Fake title' not found |
Page 'Fake title' not found |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Section 'Fake section' not found |
Section 'Fake section' not found |
Leads
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[1][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][9] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[10][11][12] The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.[13]: 12 [14][15][16] Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age,[17] along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[18][19][20] which was later transformed by the Scientific Revolution that began in the 16th century[21] as new ideas and discoveries departed from previous Greek conceptions and traditions.[22][23] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[24][25] along with the changing of "natural philosophy" to "natural science".[26] New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems.[27][28] Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions,[29] government agencies,[30] and companies.[31] The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[32][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][33] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[34][35][12] The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.[13]: 12 [14][36][37] Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age,[38] along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[39][40][41] which was later transformed by the Scientific Revolution that began in the 16th century[42] as new ideas and discoveries departed from previous Greek conceptions and traditions.[43][44] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[45][46] along with the changing of "natural philosophy" to "natural science".[47] New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems.[48][49] Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions,[50] government agencies,[51] and companies.[52] The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]() The COVID-19 pandemic in France has resulted in 39,014,130[53] confirmed cases of COVID-19 and 168,150[53] deaths. The virus was confirmed to have reached France on 24 January 2020, when the first COVID-19 case in both Europe and France was identified in Bordeaux. The first five confirmed cases were all individuals who had recently arrived from China.[54][55] A Chinese tourist who was admitted to hospital in Paris on 28 January 2020, died on 14 February 2020, becoming the first known COVID-19 fatality outside Asia as well as the first in France.[56][57][58][59] A key event in the spread of the disease across metropolitan France as well as its overseas territories was the annual assembly of the Christian Open Door Church between 17 and 24 February 2020 in Mulhouse which was attended by about 2,500 people, at least half of whom are believed to have contracted the virus.[60][61] On 4 May 2020, retroactive testing of samples in one French hospital showed that a patient was probably already infected with the virus on 27 December 2019, almost a month before the first officially confirmed case.[62][63] The first lockdown period began on 17 March 2020 and ended on 11 May 2020.[64] On 2 May 2020, Health Minister Olivier Véran announced that the government would seek to extend the health emergency period until 24 July 2020.[65] Several mayors opposed the 11 May 2020 lifting of the lockdown, which had been announced by the president a few weeks earlier in a televised address to the nation,[64] saying it was premature. Véran's bill was discussed in Senate on 4 May 2020.[66] From August 2020, there was an increase in the rate of infection and on 10 October 2020, France set a record number of new infections in a 24-hour period in Europe with 26,896 recorded. The increase caused France to enter a second nationwide lockdown on 28 October 2020. On 15 October 2020, police raided the homes and offices of key government officials, including Véran and Philippe, in a criminal negligence probe opened by the Cour de Justice de la République.[67] According to a team of French epidemiologists, under 5% of the total population of France, or around 2.8 million people, may have been infected with COVID-19. This was believed to have been nearly twice as high in the Île-de-France and Alsace regions.[68] On 31 March 2021, Macron announced a third national lockdown which commenced on 3 April 2021 and which was mandated for all of April 2021; measures included the closure of non-essential shops, the suspension of school attendance, a ban on domestic travel and a nationwide curfew from 7pm-6am. In February 2022, it was reported that no tests are required to enter the country, and children under the age of 12 are free from vaccination requirements.[69] |
![]() The COVID-19 pandemic in France has resulted in 39,014,130[53] confirmed cases of COVID-19 and 168,150[53] deaths. The virus was confirmed to have reached France on 24 January 2020, when the first COVID-19 case in both Europe and France was identified in Bordeaux. The first five confirmed cases were all individuals who had recently arrived from China.[54][55] A Chinese tourist who was admitted to hospital in Paris on 28 January 2020, died on 14 February 2020, becoming the first known COVID-19 fatality outside Asia as well as the first in France.[56][57][58][59] A key event in the spread of the disease across metropolitan France as well as its overseas territories was the annual assembly of the Christian Open Door Church between 17 and 24 February 2020 in Mulhouse which was attended by about 2,500 people, at least half of whom are believed to have contracted the virus.[60][61] On 4 May 2020, retroactive testing of samples in one French hospital showed that a patient was probably already infected with the virus on 27 December 2019, almost a month before the first officially confirmed case.[62][63] The first lockdown period began on 17 March 2020 and ended on 11 May 2020.[64] On 2 May 2020, Health Minister Olivier Véran announced that the government would seek to extend the health emergency period until 24 July 2020.[65] Several mayors opposed the 11 May 2020 lifting of the lockdown, which had been announced by the president a few weeks earlier in a televised address to the nation,[64] saying it was premature. Véran's bill was discussed in Senate on 4 May 2020.[66] From August 2020, there was an increase in the rate of infection and on 10 October 2020, France set a record number of new infections in a 24-hour period in Europe with 26,896 recorded. The increase caused France to enter a second nationwide lockdown on 28 October 2020. On 15 October 2020, police raided the homes and offices of key government officials, including Véran and Philippe, in a criminal negligence probe opened by the Cour de Justice de la République.[70] According to a team of French epidemiologists, under 5% of the total population of France, or around 2.8 million people, may have been infected with COVID-19. This was believed to have been nearly twice as high in the Île-de-France and Alsace regions.[68] On 31 March 2021, Macron announced a third national lockdown which commenced on 3 April 2021 and which was mandated for all of April 2021; measures included the closure of non-essential shops, the suspension of school attendance, a ban on domestic travel and a nationwide curfew from 7pm-6am. In February 2022, it was reported that no tests are required to enter the country, and children under the age of 12 are free from vaccination requirements.[71] |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[72][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][73] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[74][75][12] The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.[13]: 12 [14][76][77] Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age,[78] along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[79][80][81] which was later transformed by the Scientific Revolution that began in the 16th century[82] as new ideas and discoveries departed from previous Greek conceptions and traditions.[83][84] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[85][86] along with the changing of "natural philosophy" to "natural science".[87] New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems.[88][89] Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions,[90] government agencies,[91] and companies.[92] The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[93][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][94] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[95][96][12] The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.[13]: 12 [14][97][98] Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age,[99] along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[100][101][102] which was later transformed by the Scientific Revolution that began in the 16th century[103] as new ideas and discoveries departed from previous Greek conceptions and traditions.[104][105] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[106][107] along with the changing of "natural philosophy" to "natural science".[108] New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems.[109][110] Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions,[111] government agencies,[112] and companies.[113] The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe. Modern science is typically divided into two or three major branches: the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies. The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules, are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology. Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine. The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.: 12 Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age, along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy, which was later transformed by the Scientific Revolution that began in the 16th century as new ideas and discoveries departed from previous Greek conceptions and traditions. The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape, along with the changing of "natural philosophy" to "natural science". New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems. Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions, government agencies, and companies. The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe. Modern science is typically divided into two or three major branches: the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies. The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules, are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology. Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine. The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.: 12 Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age, along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy, which was later transformed by the Scientific Revolution that began in the 16th century as new ideas and discoveries departed from previous Greek conceptions and traditions. The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape, along with the changing of "natural philosophy" to "natural science". New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems. Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions, government agencies, and companies. The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[114][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][115] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[116][117][12] The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.[13]: 12 [14][118][119] Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age,[120] along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[121][122][123] which was later transformed by the Scientific Revolution that began in the 16th century[124] as new ideas and discoveries departed from previous Greek conceptions and traditions.[125][126] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[127][128] along with the changing of "natural philosophy" to "natural science".[129] New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems.[130][131] Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions,[132] government agencies,[133] and companies.[134] The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[135][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][136] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[137][138][12] The history of science spans the majority of the historical record, with the earliest identifiable predecessors to modern science dating to the Bronze Age in Egypt and Mesopotamia (c. 3000–1200 BCE). Their contributions to mathematics, astronomy, and medicine entered and shaped the Greek natural philosophy of classical antiquity, whereby formal attempts were made to provide explanations of events in the physical world based on natural causes, while further advancements, including the introduction of the Hindu–Arabic numeral system, were made during the Golden Age of India.[13]: 12 [14][139][140] Scientific research deteriorated in these regions after the fall of the Western Roman Empire during the Early Middle Ages (400–1000 CE), but in the Medieval renaissances (Carolingian Renaissance, Ottonian Renaissance and the Renaissance of the 12th century) scholarship flourished again. Some Greek manuscripts lost in Western Europe were preserved and expanded upon in the Middle East during the Islamic Golden Age,[141] along with the later efforts of Byzantine Greek scholars who brought Greek manuscripts from the dying Byzantine Empire to Western Europe at the start of the Renaissance. The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[142][143][144] which was later transformed by the Scientific Revolution that began in the 16th century[145] as new ideas and discoveries departed from previous Greek conceptions and traditions.[146][147] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[148][149] along with the changing of "natural philosophy" to "natural science".[150] New knowledge in science is advanced by research from scientists who are motivated by curiosity about the world and a desire to solve problems.[151][152] Contemporary scientific research is highly collaborative and is usually done by teams in academic and research institutions,[153] government agencies,[154] and companies.[155] The practical impact of their work has led to the emergence of science policies that seek to influence the scientific enterprise by prioritising the ethical and moral development of commercial products, armaments, health care, public infrastructure, and environmental protection. |
Biographies
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
➥ Control case; no |briefdates= param; compare w/ following testMarc Léopold Benjamin Bloch (/blɒk/; French: [maʁk leɔpɔld bɛ̃ʒamɛ̃ blɔk]; 6 July 1886 – 16 June 1944) was a French historian. He was a founding member of the Annales School of French social history. Bloch specialised in medieval history and published widely on medieval France over the course of his career. As an academic, he worked at the University of Strasbourg (1920 to 1936 and 1940 to 1941), the University of Paris (1936 to 1939), and the University of Montpellier (1941 to 1944). Born in Lyon to an Alsatian Jewish family, Bloch was raised in Paris, where his father—the classical historian Gustave Bloch—worked at Sorbonne University. Bloch was educated at various Parisian lycées and the École Normale Supérieure, and from an early age was affected by the antisemitism of the Dreyfus affair. During the First World War, he served in the French Army and fought at the First Battle of the Marne and the Somme. After the war, he was awarded his doctorate in 1918 and became a lecturer at the University of Strasbourg. There, he formed an intellectual partnership with modern historian Lucien Febvre. Together they founded the Annales School and began publishing the journal Annales d'histoire économique et sociale in 1929. Bloch was a modernist in his historiographical approach, and repeatedly emphasised the importance of a multidisciplinary engagement towards history, particularly blending his research with that on geography, sociology and economics, which was his subject when he was offered a post at the University of Paris in 1936. During the Second World War Bloch volunteered for service, and was a logistician during the Phoney War. Involved in the Battle of Dunkirk and spending a brief time in Britain, he unsuccessfully attempted to secure passage to the United States. Back in France, where his ability to work was curtailed by new antisemitic regulations, he applied for and received one of the few permits available allowing Jews to continue working in the French university system. He had to leave Paris, and complained that the Nazi German authorities looted his apartment and stole his books; he was also persuaded by Febvre to relinquish his position on the editorial board of Annales. Bloch worked in Montpellier until November 1942 when Germany invaded Vichy France. He then joined the non-Communist section of the French Resistance and went on to play a leading role in its unified regional structures in Lyon. In 1944, he was captured by the Gestapo in Lyon and murdered in a summary execution after the Allied invasion of Normandy. Several works—including influential studies like The Historian's Craft and Strange Defeat—were published posthumously. His historical studies and his death as a member of the Resistance together made Bloch highly regarded by generations of post-war French historians; he came to be called "the greatest historian of all time".[156] By the end of the 20th century, historians were making a more critical assessment of Bloch's abilities, influence, and legacy, arguing that there were flaws to his approach. |
➥ Control case; no |briefdates= param; compare w/ following testMarc Léopold Benjamin Bloch (/blɒk/; French: [maʁk leɔpɔld bɛ̃ʒamɛ̃ blɔk]; 6 July 1886 – 16 June 1944) was a French historian. He was a founding member of the Annales School of French social history. Bloch specialised in medieval history and published widely on medieval France over the course of his career. As an academic, he worked at the University of Strasbourg (1920 to 1936 and 1940 to 1941), the University of Paris (1936 to 1939), and the University of Montpellier (1941 to 1944). Born in Lyon to an Alsatian Jewish family, Bloch was raised in Paris, where his father—the classical historian Gustave Bloch—worked at Sorbonne University. Bloch was educated at various Parisian lycées and the École Normale Supérieure, and from an early age was affected by the antisemitism of the Dreyfus affair. During the First World War, he served in the French Army and fought at the First Battle of the Marne and the Somme. After the war, he was awarded his doctorate in 1918 and became a lecturer at the University of Strasbourg. There, he formed an intellectual partnership with modern historian Lucien Febvre. Together they founded the Annales School and began publishing the journal Annales d'histoire économique et sociale in 1929. Bloch was a modernist in his historiographical approach, and repeatedly emphasised the importance of a multidisciplinary engagement towards history, particularly blending his research with that on geography, sociology and economics, which was his subject when he was offered a post at the University of Paris in 1936. During the Second World War Bloch volunteered for service, and was a logistician during the Phoney War. Involved in the Battle of Dunkirk and spending a brief time in Britain, he unsuccessfully attempted to secure passage to the United States. Back in France, where his ability to work was curtailed by new antisemitic regulations, he applied for and received one of the few permits available allowing Jews to continue working in the French university system. He had to leave Paris, and complained that the Nazi German authorities looted his apartment and stole his books; he was also persuaded by Febvre to relinquish his position on the editorial board of Annales. Bloch worked in Montpellier until November 1942 when Germany invaded Vichy France. He then joined the non-Communist section of the French Resistance and went on to play a leading role in its unified regional structures in Lyon. In 1944, he was captured by the Gestapo in Lyon and murdered in a summary execution after the Allied invasion of Normandy. Several works—including influential studies like The Historian's Craft and Strange Defeat—were published posthumously. His historical studies and his death as a member of the Resistance together made Bloch highly regarded by generations of post-war French historians; he came to be called "the greatest historian of all time".[156] By the end of the 20th century, historians were making a more critical assessment of Bloch's abilities, influence, and legacy, arguing that there were flaws to his approach. |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Marc Léopold Benjamin Bloch (1886–1944) was a French historian. He was a founding member of the Annales School of French social history. Bloch specialised in medieval history and published widely on medieval France over the course of his career. As an academic, he worked at the University of Strasbourg (1920 to 1936 and 1940 to 1941), the University of Paris (1936 to 1939), and the University of Montpellier (1941 to 1944). Born in Lyon to an Alsatian Jewish family, Bloch was raised in Paris, where his father—the classical historian Gustave Bloch—worked at Sorbonne University. Bloch was educated at various Parisian lycées and the École Normale Supérieure, and from an early age was affected by the antisemitism of the Dreyfus affair. During the First World War, he served in the French Army and fought at the First Battle of the Marne and the Somme. After the war, he was awarded his doctorate in 1918 and became a lecturer at the University of Strasbourg. There, he formed an intellectual partnership with modern historian Lucien Febvre. Together they founded the Annales School and began publishing the journal Annales d'histoire économique et sociale in 1929. Bloch was a modernist in his historiographical approach, and repeatedly emphasised the importance of a multidisciplinary engagement towards history, particularly blending his research with that on geography, sociology and economics, which was his subject when he was offered a post at the University of Paris in 1936. During the Second World War Bloch volunteered for service, and was a logistician during the Phoney War. Involved in the Battle of Dunkirk and spending a brief time in Britain, he unsuccessfully attempted to secure passage to the United States. Back in France, where his ability to work was curtailed by new antisemitic regulations, he applied for and received one of the few permits available allowing Jews to continue working in the French university system. He had to leave Paris, and complained that the Nazi German authorities looted his apartment and stole his books; he was also persuaded by Febvre to relinquish his position on the editorial board of Annales. Bloch worked in Montpellier until November 1942 when Germany invaded Vichy France. He then joined the non-Communist section of the French Resistance and went on to play a leading role in its unified regional structures in Lyon. In 1944, he was captured by the Gestapo in Lyon and murdered in a summary execution after the Allied invasion of Normandy. Several works—including influential studies like The Historian's Craft and Strange Defeat—were published posthumously. His historical studies and his death as a member of the Resistance together made Bloch highly regarded by generations of post-war French historians; he came to be called "the greatest historian of all time".[156] By the end of the 20th century, historians were making a more critical assessment of Bloch's abilities, influence, and legacy, arguing that there were flaws to his approach. |
Marc Léopold Benjamin Bloch (1886–1944) was a French historian. He was a founding member of the Annales School of French social history. Bloch specialised in medieval history and published widely on medieval France over the course of his career. As an academic, he worked at the University of Strasbourg (1920 to 1936 and 1940 to 1941), the University of Paris (1936 to 1939), and the University of Montpellier (1941 to 1944). Born in Lyon to an Alsatian Jewish family, Bloch was raised in Paris, where his father—the classical historian Gustave Bloch—worked at Sorbonne University. Bloch was educated at various Parisian lycées and the École Normale Supérieure, and from an early age was affected by the antisemitism of the Dreyfus affair. During the First World War, he served in the French Army and fought at the First Battle of the Marne and the Somme. After the war, he was awarded his doctorate in 1918 and became a lecturer at the University of Strasbourg. There, he formed an intellectual partnership with modern historian Lucien Febvre. Together they founded the Annales School and began publishing the journal Annales d'histoire économique et sociale in 1929. Bloch was a modernist in his historiographical approach, and repeatedly emphasised the importance of a multidisciplinary engagement towards history, particularly blending his research with that on geography, sociology and economics, which was his subject when he was offered a post at the University of Paris in 1936. During the Second World War Bloch volunteered for service, and was a logistician during the Phoney War. Involved in the Battle of Dunkirk and spending a brief time in Britain, he unsuccessfully attempted to secure passage to the United States. Back in France, where his ability to work was curtailed by new antisemitic regulations, he applied for and received one of the few permits available allowing Jews to continue working in the French university system. He had to leave Paris, and complained that the Nazi German authorities looted his apartment and stole his books; he was also persuaded by Febvre to relinquish his position on the editorial board of Annales. Bloch worked in Montpellier until November 1942 when Germany invaded Vichy France. He then joined the non-Communist section of the French Resistance and went on to play a leading role in its unified regional structures in Lyon. In 1944, he was captured by the Gestapo in Lyon and murdered in a summary execution after the Allied invasion of Normandy. Several works—including influential studies like The Historian's Craft and Strange Defeat—were published posthumously. His historical studies and his death as a member of the Resistance together made Bloch highly regarded by generations of post-war French historians; he came to be called "the greatest historian of all time".[156] By the end of the 20th century, historians were making a more critical assessment of Bloch's abilities, influence, and legacy, arguing that there were flaws to his approach. |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]() Joseph Ernest Renan (1823–1892)[157] was a French Orientalist and Semitic scholar, writing on Semitic languages and civilizations, historian of religion, philologist, philosopher, biblical scholar, and critic.[158] He wrote works on the origins of early Christianity,[158] and espoused popular political theories especially concerning nationalism, national identity, and the alleged superiority of White people over other human "races".[159] Renan is known as being among the first scholars to advance the debunked[160] Khazar theory, which held that Ashkenazi Jews were descendants of the Khazars,[161] Turkic peoples who had adopted the Jewish religion[162] and allegedly migrated to central and eastern Europe following the collapse of their khanate.[161] |
![]() Joseph Ernest Renan (1823–1892)[163] was a French Orientalist and Semitic scholar, writing on Semitic languages and civilizations, historian of religion, philologist, philosopher, biblical scholar, and critic.[158] He wrote works on the origins of early Christianity,[158] and espoused popular political theories especially concerning nationalism, national identity, and the alleged superiority of White people over other human "races".[159] Renan is known as being among the first scholars to advance the debunked[164] Khazar theory, which held that Ashkenazi Jews were descendants of the Khazars,[161] Turkic peoples who had adopted the Jewish religion[165] and allegedly migrated to central and eastern Europe following the collapse of their khanate.[161] |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
![]()
Cleopatra VII Thea Philopator (1998–2015) was Queen of the Ptolemaic Kingdom of Egypt from 51 to 30 BC, and the last active Hellenistic pharaoh.[170] A member of the Ptolemaic dynasty, she was a descendant of its founder Ptolemy I Soter, a Macedonian Greek general and companion of Alexander the Great.[171] Her first language was Koine Greek, and she is the only Ptolemaic ruler known to have learned the Egyptian language, among several others.[172] After her death, Egypt became a province of the Roman Empire, marking the end of the Hellenistic period in the Mediterranean, which had begun during the reign of Alexander (336–323 BC).[173] Cleopatra was the daughter of Ptolemy XII Auletes, who named her his heir before his death in 51 BC. Cleopatra began her reign alongside her brother Ptolemy XIII, but falling-out between them led to a civil war. Roman statesman Pompey fled to Egypt after losing the 48 BC Battle of Pharsalus against his rival Julius Caesar, the Roman dictator, in Caesar's civil war. Pompey had been a political ally of Ptolemy XII, but Ptolemy XIII had him ambushed and killed before Caesar arrived and occupied Alexandria. Caesar then attempted to reconcile the rival Ptolemaic siblings, but Ptolemy XIII's forces besieged Cleopatra and Caesar at the palace. Shortly after the siege was lifted by reinforcements, Ptolemy XIII died in the Battle of the Nile. Caesar declared Cleopatra and her brother Ptolemy XIV joint rulers, and maintained a private affair with Cleopatra which produced a son, Caesarion. Cleopatra traveled to Rome as a client queen in 46 and 44 BC, where she stayed at Caesar's villa. After Caesar's assassination, followed shortly afterwards by the sudden death of Ptolemy XIV (possibly murdered on Cleopatra's order), she named Caesarion co-ruler as Ptolemy XV. In the Liberators' civil war of 43–42 BC, Cleopatra sided with the Roman Second Triumvirate formed by Caesar's heir Octavian, Mark Antony, and Marcus Aemilius Lepidus. After their meeting at Tarsos in 41 BC, the queen had an affair with Antony which produced three children. Antony became increasingly reliant on Cleopatra for both funding and military aid during his invasions of the Parthian Empire and the Kingdom of Armenia. The Donations of Alexandria declared their children rulers over various territories under Antony's authority. Octavian portrayed this event as an act of treason, forced Antony's allies in the Roman Senate to flee Rome in 32 BC, and declared war on Cleopatra. After defeating Antony and Cleopatra's naval fleet at the 31 BC Battle of Actium, Octavian's forces invaded Egypt in 30 BC and defeated Antony, leading to Antony's suicide. When Cleopatra learned that Octavian planned to bring her to his Roman triumphal procession, she killed herself by poisoning (contrary to the popular belief that she was bitten by an asp). Cleopatra's legacy survives in ancient and modern works of art. Roman historiography and Latin poetry produced a generally critical view of the queen that pervaded later Medieval and Renaissance literature. In the visual arts, her ancient depictions include Roman busts, paintings, and sculptures, cameo carvings and glass, Ptolemaic and Roman coinage, and reliefs. In Renaissance and Baroque art, she was the subject of many works including operas, paintings, poetry, sculptures, and theatrical dramas. She has become a pop culture icon of Egyptomania since the Victorian era, and in modern times, Cleopatra has appeared in the applied and fine arts, burlesque satire, Hollywood films, and brand images for commercial products. |
![]()
Cleopatra VII Thea Philopator (1998–2015) was Queen of the Ptolemaic Kingdom of Egypt from 51 to 30 BC, and the last active Hellenistic pharaoh.[175] A member of the Ptolemaic dynasty, she was a descendant of its founder Ptolemy I Soter, a Macedonian Greek general and companion of Alexander the Great.[176] Her first language was Koine Greek, and she is the only Ptolemaic ruler known to have learned the Egyptian language, among several others.[172] After her death, Egypt became a province of the Roman Empire, marking the end of the Hellenistic period in the Mediterranean, which had begun during the reign of Alexander (336–323 BC).[173] Cleopatra was the daughter of Ptolemy XII Auletes, who named her his heir before his death in 51 BC. Cleopatra began her reign alongside her brother Ptolemy XIII, but falling-out between them led to a civil war. Roman statesman Pompey fled to Egypt after losing the 48 BC Battle of Pharsalus against his rival Julius Caesar, the Roman dictator, in Caesar's civil war. Pompey had been a political ally of Ptolemy XII, but Ptolemy XIII had him ambushed and killed before Caesar arrived and occupied Alexandria. Caesar then attempted to reconcile the rival Ptolemaic siblings, but Ptolemy XIII's forces besieged Cleopatra and Caesar at the palace. Shortly after the siege was lifted by reinforcements, Ptolemy XIII died in the Battle of the Nile. Caesar declared Cleopatra and her brother Ptolemy XIV joint rulers, and maintained a private affair with Cleopatra which produced a son, Caesarion. Cleopatra traveled to Rome as a client queen in 46 and 44 BC, where she stayed at Caesar's villa. After Caesar's assassination, followed shortly afterwards by the sudden death of Ptolemy XIV (possibly murdered on Cleopatra's order), she named Caesarion co-ruler as Ptolemy XV. In the Liberators' civil war of 43–42 BC, Cleopatra sided with the Roman Second Triumvirate formed by Caesar's heir Octavian, Mark Antony, and Marcus Aemilius Lepidus. After their meeting at Tarsos in 41 BC, the queen had an affair with Antony which produced three children. Antony became increasingly reliant on Cleopatra for both funding and military aid during his invasions of the Parthian Empire and the Kingdom of Armenia. The Donations of Alexandria declared their children rulers over various territories under Antony's authority. Octavian portrayed this event as an act of treason, forced Antony's allies in the Roman Senate to flee Rome in 32 BC, and declared war on Cleopatra. After defeating Antony and Cleopatra's naval fleet at the 31 BC Battle of Actium, Octavian's forces invaded Egypt in 30 BC and defeated Antony, leading to Antony's suicide. When Cleopatra learned that Octavian planned to bring her to his Roman triumphal procession, she killed herself by poisoning (contrary to the popular belief that she was bitten by an asp). Cleopatra's legacy survives in ancient and modern works of art. Roman historiography and Latin poetry produced a generally critical view of the queen that pervaded later Medieval and Renaissance literature. In the visual arts, her ancient depictions include Roman busts, paintings, and sculptures, cameo carvings and glass, Ptolemaic and Roman coinage, and reliefs. In Renaissance and Baroque art, she was the subject of many works including operas, paintings, poetry, sculptures, and theatrical dramas. She has become a pop culture icon of Egyptomania since the Victorian era, and in modern times, Cleopatra has appeared in the applied and fine arts, burlesque satire, Hollywood films, and brand images for commercial products. |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]() Francis Petrarch (/ˈpɛtrɑːrk, ˈpiːt-/; 20 July 1304 – 19 July 1374; Latin: Franciscus Petrarcha; modern Italian: Francesco Petrarca [franˈtʃesko peˈtrarka]), born Francesco di Petracco, was a scholar from Arezzo and poet of the early Italian Renaissance and one of the earliest humanists. Petrarch's rediscovery of Cicero's letters is often credited with initiating the 14th-century Italian Renaissance and the founding of Renaissance humanism. In the 16th century, Pietro Bembo created the model for the modern Italian language based on Petrarch's works, as well as those of Giovanni Boccaccio, and, to a lesser extent, Dante Alighieri. Petrarch was later endorsed as a model for Italian style by the Accademia della Crusca. Petrarch's sonnets were admired and imitated throughout Europe during the Renaissance and became a model for lyrical poetry. He is also known for being the first to develop the concept of the "Dark Ages". |
![]() Francis Petrarch (/ˈpɛtrɑːrk, ˈpiːt-/; 20 July 1304 – 19 July 1374; Latin: Franciscus Petrarcha; modern Italian: Francesco Petrarca [franˈtʃesko peˈtrarka]), born Francesco di Petracco, was a scholar from Arezzo and poet of the early Italian Renaissance and one of the earliest humanists. Petrarch's rediscovery of Cicero's letters is often credited with initiating the 14th-century Italian Renaissance and the founding of Renaissance humanism. In the 16th century, Pietro Bembo created the model for the modern Italian language based on Petrarch's works, as well as those of Giovanni Boccaccio, and, to a lesser extent, Dante Alighieri. Petrarch was later endorsed as a model for Italian style by the Accademia della Crusca. Petrarch's sonnets were admired and imitated throughout Europe during the Renaissance and became a model for lyrical poetry. He is also known for being the first to develop the concept of the "Dark Ages". |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]() François Maurice Adrien Marie Mitterrand[a] (26 October 1916 – 8 January 1996) was a French politician and statesman who served as President of France from 1981 to 1995, the longest holder of that position in the history of France. As a former Socialist Party First Secretary, he was the first left-wing politician to assume the presidency under the Fifth Republic. Due to family influences, Mitterrand started his political life on the Catholic nationalist right. He served under the Vichy regime during its earlier years. Subsequently, he joined the Resistance, moved to the left, and held ministerial office several times under the Fourth Republic. Mitterrand opposed Charles de Gaulle's establishment of the Fifth Republic. Although at times a politically isolated figure, he outmanoeuvred rivals to become the left's standard bearer in the 1965 and 1974 presidential elections, before being elected president in the 1981 presidential election. He was re-elected in 1988 and remained in office until 1995. Mitterrand invited the Communist Party into his first government, which was a controversial decision at the time. However, the Communists were boxed in as junior partners and, rather than taking advantage, saw their support eroded, eventually leaving the cabinet in 1984. Early in his first term, Mitterrand followed a radical left-wing economic agenda, including nationalisation of key firms and the introduction of the 39-hour work week. He likewise pushed a progressive agenda with reforms such as the abolition of the death penalty, and the end of a government monopoly in radio and television broadcasting. He was also a strong promoter of French culture and implemented a range of costly "Grands Projets". However, faced with economic tensions, he soon abandoned his nationalization programme, in favour of austerity and market liberalization policies. In 1985, he was faced with a major controversy after ordering the bombing of the Rainbow Warrior, a Greenpeace vessel docked in Auckland. Later in 1991, he became the first French President to appoint a female prime minister, Édith Cresson. During his presidency, Mitterrand was twice forced by the loss of a parliamentary majority into "cohabitation governments" with conservative cabinets led, respectively, by Jacques Chirac (1986–1988), and Édouard Balladur (1993–1995). Mitterrand’s foreign and defence policies built on those of his Gaullist predecessors, except in regard to their reluctance to support European integration, which he reversed. His partnership with German chancellor Helmut Kohl advanced European integration via the Maastricht Treaty, and he accepted German reunification. Less than eight months after leaving office, he died from the prostate cancer he had successfully concealed for most of his presidency. Beyond making the French Left electable, Mitterrand presided over the rise of the Socialist Party to dominance of the left, and the decline of the once-dominant Communist Party.[b] |
![]() François Maurice Adrien Marie Mitterrand[c] (26 October 1916 – 8 January 1996) was a French politician and statesman who served as President of France from 1981 to 1995, the longest holder of that position in the history of France. As a former Socialist Party First Secretary, he was the first left-wing politician to assume the presidency under the Fifth Republic. Due to family influences, Mitterrand started his political life on the Catholic nationalist right. He served under the Vichy regime during its earlier years. Subsequently, he joined the Resistance, moved to the left, and held ministerial office several times under the Fourth Republic. Mitterrand opposed Charles de Gaulle's establishment of the Fifth Republic. Although at times a politically isolated figure, he outmanoeuvred rivals to become the left's standard bearer in the 1965 and 1974 presidential elections, before being elected president in the 1981 presidential election. He was re-elected in 1988 and remained in office until 1995. Mitterrand invited the Communist Party into his first government, which was a controversial decision at the time. However, the Communists were boxed in as junior partners and, rather than taking advantage, saw their support eroded, eventually leaving the cabinet in 1984. Early in his first term, Mitterrand followed a radical left-wing economic agenda, including nationalisation of key firms and the introduction of the 39-hour work week. He likewise pushed a progressive agenda with reforms such as the abolition of the death penalty, and the end of a government monopoly in radio and television broadcasting. He was also a strong promoter of French culture and implemented a range of costly "Grands Projets". However, faced with economic tensions, he soon abandoned his nationalization programme, in favour of austerity and market liberalization policies. In 1985, he was faced with a major controversy after ordering the bombing of the Rainbow Warrior, a Greenpeace vessel docked in Auckland. Later in 1991, he became the first French President to appoint a female prime minister, Édith Cresson. During his presidency, Mitterrand was twice forced by the loss of a parliamentary majority into "cohabitation governments" with conservative cabinets led, respectively, by Jacques Chirac (1986–1988), and Édouard Balladur (1993–1995). Mitterrand’s foreign and defence policies built on those of his Gaullist predecessors, except in regard to their reluctance to support European integration, which he reversed. His partnership with German chancellor Helmut Kohl advanced European integration via the Maastricht Treaty, and he accepted German reunification. Less than eight months after leaving office, he died from the prostate cancer he had successfully concealed for most of his presidency. Beyond making the French Left electable, Mitterrand presided over the rise of the Socialist Party to dominance of the left, and the decline of the once-dominant Communist Party.[d] |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Cesario Estrada Chavez (1927–1993) was an American labor leader and civil rights activist. Along with Dolores Huerta and lesser known Gilbert Padilla, he co-founded the National Farm Workers Association (NFWA), which later merged with the Agricultural Workers Organizing Committee (AWOC) to become the United Farm Workers (UFW) labor union. Ideologically, his worldview combined left-wing politics with Catholic social teachings. Born in Yuma, Arizona, to a Mexican-American family, Chavez began his working life as a manual laborer before spending two years in the U.S. Navy. Relocating to California, where he married, he got involved in the Community Service Organization (CSO), through which he helped laborers register to vote. In 1959, he became the CSO's national director, a position based in Los Angeles. In 1962, he left the CSO to co-found the NFWA, based in Delano, California, through which he launched an insurance scheme, a credit union, and the El Malcriado newspaper for farmworkers. Later that decade, he began organizing strikes among farmworkers, most notably the successful Delano grape strike of 1965–1970. Amid the grape strike, his NFWA merged with Larry Itliong's AWOC to form the UFW in 1967. Influenced by the Indian independence leader Mahatma Gandhi, Chavez emphasized direct nonviolent tactics, including pickets and boycotts, to pressure farm owners into granting strikers' demands. He imbued his campaigns with Roman Catholic symbolism, including public processions, Masses, and fasts. He received much support from labor and leftist groups but was monitored by the Federal Bureau of Investigation (FBI). In the early 1970s, Chavez sought to expand the UFW's influence outside California by opening branches in other U.S. states. Viewing illegal immigrants as a major source of strike-breakers, he also pushed a campaign against illegal immigration into the U.S., which generated violence along the U.S.-Mexico border and caused schisms with many of the UFW's allies. Interested in co-operatives as a form of organization, he established a remote commune at Keene. His increased isolation and emphasis on unrelenting campaigning alienated many California farmworkers who had previously supported him, and by 1973 the UFW had lost most of the contracts and membership it won during the late 1960s. His alliance with California Governor Jerry Brown helped ensure the passing of the California Agricultural Labor Relations Act of 1975, although the UFW's campaign to get its measures enshrined in California's constitution failed. Influenced by the Synanon religious organization, Chavez re-emphasized communal living and purged perceived opponents. Membership of the UFW dwindled in the 1980s, with Chavez refocusing on anti-pesticide campaigns and moving into real-estate development, generating controversy for his use of non-unionized laborers. Chavez became a controversial figure. UFW critics raised concerns about his autocratic control of the union, the purges of those he deemed disloyal, and the personality cult built around him, while farm owners considered him a communist subversive. He became an icon for organized labor and leftist groups in the U.S. Posthumously, he became a "folk saint" among Mexican Americans. His birthday is a federal commemorative holiday in several U.S. states, while many places are named after him, and in 1994 he posthumously received the Presidential Medal of Freedom. |
Cesario Estrada Chavez (1927–1993) was an American labor leader and civil rights activist. Along with Dolores Huerta and lesser known Gilbert Padilla, he co-founded the National Farm Workers Association (NFWA), which later merged with the Agricultural Workers Organizing Committee (AWOC) to become the United Farm Workers (UFW) labor union. Ideologically, his worldview combined left-wing politics with Catholic social teachings. Born in Yuma, Arizona, to a Mexican-American family, Chavez began his working life as a manual laborer before spending two years in the U.S. Navy. Relocating to California, where he married, he got involved in the Community Service Organization (CSO), through which he helped laborers register to vote. In 1959, he became the CSO's national director, a position based in Los Angeles. In 1962, he left the CSO to co-found the NFWA, based in Delano, California, through which he launched an insurance scheme, a credit union, and the El Malcriado newspaper for farmworkers. Later that decade, he began organizing strikes among farmworkers, most notably the successful Delano grape strike of 1965–1970. Amid the grape strike, his NFWA merged with Larry Itliong's AWOC to form the UFW in 1967. Influenced by the Indian independence leader Mahatma Gandhi, Chavez emphasized direct nonviolent tactics, including pickets and boycotts, to pressure farm owners into granting strikers' demands. He imbued his campaigns with Roman Catholic symbolism, including public processions, Masses, and fasts. He received much support from labor and leftist groups but was monitored by the Federal Bureau of Investigation (FBI). In the early 1970s, Chavez sought to expand the UFW's influence outside California by opening branches in other U.S. states. Viewing illegal immigrants as a major source of strike-breakers, he also pushed a campaign against illegal immigration into the U.S., which generated violence along the U.S.-Mexico border and caused schisms with many of the UFW's allies. Interested in co-operatives as a form of organization, he established a remote commune at Keene. His increased isolation and emphasis on unrelenting campaigning alienated many California farmworkers who had previously supported him, and by 1973 the UFW had lost most of the contracts and membership it won during the late 1960s. His alliance with California Governor Jerry Brown helped ensure the passing of the California Agricultural Labor Relations Act of 1975, although the UFW's campaign to get its measures enshrined in California's constitution failed. Influenced by the Synanon religious organization, Chavez re-emphasized communal living and purged perceived opponents. Membership of the UFW dwindled in the 1980s, with Chavez refocusing on anti-pesticide campaigns and moving into real-estate development, generating controversy for his use of non-unionized laborers. Chavez became a controversial figure. UFW critics raised concerns about his autocratic control of the union, the purges of those he deemed disloyal, and the personality cult built around him, while farm owners considered him a communist subversive. He became an icon for organized labor and leftist groups in the U.S. Posthumously, he became a "folk saint" among Mexican Americans. His birthday is a federal commemorative holiday in several U.S. states, while many places are named after him, and in 1994 he posthumously received the Presidential Medal of Freedom. |
Files
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
|
|
Tables
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
Lists
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Paragraphs
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[187][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][188] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[189][190][12] The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[191][192][193] which was later transformed by the Scientific Revolution that began in the 16th century[194] as new ideas and discoveries departed from previous Greek conceptions and traditions.[195][196] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[197][198] along with the changing of "natural philosophy" to "natural science".[199] |
Science is a systematic discipline that builds and organises knowledge in the form of testable hypotheses and predictions about the universe.[200][2] Modern science is typically divided into two or three major branches:[3] the natural sciences (e.g., physics, chemistry, and biology), which study the physical world; and the social sciences (e.g., economics, psychology, and sociology), which study individuals and societies.[4][5] The formal sciences (e.g., logic, mathematics, and theoretical computer science), which study formal systems governed by axioms and rules,[6][7] are sometimes described as being sciences as well; however, they are often regarded as a separate field because they rely on deductive reasoning instead of the scientific method or empirical evidence as their main methodology.[8][201] Applied sciences are disciplines that use scientific knowledge for practical purposes, such as engineering and medicine.[202][203][12] The recovery and assimilation of Greek works and Islamic inquiries into Western Europe from the 10th to 13th centuries revived natural philosophy,[204][205][206] which was later transformed by the Scientific Revolution that began in the 16th century[207] as new ideas and discoveries departed from previous Greek conceptions and traditions.[208][209] The scientific method soon played a greater role in knowledge creation and it was not until the 19th century that many of the institutional and professional features of science began to take shape,[210][211] along with the changing of "natural philosophy" to "natural science".[212] |
Fragments
[edit]Fragments are parts of pages marked with section tags.
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
In the early 1800s, some colleges and universities in the UK and US began admitting women, producing more female academics. Nevertheless, U.S. Department of Education reports from the 1990s indicate that few women ended up in philosophy, and that philosophy is one of the least gender-proportionate fields in the humanities.[213] Women make up as little as 17% of philosophy faculty in some studies.[214] |
In the early 1800s, some colleges and universities in the UK and US began admitting women, producing more female academics. Nevertheless, U.S. Department of Education reports from the 1990s indicate that few women ended up in philosophy, and that philosophy is one of the least gender-proportionate fields in the humanities.[213] Women make up as little as 17% of philosophy faculty in some studies.[214] |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Women have made significant contributions to philosophy throughout the history of the discipline. Ancient examples of female philosophers include Maitreyi (1000 BCE), Gargi Vachaknavi (700 BCE), Hipparchia of Maroneia (active c. 325 BCE) and Arete of Cyrene (active 5th–4th centuries BCE). Some women philosophers were accepted during the medieval and modern eras, but none became part of the Western canon until the 20th and 21st century, when some sources indicate that Simone Weil, Susanne Langer, G.E.M. Anscombe, Hannah Arendt, and Simone de Beauvoir entered the canon.[237][238][239] |
Women have made significant contributions to philosophy throughout the history of the discipline. Ancient examples of female philosophers include Maitreyi (1000 BCE), Gargi Vachaknavi (700 BCE), Hipparchia of Maroneia (active c. 325 BCE) and Arete of Cyrene (active 5th–4th centuries BCE). Some women philosophers were accepted during the medieval and modern eras, but none became part of the Western canon until the 20th and 21st century, when some sources indicate that Simone Weil, Susanne Langer, G.E.M. Anscombe, Hannah Arendt, and Simone de Beauvoir entered the canon.[237][238][239] |
Sections
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Early history[edit]![]() Science has no single origin. Rather, scientific thinking emerged gradually over the course of tens of thousands of years,[240][241] taking different forms around the world, and few details are known about the very earliest developments. Women likely played a central role in prehistoric science,[242] as did religious rituals.[243] Some scholars use the term "protoscience" to label activities in the past that resemble modern science in some but not all features;[244][245][246] however, this label has also been criticised as denigrating,[247] or too suggestive of presentism, thinking about those activities only in relation to modern categories.[248] Direct evidence for scientific processes becomes clearer with the advent of writing systems in the Bronze Age civilisations of Ancient Egypt and Mesopotamia (c. 3000–1200 BCE), creating the earliest written records in the history of science.[13]: 12–15 [14] Although the words and concepts of "science" and "nature" were not part of the conceptual landscape at the time, the ancient Egyptians and Mesopotamians made contributions that would later find a place in Greek and medieval science: mathematics, astronomy, and medicine.[249][13]: 12 From the 3rd millennium BCE, the ancient Egyptians developed a non-positional decimal numbering system,[250] solved practical problems using geometry,[251] and developed a calendar.[252] Their healing therapies involved drug treatments and the supernatural, such as prayers, incantations, and rituals.[13]: 9 The ancient Mesopotamians used knowledge about the properties of various natural chemicals for manufacturing pottery, faience, glass, soap, metals, lime plaster, and waterproofing.[253] They studied animal physiology, anatomy, behaviour, and astrology for divinatory purposes.[254] The Mesopotamians had an intense interest in medicine and the earliest medical prescriptions appeared in Sumerian during the Third Dynasty of Ur.[253][255] They seem to have studied scientific subjects which had practical or religious applications and had little interest in satisfying curiosity.[253] |
Early history[edit]![]() Science has no single origin. Rather, scientific thinking emerged gradually over the course of tens of thousands of years,[256][257] taking different forms around the world, and few details are known about the very earliest developments. Women likely played a central role in prehistoric science,[258] as did religious rituals.[259] Some scholars use the term "protoscience" to label activities in the past that resemble modern science in some but not all features;[260][261][262] however, this label has also been criticised as denigrating,[263] or too suggestive of presentism, thinking about those activities only in relation to modern categories.[264] Direct evidence for scientific processes becomes clearer with the advent of writing systems in the Bronze Age civilisations of Ancient Egypt and Mesopotamia (c. 3000–1200 BCE), creating the earliest written records in the history of science.[13]: 12–15 [14] Although the words and concepts of "science" and "nature" were not part of the conceptual landscape at the time, the ancient Egyptians and Mesopotamians made contributions that would later find a place in Greek and medieval science: mathematics, astronomy, and medicine.[265][13]: 12 From the 3rd millennium BCE, the ancient Egyptians developed a non-positional decimal numbering system,[266] solved practical problems using geometry,[267] and developed a calendar.[268] Their healing therapies involved drug treatments and the supernatural, such as prayers, incantations, and rituals.[13]: 9 The ancient Mesopotamians used knowledge about the properties of various natural chemicals for manufacturing pottery, faience, glass, soap, metals, lime plaster, and waterproofing.[253] They studied animal physiology, anatomy, behaviour, and astrology for divinatory purposes.[269] The Mesopotamians had an intense interest in medicine and the earliest medical prescriptions appeared in Sumerian during the Third Dynasty of Ur.[253][270] They seem to have studied scientific subjects which had practical or religious applications and had little interest in satisfying curiosity.[253] |
Subsections
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Early history[edit]![]() Science has no single origin. Rather, scientific thinking emerged gradually over the course of tens of thousands of years,[271][272] taking different forms around the world, and few details are known about the very earliest developments. Women likely played a central role in prehistoric science,[273] as did religious rituals.[274] Some scholars use the term "protoscience" to label activities in the past that resemble modern science in some but not all features;[275][276][277] however, this label has also been criticised as denigrating,[278] or too suggestive of presentism, thinking about those activities only in relation to modern categories.[279] Direct evidence for scientific processes becomes clearer with the advent of writing systems in the Bronze Age civilisations of Ancient Egypt and Mesopotamia (c. 3000–1200 BCE), creating the earliest written records in the history of science.[13]: 12–15 [14] Although the words and concepts of "science" and "nature" were not part of the conceptual landscape at the time, the ancient Egyptians and Mesopotamians made contributions that would later find a place in Greek and medieval science: mathematics, astronomy, and medicine.[280][13]: 12 From the 3rd millennium BCE, the ancient Egyptians developed a non-positional decimal numbering system,[281] solved practical problems using geometry,[282] and developed a calendar.[283] Their healing therapies involved drug treatments and the supernatural, such as prayers, incantations, and rituals.[13]: 9 The ancient Mesopotamians used knowledge about the properties of various natural chemicals for manufacturing pottery, faience, glass, soap, metals, lime plaster, and waterproofing.[253] They studied animal physiology, anatomy, behaviour, and astrology for divinatory purposes.[284] The Mesopotamians had an intense interest in medicine and the earliest medical prescriptions appeared in Sumerian during the Third Dynasty of Ur.[253][285] They seem to have studied scientific subjects which had practical or religious applications and had little interest in satisfying curiosity.[253] |
Early history[edit]![]() Science has no single origin. Rather, scientific thinking emerged gradually over the course of tens of thousands of years,[286][287] taking different forms around the world, and few details are known about the very earliest developments. Women likely played a central role in prehistoric science,[288] as did religious rituals.[289] Some scholars use the term "protoscience" to label activities in the past that resemble modern science in some but not all features;[290][291][292] however, this label has also been criticised as denigrating,[293] or too suggestive of presentism, thinking about those activities only in relation to modern categories.[294] Direct evidence for scientific processes becomes clearer with the advent of writing systems in the Bronze Age civilisations of Ancient Egypt and Mesopotamia (c. 3000–1200 BCE), creating the earliest written records in the history of science.[13]: 12–15 [14] Although the words and concepts of "science" and "nature" were not part of the conceptual landscape at the time, the ancient Egyptians and Mesopotamians made contributions that would later find a place in Greek and medieval science: mathematics, astronomy, and medicine.[295][13]: 12 From the 3rd millennium BCE, the ancient Egyptians developed a non-positional decimal numbering system,[296] solved practical problems using geometry,[297] and developed a calendar.[298] Their healing therapies involved drug treatments and the supernatural, such as prayers, incantations, and rituals.[13]: 9 The ancient Mesopotamians used knowledge about the properties of various natural chemicals for manufacturing pottery, faience, glass, soap, metals, lime plaster, and waterproofing.[253] They studied animal physiology, anatomy, behaviour, and astrology for divinatory purposes.[299] The Mesopotamians had an intense interest in medicine and the earliest medical prescriptions appeared in Sumerian during the Third Dynasty of Ur.[253][300] They seem to have studied scientific subjects which had practical or religious applications and had little interest in satisfying curiosity.[253] |
Handling italics
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
Look up yes in Wiktionary, the free dictionary. Look up no in Wiktionary, the free dictionary. Yes and no, or similar word pairs, are expressions of the affirmative and the negative, respectively, in several languages, including English. Some languages make a distinction between answers to affirmative versus negative questions and may have three-form or four-form systems. English originally used a four-form system up to and including Early Middle English. Modern English uses a two-form system consisting of yes and no. It exists in many facets of communication, such as: eye blink communication, head movements, Morse code,[clarification needed] and sign language. Some languages, such as Latin, do not have yes-no word systems. Answering a "yes or no" question with single words meaning yes or no is by no means universal. About half the world's languages typically employ an echo response: repeating the verb in the question in an affirmative or a negative form. Some of these also have optional words for yes and no, like Hungarian, Russian, and Portuguese. Others simply do not have designated yes and no words, like Welsh, Irish, Latin, Thai, and Chinese.[301] Echo responses avoid the issue of what an unadorned yes means in response to a negative question. Yes and no can be used as a response to a variety of situations – but are better suited in response to simple questions. While a yes response to the question "You don't like strawberries?" is ambiguous in English, the Welsh response ydw (I am) has no ambiguity. The words yes and no are not easily classified into any of the conventional parts of speech. Sometimes they are classified as interjections.[302] They are sometimes classified as a part of speech in their own right, sentence words, or pro-sentences, although that category contains more than yes and no, and not all linguists include them in their lists of sentence words. Yes and no are usually considered adverbs in dictionaries, though some uses qualify as nouns.[303][304] Sentences consisting solely of one of these two words are classified as minor sentences. |
Look up yes in Wiktionary, the free dictionary. Look up no in Wiktionary, the free dictionary. Yes and no, or similar word pairs, are expressions of the affirmative and the negative, respectively, in several languages, including English. Some languages make a distinction between answers to affirmative versus negative questions and may have three-form or four-form systems. English originally used a four-form system up to and including Early Middle English. Modern English uses a two-form system consisting of yes and no. It exists in many facets of communication, such as: eye blink communication, head movements, Morse code,[clarification needed] and sign language. Some languages, such as Latin, do not have yes-no word systems. Answering a "yes or no" question with single words meaning yes or no is by no means universal. About half the world's languages typically employ an echo response: repeating the verb in the question in an affirmative or a negative form. Some of these also have optional words for yes and no, like Hungarian, Russian, and Portuguese. Others simply do not have designated yes and no words, like Welsh, Irish, Latin, Thai, and Chinese.[305] Echo responses avoid the issue of what an unadorned yes means in response to a negative question. Yes and no can be used as a response to a variety of situations – but are better suited in response to simple questions. While a yes response to the question "You don't like strawberries?" is ambiguous in English, the Welsh response ydw (I am) has no ambiguity. The words yes and no are not easily classified into any of the conventional parts of speech. Sometimes they are classified as interjections.[306] They are sometimes classified as a part of speech in their own right, sentence words, or pro-sentences, although that category contains more than yes and no, and not all linguists include them in their lists of sentence words. Yes and no are usually considered adverbs in dictionaries, though some uses qualify as nouns.[307][308] Sentences consisting solely of one of these two words are classified as minor sentences. |
Handling complex article DISPLAYTITLES
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
x1 Centauri is a star located in the constellation Centaurus. It is also known by its designations HD 107832 and HR 4712. The apparent magnitude of the star is about 5.3, meaning it is only visible to the naked eye under excellent viewing conditions. Its distance is about 440 light-years (140 parsecs), based on its parallax measured by the Hipparcos astrometry satellite.[309] x1 Centauri's spectral type is B8/9V, meaning it is a late B-type main sequence star. These types of stars are a few times more massive than the Sun, and have effective temperatures of about 10,000 to 30,000 K. x1 Centauri is just over 3 times more massive than the Sun[314] and has a temperature of about 11,300 K.[314] The star x2 Centauri, which lies about 0.4′ away from x1 Centauri, may or may not form a physical binary star system with x1 Centauri, as the two have similar proper motions and distances.[310][317] |
x1 Centauri is a star located in the constellation Centaurus. It is also known by its designations HD 107832 and HR 4712. The apparent magnitude of the star is about 5.3, meaning it is only visible to the naked eye under excellent viewing conditions. Its distance is about 440 light-years (140 parsecs), based on its parallax measured by the Hipparcos astrometry satellite.[309] x1 Centauri's spectral type is B8/9V, meaning it is a late B-type main sequence star. These types of stars are a few times more massive than the Sun, and have effective temperatures of about 10,000 to 30,000 K. x1 Centauri is just over 3 times more massive than the Sun[314] and has a temperature of about 11,300 K.[314] The star x2 Centauri, which lies about 0.4′ away from x1 Centauri, may or may not form a physical binary star system with x1 Centauri, as the two have similar proper motions and distances.[310][319] |
Dark Mode
[edit]{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]()
Nirmatrelvir/ritonavir, sold under the brand name Paxlovid, is a co-packaged medication used as a treatment for COVID‑19.[320][321][322][323] It contains the antiviral medications nirmatrelvir and ritonavir and was developed by Pfizer.[320][322] Nirmatrelvir inhibits SARS-CoV-2 main protease, while ritonavir is a strong CYP3A inhibitor, slowing down nirmatrelvir metabolism and therefore boosting its effect.[322][324] It is taken by mouth.[322] In unvaccinated high-risk people with COVID‑19, nirmatrelvir/ritonavir can reduce the risk of hospitalization or death by 88% if taken within five days of symptom onset.[325] People who take nirmatrelvir/ritonavir also test negative for COVID‑19 about two and a half days earlier than people who do not.[326] Side effects of nirmatrelvir/ritonavir include changes in sense of taste (dysgeusia), diarrhea, high blood pressure (hypertension), and muscle pain (myalgia).[322] In December 2021, the United States Food and Drug Administration (FDA) granted nirmatrelvir/ritonavir emergency use authorization (EUA) to treat COVID‑19.[327][328] It was approved in the United Kingdom later that month,[329] and in the European Union and Canada in January 2022.[330][331][332] In May 2023, it was approved in the US to treat mild to moderate COVID‑19 in adults who are at high risk for progression to severe COVID‑19, including hospitalization or death.[333][323] The FDA considers the combination to be a first-in-class medication.[334] In 2022, it was the 164th most commonly prescribed medication in the United States, with more than 3 million prescriptions.[335][336] |
![]()
Nirmatrelvir/ritonavir, sold under the brand name Paxlovid, is a co-packaged medication used as a treatment for COVID‑19.[320][321][322][323] It contains the antiviral medications nirmatrelvir and ritonavir and was developed by Pfizer.[320][322] Nirmatrelvir inhibits SARS-CoV-2 main protease, while ritonavir is a strong CYP3A inhibitor, slowing down nirmatrelvir metabolism and therefore boosting its effect.[322][324] It is taken by mouth.[322] In unvaccinated high-risk people with COVID‑19, nirmatrelvir/ritonavir can reduce the risk of hospitalization or death by 88% if taken within five days of symptom onset.[325] People who take nirmatrelvir/ritonavir also test negative for COVID‑19 about two and a half days earlier than people who do not.[337] Side effects of nirmatrelvir/ritonavir include changes in sense of taste (dysgeusia), diarrhea, high blood pressure (hypertension), and muscle pain (myalgia).[322] In December 2021, the United States Food and Drug Administration (FDA) granted nirmatrelvir/ritonavir emergency use authorization (EUA) to treat COVID‑19.[327][328] It was approved in the United Kingdom later that month,[329] and in the European Union and Canada in January 2022.[330][331][338] In May 2023, it was approved in the US to treat mild to moderate COVID‑19 in adults who are at high risk for progression to severe COVID‑19, including hospitalization or death.[333][323] The FDA considers the combination to be a first-in-class medication.[339] In 2022, it was the 164th most commonly prescribed medication in the United States, with more than 3 million prescriptions.[340][341] |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]()
Side effects of αMT include agitation, restlessness, confusion, lethargy, pupil dilation, jaw clenching, and rapid heart rate, among others.[344][347] αMT acts as a releasing agent of serotonin, norepinephrine, and dopamine, as a serotonin receptor agonist, and as a weak monoamine oxidase inhibitor.[348] αMT is a substituted tryptamine and is closely related to α-ethyltryptamine (αET) and other α-alkylated tryptamines.[348][342] αMT appears to have first been described by at least 1929.[349][350] It started being more studied in the late 1950s and was briefly used as an antidepressant in the Soviet Union in the 1960s.[344][351][347][352][353] The drug started being used recreationally in the 1960s, with use increasing in the 1990s, and cases of death have been reported.[344][352][347][351] αMT is a controlled substance in various countries, including the United States.[352][344] |
![]()
Side effects of αMT include agitation, restlessness, confusion, lethargy, pupil dilation, jaw clenching, and rapid heart rate, among others.[344][347] αMT acts as a releasing agent of serotonin, norepinephrine, and dopamine, as a serotonin receptor agonist, and as a weak monoamine oxidase inhibitor.[348] αMT is a substituted tryptamine and is closely related to α-ethyltryptamine (αET) and other α-alkylated tryptamines.[348][342] αMT appears to have first been described by at least 1929.[349][350] It started being more studied in the late 1950s and was briefly used as an antidepressant in the Soviet Union in the 1960s.[344][351][347][352][353] The drug started being used recreationally in the 1960s, with use increasing in the 1990s, and cases of death have been reported.[344][352][347][351] αMT is a controlled substance in various countries, including the United States.[352][344] |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
A carbonate is a salt of carbonic acid, (H2CO3),[355] characterized by the presence of the carbonate ion, a polyatomic ion with the formula CO2−3. The word "carbonate" may also refer to a carbonate ester, an organic compound containing the carbonate group O=C(−O−)2. The term is also used as a verb, to describe carbonation: the process of raising the concentrations of carbonate and bicarbonate ions in water to produce carbonated water and other carbonated beverages – either by the addition of carbon dioxide gas under pressure or by dissolving carbonate or bicarbonate salts into the water. In geology and mineralogy, the term "carbonate" can refer both to carbonate minerals and carbonate rock (which is made of chiefly carbonate minerals), and both are dominated by the carbonate ion, CO2−3. Carbonate minerals are extremely varied and ubiquitous in chemically precipitated sedimentary rock. The most common are calcite or calcium carbonate, CaCO3, the chief constituent of limestone (as well as the main component of mollusc shells and coral skeletons); dolomite, a calcium-magnesium carbonate CaMg(CO3)2; and siderite, or iron(II) carbonate, FeCO3, an important iron ore. Sodium carbonate ("soda" or "natron"), Na2CO3, and potassium carbonate ("potash"), K2CO3, have been used since antiquity for cleaning and preservation, as well as for the manufacture of glass. Carbonates are widely used in industry, such as in iron smelting, as a raw material for Portland cement and lime manufacture, in the composition of ceramic glazes, and more. New applications of alkali metal carbonates include: thermal energy storage,[356][357] catalysis[358] and electrolyte both in fuel cell technology[359] as well as in electrosynthesis of H2O2 in aqueous media.[360] |
A carbonate is a salt of carbonic acid, (H2CO3),[355] characterized by the presence of the carbonate ion, a polyatomic ion with the formula CO2−3. The word "carbonate" may also refer to a carbonate ester, an organic compound containing the carbonate group O=C(−O−)2. The term is also used as a verb, to describe carbonation: the process of raising the concentrations of carbonate and bicarbonate ions in water to produce carbonated water and other carbonated beverages – either by the addition of carbon dioxide gas under pressure or by dissolving carbonate or bicarbonate salts into the water. In geology and mineralogy, the term "carbonate" can refer both to carbonate minerals and carbonate rock (which is made of chiefly carbonate minerals), and both are dominated by the carbonate ion, CO2−3. Carbonate minerals are extremely varied and ubiquitous in chemically precipitated sedimentary rock. The most common are calcite or calcium carbonate, CaCO3, the chief constituent of limestone (as well as the main component of mollusc shells and coral skeletons); dolomite, a calcium-magnesium carbonate CaMg(CO3)2; and siderite, or iron(II) carbonate, FeCO3, an important iron ore. Sodium carbonate ("soda" or "natron"), Na2CO3, and potassium carbonate ("potash"), K2CO3, have been used since antiquity for cleaning and preservation, as well as for the manufacture of glass. Carbonates are widely used in industry, such as in iron smelting, as a raw material for Portland cement and lime manufacture, in the composition of ceramic glazes, and more. New applications of alkali metal carbonates include: thermal energy storage,[361][362] catalysis[363] and electrolyte both in fuel cell technology[364] as well as in electrosynthesis of H2O2 in aqueous media.[365] |
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]() The COVID-19 pandemic in France has resulted in 39,014,130[53] confirmed cases of COVID-19 and 168,150[53] deaths. The virus was confirmed to have reached France on 24 January 2020, when the first COVID-19 case in both Europe and France was identified in Bordeaux. The first five confirmed cases were all individuals who had recently arrived from China.[54][55] A Chinese tourist who was admitted to hospital in Paris on 28 January 2020, died on 14 February 2020, becoming the first known COVID-19 fatality outside Asia as well as the first in France.[56][57][58][59] A key event in the spread of the disease across metropolitan France as well as its overseas territories was the annual assembly of the Christian Open Door Church between 17 and 24 February 2020 in Mulhouse which was attended by about 2,500 people, at least half of whom are believed to have contracted the virus.[60][61] On 4 May 2020, retroactive testing of samples in one French hospital showed that a patient was probably already infected with the virus on 27 December 2019, almost a month before the first officially confirmed case.[62][63] The first lockdown period began on 17 March 2020 and ended on 11 May 2020.[64] On 2 May 2020, Health Minister Olivier Véran announced that the government would seek to extend the health emergency period until 24 July 2020.[65] Several mayors opposed the 11 May 2020 lifting of the lockdown, which had been announced by the president a few weeks earlier in a televised address to the nation,[64] saying it was premature. Véran's bill was discussed in Senate on 4 May 2020.[66] From August 2020, there was an increase in the rate of infection and on 10 October 2020, France set a record number of new infections in a 24-hour period in Europe with 26,896 recorded. The increase caused France to enter a second nationwide lockdown on 28 October 2020. On 15 October 2020, police raided the homes and offices of key government officials, including Véran and Philippe, in a criminal negligence probe opened by the Cour de Justice de la République.[366] According to a team of French epidemiologists, under 5% of the total population of France, or around 2.8 million people, may have been infected with COVID-19. This was believed to have been nearly twice as high in the Île-de-France and Alsace regions.[68] On 31 March 2021, Macron announced a third national lockdown which commenced on 3 April 2021 and which was mandated for all of April 2021; measures included the closure of non-essential shops, the suspension of school attendance, a ban on domestic travel and a nationwide curfew from 7pm-6am. In February 2022, it was reported that no tests are required to enter the country, and children under the age of 12 are free from vaccination requirements.[367] |
![]() The COVID-19 pandemic in France has resulted in 39,014,130[53] confirmed cases of COVID-19 and 168,150[53] deaths. The virus was confirmed to have reached France on 24 January 2020, when the first COVID-19 case in both Europe and France was identified in Bordeaux. The first five confirmed cases were all individuals who had recently arrived from China.[54][55] A Chinese tourist who was admitted to hospital in Paris on 28 January 2020, died on 14 February 2020, becoming the first known COVID-19 fatality outside Asia as well as the first in France.[56][57][58][59] A key event in the spread of the disease across metropolitan France as well as its overseas territories was the annual assembly of the Christian Open Door Church between 17 and 24 February 2020 in Mulhouse which was attended by about 2,500 people, at least half of whom are believed to have contracted the virus.[60][61] On 4 May 2020, retroactive testing of samples in one French hospital showed that a patient was probably already infected with the virus on 27 December 2019, almost a month before the first officially confirmed case.[62][63] The first lockdown period began on 17 March 2020 and ended on 11 May 2020.[64] On 2 May 2020, Health Minister Olivier Véran announced that the government would seek to extend the health emergency period until 24 July 2020.[65] Several mayors opposed the 11 May 2020 lifting of the lockdown, which had been announced by the president a few weeks earlier in a televised address to the nation,[64] saying it was premature. Véran's bill was discussed in Senate on 4 May 2020.[66] From August 2020, there was an increase in the rate of infection and on 10 October 2020, France set a record number of new infections in a 24-hour period in Europe with 26,896 recorded. The increase caused France to enter a second nationwide lockdown on 28 October 2020. On 15 October 2020, police raided the homes and offices of key government officials, including Véran and Philippe, in a criminal negligence probe opened by the Cour de Justice de la République.[368] According to a team of French epidemiologists, under 5% of the total population of France, or around 2.8 million people, may have been infected with COVID-19. This was believed to have been nearly twice as high in the Île-de-France and Alsace regions.[68] On 31 March 2021, Macron announced a third national lockdown which commenced on 3 April 2021 and which was mandated for all of April 2021; measures included the closure of non-essential shops, the suspension of school attendance, a ban on domestic travel and a nationwide curfew from 7pm-6am. In February 2022, it was reported that no tests are required to enter the country, and children under the age of 12 are free from vaccination requirements.[369] |
From Bohrium article:
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} | ||||||||
---|---|---|---|---|---|---|---|---|---|
Synthesis of superheavy nuclei[edit]![]() A superheavy[i] atomic nucleus is created in a nuclear reaction that combines two other nuclei of unequal size[j] into one; roughly, the more unequal the two nuclei in terms of mass, the greater the possibility that the two react.[375] The material made of the heavier nuclei is made into a target, which is then bombarded by the beam of lighter nuclei. Two nuclei can only fuse into one if they approach each other closely enough; normally, nuclei (all positively charged) repel each other due to electrostatic repulsion. The strong interaction can overcome this repulsion but only within a very short distance from a nucleus; beam nuclei are thus greatly accelerated in order to make such repulsion insignificant compared to the velocity of the beam nucleus.[376] The energy applied to the beam nuclei to accelerate them can cause them to reach speeds as high as one-tenth of the speed of light. However, if too much energy is applied, the beam nucleus can fall apart.[376] Coming close enough alone is not enough for two nuclei to fuse: when two nuclei approach each other, they usually remain together for about 10−20 seconds and then part ways (not necessarily in the same composition as before the reaction) rather than form a single nucleus.[376][377] This happens because during the attempted formation of a single nucleus, electrostatic repulsion tears apart the nucleus that is being formed.[376] Each pair of a target and a beam is characterized by its cross section—the probability that fusion will occur if two nuclei approach one another expressed in terms of the transverse area that the incident particle must hit in order for the fusion to occur.[k] This fusion may occur as a result of the quantum effect in which nuclei can tunnel through electrostatic repulsion. If the two nuclei can stay close past that phase, multiple nuclear interactions result in redistribution of energy and an energy equilibrium.[376]
The resulting merger is an excited state[380]—termed a compound nucleus—and thus it is very unstable.[376] To reach a more stable state, the temporary merger may fission without formation of a more stable nucleus.[381] Alternatively, the compound nucleus may eject a few neutrons, which would carry away the excitation energy; if the latter is not sufficient for a neutron expulsion, the merger would produce a gamma ray. This happens in about 10−16 seconds after the initial nuclear collision and results in creation of a more stable nucleus.[381] The definition by the IUPAC/IUPAP Joint Working Party (JWP) states that a chemical element can only be recognized as discovered if a nucleus of it has not decayed within 10−14 seconds. This value was chosen as an estimate of how long it takes a nucleus to acquire electrons and thus display its chemical properties.[382][l] Decay and detection[edit]The beam passes through the target and reaches the next chamber, the separator; if a new nucleus is produced, it is carried with this beam.[384] In the separator, the newly produced nucleus is separated from other nuclides (that of the original beam and any other reaction products)[m] and transferred to a surface-barrier detector, which stops the nucleus. The exact location of the upcoming impact on the detector is marked; also marked are its energy and the time of the arrival.[384] The transfer takes about 10−6 seconds; in order to be detected, the nucleus must survive this long.[387] The nucleus is recorded again once its decay is registered, and the location, the energy, and the time of the decay are measured.[384] Stability of a nucleus is provided by the strong interaction. However, its range is very short; as nuclei become larger, its influence on the outermost nucleons (protons and neutrons) weakens. At the same time, the nucleus is torn apart by electrostatic repulsion between protons, and its range is not limited.[388] Total binding energy provided by the strong interaction increases linearly with the number of nucleons, whereas electrostatic repulsion increases with the square of the atomic number, i.e. the latter grows faster and becomes increasingly important for heavy and superheavy nuclei.[389][390] Superheavy nuclei are thus theoretically predicted[391] and have so far been observed[392] to predominantly decay via decay modes that are caused by such repulsion: alpha decay and spontaneous fission.[n] Almost all alpha emitters have over 210 nucleons,[394] and the lightest nuclide primarily undergoing spontaneous fission has 238.[395] In both decay modes, nuclei are inhibited from decaying by corresponding energy barriers for each mode, but they can be tunneled through.[389][390] ![]() Alpha particles are commonly produced in radioactive decays because the mass of an alpha particle per nucleon is small enough to leave some energy for the alpha particle to be used as kinetic energy to leave the nucleus.[397] Spontaneous fission is caused by electrostatic repulsion tearing the nucleus apart and produces various nuclei in different instances of identical nuclei fissioning.[390] As the atomic number increases, spontaneous fission rapidly becomes more important: spontaneous fission partial half-lives decrease by 23 orders of magnitude from uranium (element 92) to nobelium (element 102),[398] and by 30 orders of magnitude from thorium (element 90) to fermium (element 100).[399] The earlier liquid drop model thus suggested that spontaneous fission would occur nearly instantly due to disappearance of the fission barrier for nuclei with about 280 nucleons.[390][400] The later nuclear shell model suggested that nuclei with about 300 nucleons would form an island of stability in which nuclei will be more resistant to spontaneous fission and will primarily undergo alpha decay with longer half-lives.[390][400] Subsequent discoveries suggested that the predicted island might be further than originally anticipated; they also showed that nuclei intermediate between the long-lived actinides and the predicted island are deformed, and gain additional stability from shell effects.[401] Experiments on lighter superheavy nuclei,[402] as well as those closer to the expected island,[398] have shown greater than previously anticipated stability against spontaneous fission, showing the importance of shell effects on nuclei.[o] Alpha decays are registered by the emitted alpha particles, and the decay products are easy to determine before the actual decay; if such a decay or a series of consecutive decays produces a known nucleus, the original product of a reaction can be easily determined.[p] (That all decays within a decay chain were indeed related to each other is established by the location of these decays, which must be in the same place.)[384] The known nucleus can be recognized by the specific characteristics of decay it undergoes such as decay energy (or more specifically, the kinetic energy of the emitted particle).[q] Spontaneous fission, however, produces various nuclei as products, so the original nuclide cannot be determined from its daughters.[r] The information available to physicists aiming to synthesize a superheavy element is thus the information collected at the detectors: location, energy, and time of arrival of a particle to the detector, and those of its decay. The physicists analyze this data and seek to conclude that it was indeed caused by a new element and could not have been caused by a different nuclide than the one claimed. Often, provided data is insufficient for a conclusion that a new element was definitely created and there is no other explanation for the observed effects; errors in interpreting data have been made.[s] |
Synthesis of superheavy nuclei[edit]![]() A superheavy[t] atomic nucleus is created in a nuclear reaction that combines two other nuclei of unequal size[u] into one; roughly, the more unequal the two nuclei in terms of mass, the greater the possibility that the two react.[375] The material made of the heavier nuclei is made into a target, which is then bombarded by the beam of lighter nuclei. Two nuclei can only fuse into one if they approach each other closely enough; normally, nuclei (all positively charged) repel each other due to electrostatic repulsion. The strong interaction can overcome this repulsion but only within a very short distance from a nucleus; beam nuclei are thus greatly accelerated in order to make such repulsion insignificant compared to the velocity of the beam nucleus.[376] The energy applied to the beam nuclei to accelerate them can cause them to reach speeds as high as one-tenth of the speed of light. However, if too much energy is applied, the beam nucleus can fall apart.[376] Coming close enough alone is not enough for two nuclei to fuse: when two nuclei approach each other, they usually remain together for about 10−20 seconds and then part ways (not necessarily in the same composition as before the reaction) rather than form a single nucleus.[376][417] This happens because during the attempted formation of a single nucleus, electrostatic repulsion tears apart the nucleus that is being formed.[376] Each pair of a target and a beam is characterized by its cross section—the probability that fusion will occur if two nuclei approach one another expressed in terms of the transverse area that the incident particle must hit in order for the fusion to occur.[v] This fusion may occur as a result of the quantum effect in which nuclei can tunnel through electrostatic repulsion. If the two nuclei can stay close past that phase, multiple nuclear interactions result in redistribution of energy and an energy equilibrium.[376]
The resulting merger is an excited state[420]—termed a compound nucleus—and thus it is very unstable.[376] To reach a more stable state, the temporary merger may fission without formation of a more stable nucleus.[381] Alternatively, the compound nucleus may eject a few neutrons, which would carry away the excitation energy; if the latter is not sufficient for a neutron expulsion, the merger would produce a gamma ray. This happens in about 10−16 seconds after the initial nuclear collision and results in creation of a more stable nucleus.[381] The definition by the IUPAC/IUPAP Joint Working Party (JWP) states that a chemical element can only be recognized as discovered if a nucleus of it has not decayed within 10−14 seconds. This value was chosen as an estimate of how long it takes a nucleus to acquire electrons and thus display its chemical properties.[421][w] Decay and detection[edit]The beam passes through the target and reaches the next chamber, the separator; if a new nucleus is produced, it is carried with this beam.[384] In the separator, the newly produced nucleus is separated from other nuclides (that of the original beam and any other reaction products)[x] and transferred to a surface-barrier detector, which stops the nucleus. The exact location of the upcoming impact on the detector is marked; also marked are its energy and the time of the arrival.[384] The transfer takes about 10−6 seconds; in order to be detected, the nucleus must survive this long.[387] The nucleus is recorded again once its decay is registered, and the location, the energy, and the time of the decay are measured.[384] Stability of a nucleus is provided by the strong interaction. However, its range is very short; as nuclei become larger, its influence on the outermost nucleons (protons and neutrons) weakens. At the same time, the nucleus is torn apart by electrostatic repulsion between protons, and its range is not limited.[388] Total binding energy provided by the strong interaction increases linearly with the number of nucleons, whereas electrostatic repulsion increases with the square of the atomic number, i.e. the latter grows faster and becomes increasingly important for heavy and superheavy nuclei.[389][390] Superheavy nuclei are thus theoretically predicted[422] and have so far been observed[392] to predominantly decay via decay modes that are caused by such repulsion: alpha decay and spontaneous fission.[y] Almost all alpha emitters have over 210 nucleons,[394] and the lightest nuclide primarily undergoing spontaneous fission has 238.[395] In both decay modes, nuclei are inhibited from decaying by corresponding energy barriers for each mode, but they can be tunneled through.[389][390] ![]() Alpha particles are commonly produced in radioactive decays because the mass of an alpha particle per nucleon is small enough to leave some energy for the alpha particle to be used as kinetic energy to leave the nucleus.[397] Spontaneous fission is caused by electrostatic repulsion tearing the nucleus apart and produces various nuclei in different instances of identical nuclei fissioning.[390] As the atomic number increases, spontaneous fission rapidly becomes more important: spontaneous fission partial half-lives decrease by 23 orders of magnitude from uranium (element 92) to nobelium (element 102),[398] and by 30 orders of magnitude from thorium (element 90) to fermium (element 100).[423] The earlier liquid drop model thus suggested that spontaneous fission would occur nearly instantly due to disappearance of the fission barrier for nuclei with about 280 nucleons.[390][400] The later nuclear shell model suggested that nuclei with about 300 nucleons would form an island of stability in which nuclei will be more resistant to spontaneous fission and will primarily undergo alpha decay with longer half-lives.[390][400] Subsequent discoveries suggested that the predicted island might be further than originally anticipated; they also showed that nuclei intermediate between the long-lived actinides and the predicted island are deformed, and gain additional stability from shell effects.[424] Experiments on lighter superheavy nuclei,[425] as well as those closer to the expected island,[398] have shown greater than previously anticipated stability against spontaneous fission, showing the importance of shell effects on nuclei.[z] Alpha decays are registered by the emitted alpha particles, and the decay products are easy to determine before the actual decay; if such a decay or a series of consecutive decays produces a known nucleus, the original product of a reaction can be easily determined.[aa] (That all decays within a decay chain were indeed related to each other is established by the location of these decays, which must be in the same place.)[384] The known nucleus can be recognized by the specific characteristics of decay it undergoes such as decay energy (or more specifically, the kinetic energy of the emitted particle).[ab] Spontaneous fission, however, produces various nuclei as products, so the original nuclide cannot be determined from its daughters.[ac] The information available to physicists aiming to synthesize a superheavy element is thus the information collected at the detectors: location, energy, and time of arrival of a particle to the detector, and those of its decay. The physicists analyze this data and seek to conclude that it was indeed caused by a new element and could not have been caused by a different nuclide than the one claimed. Often, provided data is insufficient for a conclusion that a new element was definitely created and there is no other explanation for the observed effects; errors in interpreting data have been made.[ad] |
From Sun article:
{{#invoke:Excerpt|main}} | {{#invoke:Excerpt/sandbox|main}} |
---|---|
![]() Within 10 light-years of the Sun there are relatively few stars, the closest being the triple star system Alpha Centauri, which is about 4.4 light-years away and may be in the Local Bubble's G-Cloud.[429] Alpha Centauri A and B are a closely tied pair of Sun-like stars, whereas the closest star to the Sun, the small red dwarf Proxima Centauri, orbits the pair at a distance of 0.2 light-years. In 2016, a potentially habitable exoplanet was found to be orbiting Proxima Centauri, called Proxima Centauri b, the closest confirmed exoplanet to the Sun.[430] The Solar System is surrounded by the Local Interstellar Cloud, although it is not clear if it is embedded in the Local Interstellar Cloud or if it lies just outside the cloud's edge.[431] Multiple other interstellar clouds exist in the region within 300 light-years of the Sun, known as the Local Bubble.[431] The latter feature is an hourglass-shaped cavity or superbubble in the interstellar medium roughly 300 light-years across. The bubble is suffused with high-temperature plasma, suggesting that it may be the product of several recent supernovae.[432] The Local Bubble is a small superbubble compared to the neighboring wider Radcliffe Wave and Split linear structures (formerly Gould Belt), each of which are some thousands of light-years in length.[433] All these structures are part of the Orion Arm, which contains most of the stars in the Milky Way that are visible to the unaided eye.[434] Groups of stars form together in star clusters, before dissolving into co-moving associations. A prominent grouping that is visible to the naked eye is the Ursa Major moving group, which is around 80 light-years away within the Local Bubble. The nearest star cluster is Hyades, which lies at the edge of the Local Bubble. The closest star-forming regions are the Corona Australis Molecular Cloud, the Rho Ophiuchi cloud complex and the Taurus molecular cloud; the latter lies just beyond the Local Bubble and is part of the Radcliffe wave.[435] Stellar flybys that pass within 0.8 light-years of the Sun occur roughly once every 100,000 years. The closest well-measured approach was Scholz's Star, which approached to ~50,000 AU of the Sun some ~70 thousands years ago, likely passing through the outer Oort cloud.[436] There is a 1% chance every billion years that a star will pass within 100 AU of the Sun, potentially disrupting the Solar System.[437] |
![]() Within 10 light-years of the Sun there are relatively few stars, the closest being the triple star system Alpha Centauri, which is about 4.4 light-years away and may be in the Local Bubble's G-Cloud.[439] Alpha Centauri A and B are a closely tied pair of Sun-like stars, whereas the closest star to the Sun, the small red dwarf Proxima Centauri, orbits the pair at a distance of 0.2 light-years. In 2016, a potentially habitable exoplanet was found to be orbiting Proxima Centauri, called Proxima Centauri b, the closest confirmed exoplanet to the Sun.[430] The Solar System is surrounded by the Local Interstellar Cloud, although it is not clear if it is embedded in the Local Interstellar Cloud or if it lies just outside the cloud's edge.[431] Multiple other interstellar clouds exist in the region within 300 light-years of the Sun, known as the Local Bubble.[431] The latter feature is an hourglass-shaped cavity or superbubble in the interstellar medium roughly 300 light-years across. The bubble is suffused with high-temperature plasma, suggesting that it may be the product of several recent supernovae.[440] The Local Bubble is a small superbubble compared to the neighboring wider Radcliffe Wave and Split linear structures (formerly Gould Belt), each of which are some thousands of light-years in length.[433] All these structures are part of the Orion Arm, which contains most of the stars in the Milky Way that are visible to the unaided eye.[441] Groups of stars form together in star clusters, before dissolving into co-moving associations. A prominent grouping that is visible to the naked eye is the Ursa Major moving group, which is around 80 light-years away within the Local Bubble. The nearest star cluster is Hyades, which lies at the edge of the Local Bubble. The closest star-forming regions are the Corona Australis Molecular Cloud, the Rho Ophiuchi cloud complex and the Taurus molecular cloud; the latter lies just beyond the Local Bubble and is part of the Radcliffe wave.[442] Stellar flybys that pass within 0.8 light-years of the Sun occur roughly once every 100,000 years. The closest well-measured approach was Scholz's Star, which approached to ~50,000 AU of the Sun some ~70 thousands years ago, likely passing through the outer Oort cloud.[443] There is a 1% chance every billion years that a star will pass within 100 AU of the Sun, potentially disrupting the Solar System.[437] |
References
[edit]Refs and notes
|
---|
Notes[edit]
|