question
stringlengths
3
301
answer
stringlengths
9
26.1k
context
sequence
how did the anime-craze start and evolve in america?
The answer is very multifaceted. Back when anime started to trickle into american television for the first time, one of the things it did was fulfill the rather common need for youth to find media that is (generally) outside the realm of their parent's enjoyments/understanding. In short, kids were attracted to it. Combine that with the fact that many anime shows were relatively cheap to produce. Think Pokemon; most of it is one drawn frame with mouths moving. You suddenly get a marketplace where an entire generation is fascinated with an inexpensive to produce cartoon that is geared towards toy sales and boom. You've suddenly given a ton of networks a multi-million dollar excuse to start offering anime more regularly. There's way more of course. Some of it is culture shock. Anime was more often risking things that American cartoons were not. (Akira.) This created an attraction to more provocative animes as well that capture an entirely different audience than kids, which is a big factor in why you see such a range of people enjoying it. America also made the mistake of not regularly developing cartoons for adults (because how could adults like cartoons?) A lot of anime with mature themes was featured in college "Anime Clubs" which sparked a whole new subset of anime fans.
[ "Anime culture in the United States began as a niche community that had a grassroots foundation built by groups of fans on the local level. Some of the earliest televised anime to air in the United States were \"Astro Boy\", \"Speed Racer\", and \"Gigantor\", which gained popularity with many American audiences during the late 1960s. Anime shows that aired in the United States up until the 1980s were usually heavily altered and localized, such as \"Science Ninja Team Gatchaman\" becoming \"Battle of the Planets\" in the 1970s, and the mecha show \"Macross\" becoming \"Robotech\" in the 1980s. Takara's \"Diaclone\" and \"Microman\" mecha toylines also became the basis for the \"Transformers\" franchise in the 1980s.\n", "Through the last two decades the introduction of anime into American mainstream culture has furthered its popularity. Such famous titles as \"Sailor Moon\", \"Dragon Ball Z\", and most importantly \"Pokémon\" have influenced anime's appeal to young Americans. \"Anime already makes up an estimated 60% of all broadcast animation across the world.\" \"ICv2 estimates the size of the North American anime market at $275-300 million\" (in retail dollars).\" To compete with its Japanese competitors, many production companies in the US have adjusted their style to one inspired by anime to hold onto their viewers. An important example that has sparked much controversy in the animation world would be \"\". A popular show on Nickelodeon, the characters have a distinct anime style, even the expressions and mannerism drawn evoke that of anime style. \"Teen Titans\" on Cartoon Network is yet another example of anime's influence on cartoons, as well as a popular comic strip turned cartoon called \"The Boondocks\". In recent years there has also been development of Anime developed in the United States in collaboration with Japan, examples would be the Netflix original series Neo Yokio and Castlevania (TV series) based off the video game series.\n", "Following the launch of Toonami on Cartoon Network and later Adult Swim, anime saw a giant rise in the North American market. Kid-friendly anime such as \"Pokémon\", \"Yu-Gi-Oh!\", \"Digimon\", \"Doraemon\", \"Bakugan\", \"Beyblade\", and the 4Kids Entertainment adaptation of \"One Piece\" have all received varying levels of success. This era also saw the rise of anime-influenced animation, most notably \"\" and its sequel \"The Legend of Korra\", \"Ben 10\", \"Chaotic\", \"Samurai Jack\", \"The Boondocks\", \"RWBY\" and \"Teen Titans\".\n", "The beginning of 1980 saw the introduction of Japanese anime series into the American culture. In the 1990s, Japanese animation slowly gained popularity in America. Media companies such as Viz and Mixx began publishing and releasing animation into the American market. The 1988 film \"Akira\" is largely credited with popularizing anime in the Western world during the early 1990s, before anime was further popularized by television shows such \"Pokémon\" and \"Dragon Ball\" in the late 1990s. The growth of the Internet later provided Western audiences an easy way to access Japanese content. This is especially the case with net services such as Netflix and Crunchyroll. As a direct result, various interests surrounding Japan has increased.\n", "According to Mike Tatsugawa, the founder and CEO of the Society for the Promotion of Japanese Animation, the first milestone for anime in the U.S. was in the 1980s with the advent of the Internet. With the Internet, fans were able to more easily communicate with each other and thus better able to exchange fan-subtitled tapes and higher quality versions of anime. Some experts, such as Susan Napier, a Professor of Japanese Language and Literature, say that \"Akira\" marked the first milestone. However, most experts agree that the next milestone was in 1992 when U.S. Renditions, a film importer, released the first English-subtitled anime videotape that year, entitled \"Gunbuster\". According to Tatsugawa, the success of \"Gunbuster\" triggered a flurry of releases.\n", "Anime of the 2000s would include \"Bleach\", \"One Piece\", \"\", \"Sonic X\", \"Tokyo Mew Mew\", \"Ojamajo Doremi\", \"Gurren Lagann\", and \"Kodai Ōja Kyōryū Kingu\". Most of these anime shows mentioned here would also go on to relative success in North America and Europe in the 2000s. For example, \"Tokyo Mew Mew\" became \"Mew Mew Power\" while \"Ojamajo Doremi\" became \"Magical Doremi\".\n", "The 1990s, was the period in which anime reached mainstream popularity in the U.S. market and the terms \"anime\" and \"manga\" became commonly well known (ultimately replacing the former majorly known term \"Japanimation\"). Companies such as FUNimation Productions, Bandai Entertainment, 4Kids Entertainment, Central Park Media, Media Blasters, Saban Entertainment, Viz Video, Pioneer LDC and ADV Films began licensing anime in the United States.\n" ]
In The Good, The Bad and The Ugly, a character mislabels a union army because of dust on their uniform making him believe they are confederates. Did this ever happen in real life?
Misidentification wasn't uncommon, especially during the early days of the war when many units had different-colored uniforms. Notably, during the First Battle of Bull Run (July 1861), one of the turning points came when the 33rd Virginia (Confederate) regiment was mistaken for a Union unit. They were able to get close enough to decimate an infantry regiment and capture several artillery pieces, which (along with a charge by Stonewall Jackson and the arrival of some Confederate reserves) started the general Union retreat.
[ "In October 1863, Clem was captured in Georgia by Confederate cavalrymen while detailed as a train guard. The Confederates confiscated his U.S. uniform which reportedly upset him terribly—including his cap which had three bullet holes in it. He was included in a prisoner exchange a short time later, but the Confederate newspapers used his age and celebrity status for propaganda purposes, to show \"what sore straits the Yankees are driven, when they have to send their babies out to fight us.\" After participating with the Army of the Cumberland in many other battles, serving as a mounted orderly, he was discharged in September 1864. Clem was wounded in combat twice during the war.\n", "As the war progressed, the image began to shift from the \"ragged rebel\" look to a well-uniformed Army in the Eastern and Western theaters. In the last 12 months of fighting, these Confederate forces were well-uniformed, the best they had ever appeared in terms of consistency, wearing clothing made of imported blue-gray cloth, either manufactured locally or bought ready-made under contract from British manufacturers, such as Peter Tait of Limerick, Ireland who became a major supplier of uniforms for the Confederacy.\n", "The Grand Army of the Republic, the United Confederate Veterans, Sons of Confederate Veterans (SCV), and the United Daughters of the Confederacy (UDC), among others, evoked sad memories of Civil War prisoners portraying Wirz either a villain, or a martyr-hero, thus further contributing to the disputation. From 1899 to 1916, sixteen states erected monuments dedicated to the Camp Sumter's prisoners. In response, the United Daughters of the Confederacy initiated a construction of a monument honoring Henry Wirz in Andersonville, Georgia. Every year the UDC and SCV hold a memorial service at the monument. Until recently, SCV annually marched to Wirz's memorial in Andersonville along with supporters of a congressional pardon for him. The SCV posthumously awarded Wirz their Confederate Medal of Honor, created in 1977.\n", "During the American Civil War of 1861–1865, McGehee supported the Union. However, he also sold clothes made in his textile factory to the Confederate States Army. The mansion at his Bowling Green Plantation was burned down by United States Colored Troops in 1864. His wife wrote about the incident in \"Army & Navy Herald,\" a Confederate newspaper.\n", "Apparently Confederate pickets failed to respond to the presence of the Union raiding party because they assumed that the Union soldiers were a party of Wright's Cavalry that recently passed through the area were returning for some reason. It was often difficult to distinguish Union soldiers from Confederates during this period of the war. Some Confederates in Arkansas were wearing clothing manufactured at Confederate depots made out of English army cloth. Confederates had received a supply of British cloth through the Union blockade to Confederate depots in Texas, and a supply of uniforms made from this cloth had been forwarded to Arkansas. Jackets made from this material were hard to distinguish from Federal infantry and mounted service jackets. To add to the confusion, some Confederate units, such as Confederate General Joe Shelby's Missouri Cavalry Brigade had become known for dressing in captured Union uniforms.\n", "The Confederate victors of Marks Mills robbed the survivors of every valuable item they possessed, including hats, boots, watches, money and in some cases, the clothes on their backs. Overall, they were very roughly handled by their captors, according to Captain Swiggett. They were force-marched to the rebel prison at Camp Ford in Tyler, Texas, where dozens of them perished from disease over the next 12 months. Conditions at Camp Ford were primitive. Although a good spring provided cool clean water, and while the Confederate guards slaughtered cattle to supply the prisoners fresh beef regularly, the prisoners had no shelter from the sun or rain except improvised huts or blankets, and as the numbers of prisoners rose, the sanitary conditions declined precipitously, leading to many deaths from exposure, chronic diarrhea and disease. A number of the 36th Iowa's officers escaped. Captain Swiggett twice escaped but was recaptured on both occasions and was rewarded for his bad behavior by being one of the last prisoners exchanged.\n", "The Roswell Mills are best known for their role in producing supplies for the Confederacy during the Civil War. They made \"Roswell Gray\" fabric to be sewed into Confederate military uniforms. Because it was of great importance to the South’s military supply chain, General Gerrard, a Union official working under the purview of General Sherman, seized the mill on July 5, 1864. Confederate forces burned down the bridge that spanned Vickery Creek before he could get to it.\n" ]
why does eating a lot of rich food cause a headache ?
It could easily be the ingestion of large amounts of salt and/or sugar (which are often heavily used in rich foods) causing dehydration. It could also be elevated levels of insulin being released into the body as the food is processed, or your body rejecting an ingredient in some way. Lots of intolerances can be very mild and go easily unnoticed. The simple answer is, if you're going to eat rich foods, drink a good amount of plain water with, or after eating, your food. It's a good remedy for most of the side effects I mentioned. Any large meal can make you feel a bit rough, so taking your time eating can be a big help as well.
[ "Many physicians also recommend changes in diet to treat chronic headaches. Many chronic headache sufferers fail to recognize foods or beverages as headache factors, because the consumption may not consistently cause headaches or the headaches may be delayed. Many of the chemicals in certain foods can cause chronic headaches, including caffeine, nitrites, nitrates, tyramine, and alcohols. Some of the foods and beverages that chronic headache sufferers are advised to avoid include caffeinated beverages, chocolate, processed meats, cheese and fermented dairy products, fresh yeast-risen baked goods, nuts, and alcohol as well as certain fruits and vegetables. Additionally, people may have differing dietary triggers on an individualized basis, because not all foods affect people the same way. Different medical professionals suggest different ways of testing or changing diets. Some may suggest eliminating a few of the potentially headache-causing foods at a time for a short period of time, while others suggest removing all the threatening foods from a person's diet and slowly adding a couple back at a time. Yet, others may not suggest diet modification at all. The treatment of chronic headaches through changes in diet is based on personal opinion, and, therefore, controversial.\n", "Gastrointestinal disorders may cause headaches, including Helicobacter pylori infection, celiac disease, non-celiac gluten sensitivity, irritable bowel syndrome, inflammatory bowel disease, gastroparesis, and hepatobiliary disorders. The treatment of the gastrointestinal disorders may lead to a remission or improvement of headaches.\n", "Headaches can be attributed to many different substances. Some of these include alcohol, NO, carbon monoxide poisoning, cocaine, caffeine and monosodium glutamate. Chronic use of certain medications used to treat headaches can also start causing headaches, known as medication overuse headaches. Headaches may also be a symptom of medication withdrawal.\n", "Like other types of pain, headaches can serve as warning signals of more serious disorders. This is particularly true for headaches caused by inflammation, including those related to meningitis as well as those resulting from diseases of the sinuses, spine, neck, ears, and teeth.\n", "A number of different causes contribute to this class of headache. Several common chemicals may be the culprit. Nitrite compounds dilate blood vessels, causing dull and pounding headaches with repeat exposure. Nitrite is found in dynamite, heart medicine and it is a chemical used to preserve meat (ergo these being known as \"nitrite\" or \"hot dog\" headaches). Eating foods prepared with monosodium glutamate (MSG) may thus result in headache. Acetaldehyde from alcohol may also cause a headache either acutely or after a number of hours (hangover).\n", "A headache-prone CNS may have resulted from interactions with other organisms in two ways. The first possibility is that migraine offers an advantage to the organism in fighting infection by increasing blood flow to the brain. The second possibility is that certain pathogens evolved to cause headache as a way of speeding their transmission to other organisms. Finally, migraine may benefit neither the host nor the pathogen, but may simply be the result of certain infections. This last explanation is concordant with the apparent negative impact of migraine on human fitness.\n", "Medication overuse headache (MOH), also known as rebound headache usually occurs when analgesics are taken frequently to relieve headaches. Rebound headaches frequently occur daily, can be very painful and are a common cause of chronic daily headache. They typically occur in patients with an underlying headache disorder such as migraine or tension-type headache that \"transforms\" over time from an episodic condition to chronic daily headache due to excessive intake of acute headache relief medications. \n" ]
How were WWI era reparation payments handled during the post-WWII separation of Germany?
Fortunately for post-war Germany they didn't have to pay any reparations from the First World War as the payment had ceased in 1932. Reparations had always been a thorny political issue in Germany, even from the signing of the Treaty of Versailles. In the Treaty itself, Germany was not given a definite figure to repay; rather it simply agreed to pay an indefinite amount that would be decided by a yet-to-be established Reparations Committee. In the meantime, Germany had to pay for the upkeep of the occupying armies with raw materials and food. This is referred to as payment in kind. Eventually the figure was settled as 132bn Marks, or £6.6bn. Given the state of the German economy after the war, and the loss of significant coal deposits in Silesia and Alsace-Lorraine, Germany was unable to keep up with her payments either in raw materials or in money. The result of this was that Belgium and France occupied the Ruhr region of Germany and confiscated all production in payment. German workers laid down their tools in a scheme of passive resistance. However, it was during this time that hyperinflation brought on by war debts, a failing economy and decreased production reached its high point. Germany needed serious help, and passive resistance in the Ruhr wasn't really winning them any friends. The practice was ended and the so-called Dawes Plan was drawn up, whereby Germany would recieve foreign investment and loans, the French and Belgians would withdraw from the Ruhr, the schedule of repayments would be redrawn and a new currency, the *Rentenmark*, would be issued to replace the hyperinflated Goldmark. However, many viewed the Dawes Plan as only a stopgap measure and pressured for a new deal, which was drawn up in 1929. The Young Plan, as it was called, set a new schedule and reduced figure for reparations. Crucially, it set a fixed end date for payments (1988), which was the first time that such a thing had happened. It also required the French and British to withdraw from the Rheinland ahead of schedule, and provided additional loans to Germany. While economically sound, the plan drew criticism from the right wing who viewed it as selling out to international financiers and making Germany dependent on foreign aid. Last but not least, following the deepening financial crisis, a moratorium had been placed on reparation repayments in 1931. In 1932, a conference was held in Lausanne which agreed that Germany should no longer have to make payments given their dire financial situation. By this point Germany had only paid roughly one sixth of what they were originally required to do so. Therefore, German ended and indeed began the Second World War without any obligation to pay reparations. However, following the end of the Second World War, various debts and repayments mostly fell on West Germany. West Germany voluntarily signed agreements with Israel, Greece and Yugoslavia amongst others to voluntarily pay reparations for the damage done to their citizens and economies. Additionally, West Germany agreed to pay the reparations which had not been previously paid from WWI, in addition to other debts accrued. It was agreed that West Germany should not bear the entire weight of debt, and a portion was only to be repaid upon reunification, which did not occur until 1990. However, reunified Germany continued to pay this debt, with the last payment being made in 2010. In summary, Germany ceased to make WWI reparations payments in 1932 but the West German government voluntarily agreed to continue payments in 1953, with the condition that part of the debt only be paid upon reunification. For further reading: Leonard Gomes, *German Reparations, 1919-1932: A Historical Survey* (2010)
[ "World War I reparations owed by Germany were stated in gold marks in 1921, 1929 and 1931; this was the victorious Allies' response to their fear that vanquished Germany might try to pay off the obligation in paper marks. The actual amount of reparations that Germany was obliged to pay out was not the 132 billion marks cited in the London Schedule of 1921 but rather the 50 billion marks stipulated in the A and B Bonds. The actual total payout from 1920 to 1931 (when payments were suspended indefinitely) was 20 billion German gold marks, worth about or . Most of that money came from loans from New York bankers.\n", "Following the Nazi seizure of power in 1933, payments of reparations were officially abandoned. West Germany after World War II did not resume payment of reparations as such, but did resume the payment of debt that Germany had acquired in the inter-war period to finance its reparation payments, paying off the principal on those debts by 1980. The interest on those debts was paid off on 3 October 2010, the 20th anniversary of German reunification.\n", "At the conclusion of World War I, the Allied and Associate Powers included in the Treaty of Versailles a plan for reparations to be paid by Germany; 20 billion gold marks was to be paid while the final figure was decided. In 1921, the London Schedule of Payments established the German reparation figure at 132 billion gold marks (separated into various classes, of which only 50 billion gold marks was required to be paid). German industrialists in the Ruhr Valley, who had lost factories in Lorraine which went back to France after the war, demanded hundreds of millions of marks compensation from the German government. Despite its obligations under the Versailles Treaty, the German government paid the Ruhr Valley industrialists, which contributed significantly to the hyperinflation that followed. For the first five years after the war, coal was scarce in Europe and France sought coal exports from Germany for its steel industry. The Germans needed coal for home heating and for domestic steel production, having lost the steel plants of Lorraine to the French.\n", "To help make reparations payments, Germany took out various loans during the 1920s. In 1933, following the cancellation of reparations, the new German Chancellor Adolf Hitler cancelled all payments. In June 1953, an agreement was reached on this existing debt with West Germany. Germany agreed to repay 50 per cent of the loan amounts that had been defaulted on in the 1920s, but deferred some of the debt until West and East Germany were unified. In 1995, following reunification, Germany began making the final payments towards the loans. A final installment of was made on 3 October 2010, settling German loan debts in regard to reparations.\n", "World War I reparations were war reparations imposed during the Paris Peace Conference upon the Central Powers following their defeat in the First World War by the Allied and Associate Powers. Each of the defeated powers was required to make payments in either cash or kind. Because of the financial situation Austria, Hungary, and Turkey found themselves in after the war, few to no reparations were paid and the requirements for reparations were cancelled. Bulgaria, having paid only a fraction of what was required, saw its reparation figure reduced and then cancelled. Historians have recognized the German requirement to pay reparations as the \"chief battleground of the post-war era\" and \"the focus of the power struggle between France and Germany over whether the Versailles Treaty was to be enforced or revised\".\n", "The German reparations decided by the Treaty of Versailles in 1919 brought in large amounts of money which served principally to repay war loans to the United States. Reparations payments ended in 1923. In January of that year, Germany defaulted on its payments and the French president, Raymond Poincaré, invoked a clause of the Versailles Treaty and sent troops to occupy the Ruhr valley in the hope of enforcing payment. Germany responded by flooding the area with inflated money, ruining its currency and denying France any hope of full reparations. Poincaré accepted an agreement mediated by the United States in which it received smaller payments, but Poincaré's government fell soon afterward.\n", "BULLET::::- Germany accepted responsibility for the damages and losses caused by the war and would make reparation payments to the Allies; the Reparations Committee in 1921 would set total reparation payments to gold marks or $5 billion in gold.\n" ]
bone density after elongated periods of time in space.
Gravity constantly pulls on your body and your bones have to be strong enough to hold your weight, so your body streghthens them to withstand that weight. In space, you have no weight, so your body redirects your resources elsewhere. Your bones don't need to be dense, because of that, so the body weakens them so it can use its energy and materials for other things that are more important.
[ "Spaceflight osteopenia is the bone loss associated with human spaceflight. After a 3–4 month trip into space, it takes about 2–3 years to regain lost bone density. New techniques are being developed to help astronauts recover faster. Research in the following areas holds the potential to aid the process of growing new bone:\n", "The apex of the conical meniscus cannot become infinitely small. A singularity develops when the hydrodynamic relaxation time formula_16 becomes larger than the charge relaxation time formula_17. The undefined symbols stand for characteristic length formula_18 and vacuum permittivity formula_19. Due to intrinsic varicose instability, the charged liquid jet ejected through the cone apex breaks into small charged droplets, which are radially dispersed by the space-charge.\n", "Bone remodels in response to stress in order to maintain constant strain energy per bone mass throughout. To do this, it grows more dense in areas experiencing high stress, while resorbing density in areas experiencing low stress. On Mars, where gravity is about one-third that of earth, the gravitational forces acting on astronauts' bodies would be much lower, causing bones to decrease in mass and density.\n", "The muscle volume can decrease up to 20% over a six-month mission, and the bone density can decrease at a rate of approximately 1.4% at the hip in a month's time. A study done by Fitts and Trappe examined the effects of prolonged space flight (defined as approximately 180 days) on human skeletal muscle using muscle biopsies. Prolonged weightlessness was shown to cause significant loss in the mass, force, and power production in the soleus and gastrocnemius muscles. Many countermeasures to these effects exist, but thus far they are not sufficient to compensate for the detrimental effects of space travel and astronauts need extensive rehabilitation upon their return to Earth.\n", "Spaceflight osteopenia refers to the characteristic bone loss that occurs during spaceflight. Astronauts lose an average of more than 1% bone mass per month spent in space. There is concern that during long duration flights, excessive bone loss and the associated increase in serum calcium ion levels will interfere with execution of mission tasks and result in irreversible skeletal damage.\n", "Studies have shown that once a human reaches adulthood, bone density steadily decreases with age, to which loss of trabecular bone mass is a partial contributor. Loss of bone mass is defined by the World Health Organization as osteopenia if bone mineral density (BMD) is one standard deviation below mean BMD in young adults, and is defined as osteoporosis if it is more than 2.5 standard deviations below the mean. A low bone density greatly increases risk for stress fracture, which can occur without warning in those at risk. The resulting low-impact fractures from osteoporosis most commonly occur in the upper femur, which consists of 25-50% trabecular bone depending on the region, in the vertebrae which are about 90% trabecular, or in the wrist.\n", "It is believed that the reduction of human bone density in prehistoric times reduced the agility and dexterity of human movement. Shifting from hunting to agriculture has caused human bone density to reduce significantly.\n" ]
would a person on earth notice the andromeda galaxy merging with it?
Yes, if Andromeda merged with Earth it would get crowded. If Andromeda were to merge with the Milky Way, no, as long as you stayed away from news, internet and large telescopes.
[ "Current measurements suggest the Andromeda Galaxy is approaching us at . In 3 to 4 billion years, there may be an Andromeda–Milky Way collision, depending on the importance of unknown lateral components to the galaxies' relative motion. If they collide, the chance of individual stars colliding with each other is extremely low, but instead the two galaxies will merge to form a single elliptical galaxy or perhaps a large disk galaxy over the course of about a billion years.\n", "The future of the Andromeda and Milky Way galaxies may be interlinked: in about five billion years, the two could potentially begin an Andromeda–Milky Way collision that would spark extensive new star formation.\n", "Over the past 2 billion years, star formation throughout Andromeda's disk is thought to have decreased to the point of near-inactivity. There have been interactions with satellite galaxies like M32, M110, or others that have already been absorbed by Andromeda Galaxy. These interactions have formed structures like Andromeda's Giant Stellar Stream. A galactic merger roughly 100 million years ago is believed to be responsible for a counter-rotating disk of gas found in the center of Andromeda as well as the presence there of a relatively young (100 million years old) stellar population.\n", "The Andromeda Galaxy is currently approximately 2.5 million light years away from our galaxy, the Milky Way Galaxy, and they are moving towards each other at approximately 300 kilometers (186 miles) per second. Approximately five billion years from now, or 19 billion years after the Big Bang, the Milky Way and the Andromeda Galaxy will collide with one another and merge into one large galaxy based on current evidence. Up until 2012, there was no way to know whether the possible collision was definitely going to happen or not. In 2012, researchers came to the conclusion that the collision is definite after using the Hubble Space Telescope between 2002 and 2010 to track the motion of Andromeda. This results in the formation of \"Milkomeda\" (also known as \"Milkdromeda\").\n", "The estimated distance of the Andromeda Galaxy from our own was doubled in 1953 when it was discovered that there is another, dimmer type of Cepheid. In the 1990s, measurements of both standard red giants as well as red clump stars from the \"Hipparcos\" satellite measurements were used to calibrate the Cepheid distances.\n", "New images from NASA's Spitzer Space Telescope shed light on the Andromeda Galaxy's violent past. The images show that one of the latter's satellite galaxies, M32, blasted through one of the Andromeda Galaxy's spiral arms a few million years ago. Infrared pictures of the galaxy's two spiral arms demonstrate that they and the prominent star-forming ring are separate structures. The images also show a hole where the rings seem to split into arcs. This hole is where astronomers believe M32 punched through Andromeda's galactic disk.\n", "After this, most of the Galactic Empire had left our galaxy, leaving only a scattered few. This departure from the galaxy, leaving it to Vanamonde, was because contact had been made with a \"very strange and very great\" intelligent, extraterrestrial species, which called them urgently to the other side of the universe.\n" ]
why do some large retail chains (e.g. blockbusters, border's books, circuit city) grow and expand successfully til they are in every city but then go out of business suddenly without instead just scaling back down as needed?
Because the very nature of the business they provide becomes obsolete and scaling back does not provide a way toward regaining profitability. For example, let's take a retail chain that is hugely successful and everywhere today - Starbucks. They provide coffee to order ready to consume to the public. Let's now take a new business - hypothetically, let's call it eCoffeeMax - finds a new way to deliver coffee to consumers in a better, cheaper way (more selection, higher quality, faster). They find a way to deliver your coffee wherever you are using quadcopter drones - at home, at the office, at school, even on your commute - for half the price and their coffee is better in every way. They'll even accept your junk mail as currency to buy coffee. Now all of the sudden, Starbucks has lost all of its competitive advantage when it comes to capturing customers. It tries to cut prices as much as it can but any further and it will lose money. It shuts down the least profitable stores but now every story is unprofitable because of the new competition. That's what happened to every bookstore in America after Amazon showed up. There was no way to compete. Amazon was the place that got you more books than independent or chain brick and mortar bookstores, at better prices, and at the convenience of shopping without having to go to the store. Blockbusters was tied to physical media as a business model when streaming became widely used. Circuit City relied on brick and mortar stores rather than online stores (Best Buy adapted to this better and shifted business to online sales to balance their sales load). Scaling back wasn't an option - it was just hitting an evolutionary dead end in the business world after the nature of the business changed beyond the ability of the company to adapt to it. There are many businesses today that are at risk. For example, Hollywood movies rely a lot on digital visual effects - used to be mostly for flashy space battles and monsters, but now even present-time dramas rely heavily on digital "set extensions" (adding digital extras walking around a busy city street instead of hiring 100 actual people as extras, adding buildings to customize a skyline, cleaning up an environment or modifying an environment). This work used to be performed by somewhat well paid digital artists in the USA in VFX studios in the US. Now, this work is moving further and further away, just like with manufacturing. At first, it was being farmed out to cheap small operations in the US, then cheap small studios in Canada and Europe. Then, India. Now, China is getting in the game. Our manufacturing base disappeared that way. A lot of it has to do with our government's political priorities. Is it more important to keep jobs and a business here or do you call it "Free market" and let corporations pay for your campaign and make sure you vote against labor union laws and laws that help give tax incentives to keep jobs here? We're competing against nations that fund their businesses using their tax moneys - South Korean corporations got big because the governments subsidized them heavily protected them vigorously. It's not a free market economy out there. We don't order our soldiers to go fight using their own privately purchased equipment while other armies are funded by their government. But in business, that's what we do and worse.
[ "A key drawback to supplier convergence is that one of the main concepts of it is to force smaller companies into mergers or out of business by replacing or threatening to replace them with one large company offering different products or services. Wal-Mart and Borders, two of the superstores cited above, have received criticism for forcing local, independent stores out of business by offering convenience and prices that smaller retail stores would not be able to match. For many, this is a concerning trend, as it means local retail outlets will continue to be replaced with large, multinational firms.\n", "One study showed that small, local businesses are better for a local economy than the introduction of new chain stores. By opening up new national level chain stores, the profits of locally owned businesses greatly decrease and many businesses end up failing and having to close. This creates an exponential effect. When one store closes, people lose their jobs, other businesses lose business from the failed business, and so on. In many cases, large firms displace just as many jobs as they create.\n", "Stop & Shop operates a total of 416 stores in Massachusetts, Rhode Island, Connecticut, upstate and metropolitan New York, and New Jersey. Giant boasts 169 stores in Delaware, Maryland, Pennsylvania, Virginia, and the District of Columbia. The combined chain competed with regional chains such as Shaw's, Hannaford, Tops, Big Y, Wegmans, Price Chopper, ShopRite, Pathmark, Waldbaum's, King Kullen, Weis Markets, Acme, and Food Lion as well as national chains such as Aldi, big box stores such as Walmart Supercenters and Super Targets, and warehouse chains such as Sam's Club and Costco.\n", "As the 21st century takes shape, some indications suggest that large retail stores have come under increasing pressure from online sales models and that reductions in store size are evident. Under such competition and other issues such as business debt, there has been a noted business disruption called the retail apocalypse in recent years which several retail businesses, especially in North America, are sharply reducing their number of stores, or going out of business entirely.\n", "Similarly, conglomerates, such as Raheja's , Future Group, Bharti, Godrej, Reliance, and TATA, have over the last decade ventured into large-format retail chains. However, most of the stores opened in large malls and not as independent big box format stores, even though small and medium enterprises (SMEs) still account for the majority of the daily consumer transaction needs. However,the most successful consumer retail chain that took the market and penetrated also to tier 2 and tier 3 cities was D Mart, owned by Avenue Supermarkets Limited.\n", "The converse argument is that large chain stores have grown so big because their products are desirable to large numbers of people and so their arrival in towns provides convenient access to the products that the population might want. It is argued that providing locals with easy access to popular products that they want should be a higher priority than ensuring that people travelling between multiple towns experience variety. Furthermore, because they are wealthy, businesses they are more likely to consume much local services and to employ local people, thus energising the local economy.\n", "E-commerce and the innovation of the online marketplace have been slowly taking over the needs for physical locations of store brands. A prime example of this includes the decline of malls within the United States. The sales of malls within the United States have declined from $87.46 billion in 2005 to $60.65 billion in 2015. The emergence of e-commerce companies like Amazon and Alibaba (both unicorns before they went public) have decreased the need for physical locations to buy consumer goods. Many large corporations have seen this trend for a while and have tried to adapt to the e-commerce trend. Walmart recently bought Jet.com, an American e-commerce company, for $3.3 billion to try and adapt to consumer preferences.\n" ]
What slang did German soldiers use to describe soldiers from various nations during WWII?
/u/LBo87 gave a good answer in an [old thread](_URL_0_). There is /u/Astrogator's [answer](_URL_1_) as well, which adds one more term. See the last bullet point.
[ "One significant source of slang were the prisoner of war camps run by the Japanese, where Diggers sometimes ended up. These were the sources of many particularly strong expressions, such as \"white nip\" for a prisoner who collaborated with the Japanese, and \"japs\", \"nips\", \"jeeps\", \"little yellow men\", and \"little yellow bastards\" for the Japanese themselves. In the camps, \"kippers\" were the British POWs, and \"cheese-eaters\" the Dutch. The urinals were \"pissaphones\" and the stew served to prisoners was \"Danube\", a contraction of the rhyming slang \"Blue Danube\".\n", "One part of the Nazi camp slang was German language terminology for people, things and events in the camps, where many ordinary German words acquired specific meanings and associations. Another was \"ad hoc\" slang based on the various native languages of inmates from different countries.\n", "In Britain, this was additionally popularised as an adjective from the experiences of thousands of U.S., British, and other English-speaking combat personnel, primarily airmen, who were captured in the European Theater of Operations during World War II. These Allied prisoners of war were given ersatz goods such as \"Ersatzkaffee\", an inferior \"Getreidekaffee\" or \"grain coffee\" as a coffee substitute by their German captors.\n", "British soldiers employed a variety of epithets for the Germans. \"Fritz\", a German pet form of Friedrich, was popular in both World War I and World War II, with \"Jerry\", short for \"German\", but also modeled on the English name, favoured in the latter.\n", "BULLET::::- Jerry, Gerry: Rhyming slang (Jerry the German), primarily used in the first and second World Wars by the British, but also other English speakers. Based on the common given nickname Jerry, short for Jeremiah, Gerald, and other similar-sounding names.\n", "There were many other Digger slang words and phrases coined during the Second World War. Two of the most notable are \"wheelbarrow\" for a conscript (because he had to be pushed) and \"doover\", a general name for just about anything at all. Others include \"snarlers\", who were soldiers from the Middle East who were \"SNLR\" (\"Services No Longer Required\") and sent home on \"three P boats\" (troopships that contained \"pox, prisoners, and provosts\").\n", "The German term Prominenten was used in World War II to describe high-profile prisoners from various countries that were imprisoned in for example the POW camp for officers Colditz (Oflag IV-C) or the Nazi concentration camps Sachsenhausen and Dachau for possible use as hostages. These prisoners had among them former statesmen, politicians, political dissidents and priests as well as prisoners of war related, or believed to be related, to allied politicians and royalties. The Colditz POW's were liberated On 14April 1945 by the U.S. Army whereas the prisoners from Sachsenhausen and Dachau where liberated from their SS-guards by in turn by troops from the Wehrmacht and shortly thereafter by units from the 42nd Infantry Division and the 45th Infantry Division after the Transport of concentration camp inmates to Tyrol.\n" ]
how does a phone battery hold a charge? what is flowing from the charging cord, where does it go into, and how is the phone able to use that charge?
Phone batteries use an chemical reaction something like Li - > Li+ + e- Mg+ + e- - > Mg This means the the electron travels from the lithium electrode to the magnesium electrode creating electron flow and therefore energy. In all chemical reactions u can push/reverse the reaction by adding energy to the aystem. If u add electricity to the system the lithium solid will become lithium aqueous (in solution) this allows it to react again and produce power.
[ "The port can charge the battery and power the phone while it is connected to, for example, a hands-free solution in a car. The FastPort became the only way to get external power to the phones. Chargers comes in several varieties, from 12/24 volt DC to use in cars, to 100-250 volt AC to use elsewhere. Some charger-models can only charge the phone (the cable is attached at the middle), in others all the connector pins through to the plug end, thus supporting data/signal transfer while the phone is being charged.\n", "Conductive charging requires a physical connection between the electronic device's battery and the power supply. The need for a metal-to-metal connection between the charger and the device requiring charging is one of the main drawbacks of this method. To accomplish this without the use of physical cords connected to wall outlets, special attachments are made from electronic devices which are fitted with technology that can detect when the device makes connection with the power source, often a charging base. Conduction based wireless accessories may include changeable backs for cellular phones, special sleeves and attachable clips.\n", "BULLET::::- Inconvenience – When a mobile device is connected to a cable, it can be moved around (albeit in a limited range) and operated while charging. In most implementations of inductive charging, the mobile device must be left on a pad to charge, and thus can't be moved around or easily operated while charging. With some standards, charging can be maintained at a distance, but only with nothing present in between the transmitter and receiver.\n", "The charger is a small box, usually powered by a battery. It contains an electronic circuit that steps the battery voltage up to the high voltage needed for charging. The box has a fixture that requires one to press the end of the dosimeter on the charging electrode. Some chargers include a light to illuminate the measurement electrode, so that measurement, logging and recharging can occur with one routine motion.\n", "Some mobile phone companies have used the extra pin \"X\" to enforce the use of their own battery chargers. Verizon Wireless is the first company to require first-party battery chargers, having colluded with Motorola to put an arbitrary 1.4 volts on pin X. This voltage violates the USB-IF standards, as pin X should either be tied to ground or not connected at all. Without this connection, the phone will refuse to charge, displaying \"unauthorized charger\" despite receiving the proper current. The phone will charge while connected to a personal computer only if the PC is running a special device driver.\n", "Conductive power transfer uses a conductor to connect two electronic devices in order to transfer energy. It is therefore not a form of wireless power transfer, which explicitly does not use conductors. In the area of cellular phone chargers, phones are equipped with an attachment such as a sleeve which, when placed on the charging board, transfers energy to the phone's battery.\n", "The charge control consists of a pressure switch built into the cell, which disconnects the charging current when the internal cell pressure rises above a certain limit (usually 200 to 300 psi or 1.4 to 2.1 MPa). This prevents overcharging and damage to the cell.\n" ]
deli meat vs packages sliced meat
Yes. Freshly sliced meat that you get at the grocery store is, well, fresh. Prepackaged meat has to go to through the middle man, who slices and packages it before it's sent to the store. Getting the meat sliced at the deli means it skips the middle man and gets to you, more fresh than the prepackaged stuff.
[ "A meat slicer, also called a slicing machine, deli slicer or simply a slicer, is a tool used in butcher shops and delicatessens to slice meats, sausages, cheeses and other deli products. Older models of meat slicer may be operated by crank, while newer ones generally use an electric motor. While the slicer is traditionally a commercial apparatus, domestic use versions are also marketed.\n", "A steak sandwich is a sandwich prepared with steak that has been broiled, fried, grilled, barbecued or seared using steel grates or gridirons, then served on bread or a roll. Steak sandwiches are sometimes served with toppings of cheese, onions, mushrooms, peppers, tomatoes, and in some instances fried eggs, coleslaw, and french fries.\n", "A \"meat slicer\", also called a \"slicing machine\", \"deli slicer\" or simply a \"slicer\", is a tool used in butcher shops and delicatessens to slice meats and cheeses. The first meat slicer was invented by Wilhelm van Berkel (Wilhelmus Adrianus van Berkel) in Rotterdam in 1898. Older models of meat slicer may be operated by crank, while newer ones generally use an electric motor.\n", "Minced cutlet (котлета рубленая, \"kotleta rublenaya\"), or, since the late 19th century, simply \"cutlet\", is a staple of Russian cuisine. It is similar to a Salisbury steak, with the main difference being pure beef is rarely employed, usually pork or a beef-pork mixture is used. The meat is seasoned with salt and pepper, mixed with finely chopped onion (optionally fried), garlic, and a binder (eggs and breadcrumbs soaked in milk), divided into oval-shaped patties, lightly breaded and shallow-fried in a half-inch of vegetable oil. The transliterated Japanese dish, \"menchi katsu\", is always deep-fried and heavily breaded, being essentially a mincemeat croquette, while the Russian version is always shallow-fried.\n", "NAMP was best known, however, for \"The Meat Buyer's Guide\", intended for butchers and commercial meat purchasers, which was a recognized reference for cutting and grading meat. It was published annually. NAMP maintains a standard numbering system for cuts of meat. \n", "BULLET::::- Price-pack/Bonus packs deal: The packaging offers a consumer a certain percentage more of the product for the same price (for example, 25 percent extra). This is another type of deal “in which customers are offered more of the product for the same price”. For example, a sales company may offer their consumers a bonus pack in which they can receive two products for the price of one. In these scenarios, this bonus pack is framed as a gain because buyers believe that they are obtaining a free product. The purchase of a bonus pack, however, is not always beneficial for the consumer. Sometimes consumers will end up spending money on an item they would not normally buy had it not been in a bonus pack. As a result, items bought in a bonus pack are often wasted and is viewed as a “loss” for the consumer.\n", "In Indian cuisine, a cutlet specifically refers to mashed vegetables (potato, carrot, beans) or cooked meat (mutton, beef, fish or chicken) stuffing that is fried with a batter/covering. The meat itself is cooked with spices - onion, cardamom, cloves, cinnamon, coriander (cilantro), green chillies, lemon and salt. This is then dipped in an egg mix or Corn starch and then in Bread crumbs (also see breaded cutlet), and fried in ghee or vegetable oil. Mostly chicken and mutton cutlets are very popular snacks in the city of Kolkata.\n" ]
what was it about the catholic church that allowed it to become the most prominent religion in the world ?
Quite simply, the Catholic Church happened to gain a foothold in the most powerful countries in the world. Spain, Portugal, France, etc. Those countries then spread their faith through their conquests. This of course came from the apostles of Christ spreading the gospel throughout Europe starting in Greece and Turkey.
[ "Christianity had a significant impact on education and science and medicine as the church created the bases of the Western system of education, and was the sponsor of founding universities in the Western world as the university is generally regarded as an institution that has its origin in the Medieval Christian setting. Many clerics throughout history have made significant contributions to science and Jesuits in particular have made numerous significant contributions to the development of science. The Civilizing influence of Christianity includes social welfare, founding hospitals, economics (as the Protestant work ethic), politics, architecture, literature and family life.\n", "Since the Renaissance era, with western colonialism, Christianity has expanded throughout the world. Today there are more than two billion Christians worldwide, and Christianity has become the world's largest religion.\n", "Christianity had a significant impact on education and science and medicine as the church created the bases of the Western system of education, and was the sponsor of founding universities in the Western world as the university is generally regarded as an institution that has its origin in the Medieval Christian setting. Many clerics throughout history have made significant contributions to science and Jesuits in particular have made numerous significant contributions to the development of science. The cultural influence of Christianity includes social welfare, founding hospitals, economics (as the Protestant work ethic), natural law (which would later influence the creation of international law), politics, architecture, literature, personal hygiene, and family life. Christianity played a role in ending practices common among pagan societies, such as human sacrifice, slavery, infanticide and polygamy.\n", "Christianity had a significant impact on education and science and medicine as the church created the bases of the Western system of education, and was the sponsor of founding universities in the Western world as the university is generally regarded as an institution that has its origin in the Medieval Christian setting. Many clerics throughout history have made significant contributions to science and Jesuits in particular have made numerous significant contributions to the development of science. The cultural influence of Christianity includes social welfare, founding hospitals, economics (as the Protestant work ethic), natural law (which would later influence the creation of international law), politics, architecture, literature, personal hygiene, and family life. Christianity played a role in ending practices common among pagan societies, such as human sacrifice, slavery, infanticide and polygamy. \n", "The early followers of Jesus, including Saints Paul and Peter carried this new theology concerning Jesus throughout the Roman Empire and beyond, sowing the seeds for the development of the Catholic Church, of which Saint Peter is remembered as the first Pope. Catholicism, as we know it, emerged slowly. Christians often faced persecution during these early centuries, particularly for their refusal to join in worshiping the emperors. Nevertheless, carried through the synagogues, merchants and missionaries across the known world, the new internationalist religion quickly grew in size and influence. Its unique appeal was partly the result of its values and ethics.\n", "During the 20th century, the church grew substantially and became an international organization, due in part to the spread of missionaries around the globe. In 2000, the church reported 60,784 missionaries and global church membership stood at just over 11 million. Worldwide membership surpassed 13 million in 2007 and reached 14 million in July 2010, with about six million of those within the United States. However, it is estimated based on demographic studies that only one-third of the total worldwide membership (about 4.5 million people as of 2014) are regularly attending churchgoers. The church cautions against overemphasis of growth statistics for comparison with other churches because relevant factors—including activity rates and death rates, methodology used in registering or counting members, what factors constitute membership, and geographical variations—are rarely accounted for in the comparisons.\n", "Christianity has had a significant impact on education, as the church created the bases of the Western system of education, and was the sponsor of founding universities in the Western world; as the university is generally regarded as an institution that has its origin in the Medieval Christian setting. Historically, Christianity has often been a patron of science and medicine. It has been prolific in the foundation of schools, universities, and hospitals, and many Catholic clergy; Jesuits in particular, have been active in the sciences throughout history and have made significant contributions to the development of science. Protestantism also has had an important influence on science. According to the Merton Thesis, there was a positive correlation between the rise of English Puritanism and German Pietism on the one hand, and early experimental science on the other. The civilizing influence of Christianity includes social welfare, founding hospitals, economics (as the Protestant work ethic), politics, architecture, literature, personal hygiene, and family life.\n" ]
What about Amsterdam has made it such a popular refuge for radical, creative thinkers, and how did they contribute to the Dutch golden age of the late 16th and 17th century?
During that time The Netherlands was fighting a war of independence, the 80 year war. While it basically stared as a revolt against the Spanish because of infringements of privileges like protection against unreasonable taxes etc.. It soon became also a religious based war between Protestants (Netherlands) and Catholics (Spanish). Since most Dutchmen were still Catholics themselves but against the Spaniards and their inquisition the common idea was that people should not be prosecuted because of their religion. This was specifically mentioned in the [act of abjuration](_URL_0_) (incomplete dutch english translation) send to the spanish king and publicly published in 1581. Besides a certain level of religious freedom several other privileges and rights were mentioned or implied. Like that the prince is there to protect his citizens as like father does his children or a shepherd his flock and not rule them like they are slaves. Also that there should be a justice system that is impartial en equal to all citizens and that there should be a freedom to trade. All this created at situation that you could hardly be prosecuted for expressing yourself if you stayed within certain limits. Also, the merchant class had become the prominent force at the time which means there are a lot of potential supporters for any kind of art and philosophy instead of a singular/limited source of patronage. All the freedoms created in this situation was the main source that Amsterdam specifically and the Netherlands as a whole entered an Golden Age with wealth and creativity. While not a classical history source an easy to read description on how this came about is [Russell Shorto's Amsterdam: A History of the World's Most Liberal City]( _URL_1_)
[ "As the centre of trade gravitated towards Amsterdam, Haarlem declined in the 18th century. The Golden age had created a large upper middle class of merchants and well-to-do small business owners. Taking advantage of the reliability of the trekschuit connection between Amsterdam and Haarlem, many people had a business address in Amsterdam and a weekend or summer home in Haarlem. Haarlem became more and more a bedroom community as the increasingly dense population of Amsterdam caused the canals to smell in the summer. Many well-to-do gentlemen moved their families to summer homes in the Spring and commuted between addresses. Popular places for summer homes were along the Spaarne in southern Haarlem. Pieter Teyler van der Hulst and Henry Hope built summer homes there, as well as many Amsterdam merchants and councilmen. Today, it is still possible to travel by boat along the Spaarne and this has turned into a popular form of tourism in the summer months. In the 18th century Haarlem became the seat of a suffragan diocese of the Old Catholic Church of Utrecht.\n", "The 17th century is considered Amsterdam's \"Golden Age\", during which it became the wealthiest city in the western world. Ships sailed from Amsterdam to the Baltic Sea, North America, and Africa, as well as present-day Indonesia, India, Sri Lanka, and Brazil, forming the basis of a worldwide trading network. Amsterdam's merchants had the largest share in both the Dutch East India Company and the Dutch West India Company. These companies acquired overseas possessions that later became Dutch colonies.\n", "The 17th century was Amsterdam's Golden Age. Ships from the city sailed to North America, Indonesia, Brazil, and Africa and formed the basis of a worldwide trading network. Amsterdam's merchants financed expeditions to the four corners of the world and they acquired the overseas possessions which formed the seeds of the later Dutch colonies. The most influential of these merchant groups was the Dutch East India Company, founded 1602, which became the first multi-national corporation to issue stocks to finance its business. By allowing for sailors to invest in the cargo that they transported, it created an incentive for individual laborers to be vested in the goods they carried and tightened allegiances to corporations outcomes where before they sailor was a migratory agent. Rembrandt painted in this century, and the city expanded greatly around its canals during this time. Amsterdam was the most important point for the transshipment of goods in Europe and it was the leading financial centre of the world (a position later taken over by London).\n", "Amsterdam and Rotterdam operated as important cosmopolitan centres where merchant ships from many parts of the world brought people of various customs and beliefs. This flourishing commercial activity encouraged a culture relatively tolerant of the play of new ideas, to a considerable degree sheltered from the censorious hand of ecclesiastical authority (though those considered to have gone \"too far\" might have gotten persecuted even in the Netherlands). Not by chance were the philosophical works of both Descartes and Spinoza developed in the cultural and intellectual background of the Dutch Republic in the 17th century. Spinoza may have had access to a circle of friends who were unconventional in terms of social tradition, including members of the Collegiants. One of the people he knew was Niels Stensen, a brilliant Danish student in Leiden; others included Albert Burgh, with whom Spinoza is known to have corresponded.\n", "The 18th and early 19th centuries saw a decline in Amsterdam's prosperity. The wars of the Dutch Republic with the United Kingdom and France took their toll on Amsterdam. During the Napoleonic wars, Amsterdam's fortunes reached their lowest point; however, with the establishment of the Kingdom of the Netherlands in 1815, things slowly began to improve. In Amsterdam, new developments were started by people like Samuel Sarphati who found their inspiration in Paris.\n", "The European economic centre shifted from the Mediterranean to Western Europe. The city of Antwerp, part of the Duchy of Brabant, became \"the centre of the \"entire\" international economy\", and the richest city in Europe at this time. Centred in Antwerp first and then in Amsterdam, \"Dutch Golden Age\" was tightly linked to the Age of Discovery. Francesco Guicciardini, a Venetian envoy, stated that hundreds of ships would pass Antwerp in a day, and 2,000 carts entered the city each week. Portuguese ships laden with pepper and cinnamon would unload their cargo. With many foreign merchants resident in the city and governed by an oligarchy of banker-aristocrats forbidden to engage in trade, the economy of Antwerp was foreigner-controlled, which made the city very international, with merchants and traders from Venice, Ragusa, Spain and Portugal and a policy of toleration, which attracted a large Orthodox Jewish community. The city experienced three booms during its golden age, the first based on the pepper market, a second launched by New World silver coming from Seville (ending with the bankruptcy of Spain in 1557), and a third boom, after the Treaty of Cateau-Cambresis, in 1559, based on the textiles industry.\n", "In the 17th century, the economic elite of Amsterdam moved with William of Orange to England, where they helped to restart the English international trade, leaving behind in Amsterdam the more religious, and less competitive, burghers.\n" ]
I've a Ni-MH recharger. Does it matter if I constantly recharge certain batteries even though they're already full?
NiMH chemistry is much more forgiving to overcharging than other types of battery chemistry, but it's really not good for the cell. Get some extra cells, and only charge ones that have been (mostly) used.
[ "Manufacturers do not support recharging of disposable alkaline batteries, and warn that it may be dangerous. Despite this advice, alkaline batteries have been recharged, and chargers have been available. The capacity of a recharged alkaline battery declines with number of recharges, until it becomes unusable after typically about ten cycles.\n", "Unlike rechargeable alkaline batteries, NiMH batteries can endure anywhere from a few hundred to a thousand (or more) deep discharge cycles, resulting in a long useful life; their limitation is now more usually by age rather than cycles. Capacity of NiMH batteries is close to that of alkaline batteries. Unlike all alkaline batteries (rechargeable or otherwise), internal resistance is low. This makes them well suited for high current capacity applications. Self-discharge rates are comparable, at least up to six months.\n", "When not under load or charge, a Ni–Cd battery will self-discharge approximately 10% per month at 20 °C, ranging up to 20% per month at higher temperatures. It is possible to perform a trickle charge at current levels just high enough to offset this discharge rate; to keep a battery fully charged. However, if the battery is going to be stored unused for a long period of time, it should be discharged down to at most 40% of capacity (some manufacturers recommend fully discharging and even short-circuiting once fully discharged), and stored in a cool, dry environment.\n", "Ni–Cd batteries may suffer from a \"memory effect\" if they are discharged and recharged to the same state of charge hundreds of times. The apparent symptom is that the battery \"remembers\" the point in its charge cycle where recharging began and during subsequent use suffers a sudden drop in voltage at that point, as if the battery had been discharged. The capacity of the battery is not actually reduced substantially. Some electronics designed to be powered by Ni–Cd batteries are able to withstand this reduced voltage long enough for the voltage to return to normal. However, if the device is unable to operate through this period of decreased voltage, it will be unable to get enough energy out of the battery, and for all practical purposes, the battery appears \"dead\" earlier than normal.\n", "Older rechargeable batteries self-discharge relatively rapidly, and require charging before first use; some newer low self-discharge NiMH batteries hold their charge for many months, and are typically sold factory-charged to about 70% of their rated capacity.\n", "The limited self-discharge characteristics of vanadium redox batteries make them useful in applications where the batteries must be stored for long periods of time with little maintenance while maintaining a ready state. This has led to their adoption in some military electronics, such as the sensor components of the GATOR mine system. Their ability to fully cycle and stay at 0% state of charge makes them suitable for solar + storage applications where the battery must start each day empty and fill up depending upon the load and weather. Lithium ion batteries, for example, are typically damaged when they are allowed to discharge below 20% state of charge, so they typically only operate between about 20% and 100%, meaning they are only using 80% of their nameplate capacity.\n", "The active material on the battery plates changes chemical composition on each charge and discharge cycle; active material may be lost due to physical changes of volume, further limiting the number of times the battery can be recharged. Most nickel-based batteries are partially discharged when purchased, and must be charged before first use. Newer NiMH batteries are ready to be used when purchased, and have only 15% discharge in a year.\n" ]
what will keep certain isps from providing "neutral internet" and completely destroying their competitors who don't?
To start an ISP you need to invest in a lot of infrastructure. Most ISPs is therefore limited to a small geographical area. And if there are two providers who need the infrastructure to every house then the service will be twice as expensive for everyone. Just imagine if there were two main water lines running to each house and you could chose between them.
[ "The comedian says that Pai has also proposed \"laughably lacking\" alternatives to net neutrality. One alternative stipulated that ISPs simply include a voluntary statement in their terms of service indicating that they would not throttle or block content, which Oliver says would \"make net neutrality as binding as a proposal on \"The Bachelor\"\". Pai's other rationales for reclassifying ISPs was that the new rules already resulted in decreased investment in broadband networks. Brian Schatz, a Democratic U.S. Senator representing Hawaii and the ranking member of the Senate Commerce Subcommittee on Communications, Technology, Innovation, and the Internet, said that Pai's claim of decreased investment was untrue. Oliver then quotes a 2014 phone call from Francis J. Shammo, the chief financial officer of Verizon, in which the latter says that the new net-neutrality rules did not affect Verizon's business. Oliver says, \"that doesn't really sound like net neutrality was jeopardizing investment at all\". He states that Pai's actions to eliminate Title II are like \"pouring [a gallon of coffee] into your hands and trusting that you don't get burnt\". At this point, Oliver holds up a Reese's mug that is even larger than Pai's mug and says, \"I'm drinking the blood of smaller mugs\".\n", "Opponents argue that net neutrality regulations prevent service providers from providing more affordable Internet access to those who can't afford it. A concept known as \"zero-rating\", ISPs would be unable to provide Internet access for free or at a reduced cost to the poor under net neutrality rules.\n", "Yahoo! and Microsoft argue that net neutrality law is necessary because without such a law, ISPs will destroy the free and open nature of the Internet and also create a tiered, dollar-driven net that favours the wealthiest corporations over everyone else.\n", "With net neutrality, ISPs may not intentionally block, slow down, or charge money for specific online content. Without net neutrality, ISPs may prioritize certain types of traffic, meter others, or potentially block traffic from specific services, while charging consumers for various tiers of service.\n", "In the United States, however, during the Obama-era administration, under the Federal Communications Commission's (FCC) 15-24 policy, there were three bright-line rules in place to protect net neutrality: no blocking, no throttling, no paid prioritization. On 14 December 2017, the FCC voted 3-2 to remove these rules, allowing ISPs to block, throttle and give fast-lane access to content on their network.\n", "Some ISPs, such as Comcast, oppose blocking or throttling, but have argued that they are allowed to charge websites for faster data delivery. AT&T has made a broad commitment to net neutrality, but has also argued for their right to offer websites paid prioritization and in favor of its current sponsored data agreements.\n", "Some opponents of net neutrality argue that under the ISP market competition, paid-prioritization of bandwidth can induce optimal user welfare. Although net neutrality might protect user welfare when the market lacks competition, they argue that a better alternative could be to introduce a neutral \"public option\" to incentivize competition, rather than enforcing existing ISPs to be neutral.\n" ]
Didn't the Zimmerman telegram serve its purpose?
No, the main aim of the Zimmerman telegram wasn't to bring the US into WW1. It was intended as insurance against the likelihood that the US would join WW1 anyway over Germany's resumption of unrestricted submarine warfare. This is the full text of the Zimmerman telegram, sent from the German foreign ministry to the German ambassador to Mexico. English translation, obviously, and the emphasis is mine. > We intend to begin on the first of February unrestricted submarine warfare. **We shall endeavor in spite of this to keep the United States of America neutral. In the event of this not succeeding**, we make Mexico a proposal of alliance on the following basis: make war together, make peace together, generous financial support and an understanding on our part that Mexico is to reconquer the lost territory in Texas, New Mexico, and Arizona. The settlement in detail is left to you. **You will inform the President of the above most secretly as soon as the outbreak of war with the United States of America is certain**, and add the suggestion that he should, on his own initiative, invite Japan to immediate adherence and at the same time mediate between Japan and ourselves. Please call the President's attention to the fact that the ruthless employment of our submarines now offers the prospect of compelling England in a few months to make peace. Signed, ZIMMERMANN Germany had decided to resume Unrestricted Submarine Warfare (USW) in 1917 for two reasons, one military and one domestic political. USW means that submarines may sink any merchant ships in the declared exclusion area, regardless of national flag, and without warning. As opposed to the "Prize Rules" or "Cruiser Rules" traditional international law had applied to commerce raiding, where only enemy-flagged merchant ships were legitimate targets. Ships belonging to neutral countries could be stopped and searched, and seized or sunk if they were found to be carrying contraband to blockaded ports or if they tried to resist the search, but could not be summarily attacked. Any ships to be sunk under cruiser rules must first be given reasonable warning for the passengers crew to abandon ship. These rules were developed with surface warships (particularly frigates and cruisers) in mind, not submarines, and submarines of the era didn't have anywhere near enough crew to search and seize a merchant ship, and submarines relied on stealth to be effective combatants so surfacing and summoning a merchant ship to submit to search or ordering the crew to abandon ship was risky, especially after the British started disguising armed merchantmen as ordinary merchant vessels in order to sink German subs when they surfaced (this was legal under international law as long as the disguised ship revealed its true nature by raising a Navy flag before opening fire). Germany had previously attempted USW briefly in 1915, only to call it off after the US responded to the sinking of the Lusitania (a British ship sailing from New York and carrying many American and British passengers as well as a cargo that included war materiel being imported by Britain). The sinking was done without warning by a submerged U-Boat, the British denied the presence of war materiel on the ship (and the Germans had no hard evidence to the contrary), and 128 Americans (along with over a thousand British and other nationalities of passengers and crew) died. The US responded with vigorous diplomatic protests, stopping about half a step short of threatening war if USW continued. Germany acquiesced at the time and resumed cruiser rules. The military reason was that the German navy had overestimated the effectiveness of USW and underestimated the British and American ability to build new merchant ships to replace those sunk. Based on their mistaken assessments, the Germans believed that they could sink enough merchant ships to prevent Britain from importing the food and raw materials they needed to continue the war, and that this would happen before Americans could mobilize enough to have a significant effect on the war. The domestic political reason was as a response to the British blockade of Germany. Britain had been blockading Germany (with surface ships, operating under cruiser rules) since the outbreak of war, and Britain had been taking a very broad basis for what constituted war materiel. Specifically, food and fertilizer were counted as contraband, since food could be sent to the front to feed soldiers, and since fertilizer could be remanufactured into explosives. Germany had imported a large portion of its food pre-war, and German civilians had been on short rations for much of the war. The Winter of 1916-1917 (called the "Turnip Winter") was a period of particular hardship for German civilians, as the 1916 potato crop had failed, compounding the food shortage. Unrestricted Submarine Warfare was seen as a way for the German Navy to demonstrate to German civilians that they were Doing Something about the blockade and inflicting the same kinds of hardship on British civilians that the British Navy was inflicting on German civilians. Once the decision to resume USW had been made, war with the US was considered inevitable. The idea of the Zimmerman Telegram was to exploit Mexican grievances against the US (not just the lost territory from the 1846-48 US/Mexican War, but also the much more recent US punitive expedition expedition in pursuit of Pancho Villa) to bring another country into the war on the German side as counterweight to the US joining the Entente. Every division the US Army used to fight Mexico was a division that couldn't be sent to France to reinforce the British and French armies, so forcing the US to divide attention between two fronts would further delay substantive American involvement with the war effort in Europe.
[ "The Zimmermann Telegram was part of an effort carried out by the Germans to postpone the transportation of supplies and other war materials from the United States to the Allied Powers that were at war with Germany. The main purpose of the telegram was to make the Mexican government declare war on the United States in hopes of tying down American forces and slowing the export of American arms. The German High Command believed they would be able to defeat the British and French on the Western Front and strangle Britain with unrestricted submarine warfare before American forces could be trained and shipped to Europe in sufficient numbers to aid the Allied Powers. The Germans were encouraged by their successes on the Eastern Front into believing that they would be able to divert large numbers of troops to the Western Front in support of their goals. The Mexicans were willing to consider the alliance but declined the deal after Americans were informed of the Telegram.\n", "The Zimmermann Telegram (or Zimmermann Note or Zimmerman Cable) was a secret diplomatic communication issued from the German Foreign Office in January 1917 that proposed a military alliance between Germany and Mexico. In the event that the United States entered World War I against Germany, Mexico would recover Texas, Arizona and New Mexico. The telegram was intercepted and decoded by British intelligence. Revelation of the contents enraged Americans, especially after German Foreign Secretary Arthur Zimmermann publicly admitted on March 3 that the telegram was genuine, helping to generate support for the United States declaration of war on Germany in April. The decryption was described as the most significant intelligence triumph for Britain during World War I, and one of the earliest occasions on which a piece of signal intelligence influenced world events.\n", "The Zimmermann Telegram was a 1917 diplomatic proposal from the German Empire for Mexico to join an alliance with Germany in the event of the United States entering World War I against Germany. The proposal was intercepted and decoded by British intelligence. The revelation of the contents outraged the American public and swayed public opinion. President Woodrow Wilson moved to arm American merchant ships in order to defend themselves against German submarines, which had started to attack them. The news helped generate support for the United States declaration of war on Germany in April of that year.\n", "In October 2005, it was reported that an original typescript of the decoded Zimmermann Telegram had recently been discovered by an unnamed historian who was researching and preparing an official history of the United Kingdom's Government Communications Headquarters (GCHQ). The document is believed to be the actual telegram shown to the American ambassador in London in 1917. Marked in Admiral Hall's handwriting at the top of the document are the words: \"This is the one handed to Dr Page and exposed by the President.\" Since many of the secret documents in this incident had been destroyed, it had previously been assumed that the original typed \"decrypt\" was gone forever. However, after the discovery of this document, the GCHQ official historian said: \"I believe that this is indeed the same document that Balfour handed to Page.\"\n", "Any doubts as to the authenticity of the telegram were removed by Arthur Zimmermann himself. First, at a press conference on 3 March 1917, he told an American journalist, \"I cannot deny it. It is true.\" Then, on 29 March 1917, Zimmermann gave a speech in the Reichstag in which he admitted the telegram was genuine. Zimmermann hoped Americans would understand the idea was that Germany would only fund Mexico's war with the United States in the prior event of American entry into World War I.\n", "Before the United States declared war in 1917, the Woodrow Wilson administration established a propaganda department along similar lines. Propaganda experts Walter Lippmann and Edward Bernays participated in the Committee on Public Information (CPI), which was tasked with swaying popular opinion to encourage enlistment and war bond sales. The CPI deployed posters, films, and provided themes for speeches by \"four-minute men\" at public functions, and also encouraged censorship of the American press. The American press played an unwitting role too by relying on daily war news cables controlled by the British government and by spreading false stories of German atrocities in Belgium and German-occupied eastern France supplied by the British as well. Starting after World War I, propaganda had a growing negative connotation. This was due in part to the 1920 book \"How We Advertised America: the First Telling of the Amazing Story of the Committee on Public Information that Carried the Gospel of Americanism to Every Corner of the Globe\" in which the impact of the CPI, and the power of propaganda, was over-emphasised. Also, exposure of fact that the atrocity stories were false created public distrust. The CPI was so unpopular that after the war, Congress closed it down without providing funding to organise and archive its papers.\n", "Foreign Minister Pavel Milyukov sent the Allied Powers a telegram, one which would prove to be a turning point of 1917. The telegram contained the statement that the Provisional Government would continue the war with the Tsarist war aims. The telegram was leaked to the Bolsheviks by an insider, and they began protests against the Provisional Government, which had previously promised the people that the terror of Tsardom had ended.\n" ]
When photographing a light, why do streaks show up in regular angular positions around the light?
That's the shape of the camera's aperture, and the effect is called [bokeh](_URL_0_).
[ "Dark slope streaks are albedo features. They appear to the eye as a brightness difference between the streak and the lighter-toned background slope. Usually no topographic relief is visible to distinguish the streak from its surroundings, except in the very highest resolution (<1 m/pixel) images. In many cases, the original surface texture of the slope is preserved and continuous across the streak, as though unaffected by events involved in dark streak formation (pictured left). The overall effect is equivalent in appearance to a partial shadow cast down the sloping surface. These observations indicate that whatever process forms the streaks, it affects only the very thinnest layer at the surface. Slope streaks are only about 10% darker than their surroundings but often appear black in images because the contrast has been enhanced (stretched).\n", "Another special case is that of forward (or backwards) camera movement, typically with the camera focused on the distance. The streaks in the resulting image converge on the central point, giving a suggestion as if it is at the end of a long tunnel. A similar effect can be achieved by changing the focal length of a zoom lens during the exposure.\n", "The basic idea is that the illumination pattern is imaged onto a geometrically congruent cutoff pattern (essentially a multiplicity of knife edges) with focusing optics, while density gradients lying between the illumination pattern and the cutoff pattern are imaged, typically by a camera system. As in classical schlieren, the distortions produce regions of brightening or darkening corresponding to the position and direction of the distortion, because they redirect rays either away from or onto the opaque part of the cutoff pattern. While in classical schlieren, distortions over the whole beam path are visualized equally, in focusing schlieren, only distortions in the object field of the camera are clearly imaged. Distortions away from the object field become blurred, so this technique allows some degree of depth selection. It also has the advantage that a wide variety of illuminated backgrounds can be used, since collimation is not required. This allows construction of projection-based focusing schlieren systems, which are much easier to build and align than classical schlieren systems. The requirement of collimated light in classical schlieren is often a substantial practical barrier for constructing large systems due to the need for the collimating optic to be the same size as the field of view. Focusing schlieren systems can use compact optics with a large background illumination pattern, which is particularly easy to produce with a projection system. For systems with large demagnification, the illumination pattern needs to be around twice larger than the field of view to allow defocusing of the background pattern.\n", "In photography, a rectilinear lens is a photographic lens that yields images where straight features, such as the walls of buildings, appear with straight lines, as opposed to being curved. In other words, it is a lens with little or no barrel or pincushion distortion. At particularly wide angles, however, the rectilinear perspective will cause objects to appear increasingly stretched and enlarged as they near the edge of the frame. These types of lenses are often used to create forced perspective effects.\n", "In photography, lens flare is a defect where various circles can appear around highlights, and with ghosts throughout a photo, due to undesired light, such as reflection and scattering off elements in the lens.\n", "Motion compensation photography (also known as ballistic synchro photography or smear photography when used to image high-speed projectiles) is a form of streak photography. When the motion of the film is opposite to that of the subject with an inverting (positive) lens, and synchronized appropriately, the images show events as a function of time. Objects remaining motionless show up as streaks. This is the technique used for finish line photographs. At no time is it possible to take a still photograph that duplicates the results of a finish line photograph taken with this method. A still is a photograph \"in\" time, a streak/smear photograph is a photograph \"of\" time. When used to image high-speed projectiles the use of a slit (as in streak photography) produce very short exposure times ensuring higher image resolution. The use for high-speed projectiles means that one still image is normally produced on one roll of cine film. From this image information such as yaw or pitch can be determined. Because of its measurement of time variations in velocity will also be shown by lateral distortions of the image.\n", "Ring flashes are commonly used in macro (close-up) photography. When the subject is very close to the camera, the distance of the flash from the optical axis becomes significant. For objects close to the camera, the size of the ring flash is significant and so the light encounters the subject from many angles in the same way that it does with a conventional flash with soft box. This has the effect of further softening any shadows.\n" ]
When the Sun's red giant phase ends it'll lose roughly 50% of its mass to space, does this excess surface hydrogen have enough mass to create a red dwarf?
Red dwarfs are main sequence stars just like the Sun, except they only contain about 8-50% as much mass. As a consequence, the fusion in their cores proceeds slower, their surface temps are cooler, and their life spans are much longer. During its Red Giant phase, the Sun will gradually lose much of its upper atmosphere to space, enriching its general neighborhood with up to 0.5 solar mass of mostly Hydrogen. Would this be enough raw material to form a Red dwarf star? Sure, if something corralled it together until it began coalescing under its own gravity. But with the Sun still coughing and wheezing in the center of it all, that Hydrogen will instead just disperse into the blackness, until sometime, somewhere, some of it finds some like-minded hydrogen to get all gravitational with.
[ "The evolved red giant star is losing mass, since radiation pressure overcomes the low gravity on the surface. The outflow of matter is captured by the gravitational field of the white dwarf and falls on its surface in the end. At least during the active phase an accretion disk forms around the white dwarf.\n", "Due to opacity effects, as metallicity of the interstellar gas increases both the maximum and minimum masses a star can have will decrease. For the latter case, it's expected that an object with a mass of 0.04 solar masses (40 times the mass of Jupiter), that currently would become a brown dwarf unable to fuse hydrogen, could do so ending in the main sequence with a surface temperature of 0 °C (273 K, thus \"frozen\"), much cooler than the dimmest red dwarfs of today, and ice clouds forming in its atmosphere. The luminosity of these objects would be more than a thousand times lower than the faintest stars currently existing and their lifetimes would also be sensibly longer.\n", "The largest star in the Cygnus OB3 association has a mass 40 times that of the Sun. As more massive stars evolve more rapidly, this implies that the progenitor star for Cygnus X-1 had more than 40 solar masses. Given the current estimated mass of the black hole, the progenitor star must have lost over 30 solar masses of material. Part of this mass may have been lost to HDE 226868, while the remainder was most likely expelled by a strong stellar wind. The helium enrichment of HDE 226868's outer atmosphere may be evidence for this mass transfer. Possibly the progenitor may have evolved into a Wolf–Rayet star, which ejects a substantial proportion of its atmosphere using just such a powerful stellar wind.\n", "The substellar mass in orbit around the white dwarf is a star that lost all of its gas to the white dwarf. What remains is an object with a mass of 0.05 solar masses (), which is too small to continue fusion, and does not have the composition of a super-planet, brown dwarf, or white dwarf. There is no category for such a stellar remnant.\n", "The equation of state for degenerate matter is \"soft\", meaning that adding more mass will result in a smaller object. Continuing to add mass to what is now a white dwarf, the object shrinks and the central density becomes even larger, with higher degenerate-electron energies. The star's radius has now shrunk to only a few thousand kilometers, and the mass is approaching the theoretical upper limit of the mass of a white dwarf, the Chandrasekhar limit, about 1.4 times the mass of the Sun ().\n", "The white dwarf in the V445 Puppis system has an estimated mass of more than 1.3 times the mass of the Sun, and this mass is increasing because of recurring helium shell flashes from accreted material. As the mass of the white dwarf approaches the Chandrasekhar limit of about 1.38 solar masses, it will likely explode as a Type Ia supernova.\n", "After fusing helium in its core to carbon, the Sun will begin to collapse again, evolving into a compact white dwarf star after ejecting its outer atmosphere as a planetary nebula. The predicted final mass is 54.1% of the present value, most likely consisting primarily of carbon and oxygen.\n" ]
what is a constituent country and how does it differ from units like state or provinces?
Functionally they are almost identical to US State or Canadian Province. They do not have their own militaries, do not have their own passports, do not have their own treaties and trade agreements, but they do have some of their own laws and self governance. Puerto Rico is a territory. They are partially subject to the US, but are not full members of the US as they have so far chosen to not become a State. They pay some taxes, have their own laws and government, but rely on the US for military support and all trade has to come through the US first. Quebec is a Province of Canada, and Bavaria is a State in Germany. They are a somewhat culturally distinct grouping but are not separate countries, nor are they different from their other provinces or states of their nation as you imply by calling them "sub-national entities".
[ "A constituent state is a state entity that constitutes a part of a sovereign state. A constituent state holds regional jurisdiction over a defined administrative territory, within a sovereign state. Government of a constituent state is a form of regional government. Throughout history, and also in modern political practice, most constituents states are parts of complex states, like federations or confederations. Constituent state can have republican or monarchical form of government. Those of republican form are usually called \"states\" or autonomous states, \"republics\" or autonomous republics, and also cantons. Those that have monarchical form of government are often defined by traditional hierarchical rank of their ruler (usually a principality, or an emirate).\n", "States may be classified by political philosophers as sovereign if they are not dependent on, or subject to any other power or state. Other states are subject to external sovereignty or hegemony where ultimate sovereignty lies in another state. Many states are federated states which participate in a federal union. A federated state is a territorial and constitutional community forming part of a federation. (Compare confederacies or confederations such as Switzerland.) Such states differ from sovereign states in that they have transferred a portion of their sovereign powers to a federal government.\n", "In some nations, a province (or its equivalent) is a first-level administrative unit of sub-national government—as in the Netherlands—and a large constituent autonomous area, as in Argentina, Canada, South Africa, and the Democratic Republic of the Congo. It can also be a constituent element of a federation, confederation, or republic. For example, in the United States, no state may secede from the federal Union without the permission of the federal government.\n", "In many federations and confederations, the province or state is not clearly subordinate to the national or central government. Rather, it is considered to be sovereign in regard to its particular set of constitutional functions. The central- and provincial-government functions, or areas of jurisdiction, are identified in a constitution. Those that are not specifically identified are called \"residual powers.\" In a decentralized federal system (such as the United States and Australia) these residual powers lie at the provincial or state level, whereas in a centralized federal system (such as Canada) they are retained at the federal level.\n", "Administrative units that are not federated or confederated but enjoy a greater degree of autonomy or self-government than other territories within the same country can be considered constituent states of that country. This relationship is by some authors called a federacy. Autonomous republics like Karakalpakstan in Uzbekistan\n", "Constituent states united in a federal union under a federal government are more specifically known as federated states. There are numerous historical examples for constituent states within various federations. In modern political practice, among most notable examples are U.S. states and constituent entities of the Russian Federation.\n", "A federal state has a central structure with at most a small amount of territory mainly containing the institutions of the federal government, and several regions (called \"states\", \"provinces\", etc.) which compose the territory of the whole state. Sovereignty is divided between the centre and the constituent regions. The constitutions of Canada and the United States establish federal states, with power divided between the federal government and the provinces or states. Each of the regions may in turn have its own constitution (of unitary nature).\n" ]
How important have aircraft carriers been to winning superpower-on-superpower warfare?
So limiting this to the 20 year rule, and thus 1997 first off with only brief references to events and trends past it. We can actually though talk a good deal about Soviet carrier efforts, which obviously play a huge part in how Russian and PLAN Naval Aviation evolved. Simply put, from their origins to today an aircraft carrier is a tool of power projection. It allows you to take a piece of ocean that couldn't be used as an airbase and suddenly you have a fully functional one with up to +75 aircraft ready to go and which can be somewhere 100's of miles away tomorrow. It is a tool to enable expeditionary warfare and a sustained presence beyond the limits of land based aircraft for the power that has one. From Imperial Japan, to the United States and Great Britain, to the French, to other minor carrier nations, it has always been about building the ability to go someplace else and attack there using your big moving runway. To avoid too deep an overview we can then suffice to say that they were crucial to Allied victory in WW2, very important in US efforts in SE Asia in the Cold War and elsewhere, and factored heavily into plans for a NATO/Warsaw Pact confrontation. And carriers through the years have done numerous different missions, often all at the same time. They can first and foremost attack other naval assets(ships, planes, subs), can attack ground targets and support landings, and can protect sea lanes of communication from enemy forces. All 3 have been key over the years from the drive across the Pacific and Battle of the Atlantic, to supporting fighting in Korea, Vietnam, and in the Middle East, to patrolling the North Atlantic and vital NATO sea lanes. This variety of missions, and embracing of robust air power and naval forward presence helped drive the US to embrace the super carrier in the 50's, after temporarily successful efforts by the Air Force in the late 40's to kill earlier designs. While the British were faced with a shrinking empire, up and down economy, and loss of great power status, their navy including carriers slowly transitioned from a broad spectrum force to by the end of the Cold War on focused heavily on leveraging their position to focus on anti submarine warfare on the Greenland-Iceland-UK line. Though the importance of carriers as forward airfields was ably stated in the Falklands when Argentinian aircraft had only a few minutes of fuel to attack while the Brits could maintain a present CAP with their Harriers(though they had short legs too), the Argentines carrier was kept in safe water for fear of attack by RN subs after they bagged the General Belgrano cruiser. Many older US Essex class carriers experienced the same transition, as their usage in operating heavier modern jets waned their usage as anti sub carriers extended their lives by decades sometimes. While the newer super carriers could serve as the attack platforms and the centerpiece of offensive patrols or ground strikes. In a hot Cold War all in the name of ensuring that reinforcements could cross the Atlantic unmolested and fight it out on the plains of Germany. The USSR though was in a very different place, they were for the most part already where they expected to fight and didn't need to take a plane or boat to get there. Thus they didn't have long sea lanes to protect or lack of airfields to support expected hot zones, though like all airfields they were stationary and fixed unlike a carrier. And their naval strategy was markedly different. They had 2 main goals, 1 was to protect undersea "Bastions" essentially heavily defended safezones where their ballistic missile subs could head to and be safe if the order came to launch. The other being to strike at NATO Sea Lanes. This was then achieved by using numerous fast attack subs, of nuclear and diesel varieties, surface ships firing big heavy and fast anti ship cruise missiles, and crucially long ranged shore based naval bombers and patrol aircraft. By the late Cold War a theoretical attack on a US Carrier Group might involve several soviet attack subs say Victor class nuke boat or a conventional Kilo, maybe a surface group of a Kirov battlecruiser a Kresta cruiser, and a gaggle of destroyers, and most importantly several dozen long ranged bombers. This would be split between the fast jet Backfires and the big ubiquitous turboprop Bears, all with the goal of saturating and overwhelming a battlegroups ability to meat all threats thanks to huge for their job missiles like the P-270, Kh-22, P-500 and P-700. So what role for a carrier then? We see it in the limited USSR carrier projects, that aside from the ridiculous Stalin era plans that carriers were meant to simply augment this existing dynamic. To provide additional cover for the Bastions, and to protect surface task forces during the run in to missile launch. We can see this in the few classes of carriers the Soviets actually got completed, ignoring the Moskva's which are really just glorified DDH's and truly dedicated ASW platforms. That leaves us with the Kiev's and Kuznetsov's, both of which have very similar roles and the Kuznetsov's are really just enlarged improved variations on the same theme capable of operating non VSTOL aircraft from the get go. They were meant to simply defend their immediate vicinity to boil it down, and were even equipped with their own anti ship cruise missiles just like the surface combatants(if the trip is likely to be one way why not augment the striking power! this also had the benefit of allowing them to do the legal gymnastics of not calling them carriers and avoiding issues with them passing through the straits into the Black Sea). The Russians did flirt with larger designs over the years but it is a big step, and even the follow on to the Kuznetsov the Ulyanovsk would still have been a ski jump design and carry a large battery of anti surface missiles, though it was scrapped with the end of the Cold War only partly begun. China faces many of the same issues, their primary concern is first ensuring the ability to defend their territory, then to meet enemies in the South China Sea and 11 Dash line of the first island chain. While for much of its existence the ability of the PLAN to project power any further towards the second chain of islands like Guam into the Pacific was simply not possible from a numbers or capability standpoint. Thus a reliance on Anti-Access/Area Denial strategy to protect the mainland and again usage of lots of shore based aviation since the foe mostly had to come to them and they were already mostly in range of what they wanted to hit to start a war (SK, Japan, the Philippines, Taiwan, etc.). This has changed though as China has sought to assert itself as a rising power. So fast forward a few years the post Soviet slump meant no new ships and the Russians were left with what they had, hence the 1 Kuznetsov that was ready, and 4 Kiev's all of which were quickly withdrawn from service (2 sold as museums to China, 1 scrapped, and 1 heavily reconstructed and sold to India and in operation today). While the Kuznetsov has lunged from breakdown to breakdown never getting the care she really needs, her unfinished sister was taken by The Ukraine and eventually sold to China who finished her as the current Liaoning, and a near sister was just launched this week. And in the end we should not dismiss the importance of prestige when it comes to having a carrier. Its just something "Great Powers" have and do. If Brazil can operate a carrier so can Russia. If France can make a nuclear carrier work why can't China? If Britain is building a modern 2 carrier fleet why shouldn't the world's largest democracy in India do so? That said the future of carriers has often been a hot debate. No more so at any other time than today with rising costs, advancing A2/AD assets, and evolving faces of warfare. Even in the US questions about what shape the force should take remain in the open and sometimes contentious. For a better understanding of the shape of the debate I would suggest checking out professional sites and sources, one of which is the US Naval Institute, their magazine is pay for view, but their blog regularly publishes excellent work on relevant topics. For instance this was a selected essay by a Midshipman from Annapolis on the PLAN's carrier outlook nd ambitions, perfect as an introduction: _URL_0_
[ "The significance of air power grew between the wars, driven by the increased range, carrying power, and effectiveness of carrier-launched aircraft, until it became impossible to disregard its importance during World War II, following the loss of many warships to aircraft, including the sinking of \"Prince of Wales\" and \"Repulse\", the Battle of Taranto, the Attack on Pearl Harbor and numerous other incidents. Following the war, carrier operations continued to increase in size and importance.\n", "In the 21st century, the aircraft carrier is the last remaining capital ship, with capability defined in decks available and aircraft per deck, rather than in guns and calibers. The United States possesses supremacy, in both categories of aircraft carriers, possessing not only 11 active duty supercarriers each capable of carrying and launching nearly 100 tactical aircraft, but an additional 12 amphibious assault ships as capable (in the \"Sea Control Ship\" configuration) as the light VSTOL carriers of other nations.\n", "Some of the world's great powers maintain substantial naval forces to provide a global presence, with the largest nations possessing aircraft carriers, nuclear submarines and advanced anti-air defense systems. The vast majority of military ships are conventionally powered, but some are nuclear-powered. There is also a large global market in second-hand naval vessels, generally purchased by developing countries from Western governments.\n", "In the long run, Navy arguments for the supercarrier prevailed, though not for the reasons originally cited. A relative failure as a strategic nuclear deterrent, the large aircraft carrier would prove invaluable as an element of conventional rapid deployment tactical air forces, requiring neither overflight permissions or overseas basing rights with host nations. Ironically, a successor to the canceled supercarrier, the radical new USS \"Forrestal\", and later designs, continue in service with the U.S. Navy into the 21st century, forming the core of its offensive striking power.\n", "With the arrival of the fleet carriers the primary striking power of the navy was no longer in its battleship force, but with the aircraft that could be brought to battle by the carriers. The means by which the US Navy operated these carriers was developed principally by Admiral Marc Mitscher. Mitscher determined that the best defense for a carrier was its own air groups, and that carriers were more easily defended if they operated together in groups, with supporting ships along with them to aid in air defense, anti-submarine defense, and rescue of downed airmen.\n", "After the successes of carrier combat achieving air supremacy in both 1942's Battle of the Coral Sea, and, to a greater extent, Battle of Midway, the Navy was forced to shift its building focus from battleships to aircraft carriers. As a result, construction of the fleet of s had been given highest priority. In the process, the United States found the high speed of 33 knots of the \"Iowa\"s valuable; this allowed them to steam with the \"Essex\"-class while providing the carriers with maximum anti-aircraft protection.\n", "The element of air superiority has been the driving force behind the development of aircraft carriers, which allow aircraft to operate in the absence of designated air bases. For example, the Japanese attack on Pearl Harbor was carried out by aircraft operating from carriers thousands of miles away from the nearest Japanese air base.\n" ]
Did the Russians really train their troops with broomsticks instead of guns during WW1?
In 1941 the British troops retreating from Greece to Crete were forced to leave their heavy artillery behind. In an attempt to prevent the Germans from making use of those guns, the British artillerymen took the gun sights with them to Crete. When the Germans invaded Crete those artillerymen were left with sights for non-existant guns In the famous Doolittle Raid, the tailguns from the US bombers were removed (in order to keep weight down) and replaced with broomsticks painted black. Despite both of those, your history teacher is unlikely to tell you the British sent their men into battle with artillery sights but no guns, or that the US sent bombers on missions without giving them weapons to defend themselves. In war, there can be times when soldiers have to fight with less than ideal equipment. Just like the British artillerymen, there were times when Soviet infantry were forced - due to circumstances on the ground - to fight without sufficient arms or ammo. But it wasn't standard practice. The point is that it's easy (and unfortunately common) to point to examples where the Soviets were forced to fight (or in this case train) without sufficient equipment and act as though that was the norm, while almost certainly not doing the same for the Western Allies. So I'm not saying your history teacher was right or wrong (I don't know a reliable source that would confirm their story, and don't know how you could show it never happened) but it's important to remember that context is always important.
[ "The Imperial Russian Army had hunter-commando units, formed by a decree of Emperor Alexander III in 1886, which saw action in World War I prior to the Russian Revolution of 1917. Also during World War I, General Aleksei Brusilov became one of the first senior commanders to utilize the tactics of fast-action shock troops for assaults following concentrated accurate artillery fire in what would be later be known as the Brusilov Offensive of 1916. Such tactics, considered revolutionary at the time, would later inspire people like Prussian Captain Willy Rohr in the development of the Prussian Stormtroopers (founded in 1915).\n", "Despite a certain degree of standardisation, the tachanka's armament was, in most cases, improvised. In Russia, the PM M1910 heavy machine gun was often used. The Polish cavalry of the Polish-Soviet War often used all kinds of machine guns available, including the Maxim, Schwarzlose MG M.07/12, Hotchkiss machine gun and Browning machine gun. The late models of standardised tachankas of the Polish Army were all equipped with Ckm wz.30, a Polish modification of the M1917 Browning machine gun which was also suitable for anti-air fire.\n", "German field guns had proven too heavy to accompany the infantry in the assault and the Germans resorted to a variety of solutions in an effort to find something that could help the infantry deal with bunkers and other obstacles. Enormous numbers of Russian 7.62 cm Model 1910 Putilov fortress guns had been captured early in the war and Krupp was told to adapt them for use as infantry guns. They mounted the barrel and breech of the Russian guns on a new solid box-trail carriage with two narrow seats behind the gunshield, facing to the rear. The gun retained its extraordinary depression of -18.6°, which was a legacy of its original purpose to fire down into fortress ditches, although its limited elevation prevented it from ranging past without digging in the trail. It used captured Russian canister ammunition for short-range engagements, but Rheinmetall manufactured its HE shell.\n", "In Russian service they saw action on the Eastern Front, during the Russian Civil War and the subsequent Polish-Bolshevik War. Some were captured by Poland during that conflict, although nothing is known of any use by them. In June, 1941 the Red Army still possessed 25 of these guns.\n", "The tactic was used as an expedient by the Red Army during World War II. Tank desant troops () were infantry trained in the tactic in order to offer small-arms support in suppression of enemy anti-tank weapons or enemy infantry using anti-tank grenades. After the war, T-55 and T-62 tanks were built with hand-holds for this purpose. In northern areas during winter, similar tactics were used by Soviet infantry riding the skids of aerosani or towed behind them on skis. Nowadays, this tactic is very rare (outside of dire emergencies) in well-equipped armed forces, with front-line troops usually riding in armoured personnel carriers or infantry fighting vehicles.\n", "During the First World War, many combatants faced the deadlock of trench warfare. On the Western Front in 1915, the Germans formed a specialized unit called the Rohr Battalion to develop assault tactics. During the Brusilov Offensive of 1916, the Russian general Aleksei Brusilov developed and implemented the idea of shock troops to attack weak points along the Austrian lines to effect a breakthrough, which the main Russian Army could then exploit. The Russian Army also deployed hunter commando units who are specially trained forces of Russian army formed in 1886 and were used to protect against ambushes, to perform reconnaissance and for low-intensity fights in no-man's-land.\n", "Combat experience in the Russo-Japanese War of 1904–1905 had convinced the Japanese of the utility of machine guns to provide covering fire for advancing infantry. This was reinforced by the first-hand observations of European combat tactics by Japanese military attachés during the First World War, and the Army Technical Bureau was tasked with the development of a lightweight machine gun, which could be easily transportable by an infantry squad. The resultant “Type 11 light machine gun” (named after the 11th year of the reign of Emperor Taishō, or 1922) was the first light machine gun to be mass-produced in Japan and the oldest Japanese light machine gun design to see service in the Pacific War. It was superseded by the Type 96 light machine gun in 1936.\n" ]
How long ago did the modern chicken lose its ability to fly?
Technically, modern domesticated chickens can still fly. Sort of. Not well. But... for short distances and not terribly gracefully^1. They are not flightless. However, their poor flying abilities are the result of domestication, and selective breeding by farmers and the commercial poultry industry (1950s-now)^2 to be larger, faster growing, and meatier while keeping the birds sheltered from any sort of predators. The large breast is actually the flight muscle, but their wings are now too short, and those muscles too heavy to support effective wing loading and flight. While chickens have been domesticated for ~6000-8000 years, [it's pretty clear](_URL_2_)^3 that many of the significant changes occurred in the very recent past (since 1950s). The authors in this study raised a set of lines representative of 1957 (experimental/archival), 1978 (experimental/archival) and 2005 (commercial/meat selected) birds on the same diet, and under the same conditions to study growth and development. While they didn't look at flight abilities of the respective chicken lines, the breast size and conversion rate of grams of food- > grams of breast meat is about tripled. They also measured a significant shift from the 1957 and 1978 lines to the 2005 line in terms of the grams of breast meat/total body weight. So, the 2005 meat-selected line is almost certainly a worse flyer than the 1957 experimental unselected chicken line. ^1 [Forget About the Road. Why Are Chickens So Bad at Flying?](_URL_4_) by Live Science ^2 [U.S. Chicken Industry History](_URL_1_) by the National Chicken Council ^3 M. J. Zuidhof et al., (2014) [Growth, efficiency, and yield of commercial broilers from 1957, 1978, and 2005](_URL_3_) *Poultry Science* - if paywalled, see [Huffington Post article instead](_URL_0_)
[ "Many domesticated birds, such as the domestic chicken and domestic duck, have lost the ability to fly for extended periods, although their ancestral species, the red junglefowl and mallard, respectively, are capable of extended flight. A few particularly bred birds, such as the Broad Breasted White turkey, have become totally flightless as a result of selective breeding; the birds were bred to grow massive breast meat that weighs too much for the bird's wings to support in flight.\n", "This method was unsuccessful in a 1969 study attempting to control \"Triatoma infestans\" in chicken houses because even though some bugs that fed on the treated birds did die, so did the birds, and the birds that survived produced fewer eggs.\n", "The average chicken may live for five to ten years, depending on the breed. The world's oldest known chicken was a hen which died of heart failure at the age of 16 years according to the Guinness World Records.\n", "The chicken is one of the most widespread domesticated species and one of the human world's largest sources of protein. Although the chicken was domesticated in South-East Asia, archaeological evidence suggests that it was not kept as a livestock species until 400 BCE in the Levant. Prior to this, chickens had been associated with humans for thousands of years and kept for cock-fighting, rituals, and royal zoos, so they were not originally a prey species. The chicken was not a popular food in Europe until only one thousand years ago.\n", "Throughout the 20th century, modern veterinary practices and the advent of radio telemetry (transmitters attached to free-flying birds) increased the average lifespan of falconry birds and allowed falconers to pursue quarry and styles of flight that had previously resulted in the loss of their hawk or falcon.\n", "Today, more than 1,532 passenger pigeon skins (along with 16 skeletons) are in existence, spread across many institutions all over the world. It has been suggested that the passenger pigeon should be revived when available technology allows it (a concept which has been termed \"de-extinction\"), using genetic material from such specimens. In 2003, the Pyrenean ibex (\"Capra pyrenaica pyrenaica\", a subspecies of the Spanish ibex) was the first extinct animal to be cloned back to life; the clone lived for only seven minutes before dying of lung defects. A hindrance to cloning the passenger pigeon is the fact that the DNA of museum specimens has been contaminated and fragmented, due to exposure to heat and oxygen. American geneticist George M. Church has proposed that the passenger pigeon genome can be reconstructed by piecing together DNA fragments from different specimens. The next step would be to splice these genes into the stem cells of rock pigeons (or band-tailed pigeons), which would then be transformed into egg and sperm cells, and placed into the eggs of rock pigeons, resulting in rock pigeons bearing passenger pigeon sperm and eggs. The offspring of these would have passenger pigeon traits, and would be further bred to favor unique features of the extinct species. Reviving the passenger pigeon may help conserve woodland biodiversity in the eastern US.\n", "The adult chicken has also made significant contributions to the advancement of science. By inoculating chickens with cholera bacteria (Pasteurella multocida) from an overgrown, and thereby attenuated, culture Louis Pasteur produced the first lab-derived attenuated vaccine (1860s). Great advances in immunology and oncology continued to characterize the 20th century, for which we indebted to the chicken model.\n" ]
what are city states and how do they operate?
Are you asking how the mechanics of city states works in the context of Civ 5? Or are you asking what city states in real life are? Two really different questions.
[ "In California, the words \"town\" and \"city\" are synonymous by law (see Cal. Govt. Code Secs. 34500–34504). There are two types of cities in California: charter and general law. Cities organized as charter cities derive their authority from a charter that they draft and file with the state, and which, among other things, states the municipality's name as \"City of (Name)\" or \"Town of (Name).\" Government Code Sections 34500–34504 applies to cities organized as general law cities, which differ from charter cities in that they do not have charters but instead operate with the powers conferred them by the pertinent sections of the Government Code. Like charter cities, general law cities may incorporate as \"City of (Name)\" or \"Town of (Name).\"\n", "In the systems of local government in some U.S. states, a general-law municipality, general-law city, or statutory city is a municipality whose government structure and powers are defined by the general law of its state. This is contrast to a charter city or home-rule city, whose government structure and powers are defined by a municipal charter.\n", "The state is divided into counties, cities, towns, and villages. Cities, towns and villages are municipal corporations with their own governments that provide most local government services. Whether a municipality is defined as a city, town, or village is dependent not on population or land area, but rather on the form of government selected by the residents and approved by the state legislature. Each such government is granted varying home rule powers as provided by the New York Constitution. New York has various corporate entities that serve single purposes that are also local governments, such as school and fire districts.\n", "A city-state is a sovereign state, also described as a type of small independent country, that usually consists of a single city and its dependent territories. Historically, this included cities such as Rome, Athens, Carthage, and the Italian city-states during the Renaissance. , only a handful of sovereign city-states exist, with some disagreement as to which are city-states. A great deal of consensus exists that the term properly applies currently to Monaco, Singapore, and Vatican City. City states are also sometimes called microstates which however also includes other configurations of very small countries, not to be confused with micronations.\n", "Out of 53 cities in the Commonwealth, there are now eleven that are legally cities and have city councils, but retained \"Town of\" in their names. This distinction derives from provisions of state law that reference the town meeting form of government and that provide for greater self-governance authority for the class of communities governed by a form of chief executive and council which are referred to as cities in state law.\n", "State governments are sovereign entities which use their powers of taxation both to match federal grants, and provide for local transportation needs. Different states have different systems for dividing responsibility for funding and maintaining road and transit networks between the state department of transportation, counties, municipalities, and other entities. Typically cities or counties are responsible for local roads, financed with block grants and local property taxes, and the state is responsible for major roads that receive state and federal designations. Many mass transit agencies are quasi-independent and subsidized branches of a state, county, or city government.\n", "Different states use a variety of terms for their incorporated places. The designations \"city\", \"town\", \"village\", and \"borough\" are most frequent, but one or more places in Kentucky, Montana, Nevada, and Tennessee have place-type governments (usually consolidated ones) that do not have any of these designations. New Jersey is the only state that has all four kinds of incorporated places. Only two other states (Connecticut and Pennsylvania) include \"boroughs\" as incorporated places. Eleven U.S. states have only \"cities\", and the remainder of the states have various combinations of \"cities\", \"towns\", and \"villages\". \n" ]
Was it common for non-Abrahamic religions to have a "bible" of sorts?
There are actually several. The thing is, Judaism, Christianity, and Islam all emphasize a "closed canon". Buddhism, for example, has a canon (though its often more of a library than a single book). But then again, Jews don't have a single book ("Torah is only understandable with Talmud") and Islam emphasizes generally emphasizes the importance of Hadiths in addition to the Quran. Several offshoots of these (like Mormonism and Druzism, and to a degree Bahaiism) use the core texts of the parent religion, plus additional texts so there's not just "one". * The Zoroastriansn have not only the Avesta, but several other core texts beyond that like the Gathas and the Yasna. * Hindus have the Vedas, and a variety of other holy books of wide acceptance and considerable antiquity (the Vedas are surprisingly unifying--in fact, when discussing India religion it's common to say what became Hinduism was Vedic, and what became Jainism, Buddhism, etc was non-Vedic) * Sikhism has the Guru Granth Sahib * If you want to count Confucianism as a religion, you have the Analects. * If you want to count Daoism as a religion, there's the Dao De Ching of Lao Tse (please pardon my mixing and matching romanizations; I'm on a phone that's cutting in and out of service so I can't look anything up) * Buddhism is complicated. Therevada had the Pali Canon, but Mahayana schools continued producing texts for a much much longer time. * I believe Jainism has a central set of texts as well. I'm going to assume you're in the west and from a Christian, likely Protestant, milieu. While these groups all have a canon of holy books agreed on (to various degrees), it should not be mistaken that , should say Zoroastrians disappear, we could recreate Zoroastrianism only from the texts before us in the holy books. Whether or not you could actually "rederive" Protestantism (or Salafi Islam) from a given set of holy books is beside the point--the religions think they are performing the authentic textual religion (if its not in the text, it's not in the religion--this is Martin Luther's *sola scriptura*). In almost every other religious tradition, recieved tradition is openly given a much more central role (like Catholicism). These books give one part of the religion and they are understood within a particular tradition. It should not be assumed that they explain all of the religion.
[ "In the study of comparative religion, the category of Abrahamic religions consists of the three monotheistic religions, Christianity, Islam and Judaism, which claim Abraham (Hebrew \"Avraham\" אַבְרָהָם; Arabic \"Ibrahim\" إبراهيم ) as a part of their sacred history. Smaller religions such as Bahá'í Faith that fit this description are sometimes included but are often omitted.\n", "The three major Abrahamic faiths (in chronological order of revelation) are Judaism, Christianity and Islam. Some strict definitions of what constitutes an Abrahamic religion include only these three faiths. However, there are many other religions incorporating Abrahamic doctrine, theology, genealogy and history into their own belief systems.\n", "Historically, the Abrahamic religions have been considered to be Judaism, Christianity and Islam. Some of this is due to the age and larger size of these three. The other, similar religions were seen as either too new to judge as being truly in the same class, or too small to be of significance to the category.\n", "Advocates of the term \"Abrahamic religion\" since the second half of the 20th century have proposed a hyper-ecumenicism that emphasizes not only Judeo-Christian commonalities but that would include Islam as well (the rationale for the term \"Abrahamic\" being that while only Christianity and Judaism give the Hebrew Bible (Old Testament) the status of scripture, Islam does also trace its origins to the figure of Abraham as the \"first Muslim\").\n", "Judaism, Christianity, and Islam are sometimes called Abrahamic religions because they all accept the tradition of the God (known as Yahweh in Hebrew and Allah in Arabic) that revealed himself to the prophet Abraham. The theological traditions of all Abrahamic religions are thus to some extent influenced by the depiction of the God of Israel in the Hebrew Bible, and the historical development of monotheism in the history of Judaism.\n", "Abrahamic religions are those religions deriving from a common ancient tradition and traced by their adherents to Abraham (circa 1900 BCE), a patriarch whose life is narrated in the Hebrew Bible/Old Testament, where he is described as a prophet (Genesis 20:7), and in the Quran, where he also appears as a prophet. This forms a large group of related largely monotheistic religions, generally held to include Judaism, Christianity, and Islam, and comprises over half of the world's religious adherents.\n", "Abrahamic religions are those monotheistic faiths emphasizing and tracing their common origin to Abraham or recognizing a spiritual tradition identified with him. They constitute one of three major divisions in comparative religion, along with Indian religions (Dharmic) and East Asian religions (Taoic).\n" ]
Does the recipient of a heart transplant inherit the resting heart rate of the donor?
Only sort of. After a transplant, the heart is generally not connected to the nerves that regulate rate. But cardiac muscle is not like skeletal muscle, in that the nerves do not directly initiate contraction, they just speed it up or slow it down. The heart itself has autorhythmicity. Not connected to those nerves, the heart will beat at its intrinsic rate, which is usually around 100 bpm.
[ "Full recovery of donor heart function often takes place over hours or days, during which time considerable damage can occur. Other deaths to patients can occur from preexisting conditions. For example, in pulmonary hypertension the patient's right ventricle has often adapted to the higher pressure over time and, although diseased and hypertrophied, is often capable of maintaining circulation to the lungs. Barnard designed the idea of the heterotopic (or \"piggy back\" transplant) in which the patient's diseased heart is left in place while the donor heart is added, essentially forming a \"double heart\". Barnard performed the first such heterotopic heart transplant in 1974.\n", "Some respiratory patients may also have severe cardiac disease which would necessitate a heart transplant. These patients can be treated by a surgery in which both lungs and the heart are replaced by organs from a donor or donors.\n", "On 13 February 2011, Copeland and his team at the UC San Diego Center for Transplantation, performed the rare \"piggyback\" heart operation again, resulting in a man having two beating hearts in his chest. Termed a heterotopic heart transplantation, recipient Smith received a donor heart while still keeping his own diseased heart. The alternative option, which would have required two operations, was to offer a Left ventricular assist device (LVAD) as a bridge-to-transplantation.\n", "The success rate of heart–lung transplants has improved significantly in recent years. The British National Health Service states that the survival rate is now around 85%, one year after the transplant was performed.\n", "Of the twelve hundred procedures performed at University of Pittsburgh Medical Center since 1980, only fifteen to twenty fell into this category. Although many heart transplant recipients must wait several months, even years for a donor heart, after only two weeks on the heart transplant waiting list, a donor heart became available. Although the donor heart was less than perfect as it exhibited signs of a cardiovascular disease, Kim chose, in his words, ”to take a chance”.\n", "Approximately 3,500 heart transplants are performed every year in the world, more than half of which occur in the US. Post-operation survival periods average 15 years. Heart transplantation is not considered to be a cure for heart disease; with that being said, it still is a life-saving treatment intended to improve the quality, and hopefully also both the overall total duration of time and the extent of time afterwards in a more active state, of life for the recipients.\n", "BULLET::::- The first successful human heart-lung transplant was performed. A team at the Stanford University Medical Center, led by Dr. Bruce Reitz, used a new technique that retained a portion of the recipient's right atrium. The recipient was Mary Gohlke, a 45-year-old woman from Mesa, Arizona, with end-stage primary pulmonary hypertension. The donor was a 15-year-old boy who had died from severe head trauma two days earlier\n" ]
what would be the effect, both positive and negative, if nobody on the planet had kids for 2 years.
Anyone working in childcare or childbirth would need to find new jobs for a while.
[ "There is also the question as to whether having children really is such a positive contribution to the world in an age when there are many concerns about overpopulation, pollution and depletion of non-renewable resources. Some critics counter that such analyses of having children may understate its potential benefits to society (e.g. a greater labor force, which may provide greater opportunity to solve social problems) and overstate the costs. That is, there is often a need for a non-zero birth rate.\n", "Brazil displayed a remarkable reduction in the rates of child stunting under age 5, from 37% in 1974, to 7.1% in 2007. This happened in association with impressive social and economic development that reduced the numbers of Brazilians living in extreme poverty (less than $1.25 per day) from 25.6% in 1990 to 4.8% in 2008. The successful reduction in child malnutrition in Brazil can be attributed to strong political commitment that led to improvements in the water and sanitation system, increased female schooling, scale-up of quality maternal and child health services, increased economic power at family level (including successful cash transfer programs), and improvements in food security throughout the country.\n", "Intrusive fears of harming immediate children can last longer than the postpartum period. A study of 100 clinically depressed women found that 41% had obsessive fears that they might harm their child, and some were afraid to care for their children. Among non-depressed mothers, the study found 7% had thoughts of harming their child—a rate that yields an additional 280,000 non-depressed mothers in the United States with intrusive thoughts about harming their children.\n", "A 2014 report by the FYA found that people under the age of 24 were likely to be worse off than their parents, with a 30% unemployment rate and more university debt and spending most income on housing.\n", "The counter-argument is that in many poor countries people get far more children than they can reasonably feed, support and provide with a decent life under the circumstances. If people had fewer children with the aid of contraception, then it would put much less pressure on already scarce resources and enable a better life for the living.\n", "However, in 2013, child poverty reached record high levels in the United States, with 16.7 million children, more than 20%, living in food insecure households. 47 million Americans depend on food banks, more than 30% above 2007 levels. Households headed by single mothers are most likely to be affected. Worst affected are the District of Columbia, Oregon, Arizona, New Mexico and Florida, while North Dakota, New Hampshire, Virginia, Minnesota and Massachusetts are the least affected.\n", "There is no clear social impact case for or against conceiving a child. Individually, for most people, bearing a child or not has no measurable impact on person well-being. A review of the economic literature on life satisfaction shows that certain groups of people are much happier without children:\n" ]
how does red dot sights work?
The dot in a red or green dot sight is not actually a coloured point that's hovering in a fix spot in the air like an iron sight does. Instead, the red dot you see is a reflection of a little red lamp that falls onto a spherical mirror which filters out only the red (or green) spectrum. This mirror is installed in a way that no matter what angle *you* look on it, the Red dot will always point towards the spot the bullet will hit. EDIT: That way the Red dot itself is virtual and basically only exists in your eyes, two people looking through the same sight at the same time from different angles would see the spots in different 3-dimensional places, but still pointed towards the target.
[ "A red dot sight is a common classification for a type of non-magnifying reflector (or reflex) sight for firearms, and other devices that require aiming, that gives the user an aimpoint in the form of an illuminated red dot. A standard design uses a red light-emitting diode (LED) at the focus of collimating optics which generates a dot style illuminated reticle that stays in alignment with the weapon the sight is attached to regardless of eye position (nearly parallax free). They are considered to be fast acquisition and easy to use gun sights for target shooting, hunting, and in police and military applications. Aside from firearm applications, they are also used on cameras and telescopes. On cameras they are used to photograph flying aircraft, birds in flight, and other distant, quickly moving subjects. Telescopes have a narrow field of view and therefore are often equipped with a secondary \"finder scope\" such as a red dot sight.\n", "Red dot sights place the target and the reticle on nearly the same optical plane, allowing a single point of focus. This makes them fast acquisition and easy to use sights, allowing the user to keep their attention on the field of view in front of them. They are common in speed shooting sports such as IPSC. Military units and police forces have also adopted them. Red dot sights are also popular among paintball and airsoft players for similar reasons.\n", "The most common reticles used today in red dot sights both for handguns and rifles are small dots covering between 0.6 to 1.6 mil (2 to 5 moa ). The choice of red dot reticle size depends on the user needs. A larger and brighter red dot makes for faster target acquisition, but may obscure the target and thereby inhibit precise aiming, while a smaller and dim dot allows for more precise but slower aiming. The 1.6 mil (5 moa) dot is small enough not to obscure most handgun targets, and large enough for most competition shooters to quickly acquire a proper sight picture. Red dots for rifles typically have a smaller dot, often 0.6 to 0.8 mil (2 to 3 moa). When red dot sights started to appear at the practical shooting competition circuit in the 1990s, reticle sizes of up to 3, 4.5 or even 6 mil (10, 15 or 20 moa) were common in order to compensate for the lack of bright illumination. However, as red dot technology and production quality has advanced, the market trend in all types of sport shooting has gone towards the smaller dots used today.\n", "Like other reflector sights, the collimated image of the red dot is truly parallax free only at infinity, with an error circle equal to the diameter of the collimating optics for any target at a finite distance. This is compensated for by keeping the dot in the middle of the optical window (sighting down the sight's optical axis). Some manufacturers modify the focus of the LED/optical collimator combination, making models with the optical collimator set to focus the dot at a finite distance. These have a maximum amount of parallax due to eye movement, equal to the size of the optical window, at close range, diminishing to a minimal amount at the set distance (somewhere around a desired target range of 25–50 yards).\n", "BULLET::::- Three-dot: On semi-automatic handguns, the most common type of enhancement is a bright white dot painted on the front sight near the top of the blade, and a dot on each side of the rear sight notch. In low lighting conditions the front sight dot is centered horizontally between the rear sight dots, with the target placed above the middle (front) dot. Some sight vendors offer differently colored dots for the front and rear sights.\n", "Sights that use dot reticles are almost invariably measured in minutes of arc, sometimes called \"minutes of angle\" or \"moa\". Moa is a convenient measure for shooters using English units, since 1 moa subtends approximately at a distance of , which makes moa a handy unit to use in ballistics. The 5 moa (1.5 milliradian) dot is small enough not to obscure most targets, and large enough to quickly acquire a proper \"sight picture\". For many types of action shooting, a larger dot has traditionally been preferred; 7, 10, 15 or even 20 moa (2, 3, 4.5 or 6 mil) have been used; often these will be combined with horizontal and/or vertical lines to provide a level reference.\n", "Another example of devices that show the PIP are red dot sights like the M68 Aimpoint. Such sights, like on a HUD's, are collimated reflector sights, so the dot always appears over the weapon's impact point, regardless of the shooter's eye position. Red dot sights do not use internal computers and must be manually zeroed for maximum accuracy.\n" ]
What kind of tasks would fill the day of someone running a medieval castle for his lord?
Are you referring to somebody such as, say, a steward? Somebody who would manage the daily affairs of the estate while the lord was away? I'm going to base this on the holdings of a knight such a sub fief beneath a baron. A lord with one castle/stronghouse, and some moderate to small land holdings. There would have been two different sorts of people to fill this role. In the direct lands controlled by the lord, there would have been the field master, or a title similar. There may have been several depending on the size of the lands. This person would have been in charge of the other field hands, ensuring that they were doing the work that they were supposed to, doing it well, and not slacking. The second person would have been the steward. This person was in charge of "the books" as it were. Ensuring that the finances were in order. They were the medieval version of an accountant. Of course, this all assumes that the Lord had the money to pay for these people. Much like later farmers of the Antebellum South period, smaller landowners didn't always have the means to pay these individuals, who were skilled in their own right. If the Lord could read and write, he might well keep his own books, to save money.
[ "Early on, manning a castle was a feudal duty of vassals to their magnates, and magnates to their kings, however this was later replaced with paid forces. A garrison was usually commanded by a constable whose peacetime role would have been looking after the castle in the owner's absence. Under him would have been knights who by benefit of their military training would have acted as a type of officer class. Below them were archers and bowmen, whose role was to prevent the enemy reaching the walls as can be seen by the positioning of arrowslits.\n", "To regain the castle, the Goodies fight the Town Planner and his clerks by archery, swordsmanship and jousting (with the Goodies riding their trandem, instead of a horse). The Goodies and their also fight other medieval contests and, surprisingly, the Goodies end up winning the 'battle' — so they are able to hold on to the castle for Tim's relatives.\n", "The game allows players to build outlandish medieval siege engines to pit against castles or armies. Players select from a collection of mechanical parts that can be connected together to build a machine. Each level has a goal, such as \"destroy the windmill\" or \"kill 100 soldiers\". Although the goals are relatively simple, the wide variety of possible approaches allows for experimentation.\n", "The castle folk (, or \"civis\") formed a class of freemen who were obliged to provide well-specified services to a royal castle and its \"ispán\", or count, in the medieval Kingdom of Hungary. They were peasants living in villages formed in the lands pertaining to the royal castle. They tilled their estates collectively.\n", "In the game, the player takes on the role of a lord who rules over a medieval castle. With his available resources, the player places buildings or features, including many different kinds of food production, industry, civil, or military buildings and defences. Available peasants automatically choose jobs whenever a building requires one, so player micromanagement is minimal, he mostly needs to set up the various buildings in an efficient way while providing safety for his peasants. Military units are directly controlled individually or in groups, sometimes quite large with sieges or battles involving many hundreds on each side. One new addition to the original stronghold is the inclusion of estates, that a player can \"buy\" with his accumulated honor (gained by popularity, holding feasts, dances, jousting, etc.). Estates are semi-independent villages (without castle fortifications) that produce their own goods that the owner can send through carts to his castle or his allies. \n", "The players must take up control of four medieval lords who are fighting for control of the land; they have to battle through eight different castles, and try to gain control of each one. When one of the lords wins a battle, he shall receive a medal - and once he has won three medals, he will gain control of the current castle (which means that they can be up to six rooms long) and the game will proceed to the next castle. The gameplay is also similar to that of Hudson Soft's \"Bomberman\" series, the initial title of which was adapted by Irem for the arcade in 1991; the lords start out in the four corners of the screen and can only be killed by bombs, but a strike from the whips will temporarily stun them. The objective is to trap the other lords with bombs and kill them or otherwise force them to walk into the path of an exploding bomb - and treasure chests are numerous in each room, and can be broken with the whip. Each treasure chest contains a powerup (such as speed shoes, boomerangs, extra bombs, larger bombs, and more); the game's name is also a reference to the fact that it is a parody of Konami's \"Castlevania\" series or \"Akumajō Dracula\" as it was originally known in Japan.\n", "Castle-guard was an arrangement under the feudal system, by which the duty of finding knights to guard royal castles was imposed on certain manors, knight's fees or baronies. The greater barons provided for the guard of their castles by exacting a similar duty from their sub-enfeoffed knights. The obligation was commuted very early for a fixed money payment, a form of scutage known as \"castle-guard rent\", which lasted into modern times. Castle-guard was a common form of feudal tenure, almost ubiquitous, on the Isle of Wight where all manors were held from the Lord of the Isle of Wight, seated at Carisbrook Castle.\n" ]
stem cell life science research
I'd love to answer your question if you could be more specific as to what it is you're asking.
[ "Laboratory of Stem Cell Biology—studies cell fate and tissue development. It examines various types of stem cells to generate neurological disease models, to discover and test drugs, and to develop replacement therapies for neurodegenerative diseases and disorders.\n", "More specifically the way that stem cell research has been involved in neuroscience is through the treatment of neurodegenerative diseases and brain tumors. In these cases scientists are using neural stem cells to regenerate tissue and to be used as carriers for gene therapy. In general, neuroethics revolves around a cost benefit approach to find techniques and technologies that are most beneficial to patients. There has been progress in certain fields that have been shown to be beneficial when using stem cells to treat certain neurodegenerative diseases such as Parkinson's disease.\n", "There have been significant developments in stem cell research activity during this period. In 1997, Dr. John Dick, a molecular biologist at the University of Toronto, was the first to discover the existence of cancer stem cells. The Stem Cell Network was established in 2001 with headquarters at the University of Ottawa, and brings together more than 80 leading scientists, clinicians and engineers from Canadian universities and hospitals. Researchers study cellular therapeutics and their pharmacological applications as well as related technologies, public policy, ethical, legal and social issues with the goal of effectively treating cancer, heart and lung disease, macular degeneration, stroke, multiple sclerosis, spinal cord injury, Parkinson's disease, muscle degeneration, hemophilia and type 1 diabetes. It is hoped that research will lead to clinical applications for these afflictions by 2015. Stem cell research is also undertaken at the McEwen Centre for Regenerative Medicine established in 2003 in Toronto as part of the University Health Network in 2003. In 2010, the McMaster University Stem Cell and Cancer Research Institute announced that it had developed a technique for transforming skin cells into multiple blood cell types. This discovery may have applications for the treatment of leukemia, for it is anticipated that a patient with the disease may be able to receive therapeutic blood transfusions derived from his or her own skin cells, thus eliminating problems related to compatibility that are associated with treatment that involves biological material from others.\n", "Stem cells are cells in the body that can divide indefinitely in a culture and can be formulated into specialized cells. In biological research, these cells have become a subject of extensive research. With these cells, scientists will be able to better understand how these cells are differentiated by the process of turning on and off genes. Through this research, scientists will be able to better understand certain diseases such as cancer and how these diseases arise.\n", "The Canadian Stem Cell Foundation is an independent, non-profit organization established in 2008 and situated in Ottawa, Ontario. Stem Cell science is a Canadian innovation through the discovery of stem cells by Drs. James Till and Ernest McCulloch. It is globally known as the leading organization for stem cell research and support in the study of treatments and cures for diseases such as cancer, diabetes, blindness and stroke.\n", "Stem Cell Research is a peer-reviewed open access scientific journals of cell biology. The journal was established in 2007, and is currently published 8 times per year by Elsevier. The current editor-in-chief is Thomas Zwaka (Icahn School of Medicine at Mount Sinai).\n", "As stem cell research continues, scientists will need a database to store and share their information and research. Three key components are necessary for this database to be effective: cell type-specific gene expression profiles, anatomical and developmental relationships between cells and tissues and signals important for development and differentiation of stem cells to mature cell types.\n" ]
What happened to the House of Kalākaua or the Hawaiian royal family after their overthrow? Were there ever pretenders to that throne or anyone who tried to stir up political trouble afterwards?
Queen Liliʻuokalani continued to resist the overthrow with political, PR and judiciary actions for many years. The royalist rebellion of 1895 had participants from the Hawaiian Royal Family, most notably Prince Kuhio. As an interesting side note, Kuhio defeated the leader of the rebellion, Wilcox in an election for the House of Representatives in 1902. "The Rights of My People: Liliuokalani's Enduring Battle with the United States 1893-1917", Neil Thomas Proto
[ "It served as the official residence of the Hawaiian monarch until the 1893 overthrow of the Kingdom of Hawaii. Therein not only Liliuokalani, but, Queen Kapiolani and other royal retainers were evicted from the palace after the overthrow.\n", "After Kalākaua's death and the accession of Queen Liliʻuokalani, Iaukea was reappointed colonel of the queen's personal military staff, and as agent of the Crown Lands. The monarchy was overthrown on January 17, 1893, by the Committee of Safety, with the support of United States Minister John L. Stevens, and the landing of American forces from the USS \"Boston\". After a brief transition under the Provisional Government, the oligarchical Republic of Hawaii was established on July4, 1894, with Sanford B. Dole as president.\n", "On January 17, 1893, the monarchy was overthrown and Queen Liliʻuokalani was deposed by the Committee of Safety with the covert support of United States Minister John L. Stevens and the landing of American forces from the USS \"Boston\". After a brief transition under the Provisional Government, the oligarchical Republic of Hawaii was established on July 4, 1894, with Sanford B. Dole as president. During this period, the defacto government was composed largely of residents of American and European ancestry, although a few Native Hawaiians including Kūnuiākea served in political roles. He became a member of the Constitutional Convention charged with drafting a new constitution for the Republic and was one of the five Native Hawaiian signatories of the document.\n", "The House of Kamehameha \"(Hale O Kamehameha)\", or the Kamehameha dynasty, was the reigning Royal Family of the Kingdom of Hawaii, beginning with its founding by Kamehameha I in 1795 and ending with the death of Kamehameha V in 1872 and Lunalilo in 1874. The kingdom would continue for another 21 years until its overthrow in 1893 with the fall of the House of Kalakaua.\n", "Following the proroguing of the legislature and the unsuccessful attempts of the queen to promulgate a new constitution, the monarchy was overthrow on January 17, 1893. After a brief transition under the Provisional Government, the oligarchical Republic of Hawaii was established on July 4, 1894. During this period, the de facto government, which was composed largely of residents of American and European ancestry, sought to annex the islands to the United States against the wish of the Native Hawaiians who wanted to remain an independent nation and for the monarchy to continue.\n", "Following the proroguing of the legislature and the unsuccessful attempts of the queen to promulgate a new constitution, the monarchy was overthrown on January 17, 1893. After a brief transition under the Provisional Government, the oligarchical Republic of Hawaii was established on July 4, 1894. During this period, the de facto government, which was composed largely of residents of American and European ancestry, sought to annex the islands to the United States against the wish of the Native Hawaiians who wanted to remain an independent nation and for the monarchy to continue. \n", "On January 17, 1893, Queen Liliuokalani was overthrown and replaced by a provisional government composed of members of the Committee of Safety. The United States Minister to the Kingdom of Hawaii (John L. Stevens) conspired with U.S. citizens to overthrow the monarchy. After the overthrow, Lawyer Sanford B. Dole, a citizen of Hawaii, became President of the Republic when the Provisional Government of Hawaii ended on July 4, 1894. Controversy ensued in the following years as the Queen tried to regain her throne. The administration of President Grover Cleveland commissioned the Blount Report, which concluded that the removal of Liliuokalani had been illegal. The U.S. government first demanded that Queen Liliuokalani be reinstated, but the Provisional Government refused.\n" ]
Did the Romans ever used pikes?
Pike phalanxes were not standard infantry, but were adopted numerous times, often when fighting cavalry-heavy armies from the east. In preparation for his Parthian campaign Caracalla ordered 16,000 men to be trained in Macedonian drill and fight in a phalanx. Severus Alexander also had elements of six legions formed into phalangiarii, who fought in a phalanx. Arrian of Nicomedia, during his campaign against the Alani, used pike phalanxes to ward off the Alani cavalry. From Arrian's "Extaxis contra Alanos" (Section 15) The infantry should be drawn up eight ranks deep, in a closely packed formation. (Section 16) The first four ranks shall consist of pike-bearers, whose pikes will end in long, slender iron points. The men in the first rank should hold their pikes at the ready, so that if the enemy comes near, they can thrust the iron tip of their pike especially at the breasts of the horses. (Section 17) The men of the second (?), third, and fourth ranks should hold their pikes forward to jab and wound the horses where they can, and kill the riders.
[ "The pike was a long weapon, varying considerably in size, from long. It was approximately in weight, with sixteenth-century military writer Sir John Smythe recommending lighter rather than heavier pikes. It had a wooden shaft with an iron or steel spearhead affixed. The shaft near the head was often reinforced with metal strips called \"cheeks\" or langets. When the troops of opposing armies both carried the pike, it often grew in a sort of arms race, getting longer in both shaft and head length to give one side's pikemen an edge in combat. The extreme length of such weapons required a strong wood such as well-seasoned ash for the pole, which was tapered towards the point to prevent the pike from sagging on the ends, although drooping or slight flection of the shaft was always a problem in pike handling. It is a common mistake to refer to a bladed polearm as a pike; such weapons are more generally halberds, glaives or voulges.\n", "Pikes live on today only in ceremonial roles, being used to carry the colours of an infantry regiment and with the Company of Pikemen and Musketeers of the Honourable Artillery Company, or by some of the infantry units on duty during their rotation as guard for the President of the Italian Republic at the Quirinale, in Rome.\n", "A pike is a pole weapon, a very long thrusting spear formerly used extensively by infantry. Pikes were used regularly in European warfare from the Late Middle Ages to the early 18th century, and were wielded by foot soldiers deployed in close quarters, until their replacement by the bayonet. The pike found extensive use with Landsknecht armies and Swiss mercenaries, who employed it as their main weapon and used it in pike square formations. A similar weapon, the sarissa, was also used by Alexander the Great's Macedonian phalanx infantry to great effect. Generally, a spear becomes a pike when it is too long to be wielded with one hand in combat.\n", "The pike, being unwieldy, was typically used in a deliberate, defensive manner, often alongside other missile and melee weapons. However, better-trained troops were capable of using the pike in an aggressive attack with each rank of pikemen being trained to hold their pikes so that they presented enemy infantry with four or five layers of spearheads bristling from the front of the formation.\n", "A common end date for the use of the pike in infantry formations is 1700, although such armies as the Prussian and Austrian had already abandoned the pike by that date. Other armies, such as the Swedish and the Russian, continued to use it for several decades afterward (the Swedes of King Charles XII in particular using it to great effect until the 1720s and 1730s). During the American Revolution, pikes called \"trench spears\" made by local blacksmiths saw limited use until enough bayonets could be procured for general use by both Continental Army and attached militia units.\n", "The use of massed pikes by Spanish armies began in the Granada War (1482–92). During the Italian Wars, under the direction of the Spanish general Gonzalo Fernández de Córdoba, called \"the great captain\", the system of combined groups of pikemen, arquebusiers and swordsmen developed. The conflicts at the end of the 15th century and early 16th century evolved into a tactically unique combination of combined arms centered around armored infantry. To counter the French heavy cavalry, a colonelcy could theoretically have up to 6,000 men, but by 1534 this had been reduced to the tercio with a maximum of 3,000.\n", "Civilian pikeman played a similar role, though outnumbered and outgunned, in the 1798 Rising in Ireland four years later. Here, especially in the Wexford Rebellion and in Dublin, the pike was useful mainly as a weapon by men and women fighting on foot against cavalry armed with guns.\n" ]
what exactly does shifting my car into d3 do?
It's for driving in the mountains. D will allow your car to go into the highest gear... your car may not have enough power to maintain a consistent speed in this gear, so it will drop into a lower gear to get back up to speed. This cycle can be annoying. D3 prevents you from going to the highest gear. Additionally, when going down a hill, shifting into D3 will use a lower gear for more effective engine breaking.
[ "With the DVC system, the car’s bodywork moves independently from the wheels and engine. The DVC system automatically adjusts the tilt angle of the vehicle to the speed and acceleration of the vehicle in such a way that it always is in balance. It was designed as a proactive instead of a reactive system, meaning a Carver tilts before cornering, much like when an airplane tilts to one side in order to turn in that direction.\n", "BULLET::::- Reverse (R): This engages reverse gear within the transmission, permitting the vehicle to be driven backward, and operates a switch to turn on the white backup lights for improved visibility (the switch may also activate a beeper on delivery trucks or other large vehicles to audibly warn other drivers and nearby pedestrians of the driver's reverse movement). To select reverse in most transmissions, the driver must come to a complete stop, depress the shift-lock button (or move the shift lever toward the driver in a column shifter, or move the shifter sideways along a notched channel in a console shifter) and select reverse. The driver should avoid engaging reverse while the vehicle is moving forwards, and likewise avoid engaging any forward gear while travelling backwards. On transmissions with a torque converter, doing so at very low speed (walking pace) is not harmful, but causes unnecessary wear on clutches and bands, and a sudden deceleration that not only is uncomfortable, but also uncontrollable since the brakes and the throttle contribute in the same direction. This sudden acceleration, or jerk, can still be felt when engaging the gear at standstill, but the driver normally suppresses this by holding the brakes. Travelling slowly in the right direction while engaging the gear minimizes the jerk further, which is actually beneficial to the wearing parts of the transmission. Electronically controlled transmissions may behave differently, as engaging a gear at speed is essentially undefined behaviour. Some modern transmissions have a safety mechanism that will resist putting the car in reverse when the vehicle is moving forward; such a mechanism may consist of a solenoid-controlled physical barrier on either side of the reverse position, electronically engaged by a switch on the brake pedal, so that the brake pedal needs to be depressed in order to allow the selection of reverse. Some electronic transmissions prevent or delay engagement of reverse gear altogether while the car is moving.\n", "Under M mode, sequential gear shifts can be made by changing the gear to M+ and M-. Under M mode, the gear selected will be displayed as one of M1/M2/M3/M4/M5/M6 on the dashboard. Under M mode, the gear will not change even if the engine RPM reaches redline, which will probably lead to engine damage. However, the user is prevented from excessively up-shifting to a gear that exceeds the maximum revolution possible for the engine, or excessively downshifting to a gear that may lead the engine to stall. When the car is braked to a complete stop, the M1 gear is automatically selected.\n", "The three-point turn (sometimes called a Y-turn, K-turn, or broken U-turn) is the standard method of turning a vehicle around to face the opposite direction in a limited space, using forward and reverse gears. This is typically done when the road is too narrow for a U-turn.\n", "More commonly, reversing is used as an intermediate step, to complete a three point turn, J-turn, parallel park, or similar maneuver. These moves are used with the goal of positioning the vehicle in a specific way under certain space restrictions, that would not be possible to achieve whilst moving forward.\n", "From this false park position, slight movements in the vehicle, vibration, or the build up of hydraulic pressure in the transmission can then cause the vehicle to reengage powered reverse after a delay from a few seconds to longer periods of time (what is called a \"self shift\"). This will cause the vehicle to suddenly and without warning move backwards unexpectedly under engine power.\n", "Newer versions of the system allow parallel or reverse parking. When parallel parking with the system, drivers first pull up alongside the parking space. They move forward until the vehicle's rear bumper passes the rear wheel of the car parked in front of the open space. Then, shifting to reverse automatically activates the backup camera system, and the car's rear view appears on dash navigation/camera display. The driver's selection of the parallel park guidance button on the navigation/camera touchscreen causes a grid to appear (with green or red lines, a flag symbol representing the corner of the parking spot, and adjustment arrows).\n" ]
can sports at an early age effect testosterone levels and how does this effect the physical appearance of someone who didn't play sports at all?
The different sports will cause different things to happen in your body while you're growing up. Yea you're genetics make you predisposed to doing well in sports or not but the type of sport will definitely effect you. The more heavy lifting and explosive workouts you do the more testosterone you will develop and the more manly you get. If you are a endurance athlete cortisol levels spike and you get more feminine traits. I used to be a very good swimmer in highschool and college and the exercise plus my diet gave me moobs. I knew a gymnast girl who has just as broad shoulders and a square jaw like me. You train hard enough your can change your body while you're an adolescent.
[ "While a high level of testosterone is often associated with an increase in aggression, this is not a noticeable effect in most trans men. HRT doses of testosterone are much lower than the typical doses taken by steroid-using athletes, and create testosterone levels comparable to those of most cisgender men. These levels of testosterone have not been proven to cause more aggression than comparable levels of estrogen. It is assumed that the effect of the start of physical treatment is such a relief, and decreases pre-existing aggression so much, that the overall level of aggression actually decreases.\n", "Studies of testosterone levels of male athletes before and after a competition revealed that testosterone levels rise shortly before their matches, as if in anticipation of the competition, and are dependent on the outcome of the event: testosterone levels of winners are high relative to those of losers. No specific response of testosterone levels to competition was observed in female athletes, although a mood difference was noted. In addition, some experiments have failed to find a relationship between testosterone levels and aggression in humans.\n", "Recently some studies have been influenced by an evolutionary psychology perspective. This includes studies on testosterone changes in sports which at least for males are similar to those in status conflicts in non-human primates with testosterone levels increasing and decreasing as an individual's status changes. A decreased testosterone level may decrease dominant and competitive behaviors which when the status conflicts involved fighting may have been important for preventing physical injury to the loser as further competition is avoided. Testosterone levels also increase before sports competitions, in particular if the event is perceived as real challenge as compared to not being important. Testosterone may also be involved in the home advantage effect which has similarities to animal defense of their home territory. In some sports there is a marked overrepresentation of left-handedness which has similarities to left-handed likely having an advantage in close combat which may have evolutionary explanations.\n", "Even though hyperandrogenism is not common in men, there have been studies done to look at the effects of high levels of testosterone in males. A study has shown that even though many of the male participants did not have behavior changes due to the increased levels of testosterone, there were cases where the participants had instances of uncharacteristic aggression. High levels of testosterone in males have not been seen to have a direct impact on their personality, but within those studies, there have been cases of sudden aggression within the male participants.\n", "BULLET::::- There is no physiological basis for the belief that having sex in the days leading up to a sporting event or contest is detrimental to performance. In fact it has been suggested that sex prior to sports activity can elevate male testosterone level, which could potentially enhance performance.\n", "Pubertal development also affects circulatory and respiratory systems as an adolescents' heart and lungs increase in both size and capacity. These changes lead to increased strength and tolerance for exercise. Sex differences are apparent as males tend to develop \"larger hearts and lungs, higher systolic blood pressure, a lower resting heart rate, a greater capacity for carrying oxygen to the blood, a greater power for neutralizing the chemical products of muscular exercise, higher blood hemoglobin and more red blood cells\".\n", "According to a study, girls who participate in girls-only activities are far less likely to experience a teenage pregnancy and less likely to be sexually active in general. Participating in competitive sports has also shown to have an effect for girls. A study published in 1999 found that female adolescents who participated in sports were less likely than their non-athletic peers to engage in sexual activity and/or report a pregnancy. Males interested in arts are also less likely to be involved in a pregnancy situation. It is unclear whether these correlations are causal or the reflection of the underlying bias of the considered population. The study that reported these findings did not take into account the sexual orientation of the subjects.\n" ]
What happens when we overdose?
Heroin overdose is similar to any opiate overdose. Opiates depress the central nervous system causing a relaxed, "euphoric" sensation. After the initial rush, breathing becomes more shallow, decreasing oxygen to the brain and rest of the body. Without oxygen, the brain will start shutting down systems, including the nervous system. The individual will feel extremely drowsy and slip into a coma state. At this point, the nervous system is so relaxed that it fails to function. The individual goes into respiratory arrest (completely stop breathing). Once this occurs, no oxygen is being brought into the body and systems shut down and death occurs shortly after. TLDR: Opiates relax the nervous system. Heroin overdose would be the same sensation as being so drowsy that you fall asleep.
[ "Overdose is a method of suicide which involves taking medication in doses greater than the indicated levels, or in a combination that will interact either to cause harmful effects or increase the potency of one or other of the substances.\n", "A drug overdose (or simply overdose or OD) is the ingestion or application of a drug or other substance in quantities greater than are recommended. Typically it is used for cases when a risk to health will potentially result. An overdose may result in a toxic state or death.\n", "The word \"overdose\" implies that there is a common safe dosage and usage for the drug; therefore, the term is commonly only applied to drugs, not poisons, even though some poisons are harmless at a low enough dosage. Drug overdose is caused to commit suicide, the result of intentional or unintentional misuse of medication. Intentional misuse leading to overdose can include using prescribed or unprescribed drugs in excessive quantities in an attempt to produce euphoria.\n", "An overdose produces symptoms similar to alcohol poisoning and is a medical emergency due to the sedative/depressant properties which manifest in overdose as potentially lethal respiratory depression. The oral in rats is 1 g/kg. The subcutaneous LD in mice is 2.1 g/kg.\n", "A large overdose can cause asphyxia and death by respiratory depression if the person does not receive medical attention immediately. Overdose treatment includes the administration of naloxone. The latter completely reverses morphine's effects, but may result in immediate onset of withdrawal in opiate-addicted subjects. Multiple doses may be needed.\n", "Reports of overdose in medical literature are generally from abuse, and often involve other drugs as well. Symptoms include vomiting, excessive sweating, coma, periods of stopped breathing, seizures, agitation, loss of psychomotor skills, and coma. Overdose can lead to death due to respiratory depression. People who overdose may die from asphyxiation from their own vomit. People that have overdosed or suspected of overdosing may need to be made to vomit, be intubated, or/and put on a respirator.\n", "The most common effects of overdose are drowsiness and tachycardia. Rare but potentially critical complications are cardiac arrest, abnormal heart rhythms, severe low blood pressure, seizures, and neuroleptic malignant syndrome. Life-threatening overdose is rare, however, as the median lethal dose is about 338 milligrams/kilogram in mice and 425 mg/kg in rats. The potential harm is increased when central nervous system depressants and antidepressants are also used; deliberate overdose often includes alcohol among other drugs.\n" ]
In the Ku Klux Klan's heyday, did people see it more as a social club with networking opportunities, or truly as an organization that defending white American interests?
If you're referring to the Second Klan (1915-1930's), it was certainly both. William Joseph Simmons *Invisible Empire of the Ku Klux Klan* was a religious, political, business and fraternal organization. Created during the Golden Age of Fraternal Orders, it was modeled on the multitude of existing fraternal orders like the Freemasons and the Elks, Simmons would often recruit members into the new Klan wearing the badges signifying he was a member of other organizations and Klansmen who recruited new members (Kleagles) specifically sought new members through Fraternal Orders they were already a part of. In addition to this, the Klan meeting was highly ceremonial and even had [assigned room placements](_URL_2_) for each member. Even in the KKK rule manual, the Kloran, (no, I have no clue why they would have modeled it on the Qur'an) it states firstly, "We avow the disctinction between the races of mankind...and we shall ever be true in the faithful maintenance of White Supremacy" followed almost immediately by "We appreciate the intrinsic value of a real practical fraternal relationship." Politically speaking, the Second Klan was Nativist, [anti-Catholic](_URL_1_), anti-black, [anti-communist](_URL_0_), anti-Jewish and Prohibitionist. It absolutely was more dedicated to keeping White American hegemony. The Second Klan was almost certainly more a brotherhood and social organization than the First Klan (1868-1871) who were dedicated to restoring white supremacy to the Reconstruction South, and the Third Klan (1950s-1970s) who were dedicated to preventing black progress during the Civil Rights Era.
[ "Initially the KKK presented itself as another fraternal organization devoted to betterment of its members. The KKK's revival was inspired in part by the movie \"Birth of a Nation\", which glorified the earlier Klan and dramatized the racist stereotypes concerning blacks of that era. The Klan focused on political mobilization, which allowed it to gain power in states such as Indiana, on a platform that combined racism with anti-immigrant, anti-Semitic, anti-Catholic and anti-union rhetoric, but also supported lynching. It reached its peak of membership and influence about 1925, declining rapidly afterward as opponents mobilized.\n", "In the late 1910s, the Ku Klux Klan was first revived in Atlanta. It gradually expanded into northern and midwestern cities, where anxieties about migration, immigration, and social changes had heightened because of rapid industrialization and movement of peoples. The KKK promoted itself as a fraternal organization, among many that had been started since the late 19th century. In this period, it was primarily opposed to Catholic and Jewish immigrants, but kept some of its racist background. A chapter was started in Orleans. A 1918 photograph shows children at the old Opera House, a number of them dressed in KKK hoods, and others in blackface.\n", "A rapid pace of change across the country, especially in growing cities, combined with new waves of immigration and migration of rural whites and blacks to cities, all contributed to a volatile social environment and the rise of a second Ku Klux Klan (KKK) in the South and Midwest after 1915. In many areas it represented itself as a fraternal group to give aid to a community. Feldman (1999) has shown that the second KKK was not a mere hate group; it showed a genuine desire for political and social reform on behalf of poor whites. For example, Alabama Klansmen such as Hugo Black were among the foremost advocates of better public schools, effective Prohibition enforcement, expanded road construction, and other \"progressive\" measures to benefit poor whites. By 1925, the Klan was a powerful political force in the state, as urban politicians such as J. Thomas Heflin, David Bibb Graves, and Hugo Black manipulated the KKK membership against the power of the \"Big Mule\" industrialists and especially the Black Belt planters who had long dominated the state.\n", "Ku Klux Klan (KKK) is the name of three entirely different organizations (1860s, 1920s, post 1960) that used the same nomenclature and costumes but had no direct connection. The KKK of the 1920s was a purification movement that rallied against crime, especially violation of prohibition, and decried the growing \"influence\" of \"big-city\" Catholics and Jews, many of them immigrants and their descendants from Ireland as well as Southern and Eastern Europe. Its membership was often exaggerated but possibly reached as many as 4 million men, but no prominent national figure claimed membership; no daily newspaper endorsed it, and indeed most actively opposed the Klan. Membership was verily evenly spread across the nation's white Protestants, North, West, and South, urban and rural. Historians in recent years have explored the Klan in depth. The KKK of the 1860s and the current KKK were indeed violent. However, historians discount lurid tales of a murderous group in the 1920s. Some crimes were probably committed in Deep South states but were quite uncommon elsewhere. The local Klans seem to have been poorly organized and were exploited as money-making devices by organizers more than anything else. (Organizers charged a $10 application fee and up to $50 for costumes.) Nonetheless, the KKK had become prominent enough that it staged a huge rally in Washington DC in 1925. Soon afterward, the national headlines reported rape and murder by the KKK leader in Indiana, and the group quickly lost its mystique and nearly all its members.\n", "In an early 20th-century revival, the Ku Klux Klan recruited new members as a fraternal organization, opposing new immigrants from southern and eastern Europe, and becoming influential in rapidly industrializing urban areas in the Midwest and West, as well as in the South. It had numerous members in Pekin and other Illinois cities. It was during this period that leading Klansmen took over ownership of the city newspaper, the \"Pekin Daily Times\"; they used it as an organ of Klan propaganda. They sold off the paper within a few years.\n", "The KKK was a nationwide organization that grew rapidly from 1921 to 1925, then collapsed just as fast. It had millions of members, but the organizational structure was in oriented entirely toward recruiting new members, collecting their initiation fees and selling costumes, so that the actual organizations seldom achieved very much. The Klan signed up millions of white Protestant men on the basis that American society needed a moral purification against the immoral power of the Catholic Church, Jews, organized crime, speakeasies, and local adulterers. Liberal and Catholic elements fought against the Klan primarily inside the Democratic Party, where a motion to repudiated it by name was defeated by one vote at the national convention of 1924. Though members of the KKK swore to uphold American values and Christian morality, and at the local level some Protestant ministers became involved, no Protestant denomination officially endorsed the KKK. Across the country, the Klan had strength in cities, towns and rural areas. Its high visibility made its intended targets of Catholics, Jews and blacks highly uncomfortable, and many local and state elections were bitterly fought around the role of the Klan. Historians in recent decades have totally revised the traditional interpretation of the second KKK as a terrorist group, or based on frustrated marginal elements. Using newly discovered minutes of local chapters they now portray the Klansmen as ordinary Americans, primarily middle class. They were local activists and joiners who thought the nation was seriously threatened by an evil conspiracy. Scholars found the Klan was \"composed of average citizens drawn from the broader middle-class.\" It \"had been not a form of racial and religious terror but rather a means of mainstream political activism.\"\n", "In the mid-1920s, the Ku Klux Klan (KKK) reached its peak membership in the South and Midwest after a revival beginning around 1915. Its growth was due in part to tensions from rapid industrialization and social change in many growing cities; in the Midwest and West, its growth was related to the competition of waves of new immigrants from Southern and Eastern Europe. The KKK was strong in the Florida cities of Jacksonville and Tampa; Miami's chapter was influential enough to hold initiations at the Miami Country Club. The Klan also flourished in smaller towns of the South where racial violence had a long tradition dating back to the Reconstruction era. An editor of \"The Gainesville Daily Sun\" admitted that he was a member of the Klan in 1922, and praised the organization in print.\n" ]
why some of websites is not available in some countries?
It has nothing to do with the websites, but rather, the service that the website provides. Netflix can't just take people movies and make them available for everybody to watch, that's illegal copyright infringement. They need to purchase the rights to distribute them, and most distribution agreements come with a particular region.
[ "Some governments, such as those of Burma, Iran, North Korea, the Mainland China, Saudi Arabia and the United Arab Emirates restrict access to content on the Internet within their territories, especially to political and religious content, with domain name and keyword filters.\n", "The international channels and associated websites are run by BBC Studios. As a result, not all of them are the same and some channels have less extensive websites than other services. CBeebies channels in Asia, Australia, Poland, South Africa and the USA all have their own international variant.\n", "Governments are also getting online. Some countries, such as those of Cuba, Iran, North Korea, Myanmar, the People's Republic of China, and Saudi Arabia use filtering and censoring software to restrict what people in their countries can access on the Internet. In the United Kingdom, they also use software to locate and arrest various individuals they perceive as a threat. Other countries including the United States, have enacted laws making the possession or distribution of certain material such as child pornography illegal but do not use filtering software. In some countries Internet service providers have agreed to restrict access to sites listed by police.\n", "Websites can have many functions and can be used in various fashions; a website can be a personal website, a corporate website for a company, a government website, an organization website, etc. Websites are typically dedicated to a particular topic or purpose, ranging from entertainment and social networking to providing news and education. All publicly accessible websites collectively constitute the World Wide Web, while private websites, such as a company's website for its employees, are typically part of an intranet.\n", "Websites can have many functions and can be used in various fashions; a website can be a personal website, a corporate website for a company, a government website, an organization website, etc. Websites are typically dedicated to a particular topic or purpose, ranging from entertainment and social networking to providing news and education. All publicly accessible websites collectively constitute the World Wide Web, while private websites, such as a company's website for its employees, are typically a part of an intranet.\n", "No longer an international phenomenon, the Internet has become quite different for its users in different countries. One reason for differentiation across nations is that users want information presented in their local language and context. Advertisers likewise want to present information to interested audiences, which tend to be geographically specific (58–63).\n", "The U.S. is the only country that has a government-specific top-level domain in addition to its country-code top-level domain. This is a result of the origins of the Internet as a U.S. federal government-sponsored research network. Other countries typically delegate a second-level domain for this purpose; for example, .gc.ca is the second-level domain for the Government of Canada and all subdomains.\n" ]
Was it because of a change in technology or a change in tactics, that machine guns no longer need to change barrels or be watercooled?
All general purpose machine guns (GPMGs) that I am aware of retain the interchangeable barrel function - a GPMG being a roughly 7.62mm belt-fed platoon- or company-level machine gun capable of laying down large volumes of fire. The M240 and MG3, the most common GPMGs in the western world, certainly have changeable barrels, and soldiers are taught to do so when laying down large volumes of suppressing fire. Even the intermediate-caliber M249 Squad Automatic Weapon has a quick change barrel. Now, there are some lighter automatic weapons that do not have a changeable barrel. But that's nothing new. The Browning Automatic Rifle, Chauchat, and Lewis Gun did not have changeable barrels. It generally doesn't become necessary to change a barrel until you've fired several hundred rounds through the gun, or less if you're firing very rapidly. Weapons that don't have quick-change barrels just aren't intended to produce the same volume of fire as a proper heavy/medium/GP machine gun.
[ "Machineguns used in fixed defensive positions sometimes use water cooling to extend barrel life through periods of rapid fire, but the weight of the water and pumping system significantly reduces the portability of water-cooled firearms.\n", "The use of guns using blank cartridges to eject water from the barrel at high velocity was well known as a means of disrupting relatively small and thin-skinned targets with a relatively low probability of causing detonation or ignition. They were, however, easily defeated by the use of more robust cases, such as those made from structural steel tubes. The replacement of the copper liner of conventional shaped charges first by gelled water, then by low melting point salts with high proportions of water of crystallisation and finally with thin-walled plastic containers of water, provided a simple means of generating jets of water of much higher velocity and with correspondingly enhanced penetrating power and disruptive capability than was possible using gun barrel technology. Concomitant experiments demonstrated that, in contrast to metal-lined charges, shaped charges projecting liquids still functioned with their cavities filled rather than just lined with water.\n", "This type of multipurpose machine gun would be further developed, and later given names such as \"universal machine gun\", and later \"general-purpose machine gun\", and would eventually supplant the water-cooled designs. These later designs used quick-change barrel replacement to reduce overheating, which further reduced the weapon's weight, but at the cost of increasing the soldier's load due to the extra barrels. Some earlier designs like the Vickers had this feature, but it was mainly for barrel wear, as they normally used water cooling. It was in the 1920s and 1930s that quick barrel replacement for cooling purposes became more popular in weapons such as the ZB vz. 30, the Bren, the MG34 and the MG42.\n", "Most proposed advances in gun technology are based on the assumption that the solid propellant as a stand-alone propulsion system is no longer capable of delivering the required muzzle energy. This requirement has been underscored by the appearance of the Russian T-90 main battle tank. Even the elongation of current gun tubes, such as the new German 120 mm L/55, which was introduced by Rheinmetall is considered only an interim solution as it does not offer the required increase in muzzle velocity. Even advanced kinetic energy ammunition such as the United States' M-829A3 is considered only an interim solution against future threats. To that extent the solid propellant is considered to have reached the end of its usefulness, although it will remain the principal propulsion method for at least the next decade until newer technologies mature. To improve on the capabilities of a solid propellant weapon the electrothermal-chemical gun may see production as early as 2016.\n", "During the 1980s, the motorized water gun was perhaps at its most prolific. Companies such as Entertech and Larami created water guns modeled after guns popularized in movies such as \"Rambo\". At the heart of these devices was a small motor and crank shaft that converted a rotary motion into a forward-backwards pumping motion to drive a small pump akin to those found in the small spray bottle-type squirt pistols. Stream performance was often not improved, but the motor removed the need to pump, which made the toys popular. The greatest fallbacks was that they wore out batteries quickly and that many trigger action blasters could outdo them, making motorized blasters the lowest class. However, their main strength—and consequent reason for dismissal—was their realistic styling. After some of these realistic-shaped water guns caused accidental shootings by police, stricter rules regarding shapes and coloring of water guns were drafted in the United States.\n", "As the bullet enters the barrel, it inserts itself into the rifling, a process that gradually wears down the barrel, and also causes the barrel to heat up more rapidly. Therefore, some machine guns are equipped with quick-change barrels that can be swapped every few thousand rounds, or in earlier designs, were water-cooled. Unlike older carbon steel barrels, which were limited to around 1,000 shots before the extreme heat caused accuracy to fade, modern stainless steel barrels for target rifles are much more resistant to wear, allowing many thousands of rounds to be fired before accuracy drops. (Many shotguns and small arms have chrome-lined barrels to reduce wear and enhance corrosion resistance. This is rare on rifles designed for extreme accuracy, as the plating process is difficult and liable to reduce the effect of the rifling.) Modern ammunition has a hardened lead core with a softer outer cladding or jacket, typically of an alloy of copper and nickel – cupro-nickel. Some ammunition is even coated with molybdenum disulfide to further reduce internal friction – the so-called 'moly-coated' bullet.\n", "The mechanism of the Maxim gun employed one of the earliest recoil-operated firing systems in history. The idea is that the energy from recoil acting on the breech block is used to eject each spent cartridge and insert the next one, instead of a hand-operated mechanism. Maxim's earliest designs used a 360-degree rotating cam to reverse the movement of the block, but this was later simplified to a toggle lock. This made it vastly more efficient and less labor-intensive than previous rapid-firing guns, such as the Mitrailleuse, Gatling, Gardner, or Nordenfelt, that relied on actual mechanical cranking. The Maxim gun design was provided with water cooling, giving it the ability to maintain its rate of fire for far longer than air-cooled guns. The disadvantage of this was that it made the gun less flexible in attack than the lighter air-cooled weapons, being heavier and more complex, and requiring a supply of water.\n" ]
Are quarks shaped differently or are they all spheres?
As far as we know, they are all pointlike. This could understanding could one day be improved upon.
[ "Quarks are very small things that make up everything we see (matter). There are six different \"flavors\" of quarks: up, down, strange, charm, bottom, and top. Quarks also have three \"colors\": red, green, and blue. There are also antiquarks, which are the opposite of the regular quarks. In total, there are 18 different types of regular quarks, and 18 different types of antiquarks. Quarks are known as the \"building blocks of matter\" because they are the smallest thing that make up all the matter in the Universe.\n", "Quarks are fermionic particles of spin (\"S\" = ). Because spin projections vary in increments of 1 (that is 1 ħ), a single quark has a spin vector of length , and has two spin projections (\"S\" = + and \"S\" = −). Two quarks can have their spins aligned, in which case the two spin vectors add to make a vector of length \"S\" = 1 and three spin projections (\"S\" = +1, \"S\" = 0, and \"S\" = −1). If two quarks have unaligned spins, the spin vectors add up to make a vector of length \"S\" = 0 and has only one spin projection (\"S\" = 0), etc. Since baryons are made of three quarks, their spin vectors can add to make a vector of length \"S\" = , which has four spin projections (\"S\" = +, \"S\" = +, \"S\" = −, and \"S\" = −), or a vector of length \"S\" =  with two spin projections (\"S\" = +, and \"S\" = −).\n", "is a transformation that simultaneously turns all the up quarks in the universe into down quarks and vice versa. More specifically, these flavour rotations are exact symmetries if \"only\" strong force interactions are looked at, but they are not truly exact symmetries of the universe because the three quarks have different masses and different electroweak interactions.\n", "Quarks are fermions—specifically in this case, particles having spin (\"S\" = ). Because spin projections vary in increments of 1 (that is 1 \"ħ\"), a single quark has a spin vector of length , and has two spin projections (\"S\" = + and \"S\" = −). Two quarks can have their spins aligned, in which case the two spin vectors add to make a vector of length \"S\" = 1 and three spin projections (\"S\" = +1, \"S\" = 0, and \"S\" = −1), called the spin-1 triplet. If two quarks have unaligned spins, the spin vectors add up to make a vector of length S = 0 and only one spin projection (\"S\" = 0), called the spin-0 singlet. Because mesons are made of one quark and one antiquark, they can be found in triplet and singlet spin states.\n", "The Quarks were rectangular, with four arms: one pair which folded into the body, the other pair being retractable. On the end of each arm was a solitary claw. The spherical head was divided into octants; the upper four octants formed the \"sensory hemisphere\", which detected changes in light, heat and motion. At five of the corners of the octants were directional crystal beam transmitters (the sixth corner joined with the robot's extremely short neck). Quarks communicated by means of high-pitched sounds. Their tendency to run out of energy quickly was their primary weakness.\n", "There are six types, known as \"flavors\", of quarks: up, down, strange, charm, bottom, and top. Up and down quarks have the lowest masses of all quarks. The heavier quarks rapidly change into up and down quarks through a process of particle decay: the transformation from a higher mass state to a lower mass state. Because of this, up and down quarks are generally stable and the most common in the universe, whereas strange, charm, bottom, and top quarks can only be produced in high energy collisions (such as those involving cosmic rays and in particle accelerators). For every quark flavor there is a corresponding type of antiparticle, known as an antiquark, that differs from the quark only in that some of its properties (such as the electric charge) have equal magnitude but opposite sign.\n", "A quark is a type of elementary particle that has mass, electric charge, and colour charge, as well as an additional property called flavour, which describes what type of quark it is (up, down, strange, charm, top, or bottom). Due to an effect known as colour confinement, quarks are never seen on their own. Instead, they form composite particles known as hadrons so that their colour charges cancel out. Hadrons made of one quark and one antiquark are known as mesons, while those made of three quarks are known as baryons. These 'regular' hadrons are well documented and characterized; however, there is nothing in theory to prevent quarks from forming 'exotic' hadrons such as tetraquarks with two quarks and two antiquarks, or pentaquarks with four quarks and one antiquark.\n" ]
why do blue laws still exist?
Many people are still religious and support such laws, so they remain. It is legal to make a religiously motivated law, just not to establish mandatory religion.
[ "In the United States, the U.S. Supreme Court has held blue laws as constitutional numerous times, citing secular bases such as securing a day of rest for mail carriers, as well as protecting workers and families, in turn contributing to societal stability and guaranteeing the free exercise of religion. The origin of the blue laws also partially stems from religion, particularly the prohibition of Sabbath desecration in Christian Churches following the first-day Sabbatarian tradition. Both labour unions and trade associations have historically supported the legislation of blue laws. Most blue laws have been repealed in the United States, although some states ban the sale of alcoholic beverages on Sundays and many states ban selling cars on Sundays.\n", "In the United States, the U.S. Supreme Court has held blue laws as constitutional numerous times, citing secular bases such as securing a day of rest for mail carriers, as well as protecting workers and families, in turn contributing to societal stability and guaranteeing the free exercise of religion. The origin of the blue laws also partially stems from religion, particularly the prohibition of Sabbath desecration in Christian Churches following the first-day Sabbatarian tradition. Both labour unions and trade associations have historically supported the legislation of blue laws. Most blue laws have been repealed in the United States, although Indiana banned the sale of alcoholic beverages on Sundays until the law was repealed on February 28, 2018, and many states ban selling cars on Sundays.\n", "Contrary to popular belief, there is no evidence that blue laws were originally printed on blue paper. Rather, the word \"blue\" was used in the 17th century as a disparaging reference to rigid moral codes and those who observed them, particularly in \"blue-stocking\", a reference to Oliver Cromwell's supporters in the parliament of 1653. Moreover, although Reverend Peters claimed that the term \"blue law\" was originally used by Puritan colonists, his work has since been found to be unreliable. In any event, Peters never asserted that the blue laws were originally printed on blue paper, and this has come to be regarded as an example of false etymology, another version of which is that the laws were first bound in books with blue covers.\n", "BULLET::::- A blue law is a type of law, typically found in the United States and Canada, designed to enforce religious standards, particularly the observance of Sunday as a day of worship or rest, and a restriction on Sunday shopping.\n", "Another feature of blue laws in the United States restricts the purchase of particular items on Sundays. Some of these laws restrict the ability to buy cars, groceries, office supplies, and housewares among other things. Though most of these laws have been relaxed or repealed in most states, they are still enforced in some other states.\n", "In the United States, judges have defended blue laws \"in terms of their secular benefit to workers\", holding that \"the laws were essential to social well-being\". Chief Justice Stephen Johnson Field, with regard to Sunday blue laws, stated:\n", "This was one of the four cases decided in 1961 that declared blue laws to be constitutional. The other three were \"Braunfeld v. Brown\", \"Two Guys from Harrison-Allentown, Inc., v. McGinley\" and \"McGowan v. Maryland\"\n" ]
the symptoms of a common cold
A common cold has not much to do with actual temperature. It is an infection, mostly by viruses. There are dozens if not more viruses known that can lead to the symptoms commonly associated with a cold. Now, how does a virus infect us? Normally your skin is pretty tight and prevents any pathogen (something that will make you sick) from entering. So, skin drops out. Theses viruses get mostly transmitted by flying in little airborne water droplets. Therefore they can enter your body when you inhale the droplets while breating. In your mouth and and airways there is also a protection layer, the mucosa. This is a thin layer that produces mucus (who would have thought?). This slime hopefully prevents the pathogens from entering and catches them. Your airways have little hairs that sweep the mucus towards your throat and into the stomach were the acid deals with them. Sometimes this doesnt work. It is suspected that cold, dry air strains your mucosa and therefore makes it easier for the virus to enter your body and infect it. It spreads from the point of entrance, aka your airways. However, since it's not much of a threat it will not get very far. This is where your symptoms stem from. Your body sents defensive cells that prevent the virus from spreading and that fight it. This leads to an inflammation of the tissue. Some simply eat viruses or bacteria (*phagocytes*), others release enzyms that damage invaders or molecules that manage the inflammation. Acute inflammation has several effects on the body, fever and loss of appetite being some of them. They are caused on a molecular level by the proteins that are released by the defensive cells. The 5 classical signs of inflammation are pain, heat, redness, swelling, and loss of function. We can definetly find some of that in a swollen, sore throat, cant we? Now, one by one: * swelling fighting needs a lot of ressources and they come from the blood. In order to get them into the tissue blood vessels dilate and get permeable. This is nice but also leads to fluid accumulating in the tissue. * heat and redness Increased and prioritised blood flow into the area (dilated blood vessels, remember?) naturally leads to higher temperature and redness * pain That's of course the one you mostly feel. The defensive cells also release chemicals that stimulate the nerve endings in the area. Stimulated nerve endings mean you feel pain. Inflammation of organs may not always be combined with pain because there might be no sensible nerves. * loss of function This can have multiple causes and doesnt matter that much in the case of the common cold. Okay, but how does all that lead to my symptoms? Inflammated mucosa produces lots of mucus. Not a good combination together with swelling. If the mucus is thin you might just get a runny nose, otherwise a stuffed-up nose. The sinuses may fill with mucus if the nose is swollen and lead to a feeling of pressure below your eyes. A feeling of pressure within your ears can also occur. This is cause by swelling of the [Eustachian tube](_URL_0_) that connects the middle ear and throat to balance pressure differences. Coughing gets caused, who would have thought, by increased production of mucus in your airways. It gets called phlegm then. It contains dead cells, different proteins and much more. The composition determines what colour it has. Fluid and lungs dont work together very well and your body tries to get the phlegm out of your lungs by coughing. The "normal yucky felling" is just a response of your body to the inflammation. Holy shit, this got longer than I expected. I hope I got everything and that I got it all right. Had to look up a lot of things (1 year medstudent) but I think overall it's more or less okay. Ask if you got any more questions and somebody with more knowledge might chip in and correct me if necessary.
[ "The common cold, also known simply as a cold, is a viral infectious disease of the upper respiratory tract that primarily affects the nose. The throat, sinuses, and larynx may also be affected. Signs and symptoms may appear less than two days after exposure to the virus. These may include coughing, sore throat, runny nose, sneezing, headache, and fever. People usually recover in seven to ten days, but some symptoms may last up to three weeks. Occasionally those with other health problems may develop pneumonia.\n", "The typical symptoms of a cold include cough, runny nose, sneezing, nasal congestion, and a sore throat, sometimes accompanied by muscle ache, fatigue, headache, and loss of appetite. A sore throat is present in about 40% of cases and a cough in about 50%, while muscle ache occurs in about half. In adults, a fever is generally not present but it is common in infants and young children. The cough is usually mild compared to that accompanying influenza. While a cough and a fever indicate a higher likelihood of influenza in adults, a great deal of similarity exists between these two conditions. A number of the viruses that cause the common cold may also result in asymptomatic infections.\n", "It can be difficult to distinguish between the common cold and influenza in the early stages of these infections. Influenza symptoms are a mixture of symptoms of common cold and pneumonia, body ache, headache, and fatigue. Diarrhea is not usually a symptom of influenza in adults, although it has been seen in some human cases of the H5N1 \"bird flu\" and can be a symptom in children. The symptoms most reliably seen in influenza are shown in the adjacent table.\n", "Along with HCoV-229E, a species in the Alphacoronavirus genus, HCoV-OC43 are among the known viruses that cause the common cold. Both viruses can cause severe lower respiratory tract infections, including pneumonia in infants, the elderly, and immunocompromised individuals such as those undergoing chemotherapy and those with HIV-AIDS.\n", "The common cold is a viral infection of the upper respiratory tract. The most commonly implicated virus is a rhinovirus (30–80%), a type of picornavirus with 99 known serotypes. Other commonly implicated viruses include human coronavirus (≈ 15%), influenza viruses (10–15%), adenoviruses (5%), human respiratory syncytial virus, enteroviruses other than rhinoviruses, human parainfluenza viruses, and metapneumovirus. Frequently more than one virus is present. In total over 200 viral types are associated with colds.\n", "BULLET::::- Common cold: rhinovirus, coronavirus, respiratory syncytial virus, and parainfluenza virus can cause infection of the throat, ear, and lungs causing standard cold-like symptoms and often pain.\n", "Symptoms commonly include fever, shivering, chills, malaise, dry cough, loss of appetite, body aches, and nausea, typically in connection with a sudden onset of illness. In most cases, the symptoms are caused by cytokines released by immune system activation, and are thus relatively non-specific.\n" ]
nasdaq, dow jones, etc.
They are stock indexes. Basically they are each used as a measure of a certain "basket" of stocks. There are thousands of different indexes used to track different market sectors. The Dow, Nasdaq and S & P 500 are the three most well know in US
[ "The Dow Jones Industrial Average (DJIA), or simply the Dow (), is a stock market index that indicates the value of 30 large, publicly owned companies based in the United States, and how they have traded in the stock market during various periods of time.\n", "The NASDAQ is an electronic exchange, where all of the trading is done over a computer network. The process is similar to the New York Stock Exchange. One or more NASDAQ market makers will always provide a bid and ask price at which they will always purchase or sell 'their' stock.\n", "Nasdaq, Inc. is an American multinational financial services corporation that owns and operates (and is listed on) the NASDAQ stock market and eight European Stock Exchanges, including the Armenian Stock Exchange, Copenhagen Stock Exchange, Helsinki Stock Exchange, Iceland Stock Exchange, Riga Stock Exchange, Stockholm Stock Exchange, Tallinn Stock Exchange, and NASDAQ OMX Vilnius. It is headquartered in New York City, and its president and chief executive officer is Adena Friedman. The Swedish investment company Investor AB has been a major shareholder since 2011.\n", "The NASDAQ-100 (^NDX) is a stock market index made up of 103 equity securities issued by 100 of the largest non-financial companies listed on the NASDAQ. It is a modified capitalization-weighted index. The stocks' weights in the index are based on their market capitalizations, with certain rules capping the influence of the largest components. It is based on exchange, and it is not an index of U.S.-based companies. It does not have any financial companies, since these were put in a separate index. Both of those criteria differentiate it from the Dow Jones Industrial Average, and the exclusion of financial companies distinguishes it from the S&P 500.\n", "The NASDAQ Financial-100 is a specialty index operated by NASDAQ that follows stocks in the Financial services index. It was created in 1985 as the sister index to the more widely followed NASDAQ-100. As designed by NASDAQ, this index follows companies that engage in banking, insurance, mortgages, securities trading and real estate. It also has, among its components, exchanges, including NASDAQ itself.\n", "Over the years, the Nasdaq Stock Market became more of a stock market by adding trade and volume reporting and automated trading systems. It was also the first stock market in the United States to trade online, highlighting Nasdaq-traded companies and closing with the declaration that the Nasdaq Stock Market is \"the stock market for the next hundred years\". The Nasdaq Stock Market attracted new growth companies, including Microsoft, Apple, Cisco, Oracle and Dell, and it helped modernize the IPO.\n", "The Nasdaq Stock Market (, also known as Nasdaq) is an American stock exchange. It is the second-largest stock exchange in the world by market capitalization, behind only the New York Stock Exchange located in the same city. The exchange platform is owned by Nasdaq, Inc., which also owns the Nasdaq Nordic (formerly known as OMX) and Nasdaq Baltic stock market network and several U.S. stock and options exchanges.\n" ]
Were soldiers traumatized or "shell-shocked" after the Napoleonic Wars or other earlier wars? How were returning soldiers treated by the public? Were they treated for their condition?
Diagnosing mental illness retrospectively is a minefield. With respect to soldiers in history and their relationship with PTSD, I would refer you to the FAQ or past threads like [this](_URL_0_). Paging /u/Iphikrates to comply with the rules on linking past threads.
[ "There were so many officers and men suffering from shell shock that 19 British military hospitals were wholly devoted to the treatment of cases. Ten years after the war, 65,000 veterans of the war were still receiving treatment for it in Britain. In France it was possible to visit aged shell shock victims in hospital in 1960.\n", "At first, shell-shock casualties were rapidly evacuated from the front line – in part because of fear of their unpredictable behaviour. As the size of the British Expeditionary Force increased, and manpower became in shorter supply, the number of shell shock cases became a growing problem for the military authorities. At the Battle of the Somme in 1916, as many as 40% of casualties were shell-shocked, resulting in concern about an epidemic of psychiatric casualties, which could not be afforded in either military or financial terms.\n", "In World War I, shell shock was considered a psychiatric illness resulting from injury to the nerves during combat. The horrors of trench warfare meant that about 10% of the fighting soldiers were killed (compared to 4.5% during World War II) and the total proportion of troops who became casualties (killed or wounded) was 56%. Whether a shell-shock sufferer was considered \"wounded\" or \"sick\" depended on the circumstances. When faced with the phenomenon of a minority of soldiers mentally breaking down, there was an expectation that the root of this problem lay in character of the individual soldier, not because of what they experienced on the front lines during the war. These sorts of attitudes helped fuel the main argument that was accepted after the war and going forward that there was a social root to shell shock that consisted of soldiers finding the only way allowed by the military to show weakness and get out of the front, claiming that their mental anguish constituted a legitimate medical diagnosis as a disease. The large proportion of World War I veterans in the European population meant that the symptoms were common to the culture.\n", "By the Battle of Passchendaele in 1917, the British Army had developed methods to reduce shell shock. A man who began to show shell-shock symptoms was best given a few days' rest by his local medical officer. Col. Rogers, Regimental Medical Officer, 4th Battalion Black Watch wrote:\n", "1812 – Napoleonic wars give rise to the military medical practice of triage in an effort to sort wounded soldiers in those to receive medical treatment and return to battle and those whose injuries are non-survivable. Dominique-Jean Larrey, a surgeon in the French emperor’s army, not only conceives of taking care of the wounded on the battlefield, but creates the concept of ambulances, collecting the wounded in horse-drawn wagons and taking them to military hospitals.\n", "Casualties during the Boer War were generally recovered in stages. Battalion doctors located on the battlefields would take the wounded by stretcher to a dressing station for initial treatment. This included the widespread use of field dressings consisting of a sterile gauze pad stitched to the bandage and covered with waterproofing. Each pack contained two dressings and numerous safety pins. Later they would be taken to ambulances or stationary hospitals where the complex recovery by the surgeons could be carried out. However, given the number of casualties in relation to staff, it was often days before a wounded soldier could see a doctor for full treatment, increasing the extent of the infection of the wounds. As a result, primary surgical care was often conducted in ambulances. Many wounds were open fractures and often involved damage to bones. In such cases, the wounds were treated by debridement and the wound packed. The limb would then be immobilised with a splint made of canvas with strips of bamboo sewn in to support it.Plaster of Paris was also used during the Boer War after the bone had been set. The treatment of wounds was greatly enhanced by the invention of X-rays and some 9 machines were taken to South Africa during the campaign. \n", "There is also evidence to suggest that the type of warfare faced by soldiers would affect the probability of shell shock symptoms developing. First hand reports from medical doctors at the time note that rates of such afflictions decreased once the war was mobilized again during the 1918 German offences, following the 1916-1917 period where the highest rates of shell shock can be found. This could suggest that it was trench warfare, and the experience of siege warfare specifically, that led to the development these symptoms.\n" ]
What were the relationships of the various intelligence services of the Eastern Bloc with the KGB/Moscow?
There wasn't a formal hierarchy, but the KGB loomed large over all of the Warsaw Pact intelligence services, and its primacy was undisputed. Soviet intelligence personnel had a hand in creating all of the Eastern Bloc secret services, and KGB liaison officers were a constant and powerful presence in those agencies' affairs. The Warsaw Pact intelligence agencies provided a constant stream of intelligence to the Lubyanka, along with extensive operational and logistical support. The KGB maintained a shared database of contacts and intelligence called SOUD, enabling them to receive and index intelligence from the Warsaw Pact services (and to share intelligence with them in turn — something they generally did only in the most circumspect manner.) The KGB frequently used those services as their proxies, especially when it came to handling sensitive or potentially compromising relationships or operations — particularly those with terrorist and revolutionary groups. The Czechoslovak StB, for example, was a longstanding liaison between Moscow Centre and the Italian Red Brigades; the Bulgarian DS was a favourite weapon for 'special tasks'. The relationship with the Stasi (in particular the foreign intelligence department, HVA), however, was by far the closest and most important to Moscow. Many of the senior figures in the Stasi had longstanding ties to the USSR, as agents or operatives of Soviet intelligence or the Comintern: Wilhelm Zaisser, Ernst Wollweber and Erich Mielke, three of the GDR's four Ministers of State Security; Markus Wolf, the longtime chief of the HVA. As with many of the institutions of East Germany, the Soviets were heavily involved in the creation of the Stasi (which was closely patterned on the KGB.) The KGB maintained [a number of facilities](_URL_1_) in the GDR, including highly important *rezidentura* (the largest KGB installation outside the USSR) at Karlshorst. KGB officers regarded East Germany as a safe, even dull posting; as a First Chief Directorate officer in the late 1980s, Vladimir Putin was reportedly disappointed to be assigned to the *rezidentura* at Dresden, rather than a more interesting and adventurous station in the West. Part of the reason the Stasi were so important to Moscow was their proximity and access to West Germany and the rest of Western Europe, but it was also thanks to a general reputation for competence and ideological reliability, and a willingness to subordinate their foreign policy/intelligence objectives to those of Moscow. They were thus also entrusted with operations further afield. The Stasi worked alongside the KGB in Cuba, training Castro's intelligence service, DGI — which, in turn, was the preferred Soviet proxy for operations in Latin America. In 1972, when Egyptian strongman Anwar Sadat expelled Soviet (read: KGB/GRU) advisors, the Stasi acted as Soviet surrogates in that crucial Arab state. There's a reasonably large body of writing on the history of the Stasi now — largely thanks to the concerted effort by the post-unification German government to preserve and open up the Stasi archives. There's a lot of interesting information on the website of the [Federal Commissioner for the Stasi Records](_URL_0_). In terms of published work: John Schmeidel's monograph *[Stasi: The Sword and the Shield of the Party](_URL_7_)* is a good place to start. Gary Bruce's *[The Firm: The Inside Story of the Stasi](_URL_2_)* is a good grassroots history. Kristie Macrakis' *[Seduced by Secrets: Inside the Stasi's Spy-Tech World](_URL_4_)* is an interesting specialist history of the Stasi's technical methods and surveillance prowess. On Soviet intelligence, I always suggest starting with Christopher Andrew's work on the KGB: *[The Sword and the Shield: The Mitrokhin Archive and the Secret History of the KGB](_URL_5_)* and *[The World Was Going Our Way: The KGB and the Battle for the Third World](_URL_6_)*. Jonathan Haslam's more recent *[Near and Distant Neighbours](_URL_3_)* is a good history of the Soviet intelligence community as a whole.
[ "The KGB was not the only intelligence agency that Russia used to gain information around the globe. The Soviet Union formed two other well known agencies: GRU (The Main Directorate of the General Staff of the Armed Forces of the Russian Federation) and SVR (Foreign Intelligence Service). \n", "All sides in the Cold War engaged in espionage. The Soviet KGB (\"Committee for State Security\"), the bureau responsible for foreign espionage and internal surveillance, was famous for its effectiveness. A massive network of informants throughout the Soviet Union was used to monitor dissent from official Soviet politics and morals.\n", "During the Cold War, GCHQ and the NSA shared intelligence on the Soviet Union, the People's Republic of China, and several eastern European countries (known as Exotics). Over the course of several decades, the ECHELON surveillance network was developed to monitor the military and diplomatic communications of the Soviet Union and its Eastern Bloc allies.\n", "The military intelligence was known in the Soviet government for its fierce independence from the rival \"internal intelligence organizations\", such as the NKVD, later the KGB. The military intelligence headquarters were located at Prechistenka street, west of the Kremlin, whereas the NKVD was in the very centre of Moscow, next to the building that housed People's Commissariat for Foreign Affairs at the bottom of Kuznetsky Most. Due to this fact, the Soviet military intelligence came to be known in Soviet diplomats′ cant as \"distant neighbours\" () as opposed to \"near neighbours\", i.e. the NKVD/KGB intelligence officers. According to public statements of Soviet military intelligence veterans, the GRU had always been operationally subordinate to the KGB.\n", "During the first years of its existence, other branches of the U.S. Federal government did not exercise very much supervision over the Central Intelligence Agency. Supposedly justified by the desire to match and defeat Soviet actions throughout the Eastern Hemisphere, it undertook a task that many believed could be accomplished only through an approach similar to the Soviet intelligence agencies, under names including NKVD, MVD, NKGB, MGB, and KGB. Those Soviet organizations also had domestic responsibilities.\n", "The KGB (), translated in English as Committee for State Security, was the main security agency for the Soviet Union from 1954 until its break-up in 1991. As a direct successor of preceding agencies such as the Cheka, NKGB, NKVD and MGB, the committee was attached to the Council of Ministers. It was the chief government agency of \"union-republican jurisdiction\", acting as internal security, intelligence and secret police. Similar agencies were constituted in each of the republics of the Soviet Union aside from Russian SFSR, and consisted of many ministries, state committees and state commissions.\n", "While several organizations such as the CIA and KGB became synonymous with Cold War espionage, many others played key roles in the collection and protection of the section concerning detection of spying, and analysis of a wide host of intelligence disciplines.\n" ]
How does genetic testing work (in order to test for genetic diseases)?
A person with a genetic disease has a particular gene that differs from the normal, unaffected population. When a genetic disease is identified that can be tested for, the discoverers known what and where the gene is and what makes it different from the normal population. That difference is called a mutation. Mutations come in different forms (insertions, deletions, frame shifts, repeats, substitutions....) but basically all you need to know is that it's a difference from the normal. Now to test for anything genetic you need DNA. Every cell in your body, minus just a few (red blood cells, gametes...) have the full genetic code of an individual inside them. Usualy a cheek swab is done to collect a few cells from the inside of a person's mouth or a few hairs plucked to get the cells at the root. Whole blood can also be used. Once the cells are collected the genetic information can be chemically extracted, amplified, then read. The amplification process is known as polymerase chain reaction (PCR) and is one of the most important discoveries in the history of biology. It works by synthetically copying the information in DNA over and over. Since DNA is double-stranded and complementary, the information from one strand can be used to re-create the other. So taking a double-stranded DNA, unwinding it, then copying each of the two halves gives you a pair of double-stranded DNA's. Do this a bunch and you get lots of DNA to work with from a very, very small amount. Once you have the DNA you have to actually read it. You can do this the old, hard way where you actually sequence part of someone's genome to figure out exactly what their genetic code is. If there's a mutation it'll show up as a change in the DNA sequence you read out from the normal sequence. Or you can do it the new, fancy way where you have a DNA microarray, which is a chip that has lots of little strands of DNA, each coding a slightly different copy of the gene in question. Which slightly different copy is in which spot on the chip is known going in. To the chip you add the DNA from the individual, but with one extra modification - you add a fluorescent marker on the end so it lights up. The complementary nature of DNA means that the individual's labeled DNA will bind to the spots on the chip that has DNA that most closely matches. A mutation in the individual's gene means that the DNA will bind less to the normal gene on the chip and more strongly to the corresponding mutant (there can be more than 1 mutant to cause one disease, too). Using a chip reader you can find where that is on the chip (thanks to the fluorescent label) and back-check to see what gene sequence (normal, mutant) that corresponds to. The microarrays can do hundreds or thousands of genes at once so this is a much more efficient way to screen someone for a genetic disease.
[ "Genetic testing identifies changes in chromosomes, genes, or proteins. Usually, testing is used to find changes that are associated with inherited disorders. The results of a genetic test can confirm or rule out a suspected genetic condition or help determine a person's chance of developing or passing on a genetic disorder. Several hundred genetic tests are currently in use, and more are being developed.\n", "Genetic testing identifies changes in chromosomes, genes, or proteins; some are associated with human disease. There are many different clinical and non-clinical situations in which genetic testing is used.\n", "Genetic testing is \"the analysis of chromosomes (DNA), proteins, and certain metabolites in order to detect heritable disease-related genotypes, mutations, phenotypes, or karyotypes for clinical purposes.\" It can provide information about a person's genes and chromosomes throughout life. \n", "Genetic testing allows the genetic diagnosis of vulnerabilities to inherited diseases, and can also be used to determine a child's parentage (genetic mother and father) or in general a person's ancestry. In addition to studying chromosomes to the level of individual genes, genetic testing in a broader sense includes biochemical tests for the possible presence of genetic diseases, or mutant forms of genes associated with increased risk of developing genetic disorders. Genetic testing identifies changes in chromosomes, genes, or proteins. Most of the time, testing is used to find changes that are associated with inherited disorders. The results of a genetic test can confirm or rule out a suspected genetic condition or help determine a person's chance of developing or passing on a genetic disorder. As of 2011 several hundred genetic tests were in use. Since genetic testing may open up ethical or psychological problems, genetic testing is often accompanied by genetic counseling.\n", "Genetic testing identifies changes in chromosomes, genes, or proteins. The variety of genetic tests has expanded throughout the years. In the months and years , the main genetic tests searched for abnormal chromosome numbers and mutations that lead to rare, inherited disorders. Today, tests involve analyzing multiple genes to determine the risk of developing specific diseases or disorders, with the more common diseases consisting of heart disease and cancer. The results of a genetic test can confirm or rule out a suspected genetic condition or help determine a person's chance of developing or passing on a genetic disorder. Several hundred genetic tests are currently in use, and more are not being developed.\n", "Traditional genetic testing involves the analysis of DNA in order to detect genotypes related to a heritable disease or phenotype of interest for clinical purposes. However, current testing methods require days to weeks before results are available limiting the clinical applicability of genetic testing in a number of circumstances.\n", "Genetic tests are performed on a sample of blood, hair, skin, amniotic fluid (the fluid that surrounds a fetus during pregnancy), or other tissue. For example, a medical procedure called a buccal smear uses a small brush or cotton swab to collect a sample of cells from the inside surface of the cheek. Alternatively, a small amount of saline mouthwash may be swished in the mouth to collect the cells. The sample is sent to a laboratory where technicians look for specific changes in chromosomes, DNA, or proteins, depending on the suspected disorders, often using DNA sequencing. The laboratory reports the test results in writing to a person's doctor or genetic counselor.\n" ]
What varieties of memory are there and how do they work?
Very detailed question, but i'll take a stab at it. I majored in Neuroscience and am now a first year medical student so I know a fair bit about this but I'm no expert. 1. The cerebellum helps to coordinate movement and accounts for one type of memory. This structure has an extraordinarily repetitive/simple neuronal structure and almost everything that makes your cerebellum different from mine is learned. This type of memory would be active if, for instance, you're a baseball catcher and someone hits a foul ball. You would see the ball coming down and *know* where to move your glove to get in front of it. This type of memory is going to be important in actions such as throwing a football or texting without looking. 2. The hippocampus is the site of additional memory functions. There are cells in your hippocampus that form a grid, and fire whenever you are in one area of the grid. The hippocampal grid cells are always helping you to localize yourself in 3D. Other hippocampal cells deal with other aspects of spacial mapping. This type memory is very important in knowing where your favorite restaurant is and enables you to walk around your house in the dark without bumping into things. 3. Everything else: Here is where it gets complicated. Almost everything else you *know* is encoded as a representation of the internal state it evokes and in relationship to other things you know. This type of memory is diffusely located throughout the temporal cortex and is processed in the oribitofrontal prefrontal cortex. To illustrate, imagine someone says to me, "Kim Jong Il." Here's what happens in my brain: The words Kim Jong Il are processed by my auditory cortex and converted into meaning. Now this is a *person* and not a string of words. This gets loaded into my working memory (localized in dorsolateral prefrontal cortex.) It calls up everything I know about Kim Jong Il. I think to myself, "he is a bad man." This is because my ventromedial prefrontal cortex has already associated KJI with negative affect and this visceral emotion (processed in my amygdala) is bound to the *idea* in working memory. Things like North Korea, famine, Kim Il Sung, all get called up secondarily, not directly from my working memory but almost bittorent-like, from the cells that were already activated. This doesn't cover all types of memory, but is as good of a introduction as I can muster. This is as accurate as I can be, but there are still no absolutes in neuroscience. Every day we learn more about the brain!
[ "Several memory models have been proposed to account for different types of recall processes, including cued recall, free recall, and serial recall. However, to explain the recall process, the memory model must identify how an encoded memory can reside in the memory storage for a prolonged period until the memory is accessed again, during the recall process; but not all models use the terminology of short-term and long-term memory to explain memory storage; the dual-store theory and a modified version of Atkinson–Shiffrin model of memory (Atkinson 1968) uses both short-and long-term memory storage, but others do not.\n", "Recall in memory refers to the mental process of retrieval of information from the past. Along with encoding and storage, it is one of the three core processes of memory. There are three main types of recall: free recall, cued recall and serial recall. Psychologists test these forms of recall as a way to study the memory processes of humans and animals.\n", "Short-term, intermediate-term and long-term memories can overlap in practice. Within these categories, memory can further be differentiated by measures such as frequency and impact of changes made. One example of an intermediate-term memory structure is one that prohibits or encourages solutions that contain certain attributes (e.g., solutions which include undesirable or desirable values for certain variables) or a memory structure that prevents or induces certain moves (e.g. based on frequency memory applied to solutions sharing features in common with unattractive or attractive solutions found in the past). In short-term memory, selected attributes in solutions recently visited are labeled \"tabu-active.\" Solutions that contain tabu-active elements are banned. Aspiration criteria are employed that override a solution's tabu state, thereby including the otherwise-excluded solution in the allowed set (provided the solution is “good enough” according to a measure of quality or diversity). A simple and commonly used aspiration criterion is to allow solutions which are better than the currently-known best solution.\n", "Memory operates on the individual as well as the collective level. \"Memory nonetheless captures simultaneously the individual, embodied, and lived side and the collective, social, and constructed side of our relations to the past”. It allows for individuals, groups and societies to be creative as its “anachronistic quality—its bringing together of now and then, here and there—is actually the source of its powerful creativity, its ability to build new worlds out of the materials of older ones”.\n", "Memory is divided into voluntary (conscious) and involuntary (unconscious) processes that function independently of each other. Theories and research on memory dates back to Hermann Ebbinghaus, who began studying nonsense syllables. Ebbinghaus classified three distinct classes of memory: sensory, short term, and long-term memory. Sensory memory is made up of a brief storage of information within a specific medium (the line you see after waving a sparkler in your field of vision is created by sensory memory). Short term memory is made up of the information currently in use to complete the task at hand. Long term memory is composed of the systems used to store memory over long periods. It enables one to remember what happened two days ago at noon, or who called last night.\n", "Memory has two classifications—short-term memory and long-term memory. Long-term memory can store information for a long duration and is subdivided into non-declarative (implicit) memory for learned skills and habits, and declarative (explicit) memory for knowledge of facts and events. Declarative memory consists of semantic memory and episodic memory.\n", "Memory is a very broad function which includes several distinct abilities, all of which can be selectively impaired and require individual testing. There is disagreement as to the number of memory systems, depending on the psychological perspective taken. From a clinical perspective, a view of five distinct types of memory, is in most cases sufficient. Semantic memory and episodic memory (collectively called declarative memory or explicit memory); procedural memory and priming or perceptual learning (collectively called non-declarative memory or implicit memory) all four of which are long term memory systems; and working memory or short term memory. Semantic memory is memory for facts, episodic memory is autobiographical memory, procedural memory is memory for the performance of skills, priming is memory facilitated by prior exposure to a stimulus and working memory is a form of short term memory for information manipulation.\n" ]
Why does bird poo look how it does?
Birds actually urinate and defecate at the same time - their urine is viscous and highly concentrated (I believe it has a lot of ammonia in it, which is the characteristic smell). The white liquid/paste is the concentrated urine and the black pellets are the feces. Hope this enlightens you! Classic example: _URL_0_
[ "It is a well camouflaged bird, it is usually shy and conceals itself close to ground vegetation and flushes only when approached closely. When flushed, they utter a sharp note that sounds like \"scape, scape\" and fly off in a series of aerial zig-zags to confuse predators. They forage in soft mud, probing or picking up food by sight. They mainly eat insects and earthworms, also some plant material.\n", "The name \"lapwing\" has been variously attributed to the \"lapping\" sound its wings make in flight, from the irregular progress in flight due to its large wings (the Oxford English Dictionary derives this from an Old English word meaning \"to totter\"), or from its habit of drawing potential predators away from its nest by trailing a wing as if broken. The names \"peewit\", \"pewit\", \"tuit\" or \"tew-it\" are onomatopaeic and refer to the bird's characteristic call.\n", "Caiques also exhibit a unique behavior known as \"surfing\", where the bird will vigorously rub its face, wings and chest against any nearby soft item (e.g. carpets, towels, cushions, crumpled paper, curtains or human hair) while using its beak to pull itself along. During this, the bird will display jerky movements and may roll over several times. This behavior is thought to be a cleaning or bathing motion and occurs regardless of age or sex. In the wild, caiques use wet leaves for this behavior.\n", "Little is known of its behaviour, but it has an aerial display, which involves flying high in circles, followed by a dive during which the bird makes a drumming sound, caused by vibrations of its modified outer tail feathers.\n", "This bird has an aerial display, which involves flying high in circles, followed by a powerful stoop during which the bird makes a \"drumming\" sound, caused by vibrations of modified outer tail feathers.\n", "Unlike many other wading birds, black-fronted dotterels retain the same plumage all year round, which makes identification easier. They forage in a series of short running motions, holding the body horizontal, stopping to peck from time to time with a rapid bobbing motion. Their diet consists of mostly insects and other small creatures, supplemented by a few seeds.\n", "This bird has a spectacular aerial display, which involves flying high in circles, followed by a powerful stoop during which the bird makes a drumming sound, caused by vibrations of modified outer tail feathers.\n" ]
Is there a way to determine the radius of a black hole, or would anything of the sort be a guess?
The radius of the event horizon can be calculated from analyzing the orbits of nearby stars or planets. From the way it interacts gravitationally with its surroundings you can find mass, and thus the radius of its event horizon.
[ "Any object whose radius is smaller than its Schwarzschild radius is called a black hole. The surface at the Schwarzschild radius acts as an event horizon in a non-rotating body (a rotating black hole operates slightly differently). Neither light nor particles can escape through this surface from the region inside, hence the name \"black hole\".\n", "If one ignores the electron's angular momentum and charge, as well as the effects of quantum mechanics, one can treat the electron as a black hole and attempt to compute its radius. The Schwarzschild radius of a mass is the radius of the event horizon for a non-rotating, uncharged black hole of that mass. It is given by\n", "The Schwarzschild black hole is characterized by a surrounding spherical boundary, called the event horizon, which is situated at the Schwarzschild radius, often called the radius of a black hole. The boundary is not a physical surface, and if a person fell through the event horizon (before being torn apart by tidal forces), they would not notice any physical surface at that position; it is a mathematical surface which is significant in determining the black hole's properties. Any non-rotating and non-charged mass that is smaller than its Schwarzschild radius forms a black hole. The solution of the Einstein field equations is valid for any mass , so in principle (according to general relativity theory) a Schwarzschild black hole of any mass could exist if conditions became sufficiently favorable to allow for its formation.\n", "The Schwarzschild radius (sometimes historically referred to as the gravitational radius) is a physical parameter that shows up in the Schwarzschild solution to Einstein's field equations, corresponding to the radius defining the event horizon of a Schwarzschild black hole. It is a characteristic radius associated with every quantity of mass. The \"Schwarzschild radius\" was named after the German astronomer Karl Schwarzschild, who calculated this exact solution for the theory of general relativity in 1916.\n", "A black hole in general is surrounded by a surface, called the event horizon and situated at the Schwarzschild radius for a nonrotating black hole, where the escape velocity is equal to the velocity of light. Within this surface, no observer/particle can maintain itself at a constant radius. It is forced to fall inwards, and so this is sometimes called the \"static limit\".\n", "where \"r\" is the Schwarzschild radius and \"M\" is the mass of the Sun. For a black hole with nonzero spin and/or electric charge, the radius is smaller, until an extremal black hole could have an event horizon close to\n", "The surface at the Schwarzschild radius acts as an event horizon in a non-rotating body that fits inside this radius (although a rotating black hole operates slightly differently). The Schwarzschild radius of an object is proportional to its mass. Theoretically, any amount of matter will become a black hole if compressed into a space that fits within its corresponding Schwarzschild radius. For the mass of the Sun this radius is approximately 3 kilometers and for the Earth it is about 9 millimeters. In practice, however, neither the Earth nor the Sun has the necessary mass and therefore the necessary gravitational force, to overcome electron and neutron degeneracy pressure. The minimal mass required for a star to be able to collapse beyond these pressures is the Tolman–Oppenheimer–Volkoff limit, which is approximately three solar masses.\n" ]
if everyone on earth infected with a common cold got better at this exact second, would anyone else get sick? if so how would it be transmitted considering there aren't any carriers of the virus?
Viruses can survive for limited times on various surfaces which would allow reinfection.
[ "The common cold is the most common human disease and affects people all over the globe. Adults typically have two to three infections annually, and children may have six to ten colds a year (and up to twelve colds a year for school children). Rates of symptomatic infections increase in the elderly due to declining immunity.\n", "Well over 200 virus strains are implicated in causing the common cold, with rhinoviruses being the most common. They spread through the air during close contact with infected people or indirectly through contact with objects in the environment, followed by transfer to the mouth or nose. Risk factors include going to child care facilities, not sleeping well, and psychological stress. The symptoms are mostly due to the body's immune response to the infection rather than to tissue destruction by the viruses themselves. In contrast, those affected by influenza can show similar symptoms as people with a cold, but symptoms are usually more severe. Additionally, influenza is less likely to result in a runny nose.\n", "Tens of thousands worldwide were infected as a result of contaminated factor products including more than 10,000 people in the United States, 3,500 British, 1,400 Japanese, 700 Canadians, 250 Irish, and 115 Iraqis.\n", "WHO officials had acknowledged that air travellers \"within two rows of an infected person could be in danger\". However, the virus had also been found to survive for days in the environment, giving rise to the possibility of spread by contact with surfaces including armrests and tray tables.\n", "The common cold is the most frequent infectious disease in humans. The average adult gets two to three colds a year, while the average child may get six to eight. Infections occur more commonly during the winter. These infections have existed throughout human history.\n", "When someone falls ill, people they might have infected are vaccinated. Depending on how easily the disease spreads, contacts who could have been infected might include family, neighbours, and friends. Several layers of contacts may be vaccinated (the contacts, the contacts' contacts, the contact's contacts' contacts, etc.). \n", "The common cold virus is typically transmitted via airborne droplets (aerosols), direct contact with infected nasal secretions, or fomites (contaminated objects). Which of these routes is of primary importance has not been determined. The viruses may survive for prolonged periods in the environment (over 18 hours for rhinoviruses) and can be picked up by people's hands and subsequently carried to their eyes or nose where infection occurs. Transmission is common in daycare and at school due to the proximity of many children with little immunity and frequently poor hygiene. These infections are then brought home to other members of the family. There is no evidence that recirculated air during commercial flight is a method of transmission. People sitting in close proximity appear to be at greater risk of infection.\n" ]
Why does a log burn longer than the same amount of wood chopped up?
If you cut a cylindrical log in two, you increase the surface area significantly. This means more of the log can burn at once, and so it is consumed faster. This is not the only effect. If you put logs in close proximity, part of the radiant heat given off by one can be caught by the other, and structuring the wood may create air currents that replenish the oxygen in the mix.
[ "Decomposition speed of organic material depends on the carbon to nitrogen ratio of the material, among other factors. Wood breaks down relatively slowly because it has one of the highest carbon to nitrogen ratios of all organic matter that is used in composting. If the wood is not processed into smaller pieces with larger surface area to speed up chemical reactions, breakdown is even slower. The decomposition process may in the short term take more nitrogen from the soil through microbial activity (nitrogen immobilization), if not enough nitrogen is available. Thus in the short term the fertility of the soil may be decreased before eventually, perhaps after 1-2 years, the nitrogen level is increased past the original level. Traditionally therefore, it is said to be advantageous to balance \"browns\" (e.g. woodchippings) with \"greens\" (e.g. leaves) for efficient composting, and to allow compost to become well-rotted before applying it a bed to prevent competition between soil bacteria and plants for nitrogen, reducing yield.\n", "Some firewood is harvested in \"woodlots\" managed for that purpose, but in heavily wooded areas it is more often harvested as a byproduct of natural forests. Deadfall that has not started to rot is preferred, since it is already partly seasoned. Standing dead timber is considered better still, as it is both seasoned, and has less rot. Harvesting this form of timber reduces the speed and intensity of bushfires. Harvesting timber for firewood is normally carried out by hand with chainsaws. Thus, longer pieces - requiring less manual labor, and less chainsaw fuel - are less expensive and only limited by the size of their firebox. Prices also vary considerably with the distance from wood lots, and quality of the wood.\n", "BULLET::::- Wood burning: When wood is burned, it is usually best to dry it first. Damage from shrinkage is not a problem here, as it may be in the case of drying for woodworking purposes. Moisture affects the burning process, with unburnt hydrocarbons going up the chimney. If a 50% wet log is burnt at high temperature, with good heat extraction from the exhaust gas leading to a 100 °C exhaust temperature, about 5% of the energy of the log is wasted through evaporating and heating the water vapour. With condensers, the efficiency can be further increased; but, for the normal stove, the key to burning wet wood is to burn it very hot, perhaps starting fire with dry wood.\n", "Its dense, wood is resistant to the elements and is used for construction, carpentry, beams for bridges, poles and firewood. The wood is said to be unaffected by insects including termites and it can last more than 20 years unpainted.\n", "When purchasing, cutting, or collecting firewood, it is good to be aware of the difference between hardwood and softwood. Both hardwood and softwood have similar energy contents by mass, but not by volume. In other words, a piece of hardwood would usually be heavier and have more available energy than the same sized piece of softwood. Hardwoods, derived from trees such as oak and ash, may burn at a slower rate, resulting in sustained output. Many softwoods are derived from conifers, which are fast growing and may burn at a faster rate. This is one reason why softwood pellets (for pellet stoves) are popular.\n", "Some firewood is harvested in \"woodlots\" managed for that purpose, but in heavily wooded areas it is more usually harvested as a byproduct of natural forests. Deadfall that has not started to rot is preferred, since it is already partly seasoned. Standing dead timber is considered better still, for it has less humid organic material on the trunk, allowing tools to stay sharper longer, as well as being both seasoned and less rotten. Harvesting this form of timber reduces the speed and intensity of bushfires, but it also reduces habitat for snag-nesting animals such as owls and some rodents.\n", "Many softwoods will season (dry) much more quickly than many hardwoods. For example, pine that has been cut, split, stacked and topcovered will usually be ready to burn in one year; oak may be expected to take three years under the same conditions.\n" ]
where does white zit pus eventually go if you don't pop the zit?
It is consumed and carried away by amoeba-like leukocytes in your body called *macrophages*. Here's a video of macrophages doing their thing: _URL_0_
[ "Purulent sputum contains pus, composed of white blood cells, cellular debris, dead tissue, serous fluid, and viscous liquid (mucus). Purulent sputum is typically yellow or green. It is seen in cases of bronchiectasis, lung abscess, an advanced stage of bronchitis, or acute upper respiratory tract infection (common cold, laryngitis).\n", "When there is discharge of thick, cloudy white fluid (wady) (that exits before or after urinating) or unlustful discharge of thin, sticky, white fluid (madhy) caused by play or kissing, it requires ghusl. And wudu. \n", "When there is discharge of thick, cloudy white fluid (wady) (that exits before or after urinating) or unlustful discharge of thin, sticky, white fluid (madhy) caused by play or kissing, it requires ghusl. And wudu. \n", "Pus is an exudate, typically white-yellow, yellow, or yellow-brown, formed at the site of inflammation during bacterial or fungal infection. An accumulation of pus in an enclosed tissue space is known as an abscess, whereas a visible collection of pus within or beneath the epidermis is known as a pustule, pimple, or \"spot\".\n", "The Grey Goo is then washed down the drain and lands outside, where it moves to a park and grows in size. It then moves on to a picnic table where the scientist and assistant are having lunch. When they discover it, the pair throw the Grey Goo into the ocean. The Grey Goo then eats through the ocean and is launched by a whale into another park and then to a city. There, the Grey Goo faces its first war from humans, who attempt to use guns, tanks, etc. to stop it. It eventually launches itself into the sky and then into orbit around the Earth. After eating the moon, Earth, and the rest of the solar system (including Pluto), it moves on to nearby stars like Alpha Centauri and beyond the Milky Way. However, its mass becomes too great after devouring the fabric of space and time, and it implodes thus causing the universe to begin again.\n", "Although pus is normally of a whitish-yellow hue, changes in the color can be observed under certain circumstances. Pus is sometimes green because of the presence of myeloperoxidase, an intensely green antibacterial protein produced by some types of white blood cells. Green, foul-smelling pus is found in certain infections of \"Pseudomonas aeruginosa\". The greenish color is a result of the bacterial pigment pyocyanin that it produces. Amoebic abscesses of the liver produce brownish pus, which is described as looking like \"anchovy paste\". Pus from anaerobic infections can more often have a foul odor.\n", "GI stasis is the condition of food not moving through the gut as quickly as normal. The gut contents may dehydrate and compact into a hard, immobile mass (impacted gut), blocking the digestive tract of the rabbit. Food in an immobile gut may also ferment, causing significant gas buildup and resultant gas pain for the rabbit.\n" ]
In general, how accurate do you find the CaspianReport's 'History of Islam' series?
> The thing is, Islam emerged in literate times. Historians were writing journals. Scholars were writing diaries and letters. Jurists were writing bureaucratic articles. So from the 7th century onward there was a rich record of documents, and that's why we know so many historical details. Noooooooooooooooooo... Noooooooooooooooooooooooooooooooooo! Ok I'm done having a minor aneurysm. But yeah this is a wildly inaccurate statement. It massively overstates the reliability of the documents he's apparently working from (I.e. Ibn Ishaq via Tabari and Ibn Hisham), is apparently ignorant of how that information got passed down (I.e orally, not written down) and fails to mention how poorly documented the 7th-8th centuries are. /u/Shlin28 gives a great description of just how poorly documented the rise of Islam is on the Christian side here: _URL_0_ I've written quite a bit on the historiographical controversy over the use of these "traditional" sources, for instance here: _URL_1_ I only listened to the first ten minutes or so of this YouTube video, but even accounting for the fact that he's uncritically relying on this dubious traditional sources I heard a good bit of bad history. A shura council is not "like a democracy", for instance. Some effects of the unreliability of those sources as it pertains to the video: We cannot reliably quite Muhammad. We cannot reliably say who was or was not present at battles. We cannot reliably say how many people attended a speech. And so on. There are people who defend the use of traditional sources, and they are of course a matter of faith to Muslims. Its an incredibly contentious issue. Personally I'm not willing to reject their content entirely but find any specific details to be incredibly dubious. As I mentioned in the other post Robert Hoyland's *Writing the Biography of the Prophet Muhammad: Problems and Solutions* is a great essay-length overview of these issues and available as a free PDF if you do a search for it.
[ "The series tells the full biography of the four imams of the Muslims from the Sunnis and the community Ahmad ibn Hanbal and the events that took place from the beginning of his family and social life and even began his scientific life in the request of forensic science from Muhammad and the Quran until his death, and also reviews the period of the rule of the Abbasid state, Islamic events and conquests , the emergence of the Mu'tazili and the emergence of the plight of the creation of the Koran , and Imam Ahmad bin Hanbal confronted them.\n", "This two volume encyclopedia which consists of 1300 pages is covering all the aspects of Islam in dept with graphical illustrations. This encyclopedia has significant notes, four dozen full-scale and unique multi-color maps. Encyclopedia of Islam happens to be the first of its kind written by a Muslim scholar himself. Apart from the prophet Muhammad and the earliest Muslims, the encyclopedia also explores the lives of prominent figures from modern Islamic history like Jamaluddin Al-Afghani, Muhammad Iqbal, Syed Abul A’la Mawdudi, Maulana Muhammad Ilyas and Hassan Al-Banna.\n", "The series revolves around a historical drama where the series discusses the biography of the life of Imam Ahmad bin Hanbal, the fourth of the four imams in the Sunnis and the group where the series reviews many of the events that took place in his life during the era of Abbasid state in the era of Caliph Harun al-Rashid The light on the city of Baghdad, the nature of social life and also monitor the poor class in which he lived Ahmed bin Hanbal , as well as the middle class, in addition to the focus on some of the Islamic conquests in the era of Harun al-Rashid and the emergence of Mu'tazila and the struggle of the Secretary and the safe to power, In the way of narration of events and facts, which highlights the first : the personal life of the Imam and the social environment in Baghdad and the second : the system of political governance that prevailed at that time and the third : some stories and novels and historical facts that were modern in the era of the Imam.\n", "The remainder of the book, Parts II and III, discuss later developments and the larger context in which Islam originated: the Late Antique Near East, and relate it to theoretical themes of cultural history.\n", "The text provides one of the earliest external accounts of Islam, presenting a significantly different Islamic historiography than found in traditional Islamic texts. It also shows Jacob comparing the Byzantine Empire to the fourth beast of the prophecy of Daniel from Judeo-Christian eschatology. Although not unfamiliar imagery, it is part of a series of Byzantine literature, from the early stages of the Islamic religion, of trying to reconcile Islam with the apocalyptic vision. Further examples of this are contained in the pseudo-Athanasian's \"Quaestiones ad Antiochum ducem\", and the \"Quaestiones et responsiones\" attributed to Anastasius of Sinai.\n", "These observations did not fit to the Islamic narratives on Islam's beginnings which depicted Islam to come into being within a polytheistic society. Wansbrough analyzed the classical Islamic narratives which had been written 150 to 200 years after Muhammad with the historical-critical method, especially literary criticism. Thus, he claimed countless proofs that these texts are not historical accounts but later literary constructions in the sense of the concept of a \"salvation history\" (\"Heilsgeschichte\") of the Old Testament. Their historical core is meager and cannot be detected.\n", "The book contains sections on the Traditions (Hadiths) of the Prophet Muhammad (primarily from the Mishkat of Al-Baghawi of Herat), the adventures of the Mulla Nasrudin, thoughts from Omar Khayyam, meditations of Rumi, and the definitions of Mulla Do-Piaza, as well as sayings from many classical Sufi Masters like Saadi, Bahaudin Naqshband and Khwaja Ahrar. Also included are some excerpts about Islamic culture and history from Edward Gibbon's \"Decline and Fall of the Roman Empire\" and from Shah's earlier book of travels, \"Destination Mecca\". Two anecdotes are about Shah's great-great grandfather, Jan-Fishan Khan.\n" ]
How did the American Forces learn to defeat the Japanese during the Guadalcanal Campaign in WWII?
In the air, the Cactus Air Force was tying out new tactics to combat the agile Japanese Zero. "Cactus Air Force" is a nickname for the pilots of Henderson Field on Guadalcanal--codenamed Cactus. These pilots used a mix of aircraft, most notably the Grumman F4F Wildcats (there were some P-39's as well, and some dive bombers too). The Wildcat was heavily armed, heavily armored, and heavy overall compared to the Zero. If a Wildcat pilot tried to turn with a Zero in a dogfight, he would quickly end up with the agile Zero behind him. So, a naval aviator named Thatch came up with a defensive maneuvering scheme that came to be known as the [Thatch Weave](_URL_0_). This allowed pairs of Wildcats to keep their wingman covered from attack from behind. The linked wikipedia page has a good illustration of the tactic. Offensive tactics revolved around diving attacks or head-on passes, where the armament and armor of the Wildcat would make it the victor in most slugging matches. At sea, there was a curious occurrence in warfare. Control of the area around Guadalcanal generally changed hands every twelve hours. During the day, the US Navy was able to largely command the waves--especially as Henderson Field became more established. At night, the Japanese expertise and experience at night fighting meant that they could exert control over the area. There were so many ships of both sides sunk in the area that it gained the nickname "Ironbottom Sound." The clashes gave US naval forces valuable but costly lessons that were applied in the later island hopping campaigns. On land, the Japanese committed troops piecemeal to various assaults on the Marines. Again, the campaign was costly for the Americans and there were several times where the enemy nearly broke through. However, the Marines gained valuable tactical, logistical, and other lessons that applied to warfare in the theater. I don't know of any tactical breakthroughs that were made, instead there were refinements to existing ideas. Coordination between infantry and mortars/artillery was improved. Close air support was practiced (especially with the aforementioned P-39's and dive bombers). Officers and men gained experience in how the Japanese liked to fight as well as what was effective against them. *Guadalcanal: The Definitive Account of the Landmark Battle* by Richard B. Frank is an outstanding account of the battle. Robert Leckie's *Challenge for the Pacific: Guadalcanal: The Turning Point of the War* is also good if a bit shorter. However, Leckie was an infantryman in the actual battle, so his insights are incredible. They are more intimately described in his memior: *Helmet for my Pillow*.
[ "In hindsight, historians have faulted the Americans, especially Patch and Halsey, for not taking advantage of their ground, aerial, and naval superiority to prevent the successful Japanese evacuation of most of their surviving forces from Guadalcanal. Said Chester Nimitz, commander of Allied forces in the Pacific, of the success of Operation Ke, \"Until the last moment it appeared that the Japanese were attempting a major reinforcement effort. Only the skill in keeping their plans disguised and bold celerity in carrying them out enabled the Japanese to withdraw the remnants of the Guadalcanal garrison. Not until all organized forces had been evacuated on 8 February did we realize the purpose of their air and naval dispositions.\"\n", "Both sides were overextended in the exhaustive sea, air and land battles for Guadalcanal. The Japanese were better at night combat (because the American destroyers had only trained for attacks on battleships). However, the Japanese could not feed its soldiers so the Americans eventually won because of superior logistics. The Navy built up its forces in 1942-43, and developed a strategy of \"island-hopping, that is to skip over most of the heavily defended Japanese islands and instead go further on and select islands to seize for forward air bases.\n", "Guadalcanal made it clear to the Americans that the Japanese would fight to the bitter end. After brutal fighting in which few prisoners were taken on either side, the United States and the Allies pressed on the offensive. The landings at Tarawa on November 20, 1943, by the Americans became bogged down as armor attempting to break through the Japanese lines of defense either sank, were disabled or took on too much water to be of use. The Americans were eventually able to land a limited number of tanks and drive inland. After days of fighting they took control of Tarawa on November 23. Of the original 2,600 Japanese soldiers on the island, only 17 were still alive.\n", "As the Americans consolidated their gains on Guadalcanal, the critical need for reinforcements prompted Japanese Admiral Isoroku Yamamoto to send the Combined Fleet south to cover a large troop convoy. American scout planes spotted the Japanese forces on the morning of 23 August. With the enemy reported to the northwest, \"Enterprise\" and launched search and attack planes, but they failed to make contact because of deteriorating weather and the fact that the Japanese, knowing that they had been spotted, reversed course.\n", "Both sides experienced extreme difficulties in fighting in the thick jungles and tropical environment that existed in the battle area. Many of the American troops were also involved in their first combat operations. The Japanese were mostly cut off from resupply and suffered greatly from malnourishment and lack of medical care. After some difficulty, the U.S. succeeded in taking Mount Austen, in the process reducing a strongly defended position called the Gifu, as well as the Galloping Horse and the Sea Horse. In the meantime, the Japanese decided to abandon Guadalcanal and withdrew to the west coast of the island. From that location most of the surviving Japanese troops were successfully evacuated during the first week of February 1943.\n", "In 1938, Japanese forces attacked Chinese defenders over the Yangtze River at the Battle of Wuhan. Soon, the Japanese would later further improve its techniques upon seaborne assaults by the Second Sino-Japanese War. By World War II, marines such as the Special Naval Landing Force, used amphibious landings to attack and sweep across territories in South East Asia. Their technique of surprise landings in continuous success and the support from the Navy, inspired the British and American landings in World War II such as D-Day and the Pacific Campaign.\n", "The Japanese military themselves also used guerrilla warfare during the later part of the Pacific War, when Japan's resource was already dwindling and the Allies have started invading. Tadamichi Kuribayashi famously used guerrilla warfare during the Battle of Iwo Jima, where the general used network of tunnels and caves to attack American forces. His tactic was somewhat successful, delaying the Americans from taking Iwo Jima for 36 days. The same tactic was used during the Battle of Okinawa.\n" ]
why do most current-gen games on consoles require you to wait until it's fully installed, while last-gen games could be played immediately?
At a guess I would say that new games are bigger then a disc can hold, my battlefield 1 digital game was 43 GB (rip my data cap :( ). Im willing to bet that if everything has to be read off the discs the load screens would be all over the place and, with all the massive open worlds lately, that would be a game killer.
[ "Some consoles lack the ability to play games from previous generations which allow a developer to release older games again but on the new consoles. The re-released game may be unchanged and simply be the same game but run on the new technology or it can be changed by the developer to have improved graphics, sound or gameplay. Some re-releases can have added features such as Final Fantasy VII which added functions to speed the game up and turn off random enemy encounters.\n", "Games developed for video game consoles have had almost no maintenance period in the past. The shipped game would forever house as many bugs and features as when released. This was common for consoles since all consoles had identical or nearly identical hardware; making incompatibility, the cause of many bugs, a non-issue. In this case, maintenance would only occur in the case of a port, sequel, or enhanced remake that reuses a large portion of the engine and assets.\n", "Despite the closure of the multiplayer servers in 2012, it is still possible to download the game updates as of today, as they're required for DLC compatibility. While most PS3 games search for updates from the XMB or after starting them, this games requires the user to enter the now-defunct online modes in order to trigger the updating process. The most recent update is version 3.1.\n", "BULLET::::- An older Kernel revision is required on the Xbox 360 itself, which may prove to be hard to find, since connecting to the Xbox Live service applies updates to the console, and many games include updates that must be applied before the game will run. This limitation is not as important as it once was, as it is now known to be possible to downgrade a Kernel greater than the last of the two exploitable Kernels by means of a timing attack.\n", "Several game discs, both first-party and third-party games, have included system software updates so that players who are not connected to the Internet can still update their system. Additionally this can force an upgrade by requiring the player to perform the update, without which the new game cannot be played. Some online games (such as Super Smash Bros. Brawl and Mario Kart Wii) have come with specific extra updates, such as being able to receive posts from game-specific addresses, so, regardless of the version of the installed software, it will install an update.\n", "Due to the time involved in licensing and porting the product, Macintosh versions of games ported by third-party companies are usually released anywhere from three months to more than a year after their Windows-based counterparts. For example, the Windows version of \"Civilization IV\" was released on October 25, 2005, but Mac gamers had to wait eight months until June 30, 2006 for the release of the Mac version.\n", "Although it is usually possible to modify the consoles themselves to bypass regional lockouts, console modifications can cause problems with screens not being displayed correctly and games running too fast or slow, due to the fact that the console itself may not be designed to output to the correct format for the game. These problems can be overcome on emulators, as they are usually designed with their own output modules, which can run both NTSC and PAL games without issue.\n" ]
Why is it easier to remember a phrase or image than a single word?
The use of mnemonics such at those you describe are likely to work at a number of different levels. Firstly, when thinking about memory and learning, it is important to emphasise the role of attention. That is, when you *personally* create a mnemonic, you are assigning more attention to it which makes it more likely that it will be encoded. Using mnemonics also helps to link the new knowledge, to existing knowledge networks, or unique (and therefore memorable) identifiers. So instead of having to recall the new information from scratch, you can be prompted by its link with the existing knowledge structures, or those unique/memorable things that you will "never forget" (like the silly words). There is a copious body of evidence that demonstrates that using mnemonics such as those you describe can assist in the recollection of new learning. Much of these principles are outlined in the [levels of processing](_URL_0_) model of memory encoding (although I am not up to date on the current consensus on this model). In this model your use of mnemonics is an example of deep level processing. edit: spelling, emphasis
[ "An experiment from 1966 showed that people remember a group of words better if they are within the same theme category. Such words that generate recall by association are known as \"semantic cues\". If the sound of the word is emphasized during the encoding process, a cue that could be used could also put emphasis on the phonetic quality of the word.\n", "In some cases controlled vocabulary can enhance recall as well, because unlike natural language schemes, once the correct authorized term is searched, there is no need to search for other terms that might be synonyms of that term.\n", "It had been determined that recall is greater for items that were generated by the subject versus if the subject was presented with the stimuli. There seems to be a memory advantage for instances where people are able to produce an answer themselves, recall was higher when Aha! reactions occurred. They tested sentences that were initially hard to understand, but when presented with a cued word, the comprehension became more apparent. Other evidence was found indicating that effort in processing visual stimuli was recalled more frequently than the stimuli that were simply presented. This study was done using connect-the-dots or verbal instruction to produce either a nonsense or real image. It is believed that effort made to comprehend something when encoding induces activation of alternative cues that later participate in recall.\n", "Visual input creates the strongest recall value of all senses, and also allows the widest spectrum of levels-of-processing modifiers. It is also one of the most widely studied. Within visual studies, pictures have been shown to have a greater recall value than words – the picture superiority effect. However, semantic associations have the reverse effect in picture memories appear to be reversed to those in other memories. When logical details are stressed, rather than physical details, an image's recall value becomes lower. When comparing orthographic (capitalization, letter and word shape), phonological (word sound) and semantic (word meaning) encoding cues, the highest levels of recall were found with the meanings of the words, followed by their sounds and finally the written and shape-based cues were found to generate the least ability to stimulate recall.\n", "Retrieval: Because a person has played an active role in processing explicit information, the internal cues that were used in processing it can also be used to initiate spontaneous recall. When someone talks about an experience, the words they use will help when they try to remember this experience at a later date. The conditions in which information is memorized can affect recall. If a person has the same surroundings or cues when the original information is presented, they are more likely to remember it. This is referred to as encoding specificity and it also applies to explicit memory. In a study where subjects were asked to perform a cued recall task participants with a high working memory did better than participants with a low working memory when the conditions were maintained. When the conditions were changed for recall both groups dropped. The subjects with higher working memory declined more. This is thought to happen because matching environments activates areas of the brain known as the left inferior frontal gyrus and the hippocampus.\n", "Some very common words do not fully follow common phonic patterns, so those words have to be memorized. Some books refer to these words as \"sight words\", but it is probably better to refer to them as \"memory words\" because some books refer to sight words as those words which are so common they do not have to be analyzed or \"sounded-out\". It does not seem like a good idea to have \"sight-words\" mean two different things when \"memory-words\" is available.\n", "Although in certain domains, such as memorization, it appears that ironic effects of attempting to remember vary with the level of mental control over mnemonic processing and may simply be due to ineffective mental strategies.\n" ]
the three scripts of the japanese language and their usage.
Hiragana ひらがな is like the standard Japanese alphabet. Katakana カタカナ is used for words of foreign origin. Kanji 漢字 is that impossible to read shit that was imported from China. Now, why don't they just get rid of Kanji, since they are such a pain in the ass? They had this debate after WW2, and felt that kanji was part of their culture, so they didn't want to dispose of it. Also, Japanese has an insane number of homonyms, so without kanji it would be a bit tough to know what a written word is.
[ "Japanese is written with a combination of three scripts: hiragana, derived from the Chinese cursive script, katakana, derived as a shorthand from Chinese characters, and kanji, imported from China. The Latin alphabet, rōmaji, is also often used in modern Japanese, especially for company names and logos, advertising, and when inputting Japanese into a computer. The Hindu-Arabic numerals are generally used for numbers, but traditional Sino-Japanese numerals are also common.\n", "Japanese is written with a combination of three scripts: hiragana, derived from the Chinese cursive script, katakana, derived as a shorthand from Chinese characters, and kanji, imported from China. The Latin alphabet, rōmaji, is also often used in modern Japanese, especially for company names and logos, advertising, and when inputting Japanese into a computer. The Hindu-Arabic numerals are generally used for numbers, but traditional Sino-Japanese numerals are also very common.\n", "Modern Japanese is written in a mixture of three main systems: kanji, characters of Chinese origin used to represent both Chinese loanwords into Japanese and a number of native Japanese morphemes; and two syllabaries: hiragana and katakana. The Latin script (or romaji in Japanese) is used to a certain extent, such as for imported acronyms and to transcribe Japanese names and in other instances where non-Japanese speakers need to know how to pronounce a word (such as \"ramen\" at a restaurant). Arabic numerals are much more common than the kanji when used in counting, but kanji numerals are still used in compounds, such as \"tōitsu\" (\"unification\").\n", "The modern Japanese writing system uses a combination of logographic kanji, which are adopted Chinese characters, and syllabic kana. Kana itself consists of a pair of syllabaries: hiragana, used primarily for native or naturalised Japanese words and grammatical elements, and katakana, used primarily for foreign words and names, loanwords, onomatopoeia, scientific names, and sometimes for emphasis. Almost all written Japanese sentences contain a mixture of kanji and kana. Because of this mixture of scripts, in addition to a large inventory of kanji characters, the Japanese writing system is often considered to be one of the most complicated in use anywhere in the world.\n", "Japanese writing evolved from Chinese script and Chinese characters, called kanji, or ideograms, were adopted to represent Japanese words and grammar. Kanji were simplified to create two other scripts, called hiragana and katakana. Hiragana is the more widely used script in Japan today, while katakana, meant for formal documents originally, is used similarly to italics in alphabetic scripts.\n", "Japanese is normally written in a combination of logographic characters borrowed from Chinese (kanji) and syllabic scripts (kana) that also ultimately derive from Chinese characters. \"Rōmaji\" may be used in any context where Japanese text is targeted at non-Japanese speakers who cannot read kanji or kana, such as for names on street signs and passports, and in dictionaries and textbooks for foreign learners of the language. It is also used to transliterate Japanese terms in text written in English (or other languages that use the Latin script) on topics related to Japan, such as linguistics, literature, history, and culture. \"Rōmaji\" is the most common way to input Japanese into word processors and computers, and may also be used to display Japanese on devices that do not support the display of Japanese characters.\n", "English is taught in Japanese schools with an emphasis on writing and grammar over speaking. The written word is taught using Romaji (Latin script); however, the correct pronunciation of the Romaji letters is not taught. The spoken word is taught using Katakana script. Katakana is a syllabary like Hiragana, but among other uses it is used to write foreign words (not just English, but also French, German, Korean etc. In some ways it is analogous to italics in Latin script. Katakana script is used in daily life for many foreign words loaned into Japanese (\"Gairaigo\").\n" ]
what are computer glitches and why do they occur?
In really simple terms, it's when you do something the program has no idea how to respond to. In a perfect program this would never happen, but no program is perfect. So for example if you glitch through the wall in a game, the game doesn't really know that you shouldn't be able to do that because maybe you did it in a way the game developer didn't think of. I hope that kinda explains it.
[ "It frequently refers to an error which is not detected at the time it occurs but shows up later in data errors or incorrect human decisions. Situations which are frequently called computer glitches are incorrectly written software (software bugs), incorrect instructions given by the operator (operator errors, and a failure to account for this possibility might also be considered a software bug), undetected invalid input data (this might also be considered a software bug), undetected communications errors, computer viruses, Trojan attacks and computer exploiting (sometimes called \"hacking\").\n", "Glitches/bugs are software errors that can cause drastic problems within the code, and typically go unnoticed or unsolved during the production of said software. These errors can be game caused or otherwise exploited until a developer/development team repairs them with patches. Complex software is rarely bug-free or otherwise free from errors upon first release.\n", "Software errors not detected by software testers during development can find their way into released versions of computer and video games. This may happen because the glitch only occurs under unusual circumstances in the game, was deemed too minor to correct, or because the game development was hurried to meet a publication deadline. Glitches can range from minor graphical errors to serious bugs that can delete saved data or cause the game to malfunction. In some cases publishers will release updates (referred to as \"patches\") to repair glitches. Sometimes a glitch may be beneficial to the player; these are often referred to as exploits.\n", "A glitch is a short-lived fault in a system, such as a transient fault that corrects itself, making it difficult to troubleshoot. The term is particularly common in the computing and electronics industries, in circuit bending, as well as among players of video games. More generally, all types of systems including human organizations and nature experience glitches.\n", "Such glitches could produce problems such as keyboard malfunction, number key failures, screen abnormalities (turned left, right or upside-down), random program malfunctions, and abnormal program registering.\n", "A glitch, which is slight and often temporary, differs from a more serious software bug which is a genuine functionality-breaking problem. Alex Pieschel, writing for \"Arcade Review\", said: \"“bug” is often cast as the weightier and more blameworthy pejorative, while “glitch” suggests something more mysterious and unknowable inflicted by surprise inputs or stuff outside the realm of code.\"\n", "Crash to desktop bugs are considered particularly problematic for users. Since they frequently display no error message, it can be very difficult to track down the source of the problem, especially if the times they occur and the actions taking place right before the crash do not appear to have any pattern or common ground. One way to track down the source of the problem for games is to run them in windowed-mode. Windows Vista has a feature that can help track down the cause of a CTD problem when it occurs on any program. Windows XP included a similar feature as well.\n" ]
how do companies calculate and quantify the value of a new sponsorship?
Essentially, they guess. It's an educated guess based on past experiences, but a guess none the less. They know the size of their current market, and what portion of that they own. They make an educated guess on how much incremental revenue they will gain from being sponsored by whomever and then offer a deal that shares some of the anticipated revenue increase with whomever is speaking on their behalf. There is no guarantee that they will make more money, but that is a risk they are willing to take, and the deals are full of clauses that allow early termination of the contract for a myriad of reasons.
[ "Various investor classes look to the financial sponsor to generate value in a company as much as the management or operations of the company. In particular, debt providers are willing to extend credit in the form of bank loans, high-yield debt and mezzanine capital based in part on the reputation of and relationship with the financial sponsor.\n", "The sales cycle for selling sponsors is often a lengthy process that consists of researching prospects, creating tailored proposals based on a company's business objectives, finding the right contacts at a company, getting buy-in from multiple constituencies and finally negotiating benefits/price. Some sales can take up to a year and sellers report spending anywhere between 1–5 hours researching each company that is viewed as a potential prospect for sponsorship.\n", "The company claimed in 2016 that they were worth $10 million. In 2016, the site generated most of its income from advertisements. Also in 2017, there was a message at the bottom of every article stating, \"Advertise\" that directed to information for potential sponsors. The company aims to generate income through ways apart from donations or banners.\n", "Marketing not only influences net profits but also can affect investment levels too. New plants and equipment, inventories, and accounts receivable are three of the main categories of investments that can be affected by marketing decisions. According to a recent study, business partnerships with \"micro-influencers\" can bring a greater ROI than collaborations with big celebrities.\n", "ValueAct Capital is a very active investor that identifies companies that appear to be out of favor, or may be undergoing a significant transition. ValueAct makes a significant investment into the company and then works with the company to make changes with goal of increasing long-term stock value and thus creating a profit for ValueAct. \n", "The difference is profit for the firm and its shareholders after all the costs and taxes owed by the business have been paid for that financial year. Value added or any related measure may help investors decide if this is a business that is worthwhile investing on, or that there are other and better opportunities (fixed deposits, debentures).\n", "BULLET::::- Sponsors do not get full value unless their data is used appropriately – sometimes this requires quality management, dissemination and branding efforts that can best be achieved by charging fees to users.\n" ]
how does googles new ai work?
It is called machine learning and is quite complex. What it boils down to is that you break a problem into discreet parts, or steps. You show the machine all the different steps it can do. Then you tell the machine what the desired "win" condition is. It goes and compares all the different steps and series of steps to figure out the best way of getting to the win condition. As these models get better and more generic, we can let the win condition be more vague and it has more freedom to learn how to get there. It is entirely dependent upon structures called neural networks, which have an input layer, a hidden layer, and an output layer. The hidden layer is trained to give outputs to certain inputs. For a board game training data might be all the moves of a bunch of games, and which player won. We don't actually know what the algorithm inside came up with, because we didn't tell it what to do. It learned how to get to the "win" state by analyzing a bunch of data and coming up with its own pattern recognition to apply to new games.
[ "Google bought the company in September 2016 and was initially known as API.AI; it provides tools to developers building apps (\"Actions\") for the Google Assistant virtual assistant. It was renamed on 10 October 2017 as Dialogflow.\n", "Founded in 2014, the Allen Institute for Artificial Intelligence (AI2)'s main focus is to research and engineer artificial intelligence. The Institute is modeled after the Allen Institute for Brain Science and led by researcher and professor, Dr. Oren Etzioni. AI2 has undertaken four main projects, Aristo, Semantic Scholar, Euclid and Plato. Project Aristo is working to build an AI system capable of passing an 8th grade science exam.\n", "Technologies developed by AI researchers have achieved commercial success in a number of domains, such as machine translation, data mining, industrial robotics, logistics, speech recognition, banking software, medical diagnosis, and Google's search engine.\n", "In January 2019, developers from Google's DeepMind tested their latest AI software by pitting a computer against a professional Starcraft II player. The AI won ten games in a row, with the human player winning the last match.\n", "In Max Tegmark's 2017 book \"Life 3.0\", a corporation's \"Omega team\" creates an extremely powerful AI able to moderately improve its own source code in a number of areas, but after a certain point the team chooses to publicly downplay the AI's ability, in order to avoid regulation or confiscation of the project. For safety, the team keeps the AI in a box where it is mostly unable to communicate with the outside world, and tasks it to flood the market through shell companies, first with Amazon Turk tasks and then with producing animated films and TV shows. While the public is aware that the lifelike animation is computer-generated, the team keeps secret that the high-quality direction and voice-acting are also mostly computer-generated, apart from a few third-world contractors unknowingly employed as decoys; the team's low overhead and high output effectively make it the world's largest media empire. Faced with a cloud computing bottleneck, the team also tasks the AI with designing (among other engineering tasks) a more efficient datacenter and other custom hardware, which they mainly keep for themselves to avoid competition. Other shell companies make blockbuster biotech drugs and other inventions, investing profits back into the AI. The team next tasks the AI with astroturfing an army of pseudonymous citizen journalists and commentators, in order to gain political influence to use \"for the greater good\" to prevent wars. The team faces risks that the AI could try to escape via inserting \"backdoors\" in the systems it designs, via hidden messages in its produced content, or via using its growing understanding of human behavior to persuade someone into letting it free. The team also faces risks that its decision to box the project will delay the project long enough for another project to overtake it.\n", "The Allen Institute for Artificial Intelligence (abbreviated AI2) is a research institute founded by late Microsoft co-founder Paul Allen. The institute seeks to achieve scientific breakthroughs by constructing AI systems with reasoning, learning, and reading capabilities. Oren Etzioni was appointed by Paul Allen in September 2013 to direct the research at the institute.\n", "The game's artificial intelligence (AI) is based on creating a multistep \"lookahead\" decision tree from the current state of the board, evaluating possible known moves by its opponents, scoring the current and resulting situations both positively and negatively, and then selecting the option with the highest score. The artificial intelligence utilizes threading to run the main intelligence algorithm alongside three \"sub-contractors\" that evaluate the possible future states and report back to the main algorithm. Each of these instances of the AI uses a special engine that both implements the rules of \"Magic\", while also providing \"undo\" actions so that they may explore up and down the decision tree. The game's intelligence algorithm runs on a separate CPU core than the main game to avoid creating framerate issues with its display while the player is considering their actions. When the player makes an action, the AI threads are interrupted and brought to the same state as the game, then continue processing, in some cases, reusing existing branches on the decision tree that match with the player's selection. The AI itself does not employ any strategy in terms of strong card combinations, but instead, when such combinations improve the computer's situation, the chances of playing these combinations will \"ripple\" back through the decision tree. In this manner, the computer AI will play one card of such a combination should it be the best choice at that time, and follow up with the second card of the combination should that option still remain the best after any other player actions.\n" ]
how would people, before television or radio decide who to vote for?
Newspapers were the major way information was spread before radio. Political candidates also did a lot more public speaking. They did "whistle-stop tours" where they basically road trains and stopped in every town and gave a speech. Local politicians would also speak in favor of their party in more remote areas.
[ "Historically, a country's votes were decided by an internal jury, but in 1997 five countries (Austria, Switzerland, Germany, Sweden and the United Kingdom) experimented with televoting, giving members of the public in those countries the opportunity to vote \"en masse\" for their favourite songs. The experiment was a success, and from 1998 onwards all countries were encouraged to use televoting wherever possible. Back-up juries are still used by each country, in the event of a televoting failure. Nowadays members of the public may also vote by SMS, in addition to televoting. In every case, every country cannot vote for its own song From 2013, the public may also vote via a mobile app.\n", "Before there were paper ballots, people would simply call out their selection at the polling place. This polling place was typically the county courthouse or town hall. Sometimes these polls were taken outside of the venue in a more informal fashion. When the voters came to the town hall to announce their choice, they would get in line to see the judge and swear in. Once the voter put his or her hand on the Bible and swore to the judge, they would be allowed to cast one ballot per election. The judge acted as the only form of voter identification and it was up to them to be able to identify individuals that had already voted and exclude them from voting again.\n", "Every person is allowed to vote for up to ten songs. Initially (until April 1996) the votes had to be mailed to the radio station for compiling or the audience had to call the station and dictate the names of the songs. Since then voting by Internet (initially a BBS site, then a www site) is the main means of voting.\n", "First the votes were sent on postcards or dictated on the phone by the listeners, since April 1996 voting has been done by the Internet. The number of votes per person changed from two in the beginning to ten nowadays. The Charts used to include 20 songs, then 30, 40, now it is 50 divided into two parts: the so-called \"waiting room\" (only a small part of each song is presented) and the basic list (the songs going up are presented in full).\n", "In the days before radio, one effective way to get a message out into society and to have it discussed was to produce short plays that could be performed around the country, and so suffrage drama was born. Elizabeth Robins's \"Votes for Women\" and Cicely Hamilton and Christopher St. John's \"How the Vote Was Won\" are two predominant examples of the genre. Hamilton also wrote \"A Pageant of Great Women\", a highly successful women's suffrage play based on the ideas of her friend, the theatre director Edith Craig. Hamilton played Woman while Craig played the painter Rosa Bonheur, one of the 50 or so great women in the play. It was produced all over the UK from 1909 until the First World War. Hamilton was a member of Craig's theatre society, the Pioneer Players. Her play \"Jack and Jill and a Friend\" was one of the three plays in the Pioneer Players' first production in May 1911.\n", "The table below shows the order in which votes were cast during the 1960 contest along with the spokesperson who was responsible for announcing the votes for their respective country. Each national broadcaster also sent a commentator to the contest, in order to provide coverage of the contest in their own native language. Details of the commentators and the broadcasting station for which they represented are also included in the table below.\n", "The idea of an election in which no one knew the results until voting was over was not present in the 19th century. In 1855, in every province but New Brunswick, voters cast their ballots orally. This led to a system where information on who voted for who could be used for nefarious reasons. Intimidation and blackmail to force people to vote in a certain way.\n" ]
Why exactly does spinning rapidly throw off our equilibrium.
This is not my area of expertise. But just in case you don't get a more detailed answer. I'm faily certain what you are looking for is the [Vestibular System](_URL_0_) of the inner ear. It contains a fluid which moves with your head/body and helps your brain understand/maintain your balance. When you spin very rapidly you do something with the fluid in this system that your brain doesn't quite understand. Someone else will need to come in to explain how exactly that fluid is disturbed as I am not qualified.
[ "where \"θ\" is the angle between the vectors Ω and L. Thus, if the top's spin slows down (for example, due to friction), its angular momentum decreases and so the rate of precession increases. This continues until the device is unable to rotate fast enough to support its own weight, when it stops precessing and falls off its support, mostly because friction against precession cause another precession that goes to cause the fall.\n", "Since the spinning mass is balanced, the only possibility to speed up the rotation is for the sides of the groove to exert forces on the ends of the axle. Furthermore, the normal and axial forces will have no effect, so tangential force must be provided by friction. If the axle is stationary, the friction will only act to slow down the rotation, but the situation is very different if the axle is turned by applying a torque.\n", "If energy is dissipated while an object is rotating, this will cause the polhode motion about the axis of maximum inertia (also called the major principal axis) to damp out or stabilize, with the polhode path becoming a smaller and smaller ellipse or circle, closing in on the axis.\n", "Gyro precession causes another phenomenon for spinning objects such as the bicycle wheel in this scenario. If the subject holding the wheel removes a hand from one end of its axle, the wheel will not topple over, but will remain upright, supported at just the other end. However, it will immediately take on an additional motion; it will begin to rotate about a vertical axis, pivoting at the point of support as it continues spinning. If you allowed the wheel to continue rotating, you would have to turn your body in the same direction as the wheel rotated. If the wheel was not spinning, it would obviously topple over and fall when one hand is removed. The initial action of the wheel beginning to topple over is equivalent to applying a force to it at 12 o'clock in the direction toward the unsupported side (or a force at 6 o’clock toward the supported side). When the wheel is spinning, the sudden lack of support at one end of its axle is equivalent to this same force. So, instead of toppling over, the wheel behaves as if a continuous force is being applied to it at 3 or 9 o’clock, depending on the direction of spin and which hand was removed. This causes the wheel to begin pivoting at the one supported end of its axle while remaining upright. Although it pivots at that point, it does so only because of the fact that it is supported there; the actual axis of precessional rotation is located vertically through the wheel, passing through its center of mass. Also, this explanation does not account for the effect of variation in the speed of the spinning object; it only illustrates how the spin axis behaves due to precession. More correctly, the object behaves according to the balance of all forces based on the magnitude of the applied force, mass and rotational speed of the object. Once it is visualized why the wheel remains upright and rotates, it can easily be seen why the axis of a spinning top slowly rotates while the top spins as shown in the illustration on this page. A top behaves exactly like the bicycle wheel due to the force of gravity pulling downward. The point of contact with the surface it spins on is equivalent to the end of the axle the wheel is supported at. As the top's spin slows, the reactive force that keeps it upright due to inertia is overcome by gravity. Once the reason for gyro precession is visualized, the mathematical formulas start to make sense.\n", "One possible origin of the word \"tilt\" is as a reference to tilting a pinball machine. The frustration from seeing the ball follow a path towards the gap between the flippers can lead to the player physically tilting the machine in an attempt to guide the ball towards the flippers. However, in doing so, some games will flash the word \"TILT\" and freeze the flippers, causing the ball to be lost for certain; as in poker, this suggests that over-aggression due to frustration leads to severely detrimental playing techniques.\n", "Rotating unbalance is the uneven distribution of mass around an axis of rotation. A rotating mass, or rotor, is said to be out of balance when its center of mass (inertia axis) is out of alignment with the center of rotation (geometric axis). Unbalance causes a moment which gives the rotor a wobbling movement characteristic of vibration of rotating structures.\n", "It is important to note that these changes in the orientation of the body as it spins may not be due to external torques, but rather result from energy dissipated internally as the body is spinning. Even if angular momentum is conserved (no external torques), internal energy can be dissipated during rotation if the body is not perfectly rigid, and any rotating body will continue to change its orientation until it has stabilized around its axis of maximum inertia, where the amount of energy corresponding to its angular momentum is least.\n" ]
what makes an everyday expense/activity tax deductable?
My dad's an accountant. Short answer is, you have to have a business or at least individual income to deduct business expenses. If you're an employee somewhere that doesn't count. If you don't file schedule C there are not many deductions you can take like what you're taking about. If you do, but the deductions are all or most of the income, you're asking for audit. If there's plenty of income and the business requires attracting clients, there's little oversight for deductions. But remember, it's deducted from taxable income, not taxes i.e. you only avoid taxes on the spent amount. Unless you spend a lot, your reduction in tax bill is not very large.
[ "Similarly, the U.S. federal income tax law does not tax the imputed income consisting of the benefit one obtains from leisure. Professor Andrews wrote: \"It would undoubtedly be impractical to try to reflect either the value of leisure or the cost of its sacrifice directly in taxable income.\"\n", "William Nordhaus argues that allocations cost the economy as they cause the under utilisation an efficient form of taxation. Nordhaus argues that normal income, goods or service taxes distort efficient investment and consumption, so by using pollution taxes to generate revenue an emissions scheme can increase the efficiency of the economy.\n", "BULLET::::- Payroll taxes, such as FICA and Unemployment Insurance in the United States, and consumption taxes such as Value Added Tax and sales taxes are regressive in that they both raise prices of purchased goods. Lower-income earners save and invest less money, so pay a larger proportion of their income toward these taxes, directly for sales tax and as the price increase required to make revenue covering payrolls for payroll taxes.\n", "Most taxes (see below) have side effects that reduce economic welfare, either by mandating unproductive labor (compliance costs) or by creating distortions to economic incentives (deadweight loss and perverse incentives).\n", "Tax expenditures (i.e., exclusions, deductions, preferential tax rates, and tax credits) cause revenues to be much lower than they would otherwise be for any given tax rate structure. The benefits from tax expenditures, such as income exclusions for healthcare insurance premiums paid for by employers and tax deductions for mortgage interest, are distributed unevenly across the income spectrum. They are often what the Congress offers to special interests in exchange for their support. According to a report from the CBO that analyzed the 2013 data:\n", "Tax expenditures (i.e., exclusions, deductions, preferential tax rates, and tax credits) cause revenues to be much lower than they would otherwise be for any given tax rate structure. The benefits from tax expenditures, such as income exclusions for healthcare insurance premiums paid for by employers and tax deductions for mortgage interest, are distributed unevenly across the income spectrum. They are often what the Congress offers to special interests in exchange for their support. According to a report from the CBO that analyzed the 2013 data:\n", "Tax expenditures (i.e., exclusions, deductions, preferential tax rates, and tax credits) cause revenues to be much lower than they would otherwise be for any given tax rate structure. The benefits from tax expenditures, such as income exclusions for healthcare insurance premiums paid for by employers and tax deductions for mortgage interest, are distributed unevenly across the income spectrum. They are often what the Congress offers to special interests in exchange for their support. According to a report from the CBO that analyzed the 2013 data:\n" ]
Natural sleeping position without a pillow
This is kind of an impossible question to answer from an archaeological perspective. Anything early humans/hominids would have used as a pillow would have been made from perishable materials that do not survive in archaeological contexts. Anything from furs to piles of leaves could have been used by early hominids as pillows. Other great apes make nests out of leaves, it's possible we did too. There's no way to know. As for natural sleeping positions, you're probably better off asking a medical doctor or a biological anthropologist. Sorry I can't be of more help.
[ "The choice of bed pillow depends to some extent upon sleeping positions: one manufacturer recommends a thinner and softer pillow for sleeping face down, medium support for sleeping on one's back, and a thicker and firmer pillow for sleeping on the side.\n", "A pillow is a support of the body at rest for comfort, therapy, or decoration. Pillows are used by many species, including humans. Some types of pillows include throw pillows and decorative pillows. Pillows that aid sleeping are a form of bedding that supports the head and neck. Other types of pillows are designed to support the body when lying down or sitting. There are also pillows that consider human body shape for increased comfort during sleep. Decorative pillows used on beds, couches or chairs are sometimes referred to as cushions.\n", "Pillow fights are known to occur during children's sleepovers. Since pillows are usually soft, injuries rarely occur. The heft of a pillow can still knock a young person off balance, especially on a soft surface such as a bed, which is a common venue. In earlier eras, pillows would often break, shedding feathers throughout a room. Modern pillows tend to be stronger and are often filled with a solid block of artificial filling, so breakage occurs far less frequently.\n", "Beds may have a headboard for resting against, and may have side rails and footboards (or \"footers\"). \"Headboard only\" beds may incorporate a \"dust ruffle\", \"bed skirt\", or \"valance sheet\" to hide the bed frame. To support the head, a pillow made of a soft, padded material is usually placed on the top of the mattress. Some form of covering blanket is often used to insulate the sleeper, often bed sheets, a quilt, or a duvet, collectively referred to as bedding. Bedding is the removable non-furniture portion of a bed, which enables these components to be washed or aired out.\n", "The Romans and Greeks of ancient Europe mastered the creation of the softer type pillow. These pillows were stuffed with reeds, feathers, and straw in order to make them softer and more comfortable. Only upper-class people typically owned these softer pillows; however, all classes of people were allowed to use some type of pillow while sleeping, lying down or sitting in order to give them support. People in ancient Europe started to use pillows when going to church in order to kneel on while praying and to place holy books on. This is a tradition that still lives on today. In addition, the Romans and Greeks used their pillows by placing them under the head of those deceased just like the ancient Egyptians did.\n", "Decorative pillows are also found on furnishings in more public parts of the home, such as sofas, chairs and window seats. Here, their common use may overlap both orthopedic and bed pillows. For example, unless a person has some particular medical condition, they will likely use a handy decorative pillow for lumbar support, as needed, while seated on a sofa. Likewise, for the occasional nap, decorative pillows are handy for supporting the head or neck, even though they may not be covered with a pillow case, as are bed pillows.\n", "Lying-in is the term given to the European forms of postpartum confinement, the traditional practice involving long bed rest after giving birth. The term and the practice it describes are old-fashioned or archaic, but it used to be considered an essential component of the postpartum period, even if there were no medical complications during childbirth.\n" ]
Why was the Revolutionary War over after the battle of Yorktown? I never understood why the rest of the British forces didn't keep fighting, especially if they still had significant forces based in New York City, and if Cornwallis wasn't even the commander of British forces in the states.
Cornwallis had one of the armies in the South with Alexander Stewart commanding another in South Carolina around Charleston. The lack of a foothold in the states and resilience of the American army helped push the British surrender and evacuation. British forces held New York City, Charleston, Georgia, and other places but could not fully control full regions. Even though the British army held Charleston, they did not have control of the rest of South Carolina. Slowly, the British forces in SC trickled back to the Charleston area after evacuating other posts such as Camden and Ninety-Six, and losing forts such as Fort Motte. The Battle of Eutaw Springs put an end with British forces trying to gain more control of South Carolina as the army would remain there until total evacuation. Even the great victories of Camden and Charleston did not seem to quell the continuing operations of the American armies. In addition, while Henry Clinton commanded the entire British army, he had distanced himself from Cornwallis after May 1780. Clinton had a disagreement with Cornwallis and for a time he would correspond with Cornwallis for updates they became sporadic by mid-1781. Clinton remained the commander in chief until 1782 when Guy Carleton took over, but the war had just about ended by then. The British army after Saratoga had diminished substantially. Burgoyne's army became part of the Convention army. Not really prisoners of war but could not fight anymore in the war. That surrender took away 6,000 soldiers from the British army - a huge blow. Cornwallis' army numbered to almost 8,000. Clinton had a small force in New York and Stewart's army in Charleston had about 2,500. None of these armies had the capability to go against the larger American force. The British army had spread itself thin with the various theatres that popped up in 1778 including Anglo-French, Anglo-Spanish wars that made them fight not only in the Caribbean but the west including Louisiana and Kentucky (at least, what would become them). The surrender of Cornwallis' army spelled the end of the war effort in America. At that point, the American army had become a well disciplined force that showed no sign of shrinking. With the addition of Spanish and French forces, the British army had a lot to contend with and could not with the decreased amount of soldiers. Often times, army leaders plan on crushing enemy forces and occupy the capital cities to win the war. The colonies remained decentralized and if one city fell, the rest would remain in resistance. The British army also failed to fully crush enemy combatants after a battle. Most militia units in the South had horses for retreating purposes and could swiftly escape the field. American forces also tended to regroup after battles and gain reinforcements, as well as relying on more guerilla tactics. The 1781 campaign started with the battle of Cowpens where the British forces lost and began the Race to the Dan. This would in turn force Cornwallis' hand with the need to chase Nathaniel Greene or beat him to the Dan river. The lose of the battle also spelled disaster for the rest of the campaign as the British army had to leave supply wagons behind in order to march at a quick pace through North Carolina. It also halted his plans for controlling the rest of South Carolina. The long chase ended at Guilford Courthouse with Cornwallis battling a renewed American army. After suffering a Pyrrhic victory, Cornwallis decided to head to the coast in order to get reinforcements and supplies at Yorktown. Cornwallis also had some of the most experienced and seasoned regiments with him at the surrender including the 33rd, 71st, 23rd, the Guard infantry, the British Legion, and German battalions. With morale crushed, Lord North proclaimed the war to be over and the administration switched from war to peace negotiations (and prisoner exchange). Ironically, the naval campaigns in the Caribbean had been quite successful but not enough to continue the war. Matthew Spring, With Zeal and Bayonets Only. John Buchanan, The Road to Guilford Courthouse. Charles Ross (ed.), Correspondence of Charles, First Marquis of Cornwallis.
[ "New York City remained in British hands until the end of the war, behind the protection of its large garrison and the Royal Navy. However, the inability of British commanders to defeat the Revolution led the war to drag on for eight years, resulting in the Continental Army's capture of two major British armies (\"Battle of Saratoga\", \"Battle of Yorktown\") and eroding the British political will to attempt a military solution. As a result, the government of Lord North collapsed and the new British government was formed from parliamentary advocates of a negotiated peace.\n", "In April 1776, after the conclusion of the Boston campaign, General Washington and the Continental Army marched to New York City and prepared for an anticipated attempt by the British Army to occupy the city. The Royal Governor of New York, William Tryon, had been driven out of the city by revolutionary forces and was compelled to seek refuge on a ship in New York Harbor. Nevertheless, the city had many Loyalist residents who favored the British side.\n", "The British returned in force in August 1776, landing in New York and defeating the fledgling Continental Army at the Battle of Long Island in one of the largest engagements of the war. They quickly seized New York City and nearly captured General Washington and his army. The British made the city their main political and military base of operations in North America, holding it until late 1783. Patriot evacuation and British military occupation made the city the destination for Loyalist refugees, and a focal point of Washington's intelligence network.\n", "New York played a pivotal role during the American Revolution and subsequent war. The Stamp Act Congress in 1765 brought together representatives from across the Thirteen Colonies to form a unified response to British policies. The Sons of Liberty were active in New York City to challenge British authority. After a major loss at the Battle of Long Island, the Continental Army suffered a series of additional defeats that forced a retreat from the New York City area, leaving the strategic port and harbor to the British army and navy as their North American base of operations for the rest of the war. The Battle of Saratoga was the turning point of the war in favor of the Americans, convincing France to formally ally with them. New York's constitution was adopted in 1777, and strongly influenced the United States Constitution. New York City was the national capital at various times between 1785 and 1790, where the Bill of Rights was drafted. Albany became the permanent state capital in 1797. In 1787, New York became the eleventh state to ratify the United States Constitution.\n", "By 1781, the Revolutionary War had become a stalemate between Britain and the United States. The British had concentrated their military operations in the Southern theater of the war, while leaving a large force garrisoned at New York City. In August 1781, Morris met with General Washington and the comte de Rochambeau, who were planning a joint Franco-American operation against the British forces. Morris redirected government funds to purchase supplies for Washington's march against British forces in Virginia, and he pleaded with state governments and the French government for further funding. At the October 1781 Battle of Yorktown, Washington forced the surrender of the British army under the command of General Cornwallis. After the Battle of Yorktown, Britain essentially abandoned its campaign on land, but the naval war continued as Britain sought to cut the United States off from its sources of trade.\n", "In 1777, the British retreated to New York City to protect it from an expected French attack. Washington quickly ordered his soldiers to march towards the British and met them at the Battle of Monmouth.\n", "The American Revolutionary War was a qualified success for the British in 1776. After being forced to abandon Boston, they captured New York City, but were unable to hold New Jersey when General George Washington surprised them at Trenton and Princeton. The British consolidated their hold on New York City and Long Island during the winter months of early 1777, while the Continental Army established a land blockade around the city in New Jersey, southern New York, and southwestern Connecticut.\n" ]
If we could cause nuclear fusion on Earth (in a reactor, of course), could we use the products for fission?
It would be an unlimited source of energy the same way having a water turbine run on water that we can then pump back up. Law no.1: If it gave out energy, you need to provide energy in order to reverse it and vice versa. Iron is the bottom of nucleus potential energy. Fusion of lighter than iron nuclei with a lighter than iron result is going to release energy. Fission of heavier than iron nuclei into heavier than iron products will release energy. For vice versa you need to put energy in. Imagine all elements lined up in a valley shaped curve with the iron at the bottom. Energy is released only when going downhill.
[ "All current thermonuclear weapons use a fission bomb as a first stage to create the high temperatures and pressures necessary to start a fusion reaction between deuterium and tritium in a second stage. For many years, nuclear weapon designers have researched whether it is possible to create high enough temperatures and pressures inside a confined space to ignite a fusion reaction, without using fission. Pure fusion weapons offer the possibility of generating arbitrarily small nuclear yields because no critical mass of fissile fuel need be assembled for detonation, as with a conventional fission primary needed to spark a fusion explosion. Also there is an advantage of reduced collateral damage stemming from fallout because these weapons would not create the highly radioactive byproducts associated with fission-type weapons. These weapons would be lethal not only because of their explosive force, which could be large compared to bombs based on chemical explosives, but also because of the neutrons they generate. A potentially useful feature of a fissionless weapon would be that no electromagnetic pulse would be produced, because this originates from the gamma rays released by fissioning nuclei.\n", "Fusion rockets, powered by nuclear fusion reactions, would \"burn\" such light element fuels as deuterium, tritium, or He. Because fusion yields about 1% of the mass of the nuclear fuel as released energy, it is energetically more favorable than fission, which releases only about 0.1% of the fuel's mass-energy. However, either fission or fusion technologies can in principle achieve velocities far higher than needed for Solar System exploration, and fusion energy still awaits practical demonstration on Earth.\n", "Fusion–fission designs essentially replace the lithium blanket with a blanket of fission fuel, either natural uranium ore or even nuclear waste. The fusion neutrons have more than enough energy to cause fission in the U-238, as well as many of the other elements in the fuel, including some of the transuranic waste elements. The reaction can continue even when all of the U-235 is burned off; the rate is controlled not by the neutrons from the fission events, but the neutrons being supplied by the fusion reactor.\n", "This would complete the fission-fusion-fission sequence. Fusion, unlike fission, is relatively \"clean\"—it releases energy but no harmful radioactive products or large amounts of nuclear fallout. The fission reactions though, especially the last fission reactions, release a tremendous amount of fission products and fallout. If the last fission stage is omitted, by replacing the uranium tamper with one made of lead, for example, the overall explosive force is reduced by approximately half but the amount of fallout is relatively low. The neutron bomb is a hydrogen bomb with an intentionally thin tamper, allowing most of the fast fusion neutrons as possible to escape.\n", "In fusion-fission hybrid reactors, high-energy fast neutrons generated by nuclear fusion can be used to convert fertile materials such as Thorium-232 or Uranium-238 into fissionable fuels for fission reactors such as Uranium-233 or Plutonium-239. The high-energy fusion-generated neutrons can also be used to trigger nuclear fission in traditional nuclear fuels such as Uranium-235. The hybrid reactor concept has a fusion reactor at the core and a surrounding “blanket” of fertile material, either Thorium-232 or Uranium-238. In a hybrid reactor, high-energy neutrons from fusion reactions are used to produce fissionable materials such as Uranium-233 or Plutonium-239 which would need then to be reprocessed and used as fuel in a fission nuclear power plant. One fusion hybrid plant can provide enough fuel for as many as five nuclear power plants of the same power.\n", "With special designs of the nuclear explosive Ted Taylor estimated that fission product fallout could be reduced tenfold, or even to zero, if a pure fusion explosive could be constructed instead. A 100% pure fusion explosive has yet to be successfully developed, according to declassified US government documents, although relatively clean PNEs (Peaceful nuclear explosions) were tested for canal excavation by the Soviet Union in the 1970s with 98% fusion yield in the \"Taiga\" test's 15 kiloton devices, 0.3 kilotons fission, which excavated part of the proposed Pechora–Kama Canal.\n", "Research has been done into the possibility of pure fusion bombs: nuclear weapons that consist of fusion reactions without requiring a fission bomb to initiate them. Such a device might provide a simpler path to thermonuclear weapons than one that required development of fission weapons first, and pure fusion weapons would create significantly less nuclear fallout than other thermonuclear weapons, because they would not disperse fission products. In 1998, the United States Department of Energy divulged that the United States had, \"...made a substantial investment\" in the past to develop pure fusion weapons, but that, \"The U.S. does not have and is not developing a pure fusion weapon\", and that, \"No credible design for a pure fusion weapon resulted from the DOE investment\".\n" ]
Why do we have organisms on Earth that can survive in the vacuum of space?
Why not? It is a side effect of having a form to survive harsh conditions on Earth, such as drought.
[ "Curiously, for as little life as there \"exists\" in the Negative Zone, it is very \"capable\" of supporting life. There, outer space itself is permeated with an oxygen-rich atmosphere, closely approximating Earth's. Consequently, humans can exist peacefully in space without the need for cumbersome pressure suits or oxygen masks.\n", "The space environment is hostile to life. Unprotected presence in space is characterised by an intense radiation field (consisting primarily of protons and other subatomic charged particles from the solar wind, in addition to cosmic rays), high vacuum, extreme temperatures, and microgravity. Some simple forms of life called extremophiles, as well as small invertebrates called tardigrades can survive in this environment in an extremely dry state through desiccation.\n", "Most of the moons and planets in the Solar System are also extreme environments. Astrobiologists have not yet found life in any environments beyond Earth, though experiments have shown that tardigrades can survive the harsh vacuum and intense radiation of outer space. The conceptual modification of conditions in locations beyond Earth, to make them more habitable by humans and other terrestrial organisms, is known as terraforming.\n", "By understanding how extremophilic organisms can survive the Earth's extreme environments, we can also understand how microorganisms could have survived space travel and how the panspermia hypothesis could be possible.\n", "Although the vast majority of life on Earth lives in mesophyllic (moderate) environments, a few organisms, most of them microbes, have managed to colonise extreme environments that are unsuitable for most higher life forms. There are bacteria, for example, living in Lake Whillans, half a mile below the ice of Antarctica; in the absence of sunlight, they must rely on organic material from elsewhere, perhaps decaying matter from glacier melt water or minerals from the underlying rock. Other bacteria can be found in abundance in the Mariana Trench, the deepest place in the ocean and on Earth; marine snow drifts down from the surface layers of the sea and accumulates in this undersea valley, providing nourishment for an extensive community of bacteria.\n", "Molecular oxygen () can be produced by geophysical processes, as well as a byproduct of photosynthesis by life forms, so although encouraging, is not a reliable biosignature. In fact, planets with high concentration of in their atmosphere may be uninhabitable. Abiogenesis in the presence of massive amounts of atmospheric oxygen could be difficult because early organisms relied on the free energy available in redox reactions involving a variety of hydrogen compounds; on an -rich planet, organisms would have to compete with the oxygen for this free energy.\n", "Lifeforms do not live just anywhere on the surface of a planet randomly. Each species occupies a definite set of surroundings, or environment, to which it is adapted. It cannot survive for long outside the limits of that environment because it can no longer obtain what it requires to survive.\n" ]
Did the Greeks and Romans read the "classics"
> Did the Greeks and Romans read the "classics"? Absolutely. I'm much more familiar with the Romans than the Greeks, so I can only speak as to them - though I can't think of any reason that the Greeks would be different - but today's great Latin literature was definitely important back then as well. For example, *The Aeneid* was, by the 2nd century A.D., mandatory if one wanted to claim a "good" education. In fact, it'd often be memorized.[^(Source, missing one page)](_URL_2_) *The Aeneid*, like *The Illiad*, was and is regarded as the epitome of literature in its language, and was, consequently, hugely popular. Plus, it certainly didn't hurt its popularity that it painted the ruling Julio-Claudian dynasty in the best of lights. > Do we have the same Greek and Roman canon as the actual Greeks and Romans? Nope, and it's a real tragedy. Regrettably, "we've lost more of Latin literature than we posses."[^Source ^\(Abstract\)](_URL_0_)^|[Source2](_URL_1_) Entire authors have disappeared from history, and we lack an enormous amount of what the Romans and Greeks wrote. It's the bane of every classicist the world over, but there's not much to be done about it.
[ "Through the Roman Empire, Greek literature also continued to make an impact in Europe long after the Empire's fall, especially after the recovery of Greek texts from the East during the high Middle Ages and the resurgence of Greek literacy during the Renaissance. Plutarch's \"Lives of the Noble Greeks and Romans\", for instance, originally written in Greek, was widely read by educated Westerners from the Renaissance up to the 20th century. Shakespeare's play Julius Caesar takes most of its material from Plutarch's biographies of Caesar, Cato, and Brutus, whose exploits were frequently discussed and debated by the literati of the time.\n", "The Ancient Classics for English Readers series was a collection of volumes of classics of ancient Greek and Latin literature, translated into English with paraphrases and commentaries by leading classical scholars. The series was published from 1870 in Edinburgh and London by William Blackwood and Sons. From 1870 J. B. Lippincott & Co. republished the series in Philadelphia. The founding editor was W. Lucas Collins.\n", "The classics are the literature of ancient Greece and Rome, known as classical antiquity, and once the principal subject studied in the humanities. Classics (without the definite article) can refer to the study of philosophy, literature, history and the arts of the ancient world, as in \"reading classics at Cambridge\". From that usage came the more general concept of 'classic'.\n", "Classical studies (Classics for short) – earliest branch of the humanities, which covers the languages, literature, history, art, and other cultural aspects of the ancient Mediterranean world. The field focuses primarily on, but is not limited to, Ancient Greece and Ancient Rome during classical antiquity, the era spanning from the late Bronze Age of Ancient Greece during the Minoan and Mycenaean periods (c. 1600-1100 BCE) through the period known as Late Antiquity to the fall of the Western Roman Empire, c. 500 CE. The word classics is also used to refer to the literature of the period.\n", "In the early Italian Renaissance, much of the focus was on translating and studying classic works from Latin and Greek. Renaissance authors were not content to rest on the laurels of ancient authors, however. Many authors attempted to integrate the methods and styles of the ancient Greeks into their own works. Among the most emulated Romans are Cicero, Horace, Sallust, and Virgil. Among the Greeks, Aristotle, Homer, and Plato were now being read in the original for the first time since the 4th century, though Greek compositions were few.\n", "BULLET::::- Classics   (outline) – study of the languages, literature, philosophy, history, art, archaeology and all other cultural elements of the ancient Mediterranean world (Bronze Age ca. BC 3000 to Late Antiquity ca. AD 300–600); especially Ancient Greece and Ancient Rome.\n", "During the Middle Ages, ancient Greek literature was largely forgotten in Western Europe. The medieval writer Roger Bacon wrote that \"there are not four men in Latin Christendom who are acquainted with the Greek, Hebrew, and Arabic grammars.\" It was not until the Renaissance that Greek writings were rediscovered by western European scholars. During the Renaissance, Greek began to be taught in western European colleges and universities for the first time, which resulted in western European scholars rediscovering the literature of ancient Greece. The \"Textus Receptus\", the first New Testament printed in the original Greek, was published in 1512 by the Italian humanist scholar Desiderius Erasmus. Erasmus also published Latin translations of classical Greek texts, including a Latin translation of Hesiod's \"Works and Days\".\n" ]
Are there any ancient civilizations that did not build stairs?
The wheel is something that would need to be invented in order to be used, wheras stairs and steps can form naturally, so this question would be incredibly hard to answer whilst sticking to the AskHistorians top answer guidelines. The wheel was an invention. Somebody wanted to find a way to reduce the amount of force necessary to move large objects and invented the wheel and axel design to do so. its much more complicated then it sounds.
[ "It is believed the stairs were built some time before the rule of Tiberius (14–37), as they were not mentioned by name in any ancient texts that predate this period. Their first use as a place of execution is primarily associated with the paranoid excesses of Tiberius' later reign.\n", "Only a minority of the houses had evidence of staircases survive, demonstrating that they definitely had upper storeys, while for the remainder of Olynthian houses the evidence is inconclusive. \"On the Murder of Eratosthenes\" demonstrates that at least some Athenian houses also had an upper storey. Entranceways at Olynthos were designed for privacy, preventing passers-by from seeing inside the house.\n", "The list of ancient spiral stairs contains a selection of Greco-Roman spiral stairs constructed during classical antiquity. The spiral stair is a type of stairway which, due to its complex helical structure, has been introduced relatively late into architecture. Although the oldest example dates back to the 5th century BC, it was only in the wake of the influential design of the Trajan's Column that this space-saving new type permanently caught hold in Roman architecture.\n", "A description of the staircase in the 18th century noted that the staircase was singular among palaces in Rome for it size and being constructed entirely of marble steps, costing 80 scudi each, arrayed in four flights of 30 steps, ten feet long and two feet wide. Along the stairs were antique busts of emperors Hadrian and Claudius; Bacchus and Silen; Apollo; Mercury; a woman dressed as Hercules; and Aesclepius.\n", "This folly draws on architectural themes from ancient Egypt, although the form is not pyramidal. There is an obelisk nearby, carved with hieroglyphs and borne on the back of four tortoises. It was possible to access the building through an underground passage. The path to this monument was cut out of the hillside and several bridges were used to overcome differences in level. The Egyptian tomb is the only folly to survive to the present with its structure essentially intact.\n", "Project co-director Yannis Gourdon from the IFAO says that \"the system is composed of a central ramp flanked by two staircases with numerous post holes. Using a sled which carried a stone block and was attached with ropes to these wooden posts, ancient Egyptians were able to pull up the alabaster blocks out of the quarry on very steep slopes of 20 percent or more.\"\n", "Access to the upper floor is a special feature of Khmer houses and very varied. The simplest form may be a wooden ramp or ladder, whilst staircases can be of wooden filigree, spiral, or partially built in masonry. Whereas this means of access would formerly have been handmade, nowadays prefabricated concrete elements are widely used.\n" ]
Why didn't the destroyed cities of Europe change their city layouts after World War II?
Some of it has to do with the internal geography (or metageography) that the residents carried. After a war where the familiar has been smashed, people very often wanted to hold on to some of that familiarity, and spatial arrangements were an important part of social interaction. Besides, what was *below* the ground wasn't necessarily destroyed, so the infrastructure for rebuilding in the same configurations existed already even if the streets were cratered. Minor changes were probably rife: slight widening of streets and roads, tearing up of brick pavements and streetcar rails, expansion of certain facilities, the burying of power cables, et cetera. But I suspect the spatial order was persistent because people *wanted* it to be as much as because it was easier to do. If any city would have been prone to a total reconstruction, it would have been Rotterdam after the Blitz. Although its skyline certainly changed, and the oud-centrum was vastly modified, I am not aware of whether the basic layout shifted (streets, etc). There's a certain psychologically restorative quality to that continuity in the face of an undeniably external, human-created disaster (as opposed to London 1666 or Chicago 1871, which came "from within" and highlighted distinct problems with the prior order), although I'm not sure if anyone's written on it specifically off the top of my head. It would not surprise me if one or another of the spatial theorists has done so. *Environment and Planning D: Society and Space* is a great journal to farm for that, if you're so inclined, as is *Progress in Human Geography*. I'm not in my office so I don't have access to my indexes.
[ "Many eastern European countries had suffered physical damage during World War II and their economies were in a very poor state. There was a need to reconstruct cities which had been severely damaged due to the war. For example, Warsaw, Poland had been practically razed to the ground under the planned destruction of Warsaw by German forces after the 1944 Warsaw Uprising. The centre of Dresden, Germany had been totally destroyed by the 1945 Allied bombardment. Stalingrad had been largely destroyed and only a small number of structures were left standing.\n", "Many eastern European countries had suffered physical damage during World War Two and their economies were in a very poor state. There was a need to reconstruct cities which had been severely damaged due to the war. For example, Warsaw, Poland had been practically razed to the ground under the planned destruction of Warsaw by German forces after the 1944 Warsaw Uprising. The centre of Dresden, Germany had been totally destroyed by the 1945 Allied bombardment. Stalingrad had been largely destroyed and only a small number of structures were left standing.\n", "From 1944 to 1991, cities experienced an ordered development with a decline in architectural quality. Massive socialist-styled apartment complexes, wide roads, and factories were constructed, while town squares were redesigned and a number of historic buildings demolished.\n", "After the city centre was destroyed in World War II, the East German authorities attempted to rebuild it to symbolise the conceptions of urban development of a socialist city. The layout of the city centre at that time was rejected in favour of a new road network. However, the original plans were not completed. In addition, the rapid development of housing took priority over the preservation of old buildings. So in the 1960s and 1970s, both in the centre as well as the periphery, large areas were built in apartment-block style, for example . The old buildings of the period, which still existed in the Kassberg, and especially, were neglected and fell increasingly into dereliction.\n", "The city was heavily damaged by air raids in 1945, especially on 22 March. Although it had little military significance, two months before the end of the war in Europe the historic city was bombed as part of the Area Bombing Directive in order to undermine the morale of the German people. 28.5% of the houses were completely destroyed and 44.7% damaged. 26.8% of the houses remained undamaged. The centre, which had retained its medieval character until then, was almost levelled. As in many cities, priority was given to rapid building of badly needed housing, and concrete structures took the place of the destroyed buildings. Most of the major churches, two of them now UNESCO World Heritage Sites, were rebuilt in the original style soon after the war. During the war, valuable world heritage materials had been hidden in the basement of the city wall. \n", "During the Allied strategic bombing campaign of World War II, the historic city centres of most cities suffered severe losses to architectural heritage, with significant cases of almost total annihilation.\n", "Ancient cities were often highly militarized and fortified defensive settlements. In times of war they were the central focus of armed conflict and would be sacked and ruined in defeat. Although less central to modern conflict, vast areas of 20th-century cities such as Warsaw, Dresden, Coventry, Stalingrad, Königsberg, and Berlin were left in ruins following World War II, and a number of major cities around the world – such as Beirut, Kabul, Sarajevo, Grozny and Baghdad – have been partially or completely ruined in recent years as a result of more localised warfare.\n" ]
f-stop and aperture
Aperture is one of the things that determines how much light reaches the sensor on your camera. Think of it as a hole in a box. Inside the box is a kid drawing with crayons. (Don't worry, he's being fed.) The bigger the hole (the larger the aperture), the more light that gets in, the more the kid can see his paper to draw what he sees through the hole. The size of the aperture is measured in f-stops. The smaller the f-stop, the larger the aperture, or the larger the hole in the box, or the more light that gets in. Shutter speed is also important: how long does that hole in the box stay open to let the light in? The longer the shutter speed, the longer the hole stays open, the more light that gets in. That's one reason why some pictures are blurry: the subject moves while the hole in the box is still open. Sports photographers use very quick shutter speeds (which requires smaller f-stops to let more light in) so they can capture motion without blur. If the aperture was not large enough, and the shutter speed too fast, the photo would be too dark because the settings did not let enough light in to accurately capture the image to the sensor.
[ "In some contexts, especially in photography and astronomy, \"aperture\" refers to the \"diameter\" of the aperture stop rather than the physical stop or the opening itself. For example, in a telescope, the aperture stop is typically the edges of the objective lens or mirror (or of the mount that holds it). One then speaks of a telescope as having, for example, a 100-centimeter \"aperture\". Note that the aperture stop is not necessarily the smallest stop in the system. Magnification and demagnification by lenses and other elements can cause a relatively large stop to be the aperture stop for the system. In astrophotography, the aperture may be given as a linear measure (for example in inches or mm) or as the dimensionless ratio between that measure and the focal length. In other photography, it is usually given as a ratio.\n", "The aperture stop of a lens is a mechanical aperture which limits the light collection for each field point. The entrance pupil is the image of the aperture stop created by the optical elements on the object side of the lens. The light scattered by an object is collected by the entrance pupil and focused onto the image plane via a series of refractive elements. The cone of the focused light at the image plane is set by the size of the entrance pupil and the focal length of the lens. This is often referred to as the f-stop or f-number of the lens. f/# = f/D where D is the diameter of the entrance pupil.\n", "BULLET::::- Aperture. In context of photography or machine vision, aperture refers to the diameter of the aperture stop of a photographic lens. The aperture stop can be adjusted to control the amount of light reaching the film or image sensor.\n", "P-Iris (Precise Iris Control) is a type of network camera lens that together with specialized software installed in the camera itself regulates the iris opening through the use of a stepper motor for contrast, clarity, resolution and depth of field. P-Iris maintains image quality by continuously adjusting the iris position. This position, also referred to as a specific f-number, is where the lens works best and optical errors are reduced. P-Iris was developed by Axis Communications and the Japanese lens maker Kowa.\n", "The lens aperture is usually specified as an f-number, the ratio of focal length to effective aperture diameter. A lens typically has a set of marked \"f-stops\" that the f-number can be set to. A lower f-number denotes a greater aperture opening which allows more light to reach the film or image sensor. The photography term \"one f-stop\" refers to a factor of (approx. 1.41) change in f-number, which in turn corresponds to a factor of 2 change in light intensity.\n", "A device called a diaphragm usually serves as the aperture stop, and controls the aperture. The diaphragm functions much like the iris of the eye – it controls the effective diameter of the lens opening. Reducing the aperture size increases the depth of field, which describes the extent to which subject matter lying closer than or farther from the actual plane of focus appears to be in focus. In general, the smaller the aperture (the larger the f-number), the greater the distance from the plane of focus the subject matter may be while still appearing in focus.\n", "In optics, the exit pupil is a virtual aperture in an optical system. Only rays which pass through this virtual aperture can exit the system. The exit pupil is the image of the aperture stop in the optics that follow it. In a telescope or compound microscope, this image is the image of the objective element(s) as produced by the eyepiece. The size and shape of this disc is crucial to the instrument's performance, because the observer's eye can see light only if it passes through this tiny aperture. The term \"exit pupil\" is also sometimes used to refer to the diameter of the virtual aperture. Older literature on optics sometimes refers to the exit pupil as the Ramsden disc, named after English instrument-maker Jesse Ramsden.\n" ]
how does soap help remove body odor im comparison to plain water?
A lot of the stuff that makes us dirty and smelly is accumulated on our body because of oil. Oil traps in dirt and odors. And if you know anything about oil, you should know that it doesn't mix or dissolve in water. So while rubbing with just water might get rid of some surface oils, we use soap, which breaks down the oil (along with the dirt and odors in it) into smaller pieces that can be dissolved into the water and washed away.
[ "In chemistry, a soap is a salt of a fatty acid. Household uses for soaps include washing, bathing, and other types of housekeeping, where soaps act as surfactants, emulsifying oils to enable them to be carried away by water.\n", "The salting-out process used in the manufacture of soaps benefits from the common-ion effect. Soaps are sodium salts of fatty acids. Addition of sodium chloride reduces the solubility of the soap salts. The soaps precipitate due to a combination of common-ion effect and increased ionic strength.\n", "Ordinary soap is a salt of a fatty acid. Soaps are mainly used as surfactants for washing, bathing, and cleaning. Soaps for cleansing are made by treating vegetable or animal oils and fats with a strongly alkaline solution. Fats and oils are composed of triglycerides; three molecules of fatty acids are attached to a single molecule of glycerol. The alkaline solution, which is often called lye (although the term \"lye soap\" refers almost exclusively to soaps made with sodium hydroxide), brings about a chemical reaction known as saponification. In this reaction, the triglyceride fats are first hydrolyzed into free fatty acids, and then these combine with the alkali to form crude soap: a combination of various soap salts, excess fat or alkali, water, and liberated glycerol (glycerin).\n", "With hard water, soap solutions form a white precipitate (soap scum) instead of producing lather, because the 2+ ions destroy the surfactant properties of the soap by forming a solid precipitate (the soap scum). A major component of such scum is calcium stearate, which arises from sodium stearate, the main component of soap:\n", "The raw material used for these kinds of soap is olive oil. The Tripoli soap is also composed of: honey, essential oils, and natural aromatic raw materials like flowers, petals, and herbs. The soaps are dried in the sun, in a dry atmosphere, allowing the evaporation of the water that served to mix the different ingredients. The drying operation lasts for almost three months. As the water evaporates, a thin white layer appears on the soap surface, from the soda that comes from the sea salts. The craftsman brushes the soap very carefully with his hand until the powder trace is entirely eliminated.\n", "A bar of wet soap relies on ventilation to dry. A number of design elements may be used to increase ambient airflow around the soap, including vented surfaces or surfaces interspersed with bumps, ridges, or slats. Mechanical ventilation has not yet become a widespread design element in soap dishes.\n", "Soaps are formed from the reaction of glycerides with sodium hydroxide. The product of the reaction is glycerol and salts of fatty acids. Fatty acids in the soap emulsify the oils in dirt, enabling the removal of oily dirt with water.\n" ]
How do we know this is really 2014?
The most definite way is by old astronomical records. We know down to the minute when all the eclipses, periodic comets, etc of history took place, so if there was an eclipse on a certain day of the reign of the emperor Claudius, we know what year that was. Thus we know that there are some years we don't know much about, but there are no missing years.
[ "The setting of XXVC is a possible future of the real universe that we live in. In the year 1999, the Soviet Union and the United States are involved in the \"Last Gasp War.\" This is the world's first nuclear war.\n", "The story takes place in the year 2031, after a series of worldwide crises called the Year of the Domino (1996) has forced the U.S. government and the heads of major corporations to relocate to Hammarskjold Center, on Mars (\"temporarily, of course\"). In the wake of the American government leaving the planet and the Soviet Union collapsing from Islamic insurrections, there was a power shift throughout the world, with Brazilian Union of the Americas and the Pan-African League becoming the new superpowers on Earth.\n", "On August 10, 2019, Earth received a signal, which was confirmed by hundreds of different scientists to be from an alien life form. The message: \"Let us meet on Mars on July 7, 2035\". With this, a space agency representing all of Earth was formed, named ST&RS. From there on, many different events occurred quite quickly, such as the re-landing of people on the moon in 2022 and the construction of many new space stations. Later on, the Space Academy was founded as a way to train upcoming generations of ST&RS.\n", "The year is 2021, 15 years after a plague has killed nearly everyone over the age of thirteen (both the event and the virus itself are referred to as \"The Big Death\" and \"The Big D\"). Two young men, Jeremiah and Kurdy, meet up and join forces with those inside \"Thunder Mountain\" and help rebuild civilization. Jeremiah is searching for the \"Valhalla Sector\" where his father may still be alive.\n", "BULLET::::- On the HBO show 'Westworld', the \"'present day\"' timeline takes place in 2052. Video footage from the official website, which has since been removed, stated that the security data is from Jun 15, 2052. It has now been replaced with xx/xx/xxxx.\n", "11.22.63 is an American science fiction thriller miniseries based on the book \"11/22/63\" by Stephen King, and consisting of eight episodes. The series is executive-produced by J. J. Abrams, King, Bridget Carpenter and Bryan Burk, and produced by James Franco, who also has the main role. It premiered on Hulu on February 15, 2016, and was received positively by critics.\n", "The Next :15 is an American reality television series that premiered on February 10, 2016, on the TV One cable network. The show chronicles the lives of six former reality television personalities whose celebrity status has faded as they are no longer in the spotlight.\n" ]
How and when was it discovered the Mayans hadn't used wheels? Are their other questionable decisions that Mayans were 'blind' to?
The wheeled vehicle is not as intuitive an invention as most people seem to think. In fact, as /u/Daeres explains in [this submission](_URL_0_), it was only ever invented twice and possibly even only once, in the entire history of mankind. The reason the wheeled vehicle did not reach the Americas before Columbus is that they were cut off from the rest of the world where this technology had spread from its original point of origin in the 4th millennium BCE.
[ "The interpretation of Maya hieroglyphs was lost as a result of the Spanish Conquest of Central America. However, recent work by Maya epigraphers and linguists has yielded a considerable amount of information on this complex writing system.\n", "Critics argue that there is no archaeological evidence to support the use of wheeled vehicles in Mesoamerica, especially since many parts of ancient Mesoamerica were not suitable for wheeled transport. Clark Wissler, the Curator of Ethnography at the American Museum of Natural History in New York City, noted: \"we see that the prevailing mode of land transport in the New World was by human carrier. The wheel was unknown in pre-Columbian times.\"\n", "Toy animals with wheels dating from the Pre-Columbian era were uncovered by archaeologists in Veracruz, Mexico, in the 1940s. The indigenous peoples of this region did not use wheels for transportation prior to the arrival of Europeans.\n", "One Mormon researcher responds to the lack of evidence with a comparison to biblical archaeology, suggesting that though there are no archaeological evidences that any of the numerous ancient American civilizations used wheeled transportation, few chariot fragments have been found in the Middle East dating to biblical times (apart from the disassembled chariots found in Tutankhamun's tomb). Although few fragments of chariots have been found in the Middle East, there are many images of ancient chariots on pottery and frescoes and in many sculptures of Mediterranean origin, thus confirming their existence in those societies. Chariots are absent in pre-Columbian frescoes, pottery and artwork found in the New World.\n", "Toy animals with wheels dating from the Pre-Columbian era were uncovered by archaeologists in Veracruz, Mexico in the 1940s. The indigenous peoples of this region did not use wheels for transportation prior to the arrival of Europeans.\n", "Although large-scale use of wheels did not occur in the Americas prior to European contact, numerous small wheeled artifacts, identified as children's toys, have been found in Mexican archeological sites, some dating to about 1500 BCE. It is thought that the primary obstacle to large-scale development of the wheel in the Americas was the absence of domesticated large animals which could be used to pull wheeled carriages. The closest relative of cattle present in Americas in pre-Columbian times, the American Bison, is difficult to domesticate and was never domesticated by Native Americans; several horse species existed until about 12,000 years ago, but ultimately became extinct. The only large animal that was domesticated in the Western hemisphere, the llama, a pack animal but not physically suited to use as a draft animal to pull wheeled vehicles, did not spread far beyond the Andes by the time of the arrival of Columbus.\n", "A comparison of the South American Inca civilization to Mesoamerican civilizations shows the same lack of wheeled vehicles. Although the Incas used a vast network of paved roads, these roads are so rough, steep, and narrow that they were likely unsuitable for wheeled use. Bridges that the Inca people built, and even continue to use and maintain today in some remote areas, are straw-rope bridges so narrow (about 2–3 feet wide) that no wheeled vehicle can fit. Inca roads were used mainly by chaski message runners and llama caravans. Mayan paved roads at Yucatan had characteristics which could allow the use of wheeled vehicles, but there is no evidence that those highways were used by other than people on foot and nobles who were borne on litters.\n" ]
how (in north america) social conservative values became aligned with economic conservatism
So there is this thing called [Duverger's Law](_URL_0_) which basically states that our electoral system will cause USA to have two dominant political parties which will effectively shut out all but the remote possibility of a third party candidate being elected. Because politicians want to be successful, they will be forced to align themselves with one of the two major parties. This causes a dialectic relationship between the parties where they must be polar opposites of each-other to remain one of the two successful parties. [Around the time of the New Deal](_URL_1_) the political parties made a shift. Prior to this time, Republicans were a party for the fiscally libertarian, which can be seen in the economic policies of the 1920s. Democrats, on the other hand, included much of the 'socially conservative' south. However, with the great depression and the new Democratic President Roosevelt promoting both [Keynesian economic policy](_URL_2_) and policy reform which promoted helping the economically disenfranchised minorities in the south, the Republican party split over whether to support it or not while the old democratic base in the South all suddenly didn't want to be Democrats anymore. In an attempt to gain votes, the Republican Party started promoting the socially conservative message the southern democrats wanted. The cycle of elections during the 1930s saw many of the old urban republicans become new deal democrats and many old southern democrats become new republicans. Other than slight shifts during the early 1980s and early 1990s, we have held the same party alignment ever since. edit: tl;dr: two parties means they must be opposites of each-other or a third party will take one of their spots. In the 1930s, the Democrats changed, so Republicans followed suit to appeal to the voters Democrats had left behind. also, freereflection is right about the rise of christian fundamentalism being a factor.
[ "The extent to which conservative ideology was embedded in 19th and 20th century Canadian society is evidenced by the power and influence of Tory factions in pre-Confederation Canada, such as the Family Compact and the Chateau Clique, the prominence of the Conservative Party of Canada after Confederation and the pronounced stifling of extreme left-leaning or progressive views until after the Second World War due to widespread public aversion to Marxist ideologies. Even to this day, social conservatism in Canada has a wide base of support outside the major urban centers of Toronto, Montreal and Vancouver.\n", "Conservatism in North America is a political philosophy that varies in form, depending on the country and the region, but that has similar themes and goals. Academic study into the differences and similarities between conservatism in North American countries has been undertaken on numerous occasions. Reginald Bibby has asserted that the primary reason that conservatism has been so strong and enduring throughout North America is because of the propagation of religious values from generation to generation. This connection is strongest in mainstream Protestantism in the United States and both Protestantism and Roman Catholicism in Canada.\n", "Conservatism emphasized family values and obedience to the father and the state. Nationalism carried by masses of people made patriotism an important civic virtue. Liberalism combined republicanism with a belief in progress and liberalization based on capitalism. Civic virtues focused on individual behavior and responsibility were very important. Many liberals turned into socialists or conservatives in the end of the 19th century and early 20th century. Others became social liberals, valuing capitalism with a strong government to protect the poor. A focus on agriculture and landed nobility was supplanted by a focus on industry and civil society.\n", "Until the mid-1970s, the Conservative Party was mostly controlled by one-nation conservatives. The rise of the New Right in conservative politics led to a critique of one-nation conservatism. The New Right thinkers contended that Keynesian economics and the welfare state had damaged the economy and society. The Winter of Discontent of 1978–1979 in which trades unions took industrial action with a wide impact on daily life was portrayed by the New Right as illustrative of the over-extension of the state. Figures such as Margaret Thatcher believed that to reverse the national decline it was necessary to revive old values of individualism and challenge the dependency culture which they felt had been created by the welfare state.\n", "It stresses the importance of a social safety net to deal with poverty, support of limited redistribution of wealth along with government regulation of markets in the interests of both consumers and producers. Paternalistic conservatism first arose as a distinct ideology in the United Kingdom under Prime Minister Benjamin Disraeli's \"One Nation\" Toryism. There have been a variety of one nation conservative governments. In the United Kingdom, the Prime Ministers Disraeli, Stanley Baldwin, Neville Chamberlain, Winston Churchill and Harold Macmillan were one nation conservatives.\n", "Social conservatism in the United States is a political ideology focused on the preservation of traditional values and beliefs, hearkening back to values believed to be present at the American founding. It focuses on a concern with moral and social values which proponents of the ideology see as degraded in modern society by social democracy and liberalism. Many religious conservatives push for a focus on Judeo-Christian traditions as a guiding force for the country on social issues, leading them to be considered social conservatives. Social conservatives are concerned with many social issues such as abortion, sex education, the Equal Rights Amendment, school prayer, same-sex marriage, and many others. They oppose many of the cultural changes brought on by the culture wars and the sexual revolution. Summarily, this branch of conservatism is concerned with moral and social issues within the United States and uses tradition, strict morals, and religion as solutions for these problems.\n", "Compared to social conservatism in the United States, social conservatism has not been as influential in Canada. The main reason is that the neoliberal or neoconservative style of politics as promoted by leaders such as former Liberal Party of Canada Prime Minister Paul Martin and Former Conservative Party of Canada Prime Minister Stephen Harper have focused on economic conservatism, with little or no emphasis on moral or social conservatism. Without a specific, large political party behind them, social conservatives have divided their votes and can be found in all political parties.\n" ]
how do they make sure that low level employees at the white house (custodians, chefs, etc) aren't going to try to kill the president?
The same way that determine that the high level employees won't do it -- very, very thorough background checks.
[ "While the President may block certain appointments to the Council, he cannot remove a permanent member from his seat. In fact, it does not appear that any person has the power to do so. The door may be open for potential abuse with appointments based on some political agenda rather than merit. The counterpoint is that it is precisely those in power who are best placed to scrutinize bills for illegitimate differentiating measures, because \"it is often those with political affiliations who can make the biggest contribution to the discussion; if nothing else, from the folly and error of their past ways\".\n", "The core White House staff appointments, and most EOP officials generally, are not required to be confirmed by the U.S. Senate, with a handful of exceptions (e.g., the director of the Office of Management and Budget, the chair and members of the Council of Economic Advisers, and the United States trade representative). There are about 4,000 positions in the Executive Office of the President.\n", "The core White House staff appointments, and most EOP officials generally, are not required to be confirmed by the U.S. Senate, although there are a handful of exceptions (e.g., the Director of the Office of Management and Budget, the Chair and members of the Council of Economic Advisers, and the United States Trade Representative).\n", "Within the executive branch, any Presidentially-appointed \"principal officer,\" including a head of an agency such as a Secretary, Administrator, or Commissioner, is a \"civil officer of the United States\" subject to impeachment. At the opposite end of the spectrum, lesser functionaries, such as federal civil service employees, do not exercise “significant authority,” and are not appointed by the President or an agency head. These employees do not appear to be subject to impeachment, though that may be a matter of allocation of House floor debate time by the Speaker, rather than a matter of law.\n", "The White House Chief of Staff has traditionally been the highest-ranking employee of the White House. The responsibilities of the chief of staff are both managerial and advisory over the president's official business. The chief of staff is appointed by and serves at the pleasure of the president; it does not require Senate confirmation.\n", "The Senate has the unique power to impeach judges and other high officials of the executive including the Federal Auditor-General and the members of the electoral and revenue commissions. This power is, however, subject to prior request by the President. The Senate also confirms the President's nomination of senior diplomats, members of the federal cabinet, federal judicial appointments and independent federal commissions.\n", "The duties of the White House chief of staff vary greatly from one administration to another and, in fact, there is no legal requirement that the president even fill the position. However, since at least 1979, all presidents have found the need for a chief of staff, who typically oversees the actions of the White House staff, manages the president's schedule, and decides who is allowed to meet with the president. Because of these duties, the chief of staff has at various times been labeled \"The Gatekeeper.\"\n" ]
Does a colony of penguins stand s chance in the north pole? What about a polar bear in antacrtica?
Unfortunately not open access, but [this article](_URL_0_) is about those exact questions. Two answers to your questions, from that article: > If polar bears were transferred to Antarctica could they survive? And would penguins survive in the Arctic? > Polar bears would probably survive in the Antarctic, and the Southern Ocean around it, but they could devastate the native wildlife. In the Arctic polar bears feed mainly on seals, especially young pups born on ice floes or beaches. Many of the differences in breeding habits between Arctic and Antarctic seals can be interpreted as adaptations to evading predation by bears. > Polar bears would find plenty of fish-eating mammals and birds around Antarctica. Penguins would tie particularly vulnerable because they are flightless and breed on open ground, with larger species taking months to raise a single chick. Bears can only run in short bursts, but they could catch a fat, sassy penguin chick or grab an egg from an incubating parent. > In the Arctic polar bears hunt mainly on the edge of the sea ice, where it is thick enough to support their weight but thin enough for seals to make breathing holes. The numerous islands off the north coast of Canada, Alaska and north-west Europe provide plenty of suitable habitats. The Antarctic continent is colder, with only a few offshore islands, so bears would probably thrive at lower latitudes in the Southern Ocean than in the Arctic. > We can only hope that nobody ever tries what the questioner suggests. Artificially introduced predators often devastate indigenous wildlife, as it is not accustomed to dealing with them. This occurred with stoats in New Zealand, foxes and cats in Australia, and rats on many isolated islands. Large, heavy animals would also trample the slow-growing, mechanically weak plants and lichens of the Antarctic. For instance, Norwegian reindeer have decimated many native plants in South Georgia, an island in the South Atlantic Ocean, since they were introduced 80 years ago. > C. M. Pond Department of Biological Sciences, The Open University, Milton Keynes, Buckinghamshire, UK And > While, as far as I know, no one has ever been stupid enough to introduce polar bears into the Antarctic, there have been at least two practical attempts to transplant penguins to the Arctic. The original "penguin" was in fact the late great auk (Pinguinus impennis), once found in vast numbers around northern shores of the Atlantic. Although no relation to southern hemisphere penguins, it was very similar in appearance, and filled much the same ecological niche as penguins, particularly the king penguins of the subantarctic region. > With any attempt to introduce an alien species, there must actually exist an appropriate ecological niche for it to fill, and it must be vacant. For the most part, the ecological niches occupied by penguins in the south are filled by the auk family to the north. But the demise of the great auk in the mid-19th century at the hands of hungry whalers created not only a vacancy that one of the larger penguins might neatly slot into, but also a potential economic demand for the penguin's fatty meat and protein-rich eggs. > It was perhaps the possible economic opportunities that prompted two separate bids to introduce penguins into Norwegian waters in the late 1930s. The first, by Carl Schoyen of the Norwegian Nature Protection Society, released groups of nine king penguins at Røst, Lofoten, Gjesvaer and Finnmark in October 1936. Two years later, the National Federation for the Protection of Nature, in an equally bizarre operation, released several macaroni and jackass penguins in the same areas, even though these smaller birds would clearly find themselves competing directly with auks or other native seabirds. > The outcome was unhappy for the experimenters and, most particularly, for the penguins. Among those whose fate is known, one king was quickly despatched by a local woman who thought it was some kind of demon, while a macaroni died on a fishing line in 1944, although from its condition it had apparently thrived during its six years in alien waters. > And it soon became obvious that the real reason why any attempt to fill the ecological gap left by the great auk was destined to fail was the very reason that the niche was vacant in the first place — such large seabirds could not happily coexist with a large and predatory human population. Of course, it is the steadily increasing human presence in the far south that is now threatening penguins in their native habitat. > Hadrian Jeffs Norwich, Norfolk, UK
[ "Although almost all penguin species are native to the Southern Hemisphere, they are not found only in cold climates, such as Antarctica. In fact, only a few species of penguin actually live so far south. Several species live in the temperate zone; one, the Galápagos penguin, lives as far north as the Galápagos Islands, but this is only made possible by the cold, rich waters of the Antarctic Humboldt Current that flows around these islands. Also, though the climate of North Pole and South Pole is similar, there are no penguins found in the North Pole.\n", "The Humboldt penguin (\"Spheniscus humboldti\") is a South American penguin living mainly in the Pinguino de Humbold National Reserve in the North of Chile, although its habitat comprises most of coastal Chile and Peru. Its nearest relatives are the African penguin, the Magellanic penguin and the Galápagos penguin. The Humboldt penguin and the cold water current it swims in both are named after the explorer Alexander von Humboldt. The species is listed as vulnerable by the IUCN with no population recovery plan in place.\n", "There is a small colony of breeding emperor penguins on nearshore sea ice at the north end of the island. There is a large Adélie penguin colony on a raised beach called Cadwalader Beach at the south-western end of the island. The island also has several breeding colonies of south polar skua.\n", "The Galápagos penguin (\"Spheniscus mendiculus\") is a penguin endemic to the Galápagos Islands. It is the only penguin found north of the equator. The cool waters of the Humboldt and Cromwell Currents allow it to survive despite the tropical latitude. The Galápagos penguin is one of the banded penguins, the other species of which live mostly on the coasts of Africa and mainland South America.\n", "Humboldt penguins nest on islands and rocky coasts, burrowing holes in guano and sometimes using scrapes or caves. In South America the Humboldt penguin is found only along the Pacific coast, and the range of the Humboldt penguin overlaps that of the Magellanic penguin on the central Chilean coast. It is vagrant in Ecuador and Colombia.\n", "Ninety percent of Galápagos penguins live on Fernandina Island and the west coast of Isabela Island, in the western part of the archipelago, but small populations also occur on Santiago, Bartolomé, northern Santa Cruz, and Floreana. The northern tip of Isabela crosses the equator, meaning that some Galápagos penguins live in the northern hemisphere, the only penguins to do so.\n", "Because of the Galápagos penguin's small size, it has many predators. On land, the penguins may fall prey to crabs, snakes, rice rats, cats, Galapagos hawks, and short-eared owls. While in the water they are preyed on by sharks, fur seals, and sea lions. They also face the hazards of unreliable food resources and volcanic activity.\n" ]
How are primers made for PCR chosen?
If the genome of the species you are trying to make primers for is well-characterised then the job of designing primers is not hard. In humans, while there is a lot of variation between people, the vast majority of a person's DNA is identical to the known human consensus sequence. So if you design a primer to be complementary to a region of DNA, based on the consensus sequence, odds are that any given individual will not have a mutation at that site. You can increase your likelihood of success by placing your primer in regions of the gene that are highly conserved (show low levels of variation between people or even across species). Sometimes if you know that a variation exists frequently at a given site, you can even design your primers to only amplify DNA if it is present, so variation/mutation/polymorphism at that site becomes a feature not a bug.
[ "Tetra-primer amplification refractory mutation system PCR, or ARMS-PCR, employs two pairs of primers to amplify two alleles in one PCR reaction. The primers are designed such that the two primer pairs overlap at a SNP location but each match perfectly to only one of the possible SNPs. The basis of the invention is that unexpectedly, oligonucleotides with a mismatched 3'-residue will not function as primers in the PCR under appropriate conditions. As a result, if a given allele is present in the PCR reaction, the primer pair specific to that allele will produce product but not to the alternative allele with a different SNP. The two primer pairs are also designed such that their PCR products are of a significantly different length allowing for easily distinguishable bands by gel electrophoresis or melt temperature analysis. In examining the results, if a genomic sample is homozygous, then the PCR products that result will be from the primer that matches the SNP location and the outer opposite-strand primer, as well from the two outer primers. If the genomic sample is heterozygous, then products will result from the primer of each allele and their respective outer primer counterparts as well as the outer primers.\n", "Normally PCR primers are chosen from an invariant part of the genome, and might be used to amplify a polymorphic area between them. In \"allele-specific PCR\" the opposite is done. At least one of the primers is chosen from a polymorphic area, with the mutations located at (or near) its 3'-end. Under stringent conditions, a mismatched primer will not initiate replication, whereas a matched primer will. The appearance of an amplification product therefore indicates the genotype. \"(For more information, see SNP genotyping.)\"\n", "When designing primers, additional nucleotide bases can be added to the back ends of each primer, resulting in a customized cap sequence on each end of the amplified region. One application for this practice is for use in TA cloning, a special subcloning technique similar to PCR, where efficiency can be increased by adding AG tails to the 5′ and the 3′ ends.\n", "Primers can also be designed to be 'degenerate' – able to initiate replication from a large number of target locations. \"Whole genome amplification\" (or \"WGA\") is a group of procedures that allow amplification to occur at many locations in an unknown genome, and which may only be available in small quantities. Other techniques use \"degenerate primers\" that are synthesized using multiple nucleotides at particular positions (the polymerase 'chooses' the correctly matched primers). Also, the primers can be synthesized with the nucleoside analog inosine, which hybridizes to three of the four normal bases. A similar technique can force PCR to perform Site-directed mutagenesis. (also see Overlap extension polymerase chain reaction)\n", "Much like how primers are designed such that there is a forward primer and a reverse primer capable of allowing DNA polymerase to fill the entire template sequence, PCA uses the same technology but with multiple oligonucleotides. While in PCR the customary size of oligonuleotides used is 18 base pairs, in PCA lengths of up to 50 are used to ensure uniqueness and correct hybridization.\n", "Primer Premier software combines design of primers for various PCR applications under one common platform. The software supports design of degenerate primers on alignments for amplifying related set of nucleotide sequences for detecting common traits among organisms and to determine heredity.\n", "These tools are used to optimize the design of primers for target DNA or cDNA sequences. Primer optimization has two goals: efficiency and selectivity. Efficiency involves taking into account such factors as GC-content, efficiency of binding, complementarity, secondary structure, and annealing and melting point (Tm). Primer selectivity requires that the primer pairs not fortuitously bind to random sites other than the target of interest, nor should the primer pairs bind to conserved regions of a gene family. If the selectivity is poor, a set of primers will amplify multiple products besides the target of interest.\n" ]
what will happen to mickey mouse when he becomes public domain in 2023?
He won't become public domain. In 2022 they'll extend how long copyright lasts after the creator's death, just like the last two times.
[ "There have been multiple attempts to argue that certain versions of Mickey Mouse are in fact in the public domain. In the 1980s, archivist George S. Brown attempted to recreate and sell cels from the 1933 short \"The Mad Doctor\", on the theory that they were in the public domain because Disney had failed to renew the copyright as required by current law. However, Disney successfully sued Brown to prevent such sale, arguing that the lapse in copyright for \"The Mad Doctor\" did not put Mickey Mouse in the public domain because of the copyright in the earlier films. Brown attempted to appeal, noting imperfections in the earlier copyright claims, but the court dismissed his argument as untimely.\n", "BULLET::::- December 15 - Entertainment pioneer Walt Disney, best known for his creation of Mickey Mouse, breakthroughs in the field of animation, filmmaking, theme park design and other achievements, dies at the age of 65. He died while he was producing \"The Jungle Book\", \"The Happiest Millionaire\", and \"Winnie the Pooh and the Blustery Day\"; the last three films under his personal supervision.\n", "BULLET::::- \"Mickey Mouse:\" Leader of the good Disney characters. Mickey had his original sketch torn apart and scattered to keep the Overtakers from finding it, leading to him vanishing. 40 years later, he is restored, but lacks the knowledge the other characters have; he would have just behaved like a character unless the Keepers were there to guide him. With his wand in hand, Mickey can restore any Disney park to normal and summon characters when needed. He doesn't speak, but instead uses some kind of telepathy only the characters can hear. Violet translates for him to the Keepers. Until he does speak in the Return series.\n", "He never received any royalties for the invention of the mouse. During an interview, he said \"SRI patented the mouse, but they really had no idea of its value. Some years later it was learned that they had licensed it to Apple Computer for something like $40,000.\" Engelbart showcased the chorded keyboard and many more of his and ARC's inventions in 1968 at The Mother of All Demos.\n", "Director Emeritus Roy E. Disney died of stomach cancer on December 16, 2009. At the time of his death, he owned roughly 1% of all of Disney which amounted to 16 million shares. He was the last member of the Disney family to be actively involved in the company. In October 2009, Disney Channel president Rich Ross, hired by Iger, replaced Dick Cook as chairman of the company and, in November, began restructuring the company to focus more on family friendly products. Later in January 2010, Disney decided to shut down Miramax after downsizing Touchstone, but one month later, they instead began selling the Miramax brand and its 700-title film library to Filmyard Holdings. In March, ImageMovers Digital, which Disney had established as a joint venture studio with Robert Zemeckis in 2007, was shut down. In April 2010, Lyric Street, Disney's country music label in Nashville, was shut down. The following month, Haim Saban reacquired the \"Power Rangers\" franchise, including its 700-episode library. In September 2012, Saban reacquired the \"Digimon\" franchise, which, like \"Power Rangers\", was part of the Fox Kids library that Disney acquired in 2001. In January 2011, Disney Interactive Studios was downsized.\n", "Mickey Mouse is depicted in the \"Kingdom Hearts\" series as the king of Disney Castle, frequently referred to in-game as \"King Mickey\" or simply . He is also an experienced Keyblade Master, alternatively wielding a golden version of Sora's \"Kingdom Key\" retrieved from the realm of darkness, and the \"Star Seeker\" he is seen using during his apprenticeship to Yen Sid in \"Birth by Sleep\". Mickey is absent for most of the original game, which sees him departing his world to discover a solution to the Heartless' invasion, leaving instructions for Donald and Goofy telling to find and protect the Keyblade wielder. He has a more active role fighting alongside Sora and his allies in later installments, which first required the approval of Disney. He is a playable character in \"Kingdom Hearts II\", stepping in to replace Sora whenever the player is defeated during certain boss battles until he is able to revive Sora. He is also available for play in the \"Mission Mode\" of \"Kingdom Hearts 358/2 Days\" as an unlockable character.\n", "After many decades of fame following the accident, Mickey had forgotten it all until the Shadow Blot enters his home through the mirror and abducts him into the ruined forgotten world, now named by this time as the Wasteland. Oswald all the while had his will and his mind twisted from years of hiding and his jealousy of Mickey's rise to fame, unaware the enigmatic Mad Doctor (who was formerly loyal to Oswald before siding with the Blot) and the Blot formulate a plan to steal Mickey's heart away, which they plan to use to escape the ruined world, as all Wasteland Toons living there are forgotten and can't leave Wasteland since they no longer have hearts of their own. However, Mickey frees himself before they can succeed and scares off the Blot with Yen Sid's brush, forcing the Mad Doctor to flee. Oswald, who was spying on them, also flees after tampering with the Mad Doctor's machines, leaving Mickey to deal with the Mad Doctor's now-hostile mechanical arm.\n" ]
why do we accept that all the world's diverse species have a common ancestor while human beings are considered to be too complex for having only one paternal and maternal ancestors?
> while human beings are considered to be too complex That has been debunked and is only believed by people who either don't understand evolution or refuse to accept that evolution is a fact.
[ "Biologists reason that all living organisms on Earth must share a single last universal ancestor, because it would be virtually impossible that two or more separate lineages could have independently developed the many complex biochemical mechanisms common to all living organisms.\n", "Assuming that all of an individual's ancestors are otherwise unrelated to each other, that individual has 2 ancestors in the \"n\"th generation before him and a total of 2 − 2 ancestors in the \"g\" generations before him. In practice, however, it is clear that most ancestors of humans (and any other species) are multiply related (see pedigree collapse). Consider \"n\" = 40: the human species is mors, both living and dead; in contrast, some more youth-oriented cultural contexts display less veneration of elders. In other cultural contexts, some people seek providence from their deceased ancestors; this practice is sometimes known as \"ancestor worship\" or, more accurately, \"ancestor veneration\".\n", "A 2011 study published in the American Journal of Human Genetics indicates that Indian ancestral components are the result of a more complex demographic history than was previously thought. According to the researchers, South Asia harbours two major ancestral components, one of which is spread at comparable frequency and genetic diversity in populations of Central Asia, West Asia and Europe; the other component is more restricted to South Asia. However, if one were to rule out the possibility of a large-scale Indo-Aryan migration, these findings suggest that the genetic affinities of both Indian ancestral components are the result of multiple gene flows over the course of thousands of years.\n", "Furthermore, because human genetic variation is clinal, many individuals affiliate with two or more continental groups. Thus, the genetically based \"biogeographical ancestry\" assigned to any given person generally will be broadly distributed and will be accompanied by sizable uncertainties (Pfaff \"et al.\" 2004).\n", "However, this explanation cannot be applied to humans (and other species, predominantly other mammals) that live in stable, established social groupings. This is because of the social intelligence that functioning within these groups requires from the individual. Humans, while they are not the only ones, possess the cognitive and mental capacity to form systems of personal relationships and ties that extend well beyond those of the nucleus of family. The continuous process of creating, interacting, and adjusting to other individuals is a key component of many species' ecology.\n", "In many parts of the world, groups have mixed in such a way that many individuals have relatively recent ancestors from widely separated regions. Although genetic analyses of large numbers of loci can produce estimates of the percentage of a person's ancestors coming from various continental populations (Shriver \"et al.\" 2003; Bamshad \"et al.\" 2004), these estimates may assume a false distinctiveness of the parental populations, since human groups have exchanged mates from local to continental scales throughout history (Cavalli-Sforza \"et al.\" 1994; Hoerder 2002). Even with large numbers of markers, information for estimating admixture proportions of individuals or groups is limited, and estimates typically will have wide confidence intervals (Pfaff \"et al.\" 2004).\n", "Because of the common ancestry of all humans, only a small number of variants have large differences in frequency between populations. However, some rare variants in the world's human population are much more frequent in at least one population (more than 5%).\n" ]
what is sdk?what is api?
When you walk into McDonalds, you can order a happy meal & get a nice pre-packaged meal (Burger, fries, toy, etc). But say you wanted to walk into the kitchen and make your own meal - you'd still have to use the ingredients the store has. You can't walk into McD & make a Taco Supreme as part of your meal - it doesn't make sense in their context, they don't know what to do with it. The SDK & API is like the menu at mcdonalds. The SDK is the collection of tools & resources you need to build a meal. The API is the list of things you can ask the SDK to do for you - if you ask the McDonalds API for 'slivered onions' it knows how to deliver - whereas the Taco Bell API would have no idea what the hell that means. If you don't like what the McDonalds SDK has to offer. You want waffle fries. You can go closer to basics & use the bare Grocery Store SDK. It's a lot more flexible & you can build more, but you're gonna have to do a lot of the basic development yourself that the SDK had pre-packaged for you, like cutting up the potatoes for fries, not to mention the tomatoes and onions etc. And if you wanted to go really crazy you could ditch the Grocery Store SDK & go all the way back to the Subsistence Farming - the Assembly Language of making meals. But now you're talking about growing your own vegetables & slaughtering your own cow - you got a LOT of work ahead of you before you get to that burger. It's an imperfect analogy, but hopefully gives you an idea
[ "An SDK can take the form of a simple implementation of one or more application programming interfaces (APIs) in the form of on-device libraries to interface to a particular programming language, or it may be as complex as hardware-specific tools that can communicate with a particular embedded system. Common tools include debugging facilities and other utilities, often presented in an integrated development environment (IDE). SDKs may also include sample code and technical notes or other supporting documentation such as tutorials to help clarify points made by the primary reference material.\n", "The SDK is a free download for users of Mac personal computers. It is not available for Microsoft Windows PCs. The SDK contains sets giving developers access to various functions and services of iOS devices, such as hardware and software attributes. It also contains an iPhone simulator to mimic the look and feel of the device on the computer while developing. New versions of the SDK accompany new versions of iOS. In order to test applications, get technical support, and distribute apps through App Store, developers are required to subscribe to the Apple Developer Program.\n", "The SDK is a free download for users of Mac personal computers. It is not available for Microsoft Windows PCs. The SDK contains sets giving developers access to various functions and services of iOS devices, such as hardware and software attributes. It also contains an iPhone simulator to mimic the look and feel of the device on the computer while developing. New versions of the SDK accompany new versions of iOS. In order to test applications, get technical support, and distribute apps through App Store, developers are required to subscribe to the Apple Developer Program.\n", "The SDK is a free download for users of Mac personal computers. It is not available for Microsoft Windows PCs. The SDK contains sets giving developers access to various functions and services of iOS devices, such as hardware and software attributes. It also contains an iPhone simulator to mimic the look and feel of the device on the computer while developing. New versions of the SDK accompany new versions of iOS. In order to test applications, get technical support, and distribute apps through App Store, developers are required to subscribe to the Apple Developer Program.\n", "An SDI should enable the discovery and delivery of spatial data from a data repository, via a spatial service provider, to a user. As mentioned earlier it is often wished that the data provider is able to update spatial data stored in a repository. Hence, the basic software components of an SDI are: \n", "A key aspect of an SDP is that it merges information and services both within and across asset classes. An SDP will typically combine elements such as pricing, trading, research and technical analysis within each asset class, and then draw together multiple asset classes.\n", "Percent-encoding, also known as URL encoding, is a mechanism for encoding information in a Uniform Resource Identifier (URI) under certain circumstances. Although it is known as \"URL encoding\" it is, in fact, used more generally within the main Uniform Resource Identifier (URI) set, which includes both Uniform Resource Locator (URL) and Uniform Resource Name (URN). As such, it is also used in the preparation of data of the codice_1 media type, as is often used in the submission of HTML form data in HTTP requests.\n" ]
why can humans not digest fiber?
There are actually almost no animals that can digest fibers! Almost all animals who live off stuff like leaves and grass break them down with the help of symbiotic bacteria that live in their intestines. Humans only have a tiny number of bacteria than can break down cellulose, and they only break down enough of it to feed themselves. As for *why* we don't have these bacteria: probably because digesting that kind of food is not very efficient and you don't get a lot of energy out of it (for instance, I happen to have in my head that guinea pigs get about 40 kcal out of 100 g of hay). With a big energy-hungry brain like ours, it's better to have a digestive tract optimised for energy-dense foods like meat, fruit and starchy things, rather than one that'll be welcoming to cellulose-eating bacteria.
[ "Whole grains, beans and other legumes, fruits (especially plums, prunes, and figs), and vegetables are good sources of dietary fiber. Fiber is important to digestive health and is thought to reduce the risk of colon cancer. For mechanical reasons, fiber can help in alleviating both constipation and diarrhea. Fiber provides bulk to the intestinal contents, and insoluble fiber especially stimulates peristalsis – the rhythmic muscular contractions of the intestines which move digesta along the digestive tract. Some soluble fibers produce a solution of high viscosity; this is essentially a gel, which slows the movement of food through the intestines. Additionally, fiber, perhaps especially that from whole grains, may help lessen insulin spikes and reduce the risk of type 2 diabetes.\n", "Many molecules that are considered to be \"dietary fiber\" are so because humans lack the necessary enzymes to split the glycosidic bond and they reach the large intestine. Many foods contain varying types of dietary fibers, all of which contribute to health in different ways.\n", "Dietary fiber is a carbohydrate that is incompletely absorbed in humans and in some animals. Like all carbohydrates, when it is metabolized it can produce four Calories (kilocalories) of energy per gram. However, in most circumstances it accounts for less than that because of its limited absorption and digestibility. Dietary fiber consists mainly of cellulose, a large carbohydrate polymer which is indigestible as humans do not have the required enzymes to disassemble it. There are two subcategories: soluble and insoluble fiber. Whole grains, fruits (especially plums, prunes, and figs), and vegetables are good sources of dietary fiber. There are many health benefits of a high-fiber diet. Dietary fiber helps reduce the chance of gastrointestinal problems such as constipation and diarrhea by increasing the weight and size of stool and softening it. Insoluble fiber, found in whole wheat flour, nuts and vegetables, especially stimulates peristalsis – the rhythmic muscular contractions of the intestines, which move digest along the digestive tract. Soluble fiber, found in oats, peas, beans, and many fruits, dissolves in water in the intestinal tract to produce a gel that slows the movement of food through the intestines. This may help lower blood glucose levels because it can slow the absorption of sugar. Additionally, fiber, perhaps especially that from whole grains, is thought to possibly help lessen insulin spikes, and therefore reduce the risk of type 2 diabetes. The link between increased fiber consumption and a decreased risk of colorectal cancer is still uncertain.\n", "Fiber might be beneficial in those who have a predominance of constipation. In people who have IBS-C, soluble fiber can reduce overall symptoms, but will not reduce pain. The research supporting dietary fiber contains conflicting small studies complicated by the heterogeneity of types of fiber and doses used.\n", "Even though these complex polysaccharides are not very digestible, they provide important dietary elements for humans. Called dietary fiber, these carbohydrates enhance digestion among other benefits. The main action of dietary fiber is to change the nature of the contents of the gastrointestinal tract, and to change how other nutrients and chemicals are absorbed. Soluble fiber binds to bile acids in the small intestine, making them less likely to enter the body; this in turn lowers cholesterol levels in the blood. Soluble fiber also attenuates the absorption of sugar, reduces sugar response after eating, normalizes blood lipid levels and, once fermented in the colon, produces short-chain fatty acids as byproducts with wide-ranging physiological activities (discussion below). Although insoluble fiber is associated with reduced diabetes risk, the mechanism by which this occurs is unknown.\n", "Dietary fiber is a carbohydrate, specifically a polysaccharide, which is incompletely absorbed in humans and in some animals. Like all carbohydrates, when it is metabolized, it can produce four Calories (kilocalories) of energy per gram, but in most circumstances, it accounts for less than that because of its limited absorption and digestibility. The two subcategories are insoluble and soluble fiber. Insoluble dietary fiber consists mainly of cellulose, a large carbohydrate polymer that is indigestible by humans, because humans do not have the required enzymes to break it down, and the human digestive system does not harbor enough of the types of microbes that can do so. Soluble dietary fiber comprises a variety of oligosaccharides, waxes, esters, resistant starches, and other carbohydrates that dissolve or gelatinize in water. Many of these soluble fibers can be fermented or partially fermented by microbes in the human digestive system to produce short-chain fatty acids which are absorbed and therefore introduce some caloric content.\n", "Dietary fiber is defined to be plant components that are not broken down by human digestive enzymes. In the late 20th century, only lignin and some polysaccharides were known to satisfy this definition, but in the early 21st century, resistant starch and oligosaccharides were included as dietary fiber components.\n" ]
why couldn't t-rex have been both predatory and scavenging?
Language like this is usually used for disambiguation. It *can* lead to confusion in the process, but mostly it's to try and *avoid* it. There's often some overlap between behaviour. In fact, *most* carnivores will scavenge when given the chance. There just wasn't much hard evidence of this specifically existing for T-Rex, and it's easier to classify animals in 'either / or' groups, than catering to blurry, movable lines. Bear in mind, this find **doesn't** constitute proof, either (and the blogspot article seems a bit biased). There could be numerous explanations for it, not all of which imply active hunting.
[ "Obligate scavenging is rare among vertebrates, due to the difficulty of finding enough carrion without expending too much energy. In vertebrates, only vultures and possibly some pterosaurs are obligate scavengers, as terrestrial soaring flyers are the only animals able to find enough carrion.\n", "Most scavenging animals are facultative scavengers that gain most of their food through other methods, especially predation. Many large carnivores that hunt regularly, such as hyenas and jackals, but also animals rarely thought of as scavengers, such as African lions, leopards, and wolves will scavenge if given the chance. They may also use their size and ferocity to intimidate the original hunters (the cheetah is a notable exception). Almost all scavengers above insect size are predators and will hunt if not enough carrion is available, as few ecosystems provide enough dead animals year-round to keep its scavengers fed on that alone. Scavenging wild dogs and crows frequently exploit roadkill.\n", "All of the aforementioned adaptations enabled Homo to scavenge for food more effectively. Endurance running could have been used as a means of gaining access to distant carcasses or food stores faster than other scavengers and/or carnivores. Scavenging may have taken one or both of two forms: opportunistic scavenging and strategic scavenging.\n", "Research based on energetic modelling has also backed up the role of active predatory behaviors in theropods and has effectively ruled out obligate scavenging as a likely strategy. However scavenging was still likely to have been an important resource for many theropod with body size playing a key role in its importance. In general, the researchers found that small species, such as coelurosaurs, and species larger than 1000 kg, such as an adult \"Tyrannosaurs rex\", would have been poor scavengers. In contrast intermediate sized species, such as \"Dilophosaurus\", were found to have been capable of gaining significant energy though scavenging.\n", "The proposed benefit of endurance running in scavenging is the ability of early hominins to outcompete other scavengers in reaching food sources. However paleoanthropological studies suggest that the savanna-woodland habitat caused a very low competition environment. Due to low visibility, carcasses were not easily located by mammalian carnivores, resulting in less competition.\n", "Strategic scavenging involves a planned search for carcasses. This style of scavenging would have benefitted from endurance running much more than opportunistic scavenging. Strategic scavenging would have involved the use of long range cues, such as birds circling overhead. Endurance running would have been advantageous in this setting because it allowed hominins to reach the carcass more quickly. Selection pressures would have been very high for strategic scavenging, because hominins were diurnal, while their major competitors (hyenas, lions, etc.) were not. Thus, they would have had to make sure to capitalize on daytime carcasses. Selection pressure also came from the weakness of Homo. Because they were very weak, they were unlikely to drive off any large competition at the carcass. This fact led to an even higher need for a way to reach the carcass before these competitors.\n", "A debate exists, however, about whether \"Tyrannosaurus\" was primarily a predator or a pure scavenger; the debate was assessed in a 1917 study by Lambe which argued \"Tyrannosaurus\" was a pure scavenger because the \"Gorgosaurus\" teeth showed hardly any wear. This argument may not be valid because theropods replaced their teeth quite rapidly. Ever since the first discovery of \"Tyrannosaurus\" most scientists have speculated that it was a predator; like modern large predators it would readily scavenge or steal another predator's kill if it had the opportunity.\n" ]
Inspired by Interstellar: What happens to the rotation of a large planet orbiting a black hole?
The reason for the time dilation that happens isn't directly due to the black hole itself, it is more to do with how fast the planet needs to orbit the black hole to maintain a stable orbit. If the planet is moving too slow it would just get sucked into the black hole. So for a stable orbit near a black hole that would result in the time dilation as shown in interstellar the planet would need to be moving at a significant portion of the speed of light. (Maybe someone who understands how to calculate time dilation can calculate how fast it was going for 1 hour local to equal 7 years on earth) As for the time dilation difference across the planet it would be quite insignificant i.e 1 hour on close side of planet = 6.9999999999999999999 years 1 hour on far side of planet = 7.00000000000001 years Regarding the rotation of the planet, the gravity shear is pretty high on a black hole, and high gravity sheer can more easily result in an orbital body becoming tidally locked like the earths moon.
[ "In the case of stars orbiting close to a spinning, supermassive black hole, frame dragging should cause the star's orbital plane to precess about the black hole spin axis. This effect should be detectable within the next few years via astrometric monitoring of stars at the center of the Milky Way galaxy.\n", "BULLET::::- In \"Interstellar\" (2014), a wormhole leads to a supermassive black hole called Gargantua, located 10 billion light-years from Earth and is orbited by several planets. Its mass is 100 million times that of Earth's sun, and it is spinning with 99.8 percent of the speed of light. For one of the planets, named Miller's planet, one hour equals seven years on Earth due to dramatic time dilation induced by the close proximity to Gargantua.\n", "As a black hole rotates, it twists spacetime in the direction of the rotation at a speed that decreases with distance from the event horizon. This process is known as the Lense–Thirring effect or frame-dragging. Because of this dragging effect, an object within the ergosphere cannot appear stationary with respect to an outside observer at a great distance unless that object were to move at faster than the speed of light (an impossibility) with respect to the local spacetime. The speed necessary for such an object to appear stationary decreases at points further out from the event horizon, until at some distance the required speed is that of the speed of light.\n", "If the black hole is rotating, there is a second radius of influence associated with the rotation. This is the radius inside of which the Lense-Thirring torques from the black hole are larger than the Newtonian torques between stars. Inside the rotational influence sphere, stellar orbits precess at approximately the Lense-Thirring rate; while outside this sphere, orbits evolve predominantly in response to perturbations from stars on other orbits. Assuming that the Milky Way black hole is maximally rotating, its rotational influence radius is about 0.001 parsec, while its radius of gravitational influence is about 3 parsecs.\n", "A rotating black hole might provide additional assistance, if its spin axis is aligned the right way. General relativity predicts that a large spinning frame-dragging—close to the object, space itself is dragged around in the direction of the spin. Any ordinary rotating object produces this effect. Although attempts to measure frame dragging about the Sun have produced no clear evidence, experiments performed by Gravity Probe B have detected frame-dragging effects caused by Earth. General relativity predicts that a spinning black hole is surrounded by a region of space, called the ergosphere, within which standing still (with respect to the black hole's spin) is impossible, because space itself is dragged at the speed of light in the same direction as the black hole's spin. The Penrose process may offer a way to gain energy from the ergosphere, although it would require the spaceship to dump some \"ballast\" into the black hole, and the spaceship would have had to expend energy to carry the \"ballast\" to the black hole.\n", "Rotating black holes are surrounded by a region of spacetime in which it is impossible to stand still, called the ergosphere. This is the result of a process known as frame-dragging; general relativity predicts that any rotating mass will tend to slightly \"drag\" along the spacetime immediately surrounding it. Any object near the rotating mass will tend to start moving in the direction of rotation. For a rotating black hole, this effect is so strong near the event horizon that an object would have to move faster than the speed of light in the opposite direction to just stand still.\n", "Rotating black holes are formed in the gravitational collapse of a massive spinning star or from the collapse or collision of a collection of compact objects, stars, or gas with a total non-zero angular momentum. As all known stars rotate and realistic collisions have non-zero angular momentum, it is expected that all black holes in nature are rotating black holes. Since observed astronomical objects do not possess an appreciable net electric charge, only the Kerr solution has astrophysical relevance.\n" ]
When splitting water molecules by electrolysis, how would sperate the hydrogen and oxygen?
What? If I’m interpreting your question right, the two gasses will collect at the positive and negative poles in the circuit. Putting a test tube or some other container in the water over each pole will collect the gases as they bubble up.
[ "Liquid water can be split into the elements hydrogen and oxygen by passing an electric current through it—a process called electrolysis. The decomposition requires more energy input than the heat released by the inverse process (285.8 kJ/mol, or 15.9 MJ/kg).\n", "Water can be converted to its component elemental gasses, H and O through the application of an external voltage. Water doesn't decompose into hydrogen and oxygen spontaneously as the Gibbs free energy for the process at standard conditions is about 474.4 kJ. The decomposition of water into hydrogen and oxygen can be performed in an electrolytic cell. In it, a pair of inert electrodes usually made of platinum immersed in water act as anode and cathode in the electrolytic process. The electrolysis starts with the application of an external voltage between the electrodes. This process will not occur except at extremely high voltages without an electrolyte such as sodium chloride or sulfuric acid (most used 0.1 M).\n", "The electrolysis of water is a simple method of producing hydrogen. A low voltage current is run through the water, and gaseous oxygen forms at the anode while gaseous hydrogen forms at the cathode. Typically the cathode is made from platinum or another inert metal when producing hydrogen for storage. If, however, the gas is to be burnt on site, oxygen is desirable to assist the combustion, and so both electrodes would be made from inert metals. (Iron, for instance, would oxidize, and thus decrease the amount of oxygen given off.) The theoretical maximum efficiency (electricity used vs. energetic value of hydrogen produced) is in the range 88–94%.\n", "Photocatalytic Water Splitting is an artificial photosynthesis process in which water is dissociated into hydrogen (H2) and oxygen (O2), using artificial or natural light. Methods such as photocatalytic water splitting are currently being investigated to produce hydrogen as a clean source of energy. The superior electron mobility and high surface area of graphene oxide sheets suggest it may be implemented as a catalyst that meets the requirements for this process. Specifically, graphene oxide’s compositional functional groups of epoxide (-O-) and hydroxide (-OH) allow for more flexible control in the water splitting process. This flexibility can be used to tailor the band gap and band positions that are targeted in photocatalytic water splitting. Recent research experiments have demonstrated that photocatalytic activity of graphene oxide containing a band gap within the required limits has produced effective splitting results, particularly when used with 40-50% coverage at a 2:1 hydroxide:epoxide ratio. When used in composite materials with CdS (a typical catalyst used in photocatalytic water splitting), graphene oxide nanocomposites have been shown to exhibit increased hydrogen production and quantum efficiency.\n", "On the contrary, as shown by Funk and Reinstrom, multiple reactions (e.g. \"k\" steps) provide additional means to allow spontaneous water-splitting without work thanks to different entropy changes ΔS° for each reaction i. An extra benefit compared with water thermolysis is that oxygen and hydrogen are separately produced, avoiding complex separations at high temperatures.\n", "The process of water-splitting is a highly endothermic process (Δ\"H\" 0). Water splitting occurs naturally in photosynthesis when photon energy is absorbed and converted into the chemical energy through a complex biological pathway (Dolai's S-state diagrams). However, production of hydrogen from water requires large amounts of input energy, making it incompatible with existing energy generation. For this reason, most commercially produced hydrogen gas is produced from natural gas.\n", "Water splitting, in which water is decomposed into its component protons, electrons, and oxygen, occurs in the light reactions in all photosynthetic organisms. Some such organisms, including the alga \"Chlamydomonas reinhardtii\" and cyanobacteria, have evolved a second step in the dark reactions in which protons and electrons are reduced to form H gas by specialized hydrogenases in the chloroplast. Efforts have been undertaken to genetically modify cyanobacterial hydrogenases to efficiently synthesize H gas even in the presence of oxygen. Efforts have also been undertaken with genetically modified alga in a bioreactor.\n" ]
Why is it that every copy of the Bible that I've ever seen has two columns on each page?
According to the Bible publisher Crossway (publishers of the English Standard Version), the reasons are threefold: Printing in a single column increases the page count by 10 to 25%, which increases how expensive it it to produce. For the typical font size, the two columns increases readability (9-12 words per line is considered optimal) Traditionally, the first printed Bibles were printed with two columns, as were ancient manuscripts handwritten by medieval monks. However, there are some Bible editions written in a single column on the page. I have used several of them, it just depends on how the publisher is sizing the font and how thick they want to make it, and how big the book will be. _URL_0_
[ "Since the list of line totals [of the books in the Bible available] in the city of Rome is not reliable, and elsewhere because of greed is not complete, I have gone through each individual book, counting 16 syllables to the line (as used in Virgil), and recorded the number for each book in all of them. \n", "The Bible’s layout is similar to the layout of the other three early Bibles Moralisées. On each page there are two columns, each with four miniatures set in medallions, which work as pairs. The upper miniature of each pair illustrates the text of the Old Testament, the lower shows a typologically equivalent scene from the New Testament or an allegorical or mystical meaning of the Old Testament story. Besides the miniatures, there are two narrow columns with explanatory text.\n", "The number of verses per series oscillates between the 264 of the number XVII and the two verses from various others (II, V, V, etc.). It is possible that many of these cases are regarding remains of incomplete series, because the text contains many holes.\n", "The books are stacked on each other to form individual letters, but separately so that one character is built of Korans, another of Bibles, and another of the Torah. Each book is a model of an original religious book, made in wax by hand. The books, of all sizes and forms, are stacked irregularly, so the sculpture appears with a slightly shaky expression. As there are only 14 letters and 3 religions, one religion will have only 4 letters but they will be bigger than the others.\n", "Biblical and literary scholars have noted that chapter and verse numbering disguises the actual form of the biblical writings and interferes with the act of reading. Ernest Sutherland Bates wrote, \"Certainly, no literary format was ever less conducive to pleasure or understanding than is the curious and complicated panoply in which the Scriptures have come down to us. None but a work of transcendent literary genius could have survived such a handicap at all.\" Richard Moulton noted, \"We are all agreed to speak of the Bible as a supremely great literature. Yet, when we open our ordinary editions, we look in vain for the lyrics, epics, dramas, essays, sonnets, treatises, which make the other great literatures of the world; instead of these the eye catches nothing but a monotonous uniformity of numbered sentences.\" New Testament scholar Hermann von Soden urged publishers, \"It is high time, in any editions that wish to facilitate rather than impede readers' understanding of the New Testament writings, for not only the verse divisions… but also the conventional chapter divisions to disappear completely from the text. Without giving any consideration to these divisions, the text must be printed in a way that will illustrate visually what units the authors themselves created within their works.\"\n", "The division of the Bible into chapters and verses has received criticism from some traditionalists and modern scholars. Critics state that the text is often divided in an incoherent way, or at inappropriate rhetorical points, and that it encourages citing passages out of context. Nevertheless, the chapter and verse numbers have become indispensable as technical references for Bible study.\n", "The tables of the are placed before every Gospel. The text is divided according to the (\"chapters\"), which numbers are given at the left margin, and their (\"titles\") at the top of the pages. There is also a division with a references to the Eusebian Canons (partially). There is no heading to the Gospel of Luke, but a blank space is left, so that perhaps the manuscript was never finished. The manuscript has a large number of unfoliated leaves.\n" ]
Were the mentally ill revered and allowed to roam freely throughout the middle ages?
Well you probably don't consider Syria to be western Europe, but I happen to have a source that suggests this was the case in medieval Muslim Syria. The book is a discussion of Medieval Syrian Saints: "Certain men outwardly defied the ideals of Muslim piety by living in a state of ritual impurity, wearing filthy garments, and not praying. Yet, the 'marginal holy man' did not represent a distinct class of saints as such. [...] The *muwallah* may be regarded as variously a cross between a borderline mystic, saint, vagabond, charlatan, and healer." Meri translates *muwallah* as 'one who is mad with devotion to God'. It's unclear whether this refers to a person that we would today consider to be mentally ill. Meri quotes from Ibn Taymiyya: "It is not possible for anybody to say that this is a friend of God. If this [person] is not insane (*majnun*), then he is a *mutawallih* without possessing insanity, or sometimes he would lose his sense of reasoning while at others regain it, and he does not undertake the religious obligations." Meri relates a few sources regarding one particular *muwallah* called Yusuf al-Qamini, who was apparently a popular healer and miracle worker in Damascus the early 1200s. It seems this was a person who had lost touch with reality but that people saw a power or religious devotion in his ramblings. Meri quotes from a late medieval writer, Ibn al-Hawrani: "Yusuf al-Qamini, the *muwallah* whom the common people regard as a *wali* [Saint] Their proof is that he possessed illuminations and spoke words based on intuitions. This is what happens with the soothsayer, the monk, and the madman who has and the madman who has a companion of the *jinn*. This sort of occurrence has increased in our time. God is asked for support. Yusuf used to wallow in his urine, walk barefoot, and repair to the furnace of Nur al-Din's bath. He did not pray." Meri cites a few other examples and goes on to theorize that *muwallah* existed as kind of marginal-charismatic figures. They were on the border between madness and holiness. On the one hand they are recorded as often performing no religious duties (like praying) and doing things that proper religious leader would never do like going outside in soiled or insufficient clothing. Yet on the other hand they associated themselves with religious places and common people thought they might perform miracles and healings. They were beyond the boundaries of normal society, but also involved in it, and in that they possessed a kind of religious charisma. Meri writes: "Men and women did not flock to them because they challenged established norms of learning, but rather for their conventional wisdom and because they regarded them as sacred persons who dispensed miracles and possessed a charismatic personality." But, were these people mentally ill? If somebody today behaved the way these Damascenes did, we would probably regard them as such. However, our modern society doesn't have a category of person that is expected to act in that manner, but the late medieval society of Damascus did, which may have in turn elicited that behavior. So it becomes completely unclear whether a *muwallah* was a person suffering from a medical or psychological condition (schizophrenia, epilepsy?) or rather a highly devoted religious person who felt compelled to act in this manner. Perhaps a combination of both? You might find this to be off topic but a relationship between "marginal" behavior and holiness is definitely documented in folk variations of Islam, both in the middle ages and into the modern era. Meri, Josef W. *The Cult of Saints Among Muslims and Jews in Medieval Syria* Oxford University Press, 2003. The discussion of the *muwallah* can be found in chapter 2, pages 91-100.
[ "In 1349 a church honouring Dymphna was built in Geel. By 1480, so many pilgrims were coming from all over Europe, seeking treatment for the mentally ill, that the church housing for them was expanded. Soon the sanctuary for the mad was again full to overflowing, and the townspeople began taking them into their own homes. Thus began a tradition for the ongoing care of the mentally ill that has endured for over 500 years and is still studied and admired today. Patients were, and still are, taken into the inhabitants of Geel's homes. Never called patients, they are called boarders, and are treated as ordinary and useful members of the town. They are treated as members of the host family. They work, most often in menial labour, and in return, they become part of the community. Some stay a few months, some decades, some for their entire lives. At its peak in the 1930s, over 4,000 'boarders' were housed with the town's inhabitants.\n", "The most feared disease of the Middle Ages was leprosy. The terrible effects of the disease on the body were so catastrophic that those with the disease were most often expelled from human society. As a result, St. George's Hospital was located outside the city. The earliest of such foundations was in 1109 which were often dedicated to Saint George (), the patron saint of knights and protector of the weak and helpless. The first mention of Copenhagen's leper hospital was in 1261 as an existing institution. In 1275, one Olof Blok willed the contents of a house in Copenhagen for the maintenance of the leper hospital.\n", "There were both cruel and humane treatments developed in the Middle Ages for the mentally ill. Most behaviors that could not be explained were attributed to supernatural causes and that humans innately had a battle between good and evil happening inside of them all the time. People were tested to see if they were evil or with the devil using “water tests”. In Baghdad and Damascus, however, in the ninth and tenth centuries, humane treatments were being developed in which centers of care for the mentally ill were based in love and kindness. Humanity regressed again in the 16th century when hospitals known as asylums were developed to provide a place for people who were unable to care for themselves. These institutions were terrible and people were often kept in restraints and left there in their own waste. In the late 1700s, there were people who began to reform the system and developed something known as moral treatment at the time. Moral treatment included organized schedules of productive behavior, socializing, entertainment, education, exercise, and nutrition.\n", "In the Islamic world, the \"Bimaristans\" were described by European travellers, who wrote about their wonder at the care and kindness shown to lunatics. In 872, Ahmad ibn Tulun built a hospital in Cairo that provided care to the insane, which included music therapy. Nonetheless, medical historian Roy Porter cautions against idealising the role of hospitals generally in medieval Islam, stating that \"They were a drop in the ocean for the vast population that they had to serve, and their true function lay in highlighting ideals of compassion and bringing together the activities of the medical profession.\"\n", "In 13th century medieval Europe, psychiatric hospitals were built to house the mentally ill, but there were not any nurses to care for them and treatment was rarely provided. These facilities functioned more as a housing unit for the insane. Throughout the high point of Christianity in Europe, hospitals for the mentally ill believed in using religious intervention. The insane were partnered with “soul friends” to help them reconnect with society. Their primary concern was befriending the melancholy and disturbed, forming intimate spiritual relationships. Today, these soul friends are seen as the first modern psychiatric nurses.\n", "The contemporary Christian view tended to view the psychologically troubled as possessed by the devil. Pare Jofré's missions to rescue captives in Muslim Spain and North Africa are thought to have given him an insight into the different way the mentally ill were treated in Islamic communities. On 24 February 1409, as he was on his way to the Cathedral in Valencia to preach the homily for the first Sunday of Lent, he saw two young men brutally attacking a madman. After rescuing the injured victim and taking him back to his convent, he returned towards the cathedral and preached a memorable sermon that urged the establishment of a charitable institution to care for and treat the mentally ill and other outcasts. At the end of the sermon a group of 11 Valencians, headed by Lorenzo Salom, joined together to fund this initiative.\n", "These impaired were not able to work to support themselves, and young children found themselves in the same situation. Very few had skills, the Hospital had seen weavers, merchants, a surgeon, and even priests, but the largest section had no skills. This showed that some of the beggars, were not truly beggars, they were poor that had fallen victims to economic issues such as unemployment. The line of distinction between the poor and the wandering beggar became unclear and the directors did not know who they should treat with the compassion which they previously reserved for the \"deserving poor\".\n" ]
Gluon are reponsible for the mass of a proton but the higg boson are the responsible for the mass of a gluon?
Gluons are massless. Protons and neutrons get a percent or so of their mass from the Higgs effect (which gives a small mass to the up and down quarks), and the rest of their mass from the strong force fields (primarily) that bind those quarks together (and the strong force field is made of gluons). Finally: the Higgs boson does not give any particles mass. The Higgs effect causes certain particles to get a mass by their interaction with the Higgs field that fills space. The Higgs boson itself and the Higgs field that fills space are both consequences of the fact that there is a fundamental Higgs field.
[ "In quantum chromodynamics, the modern theory of the nuclear force, most of the mass of protons and neutrons is explained by special relativity. The mass of a proton is about 80–100 times greater than the sum of the rest masses of the quarks that make it up, while the gluons have zero rest mass. The extra energy of the quarks and gluons in a region within a proton, as compared to the rest energy of the quarks alone in the QCD vacuum, accounts for almost 99% of the mass. The rest mass of a proton is, thus, the invariant mass of the system of moving quarks and gluons that make up the particle, and, in such systems, even the energy of massless particles is still measured as part of the rest mass of the system.\n", "Two terms are used in referring to the mass of the quarks that make up protons: \"current quark mass\" refers to the mass of a quark by itself, while \"constituent quark mass\" refers to the current quark mass plus the mass of the gluon particle field surrounding the quark. These masses typically have very different values. As noted, most of a proton's mass comes from the gluons that bind the current quarks together, rather than from the quarks themselves. While gluons are inherently massless, they possess energy—to be more specific, quantum chromodynamics binding energy (QCBE)—and it is this that contributes so greatly to the overall mass of protons (see mass in special relativity). A proton has a mass of approximately 938 MeV/c, of which the rest mass of its three valence quarks contributes only about 9.4 MeV/c; much of the remainder can be attributed to the gluons' QCBE. \n", "The gluons play an important role in the elementary strong interactions between quarks and gluons, described by QCD and studied particularly at the electron-proton collider HERA at DESY. The number and momentum distribution of the gluons in the proton (gluon density) have been measured by two experiments, H1 and ZEUS, in the years 1996-2007. The gluon contribution to the proton spin has been studied by the HERMES experiment at HERA. The gluon density in the proton (when behaving hadronically) also has been measured.\n", "Most of the mass of hadrons is actually QCD binding energy, through mass-energy equivalence. This phenomenon is related to chiral symmetry breaking. In the case of nucleons – protons and neutrons – QCD binding energy forms about 99% of the nucleon's mass. That is if assuming that the kinetic energy of the hadron's constituents, moving at near the speed of light, which contributes greatly to the hadron mass, is part of QCD binding energy. For protons, the sum of the rest masses of the three valence quarks (two up quarks and one down quark) is approximately 9.4 MeV/c, while the proton's total mass is about 938.3 MeV/c. For neutrons, the sum of the rest masses of the three valence quarks (two down quarks and one up quark) is approximately 11.9 MeV/c, while the neutron's total mass is about 939.6 MeV/c. Considering that nearly all of the atom's mass is concentrated in the nucleons, this means that about 99% of the mass of everyday matter (baryonic matter) is, in fact, chromodynamic binding energy.\n", "The gluon is a vector boson; like the photon, it has a spin of 1. While massive spin-1 particles have three polarization states, massless gauge bosons like the gluon have only two polarization states because gauge invariance requires the polarization to be transverse. In quantum field theory, unbroken gauge invariance requires that gauge bosons have zero mass (experiments limit the gluon's rest mass to less than a few meV/\"c\"). The gluon has negative intrinsic parity.\n", "BULLET::::- The proton mass \"m\" is composed primarily of gluons, and of the quarks (the up quark and down quark) making up the proton. Hence \"m\", and therefore the ratio \"μ\", are easily measurable consequences of the strong force. In fact, in the chiral limit, \"m\" is proportional to the QCD energy scale, Λ. At a given energy scale, the strong coupling constant \"α\" is related to the QCD scale (and thus \"μ\") as\n", "In technical terms, gluons are vector gauge bosons that mediate strong interactions of quarks in quantum chromodynamics (QCD). Gluons themselves carry the color charge of the strong interaction. This is unlike the photon, which mediates the electromagnetic interaction but lacks an electric charge. Gluons therefore participate in the strong interaction in addition to mediating it, making QCD significantly harder to analyze than QED (quantum electrodynamics).\n" ]
Does a star always have to be at the center of a solar system?
It's a misconception that stars are in the center of a stellar system. In such systems, all objects are orbiting around a common center of mass, the so called barycenter. Is the system a planetary system like ours, the barycenter is inside or close to the mother star, not because "it has to be like that", but because basically all the mass of our solar sytem (99,9%) is inside the sun. Because of that, ti's usually enough to say, that the sun is the center of our solar system, although in fact it isn't (jupiters mass is enough to let the barycenter be outside of the sun). However most stellar systems aren't quite like our own one, with a single star and some planets revolving around it. Most stars are in a binary or even multiple star system and as a stars mass is far higher than a planets mass, the barycenter is much further away from each star than it is in our system from the sun. So the answer to your question is: In our own galaxy it is rather unlikely, that a star is in or close the the center of it's stellar system. It is believed that about 70% of all stars in the milky way are in a binary or multiple star system. Asking if a star could orbit a planet becomes needless, because as already said, both objects revolve around the barycenter and not one rotates around the other. A planet however can never have a mass higher than a star, because if it would, it would start fusion and therefore become a star aswell. The size doesn't matter either. It is thinkable that a gaseous planet can be bigger in size than a very small star, however the star would still contain more mass. I hope, that cleared out the topic.
[ "Each star in the system has about eighty times the mass of the Sun. It is not clear why this system is located away from the center of the cluster. It is possible that the system was formed in the core, but that it was ejected by dynamical interactions.\n", "If the Sun were to be observed from the Alpha Centauri system, the nearest star system to ours, it would appear to be a 0.46 magnitude star in the constellation Cassiopeia, and would create a \"/W\" shape instead of the \"W\" as seen from Earth. Due to the proximity of the Alpha Centauri system, the constellations would, for the most part, appear similar. However, there are some notable differences with the position of other nearby stars; for example, Sirius would appear about one degree from the star Betelgeuse in the constellation Orion. Also, Procyon would appear in the constellation Gemini, about 13 degrees below Pollux.\n", "In hierarchical systems the planets are arranged so that the system can be gravitationally considered as a nested system of two-bodies, e.g. in a star with a close-in hot jupiter with another gas giant much further out, the star and hot jupiter form a pair that appears as a single object to another planet that is far enough out.\n", "The primary star in this system spins at a rate 50 times that of the Sun, and consequently has a strong magnetic field. It has a greater number of star spots than the Sun. These can cause the luminosity of the star to appear to vary over each orbital cycle. Measurements of the spin rate of this star at its equator have shown that it varies over time due to the effect of this magnetic field.\n", "Most of the major bodies of the Solar System orbit the Sun in nearly the same plane. This is likely due to the way in which the Solar System formed from a protoplanetary disk. Probably the closest current representation of the disk is known as the \"invariable plane of the Solar System\". Earth's orbit, and hence, the ecliptic, is inclined a little more than 1° to the invariable plane, Jupiter's orbit is within a little more than ° of it, and the other major planets are all within about 6°. Because of this, most Solar System bodies appear very close to the ecliptic in the sky.\n", "The star rotates at an inclination of 13 degrees relative to Earth. It has been assumed that the planets share that inclination. However b and c are \"hot Neptunes\", and outside this system several are now known to be oblique relative to the stellar axis.\n", "In a \"physical\" triple star system, each star orbits the center of mass of the system. Usually, two of the stars form a close binary system, and the third orbits this pair at a distance much larger than that of the binary orbit. This arrangement is called \"hierarchical\". The reason for this is that if the inner and outer orbits are comparable in size, the system may become dynamically unstable, leading to a star being ejected from the system. Triple stars that are \"not\" all gravitationally bound might comprise a physical binary and an \"optical\" companion, such as Beta Cephei, or rarely, a purely \"optical\" triple star, such as Gamma Serpentis.\n" ]
Before there was Rock 'n' Roll, what music did the rebellious teens of a decade such as the 1930's (and the surrounding years) listen to?
There was a time when Jazz was not fancy, there was a time when it was considered a corrupting influence. It was linked to crime, drugs, insanity, promiscuity... Check [this article](_URL_0_) published in Ladies Home Journal, August 1921. Racism was obviously a component in the prejudice against that music (and some others that were put in the same category).
[ "By the end of the 1960s, the American and British counterculture and hippie movement had moved rock towards psychedelic rock, heavy metal, progressive rock and other styles (from which Scorpions rose to prominence), incorporating, for the first time in popular music, socially and politically incisive lyrics. The 1968 student riots in Germany, France and Italy had created a class of young, intellectual continental listeners, while nuclear weapons, pollution and war inspired protests and activism. Music had taken a turn towards electronic avant-garde in the mid-1950s.\n", "Other regional scenes of teenage bands playing R&B-oriented rock were well-established in the early 1960s, several years before the British Invasion, in places such as Texas and the Midwest. At the same time, in southern California surf bands formed, playing raucous guitar- and saxophone-driven instrumentals. Writer Neil Campbell commented: \"There were literally thousands of rough-and-ready groups performing in local bars and dance halls throughout the US \"prior\" to the arrival of the Beatles ... [T]he indigenous popular music which functioned in this way ... was the protopunk more commonly identified as \"garage rock\"\".\n", "In the 1960s, music became heavily involved in the burgeoning youth counter culture, as well as various social and political causes. The beginning of the decade saw the peak of doo wop's popularity, in about 1961, as well as the rise of surf, girl groups and the first soul singers. Psychedelic and progressive rock arose during this period, along with the roots of what would later become funk, hip hop, salsa, electronic music, punk rock and heavy metal. An American roots revival occurred simultaneously as a period of sexual liberation and racial conflict, leading to growth in the lyrical maturity and complexity of popular music as songwriters wrote about the changes the country was going through.\n", "The period of the later 1950s and early 1960s has traditionally been seen as an era of hiatus for rock and roll. More recently some authors have emphasised important innovations and trends in this period without which future developments would not have been possible. While early rock and roll, particularly through the advent of rockabilly, saw the greatest commercial success for male and white performers, in this era the genre was dominated by black and female artists. Rock and roll had not disappeared at the end of the 1950s and some of its energy can be seen in the Twist dance craze of the early 1960s, mainly benefiting the career of Chubby Checker.\n", "From its early 1950s beginnings through the early 1960s, rock and roll spawned new dance crazes including the twist. Teenagers found the syncopated backbeat rhythm especially suited to reviving Big Band-era jitterbug dancing. Sock hops, school and church gym dances, and home basement dance parties became the rage, and American teens watched Dick Clark's \"American Bandstand\" to keep up on the latest dance and fashion styles. From the mid-1960s on, as \"rock and roll\" was rebranded as \"rock,\" later dance genres followed, leading to funk, disco, house, techno, and hip hop.\n", "Between 1955 and 1957, the \"porteño\" youth was introduced to rock 'n' roll music through films like \"Blackboard Jungle\", \"Rock Around the Clock\", \"Don't Knock the Rock\" and \"Rock, Rock Rock\"; becoming acquainted with American artists such as Bill Haley & His Comets, Little Richard, Jerry Lee Lewis and Elvis Presley. Mr. Roll y sus Rockers, a band led by Eddie Pequenino, recorded the first rock 'n' roll music of the country, covering various songs by Bill Haley and also writing original compositions in English. Like Billy Cafaro, another of the first Argentine rock 'n' roll musicians, Pequenino limited himself to imitating foreign artists. As it grew in popularity, rock 'n' roll music began to be embraced by young people, who began to veer away from tango, jazz, and the dance halls that until then they shared with their parents. According to Yanko González, \"this initial rock represents, basically, a space of youthful fun that expresses itself through a dance with very provocative corporal movements for the time. There its transgression is established: in the audacity of its movements, in the expressive use of the body\". In the early 1960s, Mexican band Los Teen Tops, originally led by singer Enrique Guzmán, gained popularity in Argentina and were influential in the development of an Argentine scene by introducing rock 'n' roll songs with Spanish lyrics. Other popular Mexican acts were Los Locos del Ritmo, Los Loud Jets, and the solo work of Enrique Guzmán. Until the release of \"La balsa\", singing rock music with Spanish lyrics was generally frowned upon, being considered \"tacky\" and \"uncool\", as the style was strongly associated with the English-speaking world.\n", "Rock and Roll was arguably birthed by Bill Haley and the Comets during their regular gigs at the Twin Bar in Gloucester City NJ just across the river from Philadelphia during the early 1950s. The city and its suburbs have since been the home of a couple of influential rock artists like Joan Jett. Many bands call Philadelphia home, and the area is constantly highly regarded by rock bands and artists.\n" ]
When were governments first expected to create jobs?
I'll take your more specific question. The sources I've read about the creation of parks in cities (specifically Central Park, which I'll look through my house to find) shows that during the 19th Century there was some municipally centered demands to create jobs, but this was not to create jobs in and of itself, but for the purpose of creating, as stated, public parks. The New Deal was really the time when the idea of "job creation" came into its own. Before this time the dominant school of economic thought was the Classical School (Smith, Ricardo, etc) which denied the existence of involuntary unemployment. Basically most people believed that unemployment could only result from people refusing to work below certain wages, not that in economic crises people could be out of work and unable to find it again. However, by 1933, that was proving to be untrue, John Maynard Keynes also helped by basically disproving it completely in "The General Theory." At this point the Roosevelt Administration were looking for some way to alleviate the crisis, and eventually started the programs composing the New Deal in 1933. A whole portion of the basic program was Relief: meaning job creation through public works to decrease unemployment. After that Keynesian thought dominated economic thought, especially the idea of job creation.
[ "In around 1890, both the United States and European governments created government-funded employment offices to provide work for unemployed unskilled laborers. These services proved to be unsuccessful. In 1933 during the Great Depression, with the Wagner-Peyser Act, the USES was reinstated “to set minimum standards, develop uniform administrative and statistical procedures, publish employment information, and promote a system of \"clearing labor\" between states.” Then President Franklin D. Roosevelt had created many government-funded work projects to help boost the economy and the USES was responsible for hiring the workers on those projects. The USES operated originally in only a few states but by World War II, it was operating in all states and played a major role in providing jobs during the war. In the United States home front during World War II, the service coordinated employment of prisoners of war (e.g., using German POWs at Gettysburg for local pulpwood cutting).\n", "It was devised by economist David L. Birch at the Massachusetts Institute of Technology. In 1980 Birch published an influential study that showed 70% of job growth in the U.S. came from enterprises with fewer than 100 employees.\n", "BULLET::::- The first Federal employment service (forerunner of the United States Employment Service) was created in the Bureau of Immigration and Naturalization, Department of Commerce and Labor (1907).\n", "Birch was working as a director at MIT in 1979 when he published his report \"The Job Creation Process\", in which he showed that, contrary to the prevailing conventional wisdom, most new jobs in the US are created by small companies. This study caught the attention of politicians at home and abroad. Birch followed in 1987 with the book \"Job creation in America : how our smallest companies put the most people to work\". Birch's work has been heavily criticized but is nevertheless considered groundbreaking as it opened the field for the study of small businesses, which had been so far disregarded by economists.\n", "In the United States, the Employment Act of 1946 set the goals of achieving full employment, full production, and stable prices. It also created the Council of Economic Advisers to provide objective economic analysis and advice on the development and implementation of a wide range of domestic and international economic policy issues. In its first 7 years the CEA made five technical advances in policy making:\n", "Major employers in the second half of the twentieth century included Monsanto and Westinghouse plants at Pensacola, the St. Joe Paper Company in Port St. Joe, and Gulf Power, a major electric utility company.\n", "The Employment Act of 1946 was enacted by the government to keep the economy from plunging back into a post-war depression. The act declared the continuing policy and responsibility of the federal government to use all reasonable means to promote maximum (not full) employment, production, and purchasing power. In addition to focusing on keeping unemployment rates low, the act called for the creation of the Council of Economic Advisors. This council had the task of assisting the president in appointing members to the Joint Economic Committee in the United States Congress and continuing to develop the role of fiscal policy in the United States.\n" ]
When and how did child marriage start to be seen as inappropriate in the Western culture?
Child marriage was never really "appropriate" in Western culture; it barely happened. According to parish records, the average age for marriage was between 17 and 25 across much of Europe in the middle ages. The average age for women was 20 to 26 in Elizabethan England, and many writers of the time condemned child marriages. Child betrothals were performed between aristocratic families to cement dynasties but it was never a common practice, and they usually waited until at least early adulthood before marriages took place (since children often died). There are of course exceptions among the nobility, as with Henry VII's mother [giving birth to him when she was just 13] (_URL_0_). You'll find pedophilia apologists trying to argue that marriage and sex between adults and children was "normal" in the middle ages, but this doesn't hold up to fact. It didn't make sense for an adult to marry a child, neither socially nor economically, and marrying children to one another before they've had a chance to establish themselves and their skills didn't make much sense either. The stuff about average marriage ages in 16th- and 17th- c. England comes from *Birth, Marriage, and Death: Ritual, Religion, and the Life-Cyle in Tudor and Stuart England*, by David Cressy. *Medieval Households* by David Herlihy has something similar to say about the High Middle Ages. I'll defer to a good friend of mine who has this to say about early modern Jewish marriages, which sometimes took place between 12-14 year olds for several very important reasons: > First, unlike Christians, most Jews didn't own land. Their property was more often in the form of businesses and the goods involved in them. That means some of the most important things parents could bequeath their children also took the form of the experience, knowledge, and connections (quite) young couples would need to make their own livings. When couples married young, parents could take an active part in the early years of their marriages, show them how to manage their money and their business affairs, and set them up with houses and in their professions. Perhaps unsurprisingly when you consider how young they were, a lot of young couples spent the first years of their marriage living under the roof of the bride or groom's parents. The short answer is that for many well-off-ish Jewish couples looking to the future of their children, their inheritance required a lot more direct involvement than, say, the inheritance of a farm. I have limited knowledge and I may be wrong, but I hope this was helpful.
[ "The marriage decisions in pre-modern China traditionally were made by parents with the help of matchmakers, and the fate of the children were determined at an early age. Since the reforms in the twentieth century, and the implementation of the marriage law, such practices have been outlawed. Legally the decision to marry lies in the freedom of choice of a man or woman to choose their partners. Before the Mao Era, and during the period of late imperial China, young people had almost no choice about their own marriage. Parents or older generations decided everything for them, on who should be their mate and the amount of money spent on the wedding.\n", "Child marriage was common throughout history, even up until the 1900s in the United States, where in 1880 CE, in the state of Delaware, the age of consent for marriage was 7 years old. Still, in 2017, over half of the 50 United States have no explicit minimum age to marry and several states set the age as low as 14. Today it is condemned by international human rights organizations. Child marriages are often arranged between the families of the future bride and groom, sometimes as soon as the girl is born. However, in the late 1800s in England and the United States, feminist activists began calling for raised age of consent laws, which was eventually handled in the 1920s, having been raised to 16-18.\n", "Child marriage has been traditionally prevalent in India but is not so continued in Modern India to this day. Historically, child brides would live with their parents until they reached puberty. In the past, child widows were condemned to a life of great agony, shaved heads, living in isolation, and being shunned by society. Although child marriage was outlawed in 1860, it is still a common practice. The Child Marriage Restraint Act, 1929 is the relevant legislation in the country.\n", "The problem of child marriage is at times justified on the basis of religious foundations. Historically, it can be explained as a reaction to invasions by foreigners; desire to perpetuate the cult of the family by marrying the son early; by marrying the daughter early to escape the discredit caused to the family by the presence of grown-up maiden; or by desire of the mother to marry her son early so that she may sooner obtain the possession of a daughter-in-law in whom the mother could inculcate her habits of obedience and who could share the domestic chores with the mother. In the case of parents, sometimes it is due to their keenness to relieve themselves of the responsibility of marrying their daughter. They are also considered socially acceptable for reasons of responsibility and economically desirable for saving marriage expenses, bride price/ dowry.\n", "During the sixth century, Emperor Justinian formulated legislation that was to become the basis of Western marriage law for the next millennia. Under his laws, cohabiting couples were no longer recognised as married and their children were regarded as illegitimate, with the same status as the children of prostitutes. However, the status of illegitimate children could be updated if the parents later married.\n", "In Europe, during the late 18th century and early 19th century, the literary and intellectual movement of romanticism presented new and progressive ideas about love marriage, which started to gain acceptance in society. In the 19th century, marriage practices varied across Europe, but in general, arranged marriages were more common among the upper class. Arranged marriages were the norm in Russia before early 20th century, most of which were endogamous. Child marriages were common historically, but began to be questioned in the 19th and 20th century. Child marriages are often considered to be forced marriages, because children (especially young ones) are not able to make a fully informed choice whether or not to marry, and are often influenced by their families.\n", "In Christian societies, sex outside marriage was forbidden. Older children were themselves often punished for being complicit in sexual interaction with an adult. Western Christianity also deemed that children were born into the original sin, and, as such, were perceived as inherently immoral. Children had very few rights and were considered the chattel of the father. From the late 18th century, and especially in the 19th century, attitudes started to change. By the mid-19th century there was increased concern over child sexual abuse.\n" ]
why do artists upload their full album to their youtube channel for free?
From an economic standpoint, there is a theory that this will actually increase album sales. It is based on the idea that psychologically, you want to do something for someone who gives you something. So, naturally, the thing to do in this situation is to buy the album, even if you can listen to it for free. Additionally, singers traditionally don't make that much from album sales as opposed to what they make from concert tickets. By giving away songs, the singer hopes it will make you more likely to come see him/her in concert.
[ "They are yet to release a full-length album, because they prefer giving out their music for free. They have had millions of downloads from their site and other related pages. They did release a free multimedia CD of their singles and videos in 2001. The band does not have any policies against their music being copied and distributed. They have made their music available for download on their official website.\n", "A goal of the project is to encourage artists to release tracks under a free license, and allow users to download or purchase these tracks. Only artists releasing music under free content licenses are promoted by the site. The website will also allow users to communicate among themselves, create groups of common interests and share information on musical events.\n", "Fans can download their purchases or stream their music on the Bandcamp app. They can also send purchased music as a gift, view lyrics, and save individual songs or albums to a wish list. Uploading music to Bandcamp is free, and the company takes a 15% cut of sales made from their website (in addition to payment processing fees), which drops to 10% after an artist's sales surpass $5000.\n", "The site typically hosts around 60 songs by 17 bands per week, almost all available as downloads. The site used to not be free, with a forced membership with a monthly fee to listen to the songs or download them since 2011, but as of March 1, 2018 it has returned to having free streaming. There is an archives section of past performers, each illustrated by a staff of illustrators. In addition to music, there are sections of reviews and commentaries of musicians and their releases.\n", "Many artists are using the Internet to give away music to create awareness and liking to a new upcoming album. The artists release a new song on the internet for free download, which consumers can download. The hope is to have the listeners buy the new album because of the free download. A common practice used today is releasing a song or two on the internet for consumers to indulge. In 2007, Radiohead released an album named \"In Rainbows\", in which fans could pay any amount they want, or download it for free.\n", "Registered users of the site could stream and share millions of songs and tens of thousands of music videos free of charge, with the costs for licensing and streaming supported by advertising on the site and on imeem Mobile.\n", "During 2011 the band also made all their music available for free download or pay as you like to adapt their model to the mp3 age where most fans choose to download the music rather than buy the albums.\n" ]
What happened to the armies immediately after a battle in Medieval Europe?
This is a vague question and has many different points that can and should be considered. But in all, I will do my best to provide an answer to a general battle during the early Medieval period. First and foremost, open battles happened quite infrequently. I mean, very rarely did to opposing forces march to an open field, size each other up and then proceed to give them the business. With that being said... Any armed force that moves about is going to comprise a long chain of soldiers, wagons, tents and various supporters and retainers. Squires and horses for nobility. Each noble or knight could have a couple horses. Blacksmiths. Fletchers. Cooks and pages to prepare food. Some form of surgeon would be present for those of the upper echelon. And the best part is that almost none of these people would see a battle. They would be at a camp that was nearby. Also, we have to remember that a majority of the forces would be levies from the feudal estates of various lords, barons, knights, aka the landed gentry. The levies typically are only there fighting after the planting season and before the harvest (between spring and fall). They are more often conscripted to fight for their lord or face punishment. They don't have retainers or squires to help carry stuff or their equipment. They carry it themselves. They end up feeding themselves. They also typically walked everywhere without the chance to ride a horse. As for a battle, it depends in large part to location, weather and position. Battles do and don't last very long. Some battles are very fast and over quickly due to a complete breakdown in morale if heavy cavalry is involved or if an ambush is sprung. Battles could last days if there were an open field siege (Hussite Wars where the Bohemians created wagon forts to defend from with spears, crossbows and early gun powder weapons). I digress, so I will list some assumptions - this battle is part of a larger conflict between two European crowns; the armies are comprised of the typical feudal arrangement of relatively few nobles and knights in comparison to the levies that make up a large portion of spear troops that are lightly armed; the armies themselves are being led by a semi-important to important general with some relation to the royal family (blood or political); last, we will assume this was a straightforward battle without much in regard to trickery, flanking or betrayal - good ole fashioned battle between two armies and only a few nearby villages or towns. In medieval battle, the goal was not to destroy the enemy completely, although that would have been a bonus if you were invading or defending lands. It was much more important to route or break an enemy army and then take prisoners of worth for ransom. A member of the royal family or someone important in the aristocracy could be ransomed and pay for an entire campaign or fighting season. Lesser lords and nobles would obviously bring less in terms of ransom, but their household or the crown would actually pay the ransom. The peasant levies would still be captured as they pose a potential threat if left behind enemy lines. The levies carried no real worth for ransom except to their feudal lord - the lord would still need to have people available for harvest after the fighting season. As a battle ends, one army is broken and retreats. There will be a time during the evening that people are sent forward to locate anyone important that might have fallen (see articles regarding Charles of Burgundy being located after his last battle and identified because of his scars). Both sides of the conflict would send out these people, also comprised of monks or priests to provide last rights to the dying. Each army would have returned to their respective camps after the battle to assess damage, losses, gains, prisoners, potential ransoms, etc. This would then play into the next parlay between the forces - basically a way to say..."I have this person and these lords, what will you trade me for them?" Quick and dirty version - the armies will withdraw from the field to their own camps, prisoners have been captured and their ransom value assessed. People will check for others on the battlefield that have fallen but have not yet died.
[ "After 4 years (907-910) of heavy defeats (Pressburg, Eisenach, Augsburg, Rednitz) from the hands of the Hungarian mounted archers, every of which resulting with the annihilation of the armies (this causing a \"shortage\" in soldiers to the Germans), and the deaths of the German commanders (among them princes, dukes, counts, margraves, bishops, archbishops), the German kings (Conrad I of Germany, Henry the Fowler) and other political leaders, decided not to fight again in an open field with the obviously tactically superior Magyars, fearing to have the same fate of their predecessors, but they retreated in their castles and walled towns (knowing that the Hungarians are not very skilled in sieges, because they have no siege equipments), waiting until they left their countries filled with spoils. It is interesting to know that not only the Germans who shared borders with the Hungarians chose not to fight with them (for example in 924 the German king Henry the Fowler retreated in his castle of Werla, instead of defending his duchy with fight, when hearing that the Hungarians crossed Saxony's borders, and started to plunder his realm), but also the French too, for example in 919, when the Hungarians invaded Lotharingia and France, the king Charles the Simple wanted to gather the forces of his kingdom against them, only the archbishop of Reims appeared from the nobles of the whole kingdom, who obviously hearing about the risks of a battle with the archers from the Carpathian Basin from the news which came from Germany, decided not to participate in a war against them, so the king withdrawed together with his 1500 soldiers, letting the Magyars to pillage his country. Because the fear of the European political and military leaders of the encounter with the Hungarians, after the years until 910, when in only four years (907-910) occurred four major battles (Battle of Pressburg in 907, Battle of Eisenach in 908, Battle of Augsburg in 910 and Battle of Rednitz), between 910 and 933 just two major battles took place between the Hungarians and their enemies: in 913 the Battle of Inn (a Bavarian-Swabian victory), and in 919 the Battle of Püchen (the Hungarians defeated the German king Henry the Fowler).\n", "This was an especially important consideration during the pre-Napoleonic era, when armies were relatively small forces composed of professional career soldiers and losses were difficult to replace. Besides, weakening one's own army might expose it to other enemies, or political struggles, and even internal enemies: a pretender to the throne or a coalition of nobles might challenge the king, a province might rebel to the central government or a third nation could exploit the situation, while the main army suffered heavy casualties after the battle.\n", "In 1414, a sporadic new war broke out, known as the \"Hunger War\" from the Knights' scorched-earth tactics of burning fields and mills; but both the Knights and the Lithuanians were too exhausted from the previous war to risk a major battle, and the fighting petered out in the autumn. Hostilities did not flare up again until 1419, during the Council of Constance, when they were called off at the papal legate's insistence.\n", "In 531, according to the \"Decem Libri\" of Gregory of Tours, the decisive battle between the Franconians and Thuringians took place along the Unstrut, which resulted in the destruction and annexation of the early medieval Thuringian kingdom by the Frankish empire. In 933 the German king Henry I fought, after a ten-year truce, against a Hungarian army in the Battle of Riade, a place near the Unstrut, but which is now unknown. His victory led to a period of peace, until the Hungarians returned in 955 and were defeated again.\n", "After his death, a period of revolts and conflict for supremacy ensued between the royalty and the nobles. In 1051 armies of the Holy Roman Empire tried to conquer Hungary, but they were defeated at Vértes Mountain. The armies of the Holy Roman Empire continued to suffer defeats; the second greatest battle was at the town now called Bratislava, in 1052. Before 1052 Peter Orseolo, a supporter of the Holy Roman Empire, was overthrown by king Samuel Aba of Hungary.\n", "The end of the medieval period was a time of suffering, notably at the hands of Emperor Maximillian (1488-1489) and during the Siege of Louvain by Maarten van Rossum in 1542, and again during the religious wars of the late sixteenth century, the campaigning following the creation of the Franco-Dutch alliance of 1635 and again the French invasion of 1695.\n", "After the extinction of the Árpád dynasty in 1301, the late medieval kingdom persisted, albeit no longer under Hungarian monarchs, and gradually reduced due to the increasing pressure by the expansion of the Ottoman Empire. Hungary bore the brunt of the Ottoman wars in Europe during the 15th century. The peak of this struggle took place during the reign of Matthias Corvinus (r. 1458–1490). The Ottoman–Hungarian wars concluded in significant loss of territory and the partition of the kingdom after the Battle of Mohács of 1526.\n" ]
Why does the sun change the color of my hair? Picture in comment
Did you dye it brown? When you lighten hair with a bleach, it does damage it and makes it more vulnerable to harsh chemicals (shampoos) radiation (the sun), wind and any other stress. Also if you dyed it brown, the dye would likely fade off revealing the bleached hair underneath. Most dark hair has some red/gold/coppery undertones. A lot of the girls in Asia lighten their hair and so their hair all has the same reddish tint. Ditto Latinas. FYI, Once when I lightened my normally dark hair to a medium brown color and while I was standing in line to pay for lunch at a cafeteria, I looked up and saw myself on the surveillance camera monitor screen, and my hair looked perfectly blond! I asked someone about this and was told it probably had something to do with the "infra red" camera.
[ "Light encountering a painted surface can either alter or break the chemical bonds of the pigment, causing the colors to bleach or change in a process known as photodegradation. Materials that resist this effect are said to be lightfast. The electromagnetic spectrum of the sun contains wavelengths from gamma waves to radio waves. The high energy of ultraviolet radiation in particular accelerates the fading of the dye.\n", "In 1978 the sun color was specified to be golden yellow (), to have an inner diameter of 10 cm, and an outer diameter of 25 cm (the diameter of the sun equals the height of the white stripe. The sun's face is of its height). It features 32 rays, alternately wavy and straight, and from 1978 it must be embroidered in the \"Official Flag Ceremony\".\n", "After the hair is lightened (the natural hues are destroyed) to a sufficient level, the pigment within the artificial color will, ideally, in perfect blue-red-yellow balance, replace the recently destroyed natural pigment. Often, hair that has been lightened will exhibit a redder or more orange tone than it did previously. This is because there is not enough blue pigment present to visually balance the red and yellow pigmentation (both natural and artificial) that remains. Understandably, hair color manufacturers elect to err on the side of turning hair a little too red, in the absence of enough blue, instead of possibly giving the hair a purple or blue tint, in the absence of enough yellow and/or red pigmentation (as might be the case when applied to grey or very light blonde hair).\n", "Where formula_13 is the emission spectrum of the sun, and formula_7 is the emission spectrum of the paint. Although, by Kirchhoff's law, formula_15 in the above equations, the above \"averages\" formula_16 and formula_17 are not generally equal to each other. The white paint will serve as a very good insulator against solar radiation, because it is very reflective of the solar radiation, and although it therefore emits poorly in the solar band, its temperature will be around room temperature, and it will emit whatever radiation it has absorbed in the infrared, where its emission coefficient is high.\n", "Hair coloring, or hair dyeing, is the practice of changing the hair color. The main reasons for this are cosmetic: to cover gray or white hair, to change to a color regarded as more fashionable or desirable, or to restore the original hair color after it has been discolored by hairdressing processes or sun bleaching.\n", "Although it may seem that the Sun is the brightest spot on the canvas, it is in fact, when measured with a photometer, the same brightness (or luminance) as the sky. Dr. Margaret Livingstone, a professor of neurobiology at Harvard University, said \"If you make a black and white copy of \"Impression: Sunrise\", the Sun disappears [almost] entirely.\"\n", "It is helpful to understand the general nature of the lightening, or \"lifting\", of natural hair color. The color of natural hair is determined by the melanin present. Generally, the more melanin, the darker the hair. In order for the color of hair to be lightened, a chemical must penetrate the cuticle of the hair, enter the cortex and destroy the melanin. Permanent hair color, bleach and other decolorants can do this to varying degrees.\n" ]
How long did slaves in America retain their specific tribal identities/religions/languages?
Many of the customs survived. If you go by John Thornton's Africans and the Making of the Atlantic World, many of the Slaves, though coming from different kingdoms had many language and religious similarities. So when new Africans arrived in the United States there was little trouble in recreating a society they knew. The biggest change is that it was in the context under English masters. The result is something of a creole culture that combines many elements of AFrica and Anglo culture. Gullah is an example of a language that survives to this day and is a creation of this cultural mix. To hear it go here _URL_0_ . Gullah is still spoken today in parts of Charleston South Carolina. Ira Berlin gives a little different view in "Many Thousands Gone". He states that the Middle Passage, the carrying of slaves from Africa to America, was such a harrowing experience that it more or less shattered the slave psyche. This combined with large language differences made the reconstruction of African culture very difficult, instead entirely new cultures were developed. Out of the two of them, I would say Thornton was more correct, as he is an African historian and would be more knowledgeable about African cultural differences. Elements of African culture still pervade African American society today, I would doubt that any of it has been truly "lost".
[ "While numerous tribes used captive enemies as servants and slaves, they also often adopted younger captives into their tribes to replace members who had died. In the Southeast, a few Native American tribes began to adopt a slavery system similar to that of the American colonists, buying African American slaves, especially the Cherokee, Choctaw, and Creek. Though less than 3% of Native Americans owned slaves, divisions grew among the Native Americans over slavery. Among the Cherokee, records show that slave holders in the tribe were largely the children of European men that had shown their children the economics of slavery. As European colonists took slaves into frontier areas, there were more opportunities for relationships between African and Native American peoples.\n", "While numerous tribes used captive enemies as servants and slaves, they also often adopted younger captives into their tribes to replace members who had died. In the Southeast, a few Native American tribes began to adopt a slavery system similar to that of the American colonists, buying African American slaves, especially the Cherokee, Choctaw, and Creek. Though less than 3% of Native Americans owned slaves, divisions grew among the Native Americans over slavery. Among the Cherokee, records show that slave holders in the tribe were largely the children of European men that had shown their children the economics of slavery. As European colonists took slaves into frontier areas, there were more opportunities for relationships between African and Native American peoples.\n", "While numerous tribes used captive enemies as servants and slaves, they also often adopted younger captives into their tribes to replace members who had died. In the Southeast, a few Native American tribes began to adopt a slavery system similar to that of the American colonists, buying African American slaves, especially the Cherokee, Choctaw, and Creek. Though less than 3% of Native Americans owned slaves, divisions grew among the Native Americans over slavery. Among the Cherokee, records show that slave holders in the tribe were largely the children of European men who had shown their children the economics of slavery. As European colonists took slaves into frontier areas, there were more opportunities for relationships between African and Native American peoples.\n", "From the late 1700s to the 1860s, the Five Civilized Tribes in the American Southeast began to adopt some colonial and American customs. Some men acquired separate land and became planters, buying African-American slaves for laborers in field work, domestic service, and various trades. The 1809 census taken by Cherokee agent Colonel Return J. Meigs, Sr. counted 583 \"Negro slaves\" held by Cherokee slaveowners. By 1835, that number increased to 1,592 slaves, with more than seven percent (7.4%) of Cherokee families owning slaves. This was a higher percentage than generally across the South, where about 5% of families owned slaves.\n", "Historically, certain Native American tribes have had close relations with African Americans, especially in regions where slavery was prevalent, or where free people of color have historically resided. Members of the Five Civilized Tribes participated in holding enslaved African Americans in the Southeast, and some enslaved or formerly enslaved people migrated with them to the West on the Trail of Tears in 1830 and later during the period of Indian Removal. In peace treaties with the US after the American Civil War, the slaveholding tribes, which had sided with the Confederacy, were required to emancipate slaves and give them full citizenship rights in their nations. In controversial actions, since the late 20th century, the Cherokee, Creek, and Seminole nations tightened their rules for membership and at times excluded Freedmen who did not have at least one ancestor listed as Native American on the early 20th-century Dawes Rolls. This exclusion was later appealed in the courts, both because of the treaty conditions and in some cases because of possible inaccuracies in some of the Rolls. The Chickasaw Nation never extended citizenship to Chickasaw Freedmen.\n", "Slavery of Native Americans was organized in colonial and Mexican California through Franciscan missions, theoretically entitled to ten years of Native labor, but in practice maintaining them in perpetual servitude, until their charge was revoked in the mid-1830s. Following the 1847–48 invasion by U.S. troops, the \"loitering or orphaned Indians\" were de facto enslaved in the new state from statehood in 1850 to 1867. Slavery required the posting of a bond by the slave holder and enslavement occurred through raids and a four-month servitude imposed as a punishment for Indian \"vagrancy\".\n", "As the United States Constitution and the laws of several states permitted slavery after the American Revolution (while northern states prohibited it), Native Americans were legally allowed to own slaves, including those brought from Africa by Europeans. In the 1790s, Benjamin Hawkins was the federal agent assigned to the southeastern tribes. Promoting assimilation to European-American mores, he advised the tribes to take up slaveholding so that they could undertake farming and plantations as did other Americans. The Cherokee tribe had the most members who held black slaves, more than any other Native American nation.\n" ]
how is there a slow lane and a fast lane on the highway if the speed limit never changes?
Some people drive faster than others. Passing on the driver's side is usually safer, which usually means passing on the left. All the people merging in and out tend to slow down the outside lanes a bit, too.
[ "A 65 mph left lane minimum speed limit is sometimes indicated on 75 mph roads with steep grades, \"slower traffic keep right\" is also in effect. On one-way roadways state law reserves the left and center lanes of two or more lanes for passing. There are reduced advisory speed limits for some roads during poor weather. Speeding fines are doubled in construction zones and designated safety corridors, with signs often stating this. There are no longer night speed limits, nor are there any differential speed limits for heavy trucks.\n", "Until around 2005, the roads had the original speed limit in use on most highways. As a result of this, many people drove at 90 km/h at 1-lane parts but at 2-lane parts, this being the speed limit on motorways. The speed limit has now been changed to with a notably smoother traffic flow. There is a problem that some people want to overtake as many slow cars as possible in two-lane section, sometimes with small margins at the end of the section.\n", "In some areas, such as the U.S. states of Colorado and Kentucky, vehicles in the left lane are required to yield to faster traffic only if the speed limit is above 65 miles per hour. In other areas, like Alaska, there is no law requiring slower traffic to move over for faster traffic.\n", "Speed limits on 4-lane divided highways are normally 65 mph although some stretches within cities are posted as low as 50 mph. Open country highways have a statutory limit of 55 mph, which includes most rural two-lane highways and even includes some one lane back country roads or any road without a posted speed limit. Cities and towns set their own speed limits, which are usually between 25 and 55 mph. School zones have a statutory speed limit of 20 mph. Speed limits are commonly reduced by 15 mph in work zones. \n", "Due to legal requirements for speed measuring, the speed limits for expressways and freeways are de facto and , respectively. The limits shown above apply only if there are no other signs present, as the signs may prescribe a lower or a higher speed limit (limits of 100 km/h or higher can also be found within inhabited places). The lowest legal speed limit under normal traffic conditions is .\n", "Speed limits are set by each state or territory, as well as counties or municipalities, on the roads within their jurisdiction. The maximum speed limit on rural two-lane roads ranges from 50 mph (80 km/h) in parts of the northeast to 75 mph (120 km/h) in parts of Texas. On rural Interstate Highways and other freeways, the speed limit ranges from 60 mph (96 km/h) in Hawaii to 85 mph (136 km/h) in parts of Texas. All roads in the United States have a speed limit, but it is not always posted (especially in rural areas).\n", "Some roads also have minimum speed limits, usually where slow speeds can impede traffic flow or be dangerous. The use of minimum speed limits is not as common as maximum speed limits, since the risks of speed are less common at lower speeds. In some jurisdictions, laws requiring a minimum speed are primarily centered around red-light districts or similar areas, where they may colloquially be referred to as \"kerb crawling laws\".\n" ]
how can a solution to a problem suddenly pop up?
Subconscious mind solves it for you with all of the acquired knowledge and experience. This won't happen if you've no knowledge.
[ "BULLET::::- Pop : When, during a solve, one or more cubies come out of contact with the puzzle, usually causing the puzzle to be unstable, in which, upon turning, more pieces may become loose and possibly pop out too.\n", "In a \"fixes that fail\" scenario the encounter of a problem is faced by a corrective action or fix that seems to solve the issue. However, this action leads to some unforeseen consequences. They form then a feedback loop that either worsens the original problem or creates a related one.\n", "BULLET::::- \"Top-down approach\": This is the direct fall-out of the recursive formulation of any problem. If the solution to any problem can be formulated recursively using the solution to its sub-problems, and if its sub-problems are overlapping, then one can easily memoize or store the solutions to the sub-problems in a table. Whenever we attempt to solve a new sub-problem, we first check the table to see if it is already solved. If a solution has been recorded, we can use it directly, otherwise we solve the sub-problem and add its solution to the table.\n", "The problem starts with an empty vase and an infinite supply of balls. An infinite number of steps are then performed, such that at each step 10 balls are added to the vase and 1 ball removed from it. The question is then posed: \"How many balls are in the vase when the task is finished?\"\n", "If answered incorrectly or not answered within the time limit, the answer is marked incorrect. If one player answers correctly twice in one round, the round will end. When a round ends, a block will be removed for each society's incorrect answers. The quizzes and floor removal will continue until a team is no longer able to stand on the shrinking floor. If a part of a someone's body touches the ground, their society will lose and the game will end.\n", "In order to solve a puzzle, the player must input an answer on the bottom screen. Sometimes, this will involve tapping a button labelled \"Input Answer\" and writing a word or number that solves the given puzzle. Other times, they might be asked to circle an answer or tap an area on the touch screen, and then hit \"Submit\". Still others are solved automatically once the player finds the solution by interacting with objects on the bottom screen. If the player has made a mistake in solving a puzzle, they can hit the button labelled \"Restart\" to reset the puzzle. Deducted Picarats and purchased hints are not reset if a puzzle is restarted. If the player submits an incorrect answer, they will be asked to repeat the puzzle, and a varying amount of Picarats will be deducted from the maximum possible for that puzzle. If the player submits a correct answer, they are rewarded with the amount of Picarats the puzzle is worth, and the plot will progress. After completing a puzzle, it can be played again at any time from the Puzzle Index, a section in the game's interface that lists the puzzles a player has completed. Occasionally, after completing a puzzle, the player will receive small items that can be used in minigames. Fulfilling certain criteria, such as clearing all levels in a minigame, unlock additional puzzles in the Bonuses section.\n", "This problem can be quickly solved with a dawning of realization, or \"insight\". A few minutes of struggling over a problem can bring these sudden insights, where the solver quickly sees the solution clearly. Problems such as this are most typically solved via insight and can be very difficult for the subject depending on either how they have structured the problem in their minds, how they draw on their past experiences, and how much they juggle this information in their working memories In the case of the nine-dot example, the solver has already been structured incorrectly in their minds because of the constraint that they have placed upon the solution. In addition to this, people experience struggles when they try to compare the problem to their prior knowledge, and they think they must keep their lines within the dots and not go beyond. They do this because trying to envision the dots connected outside of the basic square puts a strain on their working memory.\n" ]
If a king was presumed dead and an heir took the throne, what happens when the old king returns?
A slightly different scene. The Ottoman emperor Murad II abdicated his throne in favour of his son Mehmed II (aged 12). Murad wanted to pursue the beauty of the life, poetry and was fed up with government work. But when his son Mehmed II took over, there was an imminent danger of a crusade (christian coalition) to take over the Balkans. Now after this point there are two different view points as to what exactly happened. 1. Some scientists claim that Mehmed II wrote a letter to his father stating "If you are the Sultan, come and lead your armies. If I am the Sultan I hereby order you to come and lead my armies." 2. Some write that the prime minister (vezir) at the time convinced Murad to come back and take over the throne because of the imminent danger. And so Murad just came back to capital and continued his rule until his death. The same prime minister (vezir) was executed shortly after Mehmed conquered Constantinople because it is claimed that Mehmed always held a grudge against him. In case of Ottomans they just continued their rule.
[ "When a monarch dies without a clear successor, a succession crisis often results. For example, when King Charles IV of France died, the Hundred Years War erupted between Charles' cousin, Philip VI of France, and Charles' nephew, Edward III of England, to determine who would succeed Charles as the King of France. Where the line of succession is clear, it has sometimes happened that a pretender with a weak or spurious claim but military or political power usurps the throne.\n", "Upon a \"demise of the Crown\" (the death or abdication of a sovereign), the late sovereign's heir immediately and automatically succeeds, without any need for confirmation or further ceremonyhence arises the phrase \"The King is dead. Long live the King!.\" Following an appropriate period of mourning, the monarch is also crowned in the United Kingdom in an ancient ritual, but one not necessary for a sovereign to reign. After an individual ascends the throne, he or she typically continues to reign until death, being unable to unilaterally abdicate per the tenets of constitutional monarchy.\n", "The following monarchs either lost their thrones through deposition by a coup d'état, by a referendum which abolished their throne, or chose to abdicate during the 21st century. A list of surviving former monarchs appears at the end of the article.\n", "The following monarchs either lost their thrones through deposition by a coup d'état, by a referendum which abolished their throne, or chose to abdicate during the 20th century. A list of surviving former monarchs appears at the end of the article.\n", "BULLET::::- Article 52. Should the Heir die or lose his rights to the Throne, his eldest son shall succeed. Should the Heir to the Throne die or lose his rights and leave no son, the succession shall pass to the brother coming after him.\n", "A person does not become an heir before the death of the deceased, since the exact identity of the persons entitled to inherit is determined only then. Members of ruling noble or royal houses who are expected to become heirs are called heirs apparent if first in line and incapable of being displaced from inheriting by another claim; otherwise, they are heirs presumptive. There is a further concept of joint inheritance, pending renunciation by all but one, which is called coparceny.\n", "The eldest son of a king who died before that was barred from direct inheritance but possessed the right to retake the throne himself on the attainment of majority. The number of princes rebelling just prior to their 30th year may even indicate that they were \"required\" to assert their rights at that point or lose them. If the eldest son were ineligible for whatever reason, his brothers, uncles, and first and second cousins were all considered legitimate substitutes. Likewise, even when the eldest son did inherit, other descendants of his great-grandfather were considered legitimate rulers and not usurpers if they were able to wrest control away from him.\n" ]
Does muscle repair faster when asleep, rather than lying completely motionless?
Sleep involves increased secretion of growth hormone, which does stimulate increased protein synthesis in skeletal muscle. _URL_0_
[ "In addition, as a result of continuous muscular activity without proper rest time, effects such as cramping are much more frequent in sleep-deprived individuals. Extreme cases of sleep deprivation have been reported to be associated with hernias, muscle fascia tears, and other such problems commonly associated with physical overexertion.\n", "In physiology, medicine, and anatomy, muscle tone (residual muscle tension or tonus) is the continuous and passive partial contraction of the muscles, or the muscle's resistance to passive stretch during resting state. It helps to maintain posture and declines during REM sleep.\n", "Muscle fatigue may be due to precise molecular changes that occur \"in vivo\" with sustained exercise. It has been found that the ryanodine receptor present in skeletal muscle undergoes a conformational change during exercise, resulting in \"leaky\" channels that are deficient in calcium release. These \"leaky\" channels may be a contributor to muscle fatigue and decreased exercise capacity.\n", "In general, muscle relaxants are not approved by FDA for long-term use. However, rheumatologists often prescribe cyclobenzaprine nightly on a daily basis to increase stage 4 sleep. By increasing this sleep stage, patients feel more refreshed in the morning. Improving sleep is also beneficial for patients who have fibromyalgia.\n", "A phenomenon of REM sleep, muscular paralysis, occurs at an inappropriate time. This loss of tonus is caused by massive inhibition of motor neurons in the spinal cord. When this happens during waking, the victim of a cataplectic attack loses control of his or her muscles. As in REM sleep, the person continues to breathe and is able to control eye movements.\n", "Muscle relaxants do not render patients unconscious or relieve pain. Instead, they are sometimes used after a patient is rendered unconscious (induction of anesthesia) to facilitate intubation or surgery by paralyzing skeletal muscle.\n", "Intercostal muscle activity decreases in REM sleep and contribution of rib cage to respiration decreases during REM sleep. This is due to REM related supraspinal inhibition of alpha motoneuron drive and specific depression of fusimotor function. Diaphraghmatic activity correspondingly increases during REM sleep. Although paradoxical thoracoabdominal movements are not observed, the thoracic and abdominal displacements are not exactly in phase. This decrease in intercostal muscle activity is primarily responsible for hypoventilation that occurs in patients with borderline pulmonary function.\n" ]
Were Gun Control Laws in the USA originally motivated by racism/Jim Crow?
The bottom line is, yes, there was definitely a racist root to much of it. One of the concerns of the Supreme Court in the infamous Dredd Scott case was that if blacks were citizens, then it would give, horror of horrors, "to persons of the negro race ... the full liberty of speech ...; to hold public meetings upon political affairs, and to keep and carry arms wherever they went." You can imagine how well the concept of armed blacks went over down South. Fortunately, there were plenty of laws on the books prohibiting certain people from being armed, such as one to "prohibit any negro or mulatto from having fire-arms.", as addressed by the deliberation for the Civil Rights Act 1866. For more details on the problems occasioned by the 14th Amendment and the fact that the white government could no longer blatantly say that blacks couldn't have guns, check here. _URL_0_ As a result, modern gun laws came to be crafted a little more creatively. If it was unconstitutional to blatantly say "Whites can carry guns, blacks can't", then they had to find more ethereal ways around it. For example, laws on the carriage of concealed weapons could be...discretionary. California's current one is a case in point. The San Francisco Chronicle article on the matter back in 1923 (available about half-way down this PDF _URL_1_ ) is quite open about how the law was carefully crafted to try to avoid unConstitutionality issues. The target in this case was not blacks, but Asians and Latinos. The still-extant law requires that the Sheriff decide that the applicant for a firearm show 'good cause' and 'be of good moral character', a very subjective assessment which, back in the days, effectively meant 'you're white', but today the application of the law has gone more to a race-neutral position based on whether or not the Sheriff likes the idea of an armed private citizenry. Of course, not all gun control laws were racist. For example, back in the day, it was considered that carrying a firearm concealed was the act of a scourge and scoundrel, and that honest gentlemen carried them openly, this applied to all regardless of race.
[ "Jim Crow laws arose directly from a Supreme Court ruling which validated a \"states' rights\" notion that blacks and whites could be equally well served using separate but equal public facilities. With \"Plessy v. Ferguson\" (USSC 1896) the United States Supreme Court confirmed the right of state legislatures to enact discriminatory legislation. With this authority, civic organizations throughout the American South moved to restrict citizen access and limit citizens from exercising their civil rights based on the basis of their social and economic status, and on their personal history as descended from a former slave.\n", "Adam Winkler, professor of constitutional law at the UCLA School of Law, has argued that there is historical precedence to the NRA's lack of advocacy for black gun owners, noting that the NRA promoted gun control legislation in the 1920s, 1930s and 1960s with the intent to reduce gun ownership by racial minorities.\n", "These Jim Crow laws revived principles of the 1865 and 1866 Black Codes, which had previously restricted the civil rights and civil liberties of African Americans. Segregation of public (state-sponsored) schools was declared unconstitutional by the Supreme Court of the United States in 1954 in \"Brown v. Board of Education\". In some states it took many years to implement this decision. Generally, the remaining Jim Crow laws were overruled by the Civil Rights Act of 1964 and the Voting Rights Act of 1965, but years of action and court challenges have been needed to unravel the many means of institutional discrimination.\n", "During the late 19th century and well into the early 20th century, there was a growing trend among the governments of southern U.S. states, controlled by Democrats, in implementing stricter and more stringent regulations restrictions on the right of firearms ownership. This was done in order to prevent African Americans from being able to bear arms for the purpose of defending themselves against the ever-growing threat of lynchings, terrorism, and extrajudicial paramilitary violence at the hands of Democrats and white supremacists, designed to prevent black voters from electing Republicans. Mississippi was no stranger to such restrictive laws; an 1817 law even forbade an African American person from being in possession of a canine. As with the sections of the 1890 constitution that disenfranchised black voters, the laws regarding weapons ownership used ostensibly non-discriminatory wording, but were enforced by the state government in an arbitrarily and subjective discriminatory manner, as in many southern U.S. states at the time.\n", "The Jim Crow system employed \"terror as a means of social control,\" with the most organized manifestations being the Ku Klux Klan and their collaborators in local police departments. This violence played a key role in blocking the progress of the civil rights movement in the late 1950s. Some black organizations in the South began practicing armed self-defense. The first to do so openly was the Monroe, North Carolina, chapter of the NAACP led by Robert F. Williams. Williams had rebuilt the chapter after its membership was terrorized out of public life by the Klan. He did so by encouraging a new, more working-class membership to arm itself thoroughly and defend against attack. When Klan nightriders attacked the home of NAACP member Dr. Albert Perry in October 1957, Williams' militia exchanged gunfire with the stunned Klansmen, who quickly retreated. The following day, the city council held an emergency session and passed an ordinance banning KKK motorcades. One year later, Lumbee Indians in North Carolina would have a similarly successful armed stand-off with the Klan (known as the Battle of Hayes Pond) which resulted in KKK leader James W. \"Catfish\" Cole being convicted of incitement to riot.\n", "Dick Heller, the plaintiff in District of Columbia v. Heller, in an interview with radio talk show host Mark Walters, said that the GCO brief really opened his mind that the nascence of Jim Crow laws as gun control in the South. He said that when talking with the press or testifying before the city council, he will be sure to bring up that Jim Crow is alive in Washington, D.C.\n", "The Jim Crow laws were state and local laws in the United States enacted between 1876 and 1965. They mandated de jure segregation in all public facilities, with a supposedly \"separate but equal\" status for black Americans. In reality, this led to treatment and accommodations that were usually inferior to those provided for white Americans, systematizing a number of economic, educational and social disadvantages.\n" ]
What is the correllation between the name of chinese currency yuan and the dynasty?
They use the same character, 元. That's it. It's a common character, meaning "round" or "round coin". To elaborate, the Yuan dynasty was a period of Mongol rule over China, from c. 1260 (de facto, 1271 officially) to 1368 (1380 if you count the Northern Yuan dynasty when the Mongols retreated to Mongolia after Kublai Khan's death and subsequently being driven out of China. *** The Chinese Yuan, referring to the currency, 圆 formally but commonly the above character, is the basic unit of Chinese currency, much like the dollar is to many other currencies of the world. It was first minted during the Qing dynasty between 1888 and 1898 in various provinces. While the modern rendering is officially rénmínbì (人民币) or People's Currency, in daily use 元 is far more common, or kuài (块) or "piece" in certain parts of China. References: _URL_0_ _URL_1_ _URL_3_ _URL_2_ Edited for formatting.
[ "The first Chinese yuan coins had the same specification as a Spanish dollar, leading to a continuing equivalence in some respects between the names \"yuan\" and \"dollar\" in the Chinese language. Other currencies also derived from the dollar include the Japanese yen, Korean won, Philippine peso and the Malaysian ringgit.\n", "or yuan sign (¥) is a currency sign used by the Japanese yen and the Chinese yuan currencies. This monetary symbol resembles a Latin letter Y with a single or double horizontal stroke. The symbol is usually placed before the value it represents, for example \"¥50\", unlike the kanji/Chinese character, which is more commonly used in Japanese and Chinese and is written following the amount: 50円 in Japan and 50元 in China.\n", "The official title of the dynasty, Da Yuan (, \"Great Yuan\"), originates from a Chinese classic text called the \"Commentaries on the Classic of Changes (I Ching)\" whose section regarding Qián () reads \"\" (dà zai Qián Yuán), literally translating to 'Great is Qián, the Primal', with \"Qián\" being the symbol of the Heaven, and the Emperor. Therefore, Yuan was the first dynasty in China to use Da (, \"Great\") in its official title, as well as being the first dynasty to use a title that did not correspond to an ancient region in China. In 1271, Khanbaliq officially became the capital of the Yuan dynasty. In the proclamation of the dynastic name, Kublai announced the name of the new dynasty as Yuan and claimed the succession of former Chinese dynasties from the Three Sovereigns and Five Emperors to Tang dynasty.\n", "The symbol for the yuan (元) is also used in Chinese to refer to the currency units of Japan \"(yen)\" and Korea \"(won)\", and is used to translate the currency unit \"dollar\" as well as some other currencies; for example, the United States dollar is called \"Meiyuan\" () in Chinese, and the euro is called \"Ouyuan\" (). When used in English in the context of the modern foreign exchange market, the Chinese yuan (CNY) refers to the renminbi (RMB), which is the official currency used in mainland China.\n", "Historically, the name has been fast growing amongst Han Chinese, and has also been taken up by various non-Chinese ethnic groups. The surname is now held by more than 6.5 million people worldwide, and makes up 0.54% of the population of mainland China. Although growth has tapered off in the past six centuries, the Yuan name is still relatively widespread throughout China, as well as among overseas Chinese, with heaviest per capita concentrations in the Yangtze Delta region of central coastal China.\n", "A variety of currencies circulated in China during the Republic of China (ROC) era, most of which were denominated in the unit \"yuán\" (Mandarin pronunciation in IPA: ). Each was distinguished by a currency name, such as the \"fabi\" (\"legal tender\"), the \"gold yuan\", and the \"silver yuan\".\n", "The name \"yuanbao\" is the pinyin romanization of the Mandarin pronunciation of the characters for \"inaugural treasures\". Under China's Tang dynasty, coins were inscribed \"Kai tong\" (, \"Circulating Treasure of the Beginning of an Era\"), later abbreviated to yuanbao. The name was also applied to other non-coin forms of currency. Yuanbao was spelt yamboo\n" ]
Why are trees in the Sahara flat and wide?
More sunlight is a good thing for the tree, as long as it can get enough water. Other trees don't spread out so much because the competition makes it a wasted effort, but in the Savannah the chance of two trees thriving right next to each other is lower. It's easier for a single tree to reach deep enough with its roots to hit water, and then take advantage of the clear sky above.
[ "A possibility causing the variation is the Sahel, a strip of semi-arid land on the southern border of the Sahara. When rain amounts in the Sahel are higher, the volume of dust is lower. The higher rainfall could make more vegetation grow in the Sahel, leaving less sand exposed to winds to blow away.\n", "The Mediterranean \"Acacia-Argania\" dry woodlands and succulent thickets is a Mediterranean forests, woodlands, and scrub ecoregion in North Africa centered mainly on Morocco but also including northwestern Western Sahara and the eastern Canary Islands. \n", "cloud belt. Tree species with laurel-shaped leaves are predominant, forming a dense canopy up to 40 m high that can be hardly trespassed by light, which results in scant vegetation in the understory. Most of these tree species in Africa are ancient paleoendemic species of the genera Laurus, Ocotea, Persea, and Picconia, which in ancient times were widely distributed on the African and European continents.\n", "Botanically speaking, Morocco enjoys a great variety of vegetation, from lush large forests of conifer and oak trees typical of the western Mediterranean countries (Morocco, Algeria, Italy, Spain, France and Portugal), to shrubs and acacias further south. This is due to the diversity of climate and the precipitation patterns in the country.\n", "In the northern Sahel, dunes are covered with scrub grasses and spiny acacia trees. Farther south, greater rainfall permits denser vegetation. Sands begin to give way to clay. Large date palm plantations are found on the Tagant Plateau, and savanna grasses, brushwood, balsam, and spurge cover fixed dunes. Occasional baobab trees dot the flat savanna grasslands of the southern Sahel. Forest areas contain palm trees and baobabs. Vast forests of gum-bearing acacia grow in Trarza and Brakna regions. Farther south, particularly in Assâba and the northern portion of Guidimaka regions, rainfall is high enough to support forms of sedentary agriculture.\n", "BULLET::::- Desert grassland: Characterized by varying densities of low lying vegetation, grasslands zones cannot support trees due to extreme aridity. Some desert regions may support trees at base of mountains however, and thus distinct grasslands zones will not form in these areas.\n", "Vegetation is extremely sparse in the hamada landscape due to the heavily weathered soil. A native species of tree, \"Vachellia tortilis\" (known locally as \"samr\") is well adapted to the desert environment and one of the most common forms of vegetation in the country. \"Tetraena qatarensis\" and \"Lycium shawii\" also grow in this landscape.\n" ]