text
stringlengths
222
548k
id
stringlengths
47
47
dump
stringclasses
95 values
url
stringlengths
14
7.09k
file_path
stringlengths
110
155
language
stringclasses
1 value
language_score
float64
0.65
1
token_count
int64
53
113k
score
float64
2.52
5.03
int_score
int64
3
5
The coronavirus has upended our lives and the news cycle, which means many important stories have fallen through the cracks. Perhaps the most important of these is former President Trump’s decision to open up the country’s largest wildlife refuge, the ANWR (Arctic National Wildlife Refuge) for oil drilling. The decision comes after the U.S. elections on November 3rd, in which President Trump lost his bid for reelection. This is part of a long string of the president’s dismantling of environmental protections, including a limiting of abilities and proposed 31% budget cut to the EPA, revoking of Obama-era environmental policies, including the Clean Air Act, and pulling out of the Paris Accord. The controversial decision comes much to the chagrin of climate/environmental activists and organizations such as the Sierra Club and Morristown’s very own Climate Strike Bike, and a reported more than half of Americans to disagree with the decision. The drilling will upend the indigenous people’s way of life, as well as the wildlife residing in the reserve, in addition to causing an increase in pollution affecting the whole world, including Morristown, and the acceleration of climate change. Proponents, however, say the financial advantages of the drilling will outweigh these effects, and that the damage will be negligible. On such a controversial topic, there is bound to be heated debate on both sides, with the unbiased facts being lost in the fray, so let’s take an objective look at the advantages and disadvantages of the decision. The negative consequences of the ANWR drilling would be detrimental to the environment, wildlife, and indigenous people of the wildlife reserve and the world while yielding little to no results until far later. To start, the decrease in air quality would be up to 32% for nearby areas, which would cause irreparable damage to not only the people and wildlife living there but also the flora, many of which are only found in the ANWR and surrounding areas. The health effects on the region should not be taken lightly. Another impact of the drilling is the unavoidable trampling and destruction of the plants in the reserve, which are not only unique species but also food for the wildlife, which are food for the indigenous people. The hauling of supplies, resources, and oil to and from the drilling sites would disrupt and damage the ecosystem. Accidents would also let oil and other toxic contaminants seep into and poison the groundwater, which is consumed by the flora, fauna, and people of the refuge. Drilling is also known to cause an increase in the rate and severity of earthquakes and sinkholes. These local effects would be accompanied by effects felt by the whole world, as the pollution caused by the drilling would spread, even as far as the east coast*, so this decision is not something far away, but rather something that affects our own community in Morristown. In addition, oil spills that occur from the transportation of the oil would affect wide swaths of the earth, and be able to spread for a while before being contained, as the nearest coast guard station is over 1,000 miles away. The drilling and subsequent burning of the oil would contribute massively to climate change, which is “the biggest threat to the existence of humanity” according to the United Nations. All of this would be for nearly nothing in the short-term, as optimistic estimates report that oil production would not begin until late 2030, or as long as early 2034, by when oil may be obsolete as we turn toward more sustainable renewable energy sources such as solar and wind power. Overall, the factual and truthful conclusion is that the effects of the drilling are much too large to be ignored, and the decision will affect humanity worldwide for years to come. This may seem like a debate among far-away people in faraway places, but the effects of this debate will be felt by everyone, including you. This decision will not only change the standard for oil drilling on wildlife reserves, but it will also contribute massively to climate change, which won’t stay in Alaska. Air pollution triggers asthma attacks, is proven to increase cancer, and that’s not the only worry. Oil spills and other contamination events would poison waterways all the way on the east coast. In addition to this, allowing the drilling would set a new standard of allowing companies to drill on protected land, so not only would the ANWR be affected, but reserves and national parks everywhere could be opened to fracking, including our very own Great Swamp. These consequences would affect people of all age ranges, not just adults. Though the benefits would also affect our town, with a new (and much-needed) economic revitalization coming from the income produced by the drilling. People who are struggling financially right now would be given a new boost from the profits of selling the products, again affecting kids and adults alike. However, the benefits of the drilling also must not be ignored, especially the possibility of financial gain. The drilling has the possibility of making up to an estimated $296 billion in federal revenue (money for the federal government), with conservative estimates at $150 billion. This is not a small sum, and the money could be used to pay down the national debt and/or decrease taxes. This would greatly benefit the economy, especially after the expected COVID-19 pandemic’s financial crisis. While oil production would not begin for at least ten years, oil companies that would like to drill on the land would have to pay at least partially up-front, so the positive effects for taxpayers could be felt immediately, providing necessary monetary relief amid the pandemic. Another way the drilling would help the American people would be lower gas prices, keeping more money in people’s pockets. To summarize, drilling in the ANWR would help the economy and Americans in ways and to an extent that other solutions cannot. In conclusion, former President Trump’s controversial decision to drill for oil in the ANWR is backed up by strong evidence on both sides, and while the environmental and health effects should not be taken lightly, the much-needed monetary relief the drilling could provide cannot be ignored. The political fighting over the decision has drowned out the facts of the matter, and so it has become hard for Americans to have a non-biased and informed opinion on the topic. When push comes to shove, the decision affects the people, so we need to form our own opinions on the topic.
<urn:uuid:4023cbd6-aabd-471f-84c5-78758ab6623c>
CC-MAIN-2021-49
https://fmsfalconpress.org/2571/world-national-news/drilling-for-oil-in-the-anwr-why-its-good-why-its-bad-and-why-you-should-care-a-factual-view/?print=true
s3://commoncrawl/crawl-data/CC-MAIN-2021-49/segments/1637964358180.42/warc/CC-MAIN-20211127103444-20211127133444-00320.warc.gz
en
0.971561
1,314
2.90625
3
Ted Simons: A new report by ASU researchers shows that nighttime temperatures in the Phoenix area increase one to two degrees because of excess heat being released by air conditioners. Joining us now to talk about their report, ASU mathematics professor Alex Mahalov, and Francisco Salamanca, post-doctoral research scientist at ASU's school of mathemetical and statistical sciences. Good to have you both here. Are we talking about air conditioning units contributing to global warming? Alex Mahalov: Yes. Air conditioning is sort of like refrigerator. So you cool your house, but then you have waste heat, and it increases the heat island by one degrees. Ted Simons: And anyone who stood by the air conditioning unit outside or if they've been on the roof when it's on, they know warm air does come out. This influences more at night than the day? Why is that? Francisco Salamanca: because during the day the energy coming from the sun dominates, so even though you aren’t putting any heat outside, they think the temperature is more. However, at night it’s definitely stronger. Ted Simons: How did you do this research? How was it conducted? Alex Mahalov: Research was funded by National Science Foundation, and so the model is a meteorological model coupled with building chemistry models. So we take into account everything. Ted Simons: Ok, and -- But again, it was over a certain period of time, were certain areas focused on? What did you get there? Alex Mahalov: We used one period in July 2009 just to validate our model. But our research is fundamental and can be applied to Phoenix and other metropolitan areas. Ted Simons: It sounds like a 10-day period there in 2009. And again, when you went into the research, what were you specifically looking for? Francisco Salamanca: Well, we were looking for the average in air conditioning temperature in the day time. At the beginning we were not expecting this magnitude, so it was a surprise this significant affect-- Really our focus was the effect of air conditioning on the temperature. Ted Simons: It's basically like the urban heat isle in effect as well. You’ve got more air conditioning units together, more of an effect I would imagine. Francisco Salamanca: Yes. The air conditioning contributes to the heat island. So in July the head island in Phoenix it's around 10, 12 degrees. So two degrees are responsible from air conditioners. Ted Simons: As far as just doing the study in the first place, whose idea was this? Who said, let's find out? Everyone knows there is waste heat coming out of their air conditioning units. Whoever thought to say, let's see if this is affecting the atmosphere? Alex Mahalov: Francisco. Ted Simons: Oh Francisco does, all right. What made you go out there and look for this? Francisco Salamanca: Some years ago during my disease, I was looking, trying to understand the effect, so in the past I applied it to Madrid, so why not in Phoenix? Ted Simons: Yeah, and you said were you surprised by the findings. Francisco Salamanca: Yes, I was waiting for a smaller effect. Phoenix is a big city, temperature is very high in the summer, so the effect is stronger. Ted Simons: And you also now look at electricity use as well and how did that factor in here? Alex Mahalov: Yes, actually one or two degree is a lot, one or two degrees matters. I brought with me my air conditioning bill, and I can compare June and July, so in July average temperature was one degree higher and my bill is $30 more. So now if we have 1 million households in the state of Arizona, multiplied by 30, $30 million per month. It wastes heat. Ted Simons: It wastse heat. But what -- It seems like it's almost a closed loop here in the sense that it's getting warmer, I need to cool down. I'm cooling down, I'm making it warmer. It just goes round and round. Alex Mahalov: Exactly. It’s a negative feedback loop, but can actually we can turn it around. Ted Simons: How can we turn it around? Alex Mahalov: It's called waste heat. When you have waste, what you do? You recycle. One solution is actually, instead use this, capture this heat from air conditioning, and run it through water heater. So if you do it, so you get hot water, and then you actually reduce your electric bill. So it's a win-win situation. First you don’t contribute to the heat island, and then you reduce your water heater bill. Ted Simons: Is that the kind of solution you're looking for when you start looking at this kind of thing? Francisco Salamanca: No, I was not thinking in this solution, just I was thinking to quantify the effect of the air conditioning, and that’s it. Ted Simons: So nothing more, but it seems to make sense doesn’t it? Francisco Salamanca: Yeah, it made sense, the important point we can do is if our indoor temperature is working at 75, just increase to 80, for example, and we save energy, we reduce the impact, and our bill will be lower. Ted Simons: Yeah, so are there other things, besides perhaps water heaters, are there other things this excess waste heat can be used for? Alex Mahalov: Well cooking, and maybe some others – well think about it. Ted Simons: You have to find a way to capture that and bring it back into the house. Alex Mahalov: Right. But it's probably not very difficult. But if you think about the magnitude of this problem, the magnitude of the solution, right as I said it’s like, 300-250 million dollars. You know, you’re moving this waste heat for good purposes and plus on top of that, say on your electric bill on water heater right. So it's a very significant. Ted Simons: If it's that significant, are you working with anyone now to say, hey, I got something significant here. Let's work on -- Let's find a way to capture this stuff and use it again. Alex Mahalov: Well we are working, but we’d like to work more, hopefully after your Arizona Horizon we'll have some phone calls and we'll redouble our efforts. Ted Simons: We'll do copyright discussion after the show there. But yeah I understand, it makes sense, and again I wonder if that closed loop idea was something that you thought about as well when you started on this research, when you embarked on this research. Or was it basically a research project saying, if X, then Y? Francisco Salamanca: People underestimated nighttime temperatures in the cities, so the study is important for two points. One, we can quantify if the effect of the air conditioner. Another point is we are improving the weather forecasting models. When we compare the temperature you can consider the heat, that is much better than if you don’t consider this heat. Ted Simons: Last question, what do we take from this study? Alex Mahalov: It's to find sustainable solution. We identify problem, and then Arizona is growing, and so this kind of problems we need to transform this into solutions. Positive feedback loop, always positive feedback loop. I think the future is very bright, because we could use energy for good purposes. Ted Simons: All right. Well, it's an interesting study. I think anyone who stood next to one of those blowers would think that's warm temperature, and you're saying it does make a difference. Thank you for your research and thank you for joining us on the show. Alex Mahalov: Thank you. Ted Simons: Tomorrow on "Arizona Horizon," we’ll hear from the attorney representing a former Tom Horne aide who alleges that Horne broke campaign laws. Ted Simons: And find out about a survey on young people's attitudes toward voting. That's Tuesday evening at 5:30 and 10:00 on "Arizona Horizon." Reminder, www.azpbs.org, that's where you can find us on the web, you can see past shows and you can see what we have planned in the future, azpbs.org/horizon. That is it for now, I’m Ted Simons. Thank you so much for joining us and you have a great night.
<urn:uuid:a723f98e-47bb-44a2-b192-fb04e3ef6bc1>
CC-MAIN-2015-40
http://www.azpbs.org/arizonahorizon/detailvid.php?id=14966
s3://commoncrawl/crawl-data/CC-MAIN-2015-40/segments/1443736676977.21/warc/CC-MAIN-20151001215756-00036-ip-10-137-6-227.ec2.internal.warc.gz
en
0.965244
1,828
2.765625
3
Evidence for quaternary seismic activity of the La Alberca-Teremendo fault, Morelia region, Trans-Mexican Volcanic Belt The La Alberca-Teremendo fault is a 26 km-long, complex fault composed of an en échelon array of short crustal fault segments, belonging to the Morelia-Acambay fault system. This fault system shows parallel scarps with morphological evidence of recent activity such as drainage alteration, maximum throws of 50 m and minimum throws of 1.4 m that displace the recent soils. The fault acted as a conduit for the formation of the La Alberca de Guadalupe maar (23000 to 21000 years ago) and displaced afterwards its phreatomagmatic sequences. The paleoseismic analysis indicates that the La Alberca-Teremendo fault moved three times in the past 23000 years (age of the maar); this activity caused an average vertical displacement of 87 cm, and might have generated earthquakes with magnitudes Mw between 6.6 and 7, as well as volcano-tectonic earthquakes with magnitudes Mw between 4 and 5.5. The displacements were identified on the fault through the superposition of soils differentiated by a disconformity and an anomalous increase in the percentage of clay and organic matter. The La Alberca-Teremendo fault has dominant dip slip with a minor left-lateral component, a slip rate of 0.114 mm/year, and an average recurrence interval of 7726 ± 68 years. According to scaling relations that use the surface rupture length, if we assume that the La Alberca-Teremendo fault moves tectonically, it could generate earthquakes with maximum magnitudes of Mw between 6.7 and 7.3, however because of the active volcanic processes in the area, we could expect moderate volcano-tectonic earthquakes (Mw 4–5.5) rather than catastrophic ones.
<urn:uuid:30551f85-4d49-421e-ba8e-32789030099d>
CC-MAIN-2019-39
http://rmcg.geociencias.unam.mx/ojs/index.php/rmcg/article/view/1092
s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514573284.48/warc/CC-MAIN-20190918110932-20190918132932-00354.warc.gz
en
0.896621
414
3.21875
3
Our heart is controlled by a very impressive system. At the top of the heart muscle, a signal is produced that travels down and because of it, the heart pulsates and constantly pumps blood. With a healthy heart, 60 to 100 signals are generated per minute. For some reason, these numbers can be disturbed and there are abnormalities in the heartbeat, which in one word is called arrhythmia. Although arrhythmia in most people is not dangerous and does not require therapy, the body after a series of warnings, if nothing is done, can be seriously endangered. Natural treatment of arrhythmia, with its unique methods, can improve the patient’s condition or act as prevention in maintaining the rhythm of the heart muscle. Depending on the problem, this type of treatment primarily changes your lifestyle, and also includes the use of numerous effective medicinal herbs. In the text that follows, we will deal with cardiac arrhythmia and treatment and present the 5 best natural remedies. Here are some quick links to what we’ll cover in this article. If you’re in hurry, or for any other reason, feel free to use the quick links to jump straight to the section you want to read: - What is cardiac arrhythmia? - The main causes of cardiac arrhythmia - The main symptoms - Cardiac arrhythmia and treatment – 5 all-natural foods - Prevention is key - Don’t panic if your heart skips a beat sometimes If we translate literally, the word arrhythmia means “without rhythm”. This health disorder occurs when electrical impulses, which are practically the main drivers of the heartbeat, do not work properly, which results in the heart muscle losing its normal rhythm. Then the person has a feeling of trembling in the chest or that the heart is beating very hard. When the heart muscle does not work properly, the whole organism is endangered, and damage to the heart itself can occur. This disorder usually affects people over the age of 50, but it is also becoming more common in young people due to poor immunity. Depending on the rhythm of the heartbeat, we divide the arrhythmia into several groups: - Bradycardia is a slowed heart rate. - Tachycardia is a rapid heartbeat. - Fibrillation or flatulence are irregular heartbeats. In the case of turbulent blood flow in the atria, blood clotting occurs, so there is a risk of embolism. This means that the clot (thrombus) can disintegrate and together with the blood continue its way through the body all the way to the place where due to its size it cannot pass, so it stops the entire circulation. With the normal functioning of the cardiovascular system, the heart can occasionally skip one or two beats, but quickly only return to its previous state. With a heart arrhythmia, the heartbeat itself cannot be normalized, which can result in a fatal heart attack. The main causes of heart failure are excessive mental work, excessive physical activity, and recovery of the body after a severe illness. Many medical conditions and diseases can lead to arrhythmias, including changes in heart structure, blocked arteries in the heart, high blood pressure, diabetes, thyroid disorders, stress, drug use, smoking, excessive coffee and alcohol consumption, certain dietary and herbal supplements treatments, as well as certain prescription drugs. Most people are not even aware that they have problems with arrhythmia, associating its symptoms mainly with chronic fatigue. Only when more serious complications occur does it turn out that the problem is an irregular heartbeat. In addition to the patient noticing that something is wrong with the heartbeat rhythm, signs such as mild headache, chest pain, possible dizziness and fainting, shortness of breath, or cold sweats may also occur. 1. Hawthorn is one of the safest and most effective herbs when it comes to treating diseases of the cardiovascular system, and that also applies to arrhythmia problems. Medicinal parts of hawthorn are dark red berries, small white flowers, and leaves. Hawthorn strengthens the heart muscle and improves the flow of nutrients and oxygen through the bloodstream to every part of the body. It dilates the arteries that carry blood to the heart so that it also receives the necessary nutrients for normal work. Hawthorn therapy should last at least three months to feel the effects. 2. Angelica is also known as angel root, and it got its name in the Middle Ages when it was used to treat vicious infectious diseases such as cholera and plague. This medicinal plant contains as many as 14 compounds that quickly alleviate all the symptoms that accompany arrhythmia. Angelica also treats other cardiovascular diseases and disorders such as angina pectoris. Successfully lowers high blood pressure. It purifies the blood and improves circulation while expelling toxins through the patient’s skin. Angelica roots, leaves, and seeds are used for healing. 3. Nettle is full of vitamins and minerals that protect the heart, and also regulate blood pumping and heart rate. The high concentration of vitamin C and iron is especially emphasized, due to which the production of red blood cells increases, and thus the transport of other nutrients and oxygen through the bloodstream is facilitated. Vitamin C accelerates the absorption of iron, and iron is a crucial component of hemoglobin. The effect of relaxing blood vessels also contributes to healthy circulation. Flavonoids in nettle, such as quercetin and kaempferol, lower blood pressure and prevent the development of dangerous cardiovascular diseases. 4. Hops is a well-known ingredient in beer, but it is also used in the treatment of disorders of the cardiovascular system. Hops can treat arrhythmia and improve circulation. Cone hop flowers are used to treat abnormal heart rhythm. This plant is especially important for arrhythmia caused by stress. It also reduces the possibility of blood clots. It increases the level of good cholesterol and reduces bad cholesterol. Hops contain beta-glucans, soluble fibers that also have a beneficial effect on the heart, but also blood vessels. 5. Asparagus is used in the diet as a vegetable, and many are not aware of how valuable asparagus is as a natural remedy. It protects your heart and blood vessels, and among other things, it successfully solves the problem of arrhythmia. It can be taken in unlimited quantities, primarily due to its low caloric value and the absence of side effects on the patient’s body. Regular use of asparagus in treatment and nutrition will help you get rid of harmful substances from your bloodstream. It cleanses the blood and enables unhindered transport of nutrients and oxygen to every tissue in the body. * Take care of your diet. Eat more fresh fruits and vegetables, foods that contain omega-3 fatty acids (fish, nuts), and foods rich in calcium (milk and dairy products). * Avoid fatty foods and stop eating nicotine, alcohol, coffee, black tea, green tea, and energy drinks. * It is recommended to set aside about 30 minutes a day for any physical activity, but be careful not to get too tired and not to exceed your own limits. * Avoid stressful situations. Stress can suddenly raise blood pressure levels and increase the tendency to form blood clots. * Relax and think positively to regulate the health of the cardiovascular system. * Check your blood pressure and blood cholesterol levels regularly. According to your doctor’s recommendations, measure your blood pressure periodically, and it would be great if you could get a home meter. * You will also receive instructions from the laboratory in order to detect a problem with cholesterol in time through a sample of your blood. You may occasionally feel that your heart has “skipped one beat” or “one beat too many”, but that’s not a bad thing. Arrhythmia attacks may not be severe, but they are usually very unpleasant. Pay attention if they start to recur and go to the doctor on time for an examination. He will determine if your condition needs to be treated in any way. When someone has an arrhythmia, it does not mean that they need immediate treatment. Therapy is performed only if the arrhythmia increases the risk of more severe types of arrhythmias or to prevent potential complications. Try the suggested natural remedies. We are sure that you will soon notice how your heart rate is slowly normalizing. We hope this post helped you learn what is cardiac arrhythmia as well as its treatment and the best and completely natural remedies. We think that it will be very interesting and important for you to look at our texts about which number is OK for normal heartbeats per minute as well as text about what happens when you have a heart attack. And if you are looking for more ways to improve your heart health, you should check our other great Majota Health Care Blog posts. Also, if you have any questions or feel that there is some other important symptom, the remedy, or treatment for a cardiac arrhythmia we have not listed, feel free to leave a comment, it will mean a lot to us and we will be grateful for it.
<urn:uuid:50727d31-4783-466a-9cc8-018d62fd2da9>
CC-MAIN-2022-21
https://majotahealthcare.com/cardiac-arrhythmia-and-treatment-the-best-natural-remedies/
s3://commoncrawl/crawl-data/CC-MAIN-2022-21/segments/1652662573189.78/warc/CC-MAIN-20220524173011-20220524203011-00078.warc.gz
en
0.935185
1,943
3.125
3
Named for the turquoise iridescent sheen of its leaves, the peacock fern (Selaginella uncinata) -- like the peacock -- owes its color to the refraction of light rather than to pigments. The plant is not actually a fern, but a type of club moss, also known as rainbow moss. Peacock fern can be grown as a groundcover in U.S. Department of Agriculture plant hardiness zones 6 to 10 or as a houseplant. Choose a location that is entirely shady or only receives morning light. Dig in a couple inches of compost if the soil is not already well-drained and rich in humus. Add a little peat moss as well if the ground is alkaline since peacock fern prefers acidic soil. Plant your peacock ferns a foot apart, measuring the roots first. Water the ferns to the depth of their roots, checking how far the moisture has penetrated by inserting a soil probe, which will slip smoothly through damp soil, but become more resistant when it hits dry ground. Continue watering until the probe verifies that the moisture has gone as deep as the roots do. Let the soil dry about halfway to the surface, as measured by the probe, before watering the ferns again. Never allow their roots to dry out completely. Fertilize the ferns in early spring with a high-nitrogen organic fertilizer, using the amount indicated for foliage plants. Trim off all dead growth from the winter. Grow as Houseplant Plant peacock fern in the soil beneath taller houseplants that tend to get leggy and need their "ankles" covered. Make sure that the fern will remain shaded for most of the day. Place the peacock fern on an east or west windowsill, where it will receive only morning or late afternoon sun. Water the plant until water begins to seep out of the pot's drainage hole. Discard the liquid in the pot's saucer. Refrain from watering the fern again until a forefinger pressed against the surface of its soil comes up clean. Don't let the soil in the pot entirely dry out. Fertilize the peacock fern once a week from spring to autumn with liquid plant food, but use only half the recommended dose. Keep it cool, if possible, over the winter -- with temperatures no higher than 65 F. Find a shady spot for it outdoors during the summer. Things You Will Need - Peat moss - Watering can - Soil probe - High-nitrogen organic fertilizer - Pruning shears - Liquid plant food - Use peacock fern in a hanging basket under a porch roof or to cover up the bare soil in outdoor containers of shade-loving plants, such as alocasias and begonias. If the container is exposed to sunlight, make sure the other plants will shade the fern during the brightest hours. - Peacock fern will root as it creeps along the ground, so it can be divided frequently and replanted without any harm to the original plant. - Peacock fern loses its turquoise coloring -- and can even burn -- if it receives too much direct sunlight. Its foliage will often turn reddish over the winter. - Groundcovers for the South; Marie Harrison - The Unexpected Houseplant; Tovah Martin - Herbaceous Perennial Plants; Allan M. Armitage - PlantoftheWeek.org: Selaginella uncinata - Peacock Spikemoss - Southern Gardening: An Environmentally Sensitive Approach; Marie Harrison - University of Arkansas: Plant of the Week: Peacock Clubmoss - AskNature: Leaves Change Colors under Different Lighting: Selaginella ferns - University of Arizona: Watering Plants
<urn:uuid:7eb304d9-7708-4b42-ad43-cde643a3ad21>
CC-MAIN-2016-18
http://homeguides.sfgate.com/grow-peacock-fern-65006.html
s3://commoncrawl/crawl-data/CC-MAIN-2016-18/segments/1461861735203.5/warc/CC-MAIN-20160428164215-00092-ip-10-239-7-51.ec2.internal.warc.gz
en
0.915914
814
2.65625
3
During the first trimester, the fetus undergoes significant development and you will undergo profound emotional and physical changes. During pregnancy first trimester as your body undergoes significant changes, it is very important to eat well in order to ensure the safety of you and your baby. Here are the 15 best foods to include in your diet during pregnancy first trimester: Spinach is rich in folic acid content which is very much essential for the development of baby’s neural system. Insufficient folate can cause serious birth defects such as cleft lip, cleft palate and spina bifida. Spinach is also a very rich source of beta carotene which the body easily converts to vitamin A. It is a good source of protein and calcium. Also the contents of yogurt help to reduce the risk of yeast infection, which is very common during pregnancy. Lentils are rich in protein and is very important for each stage of pregnancy. Sufficient protein consumption ensures proper growth of baby’s muscles and tissues. They are great source of folate and also rich in iron. Also they are rich in fiber which helps to prevent constipation. They are rich in fiber than any other fruit and vegetable and contain good amounts of calcium, potassium and iron. 11. Orange juice Oranges are not only rich in vitamin C and folic acid but also they are good sources of potassium. Proper consumption of orange juice can help to lower blood pressure during pregnancy. This veggie is known for being a great source of calcium and also it contains vitamin B6, vitamin C and folate. They are good source of protein, vitamin E, iron, and heart-healthy unsaturated fats. All these nutrients will support a healthy pregnancy. Consuming low fat cottage cheese can give good amounts of proteins and calcium, which are very essential for healthy pregnancy. Also, it is a good idea to avoid unpasteurized cheese. Not only eggs are good source of protein, but also they are rich in calcium and vitamin D, which are essential for proper bone formation in your baby. It is one of the safe fish to eat during pregnancy. It is high in vitamin D and calcium. Salmon fish is a great source of DHA, a type of Omega-3 fatty acid, which is the primary source for brain development. They are good foods for brain and they boost memory. Walnuts are rich in omega-3 and omega-6 fatty acids and also antioxidants. It is rich in vitamin D of which most pregnant women are deficient. Also it is a good source of folic acid, which is essential for neural development in your baby. It gives folic acid for pregnant woman to prevent birth defects. Not only they are rich in folic acid but also they are good sources of fiber, protein, vitamin A, vitamin C, calcium, potassium and magnesium. It is rich in iron, which forms red blood cells and allows your body gets sufficient oxygen levels. But pregnant women should ensure to eat organic and hormone free meat. 1. Collard greens They are great source of iron which most pregnant women are deficient.
<urn:uuid:35956ffc-9b11-41b9-897d-a143210163aa>
CC-MAIN-2019-26
https://www.pregnancyweekbyweekcalendar.com/pregnancy-diet/foods-to-eat-during-pregnancy-first-trimester.html
s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560628000175.78/warc/CC-MAIN-20190626053719-20190626075719-00388.warc.gz
en
0.947845
646
2.65625
3
WASHINGTON, September 4, 2013 – A review of health systems around the world found it is easier to get adequate medical care in Libya, Bosnia, and Romania than it is in the United States, according to a report released by World Vision on Wednesday. This gap between the “health rich” and “health poor” is contributing to the deaths of thousands of children every day. The Killer Gap: A Global Index of Health Inequality for Children (PDF) assesses 176 countries around the world according to the size of the gap between those who have access to good health care and those who don’t. The United States is ranked a surprising 46th on the global index, ranking below lower-income countries like Libya (21st), Bosnia (36th) and Romania (25th). “The fact that the U.S. is ranked lower than many lower-income countries on the index proves that a country’s wealth does not guarantee its people access to good health care and quality health,” said Lisa O’Shea, World Vision’s Campaign Director for Child Health Now. “It’s a horrifying reality that in today’s world, when we have the knowledge, resources and tools to provide everyone with quality maternal, newborn and child health, we still fall so short.” The following health outcome indicators were used to rank countries in the report: Health Inequality-adjusted Human Development Index representing a national average of human development achievements in three areas: health (life expectancy), education and income. Financing for health: Out-of-pocket expenditure as a percentage of total health expenditure Health outcome: Adolescent Fertility Rate Coverage of health services: Density of physicians and density of nursing and midwifery staff The ten countries with the smallest gaps, according to World Vision’s index, are (in order from smallest to widest gap) France, Denmark, Norway, Luxembourg, Finland, Germany, Sweden, Slovenia, Cuba and Switzerland. The ten countries with the largest gaps are Chad, Sierra Leone, Guinea, Mali, Equatorial Guinea, Niger, the Democratic Republic of Congo, Afghanistan, Cameroon and Cote d’Ivoire. Seven of the ten countries with the greatest health gaps are among the poorest countries in the world, but Equatorial Guinea, Cote d’Ivoire and Cameroon are middle income. “Over the past 20 years, we’ve made a lot of progress – the number of children under the age of five dying every year has fallen dramatically. But it’s still too high – 19,000 every day – and this report looks at one reason why this is,” said O’Shea. “In achieving this and tackling global poverty and poor health, governments and organizations have reached those who are easiest to get to, but in many cases this has meant a devastating increase in the gap between the health rich and poor, with the most vulnerable children bearing the brunt.” With two years until the deadline for the Millennium Development Goals, World Vision is using the report to urge governments to finish the job and leaders to take every step possible to close the health gap in their countries and to achieve Millennium Development Goals 4 and 5, tackling child and maternal health. - END - NOTES TO EDITORS: For pictures, footage, infographics or interviews, please contact Holly Frew at +1.770.842.6188.
<urn:uuid:164a287b-7e48-4022-b202-b1f0beadaf2e>
CC-MAIN-2015-18
http://www.worldvision.org/press-release/new-study-finds-access-health-care-us-worse-libya-bosnia-and-romania-1
s3://commoncrawl/crawl-data/CC-MAIN-2015-18/segments/1429246658116.80/warc/CC-MAIN-20150417045738-00019-ip-10-235-10-82.ec2.internal.warc.gz
en
0.934886
722
2.875
3
- Express Letter - Open Access A new miniaturized magnetometer system for long-term distributed observation on the seafloor Earth, Planets and Space volume 70, Article number: 111 (2018) We have developed a new magnetometer system specialized to multipoint and long-term observations on the seafloor to promote marine or ocean-bottom experiments mainly for the electromagnetic sounding of the Earth’s interior. In situ magnetic field observation on the seafloor is an essential geophysical technique to investigate structures of the oceanic crust and the upper mantle, many of which are still frontier as to observational evidences. The in situ and long-term observations require long-term-operable and small-size magnetometer systems, which are placed on the seafloor over a year in pressure-resistant cases without external power supply and communication. We have designed and developed a new electric circuit board of small size and lower power consumption for the magnetometer system. Our new magnetometer system, what we call “DOKODEMO MAG,” is suitable to be installed in a pressure-resistant cylinder of 36 mm diameter and can operate independently over 2 years with a smaller amount of batteries than the conventional system because its power consumption was saved to ~ 33 mW. This magnetometer system is capable of observing orthogonal three-axis magnetic fields continuously with sampling frequency of 5 Hz at maximum and an accuracy of ~ 0.1 nT. The system also records tilt and temperature of the system and voltage of the batteries. Prototype models of this magnetometer system were tested for in situ operation for 5 months on the seafloor around the Kikai caldera in the south of the Kyusyu Island, SW Japan. The results of the test showed sufficient performance of our new magnetometer system and its potential of future usage for every type of marine or ocean-bottom operations. Geophysical observations are fundamental tools to probe the internal structure of the Earth, and the increase in the number of observation sites would result in higher resolution of the probing. Examples at stationary observation sites are GNSS geodetic observations in GEONET Japan, seismic observations in Hi-net Japan, and multi-geophysical observations in US EarthScope, which have performed important roles to understand crustal deformation, seismic and volcanic activities, and internal structure in the Earth (e.g., Ozawa et al. 2011; Ito et al. 2007; Schmandt and Humphreys 2010). Land-based multipoint electromagnetic investigations have been enabled by the development of simplified and low-cost instruments, and they have also revealed detailed electrical resistivity distributions than ever (e.g., Ichiki et al. 2015; Yang et al. 2015). All these previous studies indicate that in situ observations with the increasing number of geophysical observation sites are a powerful approach to reveal geodynamics. Multipoint electromagnetic investigations on the seafloor at temporal observation sites also have a high potential to probe crustal and upper mantle structure. There were a few experiments with relatively large number of observation sites across the East Pacific Rise (Baba et al. 2006; Key et al. 2013), the Mariana subduction system (Matsuno et al. 2010), and the Cocos Plate subduction zone (Naif et al. 2013), which provide critical geophysical constraints for understanding their geodynamics beneath the seafloor. Nevertheless, the number of experiments on the seafloor is limited comparing to those on land because of high cost for multipoint seafloor surveys. In fact, a large number of observation units are required for dense three-dimensional observations, and these units usually become large in size because the system requires waterproof containers and deployment/retrieval mechanisms, which results in occupations on research vessels and high costs. Such properties of the seafloor instruments crucially hamper marine observations at multipoint and in large scale. One important way to enhance opportunities for the seafloor multipoint observations is miniaturization of the observation systems. In this paper, we introduce a magnetometer system that was newly designed and developed with specializing for multipoint seafloor observations. The main concept of the system design is simplification and miniaturization. The new magnetometer system is suitable to be installed in a small submarine package. This small system is capable of operating in the isolated seafloor environment for over 2 years by low power consumption, which allows us to attach it to other seafloor instruments such as an ocean-bottom seismometer for example. We also introduce the first results of an in situ operational test on the seafloor around the Kikai caldera, Japan, by using prototype models of the new magnetometer system. These results showed desired performance and sufficient potential of the system for future multipoint simultaneous observations. System requirements for observing temporal geomagnetic field variations on seafloor The long-term distributed geomagnetic observations on the seafloor require four main characteristics for the individual magnetometer systems: high measurement accuracy, small size, easy handling, and low cost. We primarily targeted on miniaturization in our developments of the new magnetometer system with keeping its measurement quality as conventional fluxgate-type magnetometers. The magnetometer system has to be enclosed in water-and-pressure-tight spheres or cylinders of non-magnetic materials for the seafloor measurements. A positive buoyancy and a detachable weight are also required for recovery. The whole system is deployed from vessels to be placed on the seafloor, and it is recovered at the end of an observation period onto the vessel using a self-pop-up instrument by releasing the weight to accent it with the positive buoyancy. When the size of the magnetometer system becomes smaller, the pressure container and the positive buoyancy should become smaller, which leads the whole system to be easier to handle and also to reduce costs because only smaller pressure container and less positive buoyancy are required. The miniaturization of the magnetometer system can be achieved by downsizing the magnetometer system itself and also by saving power consumption in the electrical circuit. Since the magnetometer systems are powered by batteries encapsulated together in the containers, saving power consumption directly reduces the system size because the entire volume of batteries occupies a large part in the pressure container. We assumed a titanium pressure cylinder of inner diameter 36 mm for a standard model of our magnetometer system as the goal of the miniaturization, because it is the minimum size to contain the commercially available high-capacity DD-size lithium battery, which has been commonly used in conventional seafloor systems. A currently available pressure cylinder with its inner diameter of 55 mm was assumed for a prototype model that was used for operational tests. The new magnetometer system measures triaxial geomagnetic field strength, and temperature and tilt of the magnetometer sensor. Performances required for the geomagnetic field measurements are derived from the strength of Earth’s main field as background, and fluctuation range and frequency of the local geomagnetic variation (e.g., Jacobs 1987). These are roughly defined as dynamic range of ± 60,000 nT, accuracy of ~ 0.1 nT, and sampling frequency of > 2 Hz. The sampling frequency of > 2 Hz is somewhat over specification for natural source marine electromagnetic sounding methods (e.g., Baba et al. 2017), but we chose it for possible applications of electromagnetic sounding on land and of future control source marine electromagnetic sounding, and for possible observation of tsunami induced magnetic field (e.g., Ichihara et al. 2013). Temperature measurements are also necessary to correct temperature-dependent drifts of fluxgate magnetometer outputs. We estimated that temperature accuracy of better than 1 °C is sufficient for our magnetometer system, while it depends on types of the magnetometer sensor. A clinometer in the system should have a resolution of better than 1° for three axes, which allows us to transform coordinate system of the geomagnetic field data from the magnetometer sensor coordinate system to the earth coordinate system. We note that this accuracy of the clinometer is not capable of correcting attitude changes of the magnetometer sensor due to perturbation by submarine water flow and crustal movements, but its data can be used for detection of such attitude changes. We assumed that 2 years of operation time is long enough for the magnetometer system, because the temporal geomagnetic field variation should be continuously observed for over 1 year for electrical resistivity imaging down to deeper depths like the bottom of the upper mantle with higher accuracy. Operable duration of the magnetometer system is constrained by a lifetime of batteries and a memory capacity for data recording. The battery voltage should be monitored and recorded constantly as the battery voltage slowly decreases with power consumption during the long-term operation. The voltage data are useful after the recovery to investigate health condition of the magnetometer system during the operation. System configuration, functions, and specifications We have designed and manufactured two types of new electrical circuit boards, the prototype model and the standard model: The prototype model was used for an in situ operation test on the seafloor, and the standard model finally satisfied all the requirements mentioned in the previous sections. In this section, we describe configuration, functions, and specifications of the new magnetometer system. The instrumental concept and some parts of detailed electrical designs of this system were derived from a space instrument, the deployable camera system (DCAM3) onboard asteroid probe Hayabusa2 (Ogawa et al. 2017; Watanabe et al. 2017) because they share some system requirements and functions in common, such as miniaturization and low power consumption techniques. A functional block diagram of our new magnetometer system is shown in Fig. 1a, and a setup of the magnetometer system is shown in Fig. 1b, c. The system consists of an internal part and an external part of the pressure cylinder. The control board in the internal part, which is the main development object in this study, has the following major functions: (1) collecting measurement signals and storing these data in the memory, (2) power control, and (3) clock control. The external part consists of an external board and a GPS board, and it allows us to set up the magnetometer system and check its status before deployment and after recovery via a water-resistant connector. Since the connection between the external part and the internal part can be achieved without opening the pressure cylinder, we can easily handle the magnetometer system even on swaying vessels. Furthermore, the external part is available for maintenance and test run of the control board by connecting directly to the control board in laboratory. The main controller collects measurement signals of the magnetometer sensor, the clinometer sensor, the temperature gauge, and the battery voltmeter. We chose three-axis fluxgate sensor, Bartington Mag648, for the magnetometer sensor because this is capable of satisfying the above-mentioned requirements: dynamic range ± 60,000 nT, accuracy better than 0.1 nT, power consumption ~ 15 mW, and the small size of 29 × 23 × 66 mm (packaged model). The 24-bit A/D converters in front of the CPU receive signals from the magnetometer sensor that is originally generated as ± 3 V for each axis magnetic field and convert the analog signals to digital. One issue in the magnetic field sampling was a selection of the voltage reference IC for these A/D converters. The choice of the IC affects performance of the A/D converters and the total power consumption by several mW. As a result of a trade-off study for several types of ICs, we selected two types of them, MAX6126 and LT1461, and manufactured three prototype boards as one board for MAX6126 and other two boards for LT1461. After performance tests of the prototype models, MAX6126 was finally selected for the standard model. The sampling frequency of the magnetometer is 5 Hz at maximum, that of the clinometer is every hour, and those of the other sensors are every 60 s by default. These sampling frequencies are also configurable by presets for target observations. The acquired data are intermittently stored in the microSD card on the control board, which has sufficient capacity (> 5 GByte) for 2-year continuous observations. Ancillary data for flags that indicate whether the control board is just sampling tilt, temperature, and battery voltage, and whether the board is just writing data in the microSD card. We noticed that these flags are essential because these actions often make noises in the magnetic field samplings. The power control includes the following three functions: (1) receiving electrical power from batteries or an external power supply and providing power to the whole system, (2) a switch for the battery power line, and (3) monitoring the battery voltage. While the primary batteries work for the main electrical power source during observations, the power also can be supplied by an external power source connected to the control board. The latching relay on the battery power line can be switched by a voltage signal from the external board to start using the batteries just before deployment of the magnetometer system from the vessel. While battery voltage gradually decreases during seafloor observations, the controller board monitors the battery voltage and quits own operations with appropriate shutdown actions at the time of that the battery voltage reaches a configurable lower limit. The time control is an important issue for observations. In particular, the sampling timing of the magnetic field is strictly managed with less than 0.1 ms error. The three A/D converters for three axes sample analog signals from the magnetometer at the same timing in parallel and hold them in registers until they are processed. The CPU keeps internal clock time by signals from the crystal oscillator counter, and it controls sampling timing for the sensors. The internal clock time is corrected by synchronizing with 1 PPS signals from the GPS board with its accuracy of less than 0.1 ms, and the time difference before and after the correction is also recorded in the memory. The corrections are conducted before and after every seafloor observation to estimate and correct clock drifts in the entire operation period. The crystal oscillator counter has an additional button battery on the control board to keep the counter survived in case the primary battery power is lost during operations. Even if it would happen, the CPU could automatically recover the internal clock time using the surviving counter when power is supplied again. Our new magnetometer system was assembled through integrating the control board, the magnetometer sensor, and the clinometer onto an aluminum plate, and this system was installed into the pressure cylinder with primary batteries (Fig. 1b). The envelope size of the system except for batteries is the followings; the prototype model is 33 × 35 × 293 mm for a cylinder of 55 mm inner diameter, and the standard model is Φ35 × 180 mm for a cylinder 36 mm inner diameter, respectively. The standard model has the smaller volume because its control board has simply a two-tiered structure with a cable binding them. Typical average of total power consumption of the prototype model was measured as ~ 36 mW (~ 320 Wh/year) including power for the sensors. Power consumption of the standard model is slightly lower at low voltage input, 33 mW at 3.8 V (~ 290 Wh/year). These low power consumption characteristics allow the magnetometer system to operate for 1 year by three cells of DD-size (33.5 mm diameter × 111.5 mm) lithium battery. The entire volume of the system including batteries is sufficiently small to be installed in small pressure cases that are readily deployed at multiple points in the ocean. The system specifications are summarized in Table 1. Test operation of new magnetometer system We conducted the first in situ operation test of the new magnetometer systems by using the prototype models to measure geomagnetic field variation for 5 months on the seafloor. This operation test was performed in the Kikai caldera, 50 km south from Kyusyu Island in Japan, as a part of a large exploration project (Tatsumi et al. 2018). The three prototype models were attached to one ocean-bottom electromagnetometer (OBEM) at observation site OBM1 and to two ocean-bottom seismometers (OBSs, in Fig. 1c) at sites OBM2 and OBM3. The locations of each observation site are shown on a map in Fig. 2. The water depths at sites OBM1, OBM2, and OBM3 are 390, 450, and 310 m, respectively. The OBEM at site OBM1 also included a conventional magnetometer (Tierra Tecnica Corp.) for comparison of data sets; this magnetometer has been widely used among the Japanese marine electromagnetic (EM) research group in many projects including international projects (e.g., Baba et al. 2017; Matsuno et al. 2010). The prototype models were installed in commercially available pressure cylinders of inner diameter 55 and 70 mm with two primary battery cells of DD-size lithium battery. The geomagnetic field observations were conducted for 5 months from October 17, 2016 (deployed), to March 3, 2017 (recovered). The three prototype models of the magnetometer system successfully measured the geomagnetic fields for the entire observational period. Their measured magnetic field variations of the geomagnetic north and its perpendicular components (H, D, and Z components) for 20 days are shown in Fig. 3a. The triaxial geomagnetic field data measured by each magnetometer system in its magnetometer sensor coordinate system were converted to the H, D, and Z components of the geomagnetic coordinate system by using available tilt data and comparing with the 12th generation of IGRF model (International Geomagnetic Reference Field; Thébault et al. 2015). These geomagnetic field variations show similar features to observation data of the conventional magnetometer (OBEM) at site OBM1 and that of the ground-based geomagnetic observatory of Kakioka, Japan (also shown in Fig. 3a). For more detail, slight differences in geomagnetic field variations among the sites were observed, and they should be due to difference in local and regional conductivity structure under each observation site. The two data sets at site OBM1 from our new magnetometer system and from the conventional magnetometer show good agreements. We compared them by induction vectors and a transfer function of each horizontal component between them. They were calculated by BIRRP (Chave and Thomson 2004) using data between October 25, 2016, and November 3, 2016, together with those from the Kakioka geomagnetic observatory as remote reference data. The induction vectors (shown in Fig. 2) are well fitted with 95% confidence limits (two standard deviations), and the error of the induction vector for our new magnetometer system is slightly smaller than those for the conventional magnetometer. The transfer functions and the coherences (Fig. 4) indicate that the two data sets are nearly identical with their period of above 16 s that are generally used for natural source marine electromagnetic sounding methods (e.g., Baba et al. 2017). In contrast, these become less similar as the period becomes shorter than 16 s, probably indicating that the geomagnetic variation signals in the shorter period become too low to be detected by the two magnetometers due to high attenuation of the electromagnetic field in the conductive seawater. The results from the first in situ operation test also provided useful information to improve the standard model design: (1) choice of the voltage reference IC for the A/D converters, (2) affection to magnetic field measurement by the other measurements, and (3) connection to clinometers. First, the choice of the voltage reference IC for the A/D converters was well related to noise floor of ~ 1 Hz in the magnetic field measurements. Three-axis-total standard deviation of 60-s magnetic field variation (in October 22 for example) was estimated to be 0.054 nT at site OBM3 that is much smaller than those at sites OBM1 and OBM2 (0.117 and 0.115 nT, respectively) and that of the conventional magnetometer at site OBM1 (0.105 nT). We have found this difference is caused by the voltage reference IC for the A/D converters; the prototype models at sites OBM1 and OBM2 have LT1461 and that at site OBM3 has MAX6126. The test results suggested the latter one is suitable in this design and the prototype model with this IC showed better performance than the others. Second, magnetic field measurement was affected by the other measurements within the same magnetometer system. The geomagnetic data have a bias of several nT when the control boards were just measuring tilt, temperature, and primary battery voltage, and writing data on the microSD card (Fig. 3b). Such noisy data can be simply removed with referring to the flags in the ancillary headers, because it leads small data loss rate (1.7% at maximum) that is acceptable for most marine electromagnetic sounding purposes using long-term observations. We note that this data corruption rate can be mitigated by reducing frequency of sampling ancillary data and storing data. Third, the two prototype models lost tilt data of one axis at site OBM1 and all axes at site OBM3. We found that the loss of the tilt data was due to bad electrical contacts at the connectors between the control boards and the clinometers. All these results were used to improve the design of the standard model. The miniaturized magnetometer system has been developed for long-term in situ observations and for future multipoint observations of geomagnetic field variation on the seafloor. This system can record geomagnetic field with sampling frequency of 5 Hz as well as tilt, temperature, and primary battery voltage every 60 s (configurable) with hardware statuses. The three prototype models of the new magnetometer system were used for the first in situ operation test for 5 months, and they successfully measured and recorded geomagnetic field variations during the test period. The data indicated the desired performance of the new magnetometer system and provided useful information to improve the design of the standard model. The observed geomagnetic field variations showed good agreements with observation data of the conventional magnetometer and of the ground-based geomagnetic observatory at Kakioka. The measurement accuracy of magnetic field satisfied the requirement of ~ 0.1 nT in the best model with the appropriate voltage reference IC for the A/D converters. The standard model was basically developed by the same electrical design to the prototype model for site OBM3, but some parts were optimized using the results from the first in situ operation test. It has the smaller envelope size Φ35 × 180 mm without batteries due to two-tiered electrical circuit boards, and the lower power consumption of 33 mW at 3.8 V input voltage which results from an optimization of DC/DC converters. The standard model can be installed with batteries in a small pressure cylinder of 36 mm inner diameter and 550 mm length for 1-year observation for example, and this size is highly effective to deploy a number of the magnetometers on multiple sites simultaneously. Since this new magnetometer system is aimed to deploy in all future marine or ocean-bottom geophysical missions both by stand-alone use and by attached to any kind of ocean-bottom instruments such as OBS, we named this system “DOKODEMO MAG” meaning “magnetometer everywhere” in Japanese. analog to digital central processing unit direct current to direct current GNSS Earth Observation Network System global navigation satellite system global positioning system high-sensitivity seismograph network International Geomagnetic Reference Field pulse per second Baba K, Chave AD, Evans RL, Hirth G, Mackie RL (2006) Mantle dynamics beneath the East Pacific Rise at 17 degrees S: Insights from the Mantle Electromagnetic and Tomography (MELT) experiment. J Geophys Res Solid Earth 111:B02101. https://doi.org/10.1029/2004jb003598 Baba K, Tada N, Matsuno T, Liang P, Li R, Zhang L, Shimizu H, Abe N, Hirano N, Ichiki M, Utada H (2017) Electrical conductivity of old oceanic mantle in the northwestern Pacific I: 1-D profiles suggesting differences in thermal structure not predictable from a plate cooling model. Earth Planets Space 69:111. https://doi.org/10.1186/s40623-017-0697-0 Chave AD, Thomson DJ (2004) Bounded influence magnetotelluric response function estimation. Geophys J Int 157:988–1006. https://doi.org/10.1111/j.1365-246X.2004.02203.x Ichihara H, Hamano Y, Baba K, Kasaya T (2013) Tsunami source of the 2011 Tohoku earthquake detected by an ocean-bottom magnetometer. Earth Planet Sci Lett 382:117–124. https://doi.org/10.1016/j.epsl.2013.09.015 Ichiki M, Ogawa Y, Kaida T, Koyama T, Uyeshima M, Demachi T, Hirahara S, Honkura Y, Kanda W, Kono T, Matsushima M, Nakayama T, Suzuki S, Toh H (2015) Electrical image of subduction zone beneath northeastern Japan. J Geophys Res Solid Earth 120:7937–7965. https://doi.org/10.1002/2015jb012028 Ito Y, Obara K, Shiomi K, Sekine S, Hirose H (2007) Slow earthquakes coincident with episodic tremors and slow slip events. Science 315:503–506. https://doi.org/10.1126/science.1134454 Jacobs JA (ed) (1987) Geomagnetism, vol 1. Academic Press Ltd., London Key K, Constable S, Liu LJ, Pommier A (2013) Electrical image of passive mantle upwelling beneath the northern East Pacific Rise. Nature 495:499–502. https://doi.org/10.1038/nature11932 Matsuno T, Seama N, Evans RL, Chave AD, Baba K, White A, Goto T, Heinson G, Boren G, Yoneda A, Utada H (2010) Upper mantle electrical resistivity structure beneath the central Mariana subduction system. Geochem Geophys Geosyst 11:Q09003. https://doi.org/10.1029/2010gc003101 Naif S, Key K, Constable S, Evans RL (2013) Melt-rich channel observed at the lithosphere-asthenosphere boundary. Nature 495:356–359. https://doi.org/10.1038/nature11939 Ogawa K, Shirai K, Sawada H, Arakawa M, Honda R, Wada K, Ishibashi K, Iijima Y, Sakatani N, Nakazawa S, Hayakawa H (2017) System configuration and operation plan of Hayabusa2 DCAM3-D camera system for scientific observation during SCI impact experiment. Space Sci Rev 208:125–142. https://doi.org/10.1007/s11214-017-0347-7 Ozawa S, Nishimura T, Suito H, Kobayashi T, Tobita M, Imakiire T (2011) Coseismic and postseismic slip of the 2011 magnitude-nine Tohoku-Oki earthquake. Nature 475:373–376. https://doi.org/10.1038/nature10227 Schmandt B, Humphreys E (2010) Complex subduction and small-scale convection revealed by body-wave tomography of the western United States upper mantle. Earth Planet Sci Lett 297:435–445. https://doi.org/10.1016/j.epsl.2010.06.047 Tatsumi Y, Suzuki-Kamata K, Matsuno T, Ichihara H, Seama N, Kiyosugi K, Nakaoka R, Nakahigashi K, Takizawa H, Hayashi K, Chiba T, Shimizu S, Sano M, Iwamaru H, Morozumi H, Sugioka H, Yamamoto Y (2018) Giant rhyolite lava dome formation after 7.3 ka supereruption at Kikai caldera, SW Japan. Sci Rep 8:2753. https://doi.org/10.1038/s41598-018-21066-w Thébault E, Finlay CC, Beggan CD et al (2015) International geomagnetic reference field: the 12th generation. Earth Planets Space 67:79. https://doi.org/10.1186/s40623-015-0228-9 Watanabe S, Tsuda Y, Yoshikawa M, Tanaka S, Saiki T, Nakazawa S (2017) Hayabusa2 mission overview. Space Sci Rev 208:3–16. https://doi.org/10.1007/s11214-017-0377-1 Yang B, Egbert GD, Kelbert A, Meqbel NM (2015) Three-dimensional electrical resistivity of the north-central USA from Earth Scope long period magnetotelluric data. Earth Planet Sci Lett 422:87–93. https://doi.org/10.1016/j.epsl.2015.04.006 The authors conducted all of this study and wrote the manuscript. All authors read and approved the final manuscript. The authors gratefully acknowledge the development efforts of Mr. Osamu Nara, Mr. Hirokazu Sato, Mr. Kaneyoshi Kuwata, and the engineers of Meisei Electric Co., Ltd. The authors thank the crews (Prof. Yoshiji Yano, captain) of the T/V Fukae-maru and the onboard research members for their valuable help to develop and recover the instruments during two cruises. The Kakioka geomagnetic field data were downloaded from the digital data service site of the observatory of Japan Meteorological Agency (http://www.kakioka-jma.go.jp/obsdata/metadata/en/products). The authors thank two anonymous reviewers for helpful comments that improved this paper. The authors declare that they have no competing interests. Availability of data and materials Consent for publication Ethics approval and consent to participate This work was partly supported by Grant-in-Aid for Scientific Research (B) (No. 15H03717), Japan Society for the Promotion of Science. Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. About this article Cite this article Ogawa, K., Matsuno, T., Ichihara, H. et al. A new miniaturized magnetometer system for long-term distributed observation on the seafloor. Earth Planets Space 70, 111 (2018). https://doi.org/10.1186/s40623-018-0877-6 - New magnetometer system - Small size - Low power consumption - Long-term observation - Electromagnetic sounding
<urn:uuid:e62b6280-7489-4d16-b0a1-11fddd678980>
CC-MAIN-2020-16
https://earth-planets-space.springeropen.com/articles/10.1186/s40623-018-0877-6
s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370505731.37/warc/CC-MAIN-20200401130837-20200401160837-00163.warc.gz
en
0.898353
6,558
2.6875
3
hi there, this problem has really confused me, could someone please help me. the distribution function of random variable A is given, and i must find a distribution function of new random variable B=1/A. what is your question? First of all that distribution function is wrong. It's ok that F(x)=0 when x is less than 1 and F(x)=1 when x is greater than 2 BUT it cannot be x from 1 to 2. MAYBE your instructor meant F(x)=x-1, since F cannot exceed 1. This says that F(1.5)=1.5 which means that which is a no-no. BUT in order to transform, I use calc 1 which can easily be proved by differentiating the cdfs, which was the approach of your teacher.
<urn:uuid:1af1d3a4-edb4-4b51-a01e-3daa71742bbb>
CC-MAIN-2015-14
http://mathhelpforum.com/advanced-statistics/147861-problem-distribution-function.html
s3://commoncrawl/crawl-data/CC-MAIN-2015-14/segments/1427131298889.70/warc/CC-MAIN-20150323172138-00110-ip-10-168-14-71.ec2.internal.warc.gz
en
0.967167
176
3.296875
3
Exercise Discovery Could Save Lives of Sickest Patients Source: University of Virginia Summary: In a new discovery researchers have found that how exercise protects us from diseases among the sickest patients and most gravely injured. People who suffer from severe trauma or sepsis (full body infection) often develop multiple organ dysfunction syndrome (MODS). It’s as if the immune system instead of attacking the infection, turns against the patient and attacks the vital organs. In intensive care units, MODS is considered to be the primary cause of death and killing almost 80% of the people who develop it. In a new discovery, researchers at the University of Virginia School of Medicine have found that how exercise protects from diseases among the sickest patients and most gravely injured. The findings were published in the journal Free Radical Biology & Medicine. Researchers suggest that skeletal muscles in our body naturally make an antioxidant which helps to get rid of excessive free radicals. The antioxidant, EcSOD is one of the mechanisms by which exercise protects the body against diseases. The antioxidant helps by acting as the first line of defense and prevents inflammatory cells from accumulating inside the vital organs. More clear understanding of the mechanism will help the researchers to strategically design better pharmacological and genetic interventions. Clinically this has huge implications which indicate the importance of regular exercise in keeping us more resistant to many disease conditions. Zhen Yan said, “In the condition of sepsis and severe trauma, our defense system becomes exaggerated to the point that [the body’s immune defenses] misunderstand the signal, so that they begin to attack the organs”, “Our body is mistakenly thinking there is a danger signal coming from the organs themselves.” More Information: Jarrod A. Call et al. “Muscle-derived extracellular superoxide dismutase inhibits endothelial activation and protects against multiple organ dysfunction syndrome in mice”, Free Radical Biology & Medicine (2017).
<urn:uuid:c479702f-5e71-4356-9c7d-4ba10b0c7fd6>
CC-MAIN-2018-09
http://www.healthinformative.com/news-research/exercise-discovery-save-lives-sickest-patients/
s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891812259.30/warc/CC-MAIN-20180218212626-20180218232626-00218.warc.gz
en
0.931064
399
2.9375
3
Mourning prints are part of Victorian fashion in the post Civil War Era. When a woman became widowed, fashion dictated that she wear all black, without ornament and a crepe weeping veil for a year and a day. Beyond the first year, tradition stated that the widow should wear half mourning for a period of at least 6 months. This was called Mourning Grays, or sometimes called Shaker Grays,. The fabrics were finely printed black on a white ground that appeared to be gray. Gradually colors were added, starting with more somber colors such as deep purple, burgundy and brown. Purples from this era are often called "Fugitive", because over time the purple coloring faded to a soft brown. The first stable purple dye wasn't widely used until the 1920's or 30's. Prior to that time, the purples were produced by using natural dyes such as madder and logwood. Placing metal salts, or mordents in specific areas of the fabric would produce different colors of purple, red or brown. It must have been fun experimenting and coming up with such beautiful results! 856-2, 856-20 857-13, 857-20 858-2, 858-20 859-13, 859-20 860-13, 860+20 861-13, 861-20 862-13, 862-20 About the Designer Join The Newcastle Mailing List! Please take a moment to fill out the following form to join our mailing list. Don't worry... we promise we'll never share your info with anyone else. Be sure to follow us on Pinterest and Facebook too! Use this form to make a public comment on this fabric collection. If you're trying to contact Newcastle Fabrics, please use the contact us form instead. Make sure you enter the (*) required information where indicated. Basic HTML code is allowed.
<urn:uuid:96df013d-2e4b-4eef-8b39-2597d36618dc>
CC-MAIN-2014-10
http://www.newcastlefabric.com/gallery/item/54-mourning-grays
s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1393999678772/warc/CC-MAIN-20140305060758-00044-ip-10-183-142-35.ec2.internal.warc.gz
en
0.945066
400
3.203125
3
Although the country is getting over gender discrimination as more females are being educated and women are being allowed to enter politics and the workforce, Afghanistan still faces security challenges. The United States, after accusing former Taliban rulers for sheltering Osama Bin Laden and other al-Qaeda leaders for the 9/11 attacks, invaded Afghanistan. In 2001, the invasion toppled Taliban from power but the country went into a civil war and since then, countless numbers of both the military and the civilians have been murdered. Young Afghans particularly have suffered the effect of the war, growing up with a background of conflict, anger, and war, blaming both the Talibans and the government of the United States. Mohammad Tahir Basharyar, who recently turned 18, lost his family and friend in the war laments on how the war has affected everyone. However, he admits that compared to the life they had lived under the Taliban, things were more bearable and generally better. According to him, many people lived from hands to mouth during the Taliban, and women were not permitted to move beyond their homes. People lacked access to basic amenities and science and freedom of expression were porno limited. A connection was also denied to the outside world. Somehow, the invasion by the United States ended the brutality of the Taliban rule. Unlike in previous years, there was the chance again to speak freely and the women were allowed to exercise more rights. The older group of people appeared to be even more disappointed. They reminisced on the days they were faced with economic hardship and religious coercion. There was hardly enough to eat all day but they were always forced to pray all five times every day. However, they expected the US invasion to lead the society to develop better, but they were disappointed. No vital changes were seen in the living of the citizens. The conditions of living had not significantly improved and the country was still suffering from endemic corruption. There was still no constant power supply and the rate of violence was increasing. Many students are in fear of not completing their studies. Many Afghan youths cannot make plans for the future because they spend time and energy worrying about conflict and the safety of their families. Many students have lost the ability to concentrate on studies in the classrooms. Although the rights of humans are more recognised and people could fight for them, the deteriorating security in the State is a source of common concern to all the citizens.
<urn:uuid:08d474d6-48ed-48f0-9d5e-51911b305c30>
CC-MAIN-2022-27
https://www.pecob.eu/tag/war-time/
s3://commoncrawl/crawl-data/CC-MAIN-2022-27/segments/1656103920118.49/warc/CC-MAIN-20220701034437-20220701064437-00189.warc.gz
en
0.987582
484
2.78125
3
The small Walker's Epinotia Moth holds its wings closed tightly around its body when resting. * MAP NOTES: The territorial heat map above showcases (in red) the states and territories of North America where the Walker's Epinotia Moth may be found (but is not limited to). This sort of data is useful when attempting to see concentrations of particular species across the continent as well as revealing possible migratory patterns over a species' given lifespan. Some insects are naturally confined by environment, weather, mating habits, food resources and the like while others see widespread expansion across most, or all, of North America. States/Territories shown above are a general indicator of areas inhabited by the Walker's Epinotia Moth. Insects generally go where they please, typically driven by diet, environmental changes, and / or mating habits.
<urn:uuid:ccd2df64-c536-41db-a167-3e902a58b9a4>
CC-MAIN-2022-40
https://www.insectidentification.org/insect-description.php?identification=Walkers-Epinotia-Moth
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335257.60/warc/CC-MAIN-20220928145118-20220928175118-00569.warc.gz
en
0.926896
176
3.21875
3
In Hitler’s Furies, the concept of female complicity and their involvement in the war and Holocaust was a prominent theme. The question remains, are women the victims or perpetrators of this violence? I can say that neither myself nor our group had a definitive answer. There were many perspectives presented by the author about the intentions behind the range of actions presented by the women. The author includes examples from the disregard and differentiation of Jewish people, being bystanders to the death camps around them, and to extreme of actually killing Jews as part of firing squads. Our group discussed some of the potential underlying intentions for the various actions, including: - Whether this was due to the systemic and overarching rules of the Nazi regime – Did the culture of fear and racism corrupt thoughts and actions where they wouldn’t normally outside of these circumstances. - The ability to use these structures to move up the social ladder. At the time of the Holocaust and the war, women in Germany did not have social and political opportunity. Through being mothers, wives, and members of the Nazi party, they were able to exploit the system to gain opportunities and jobs. - Did complicity come from being strictly in a “mother” role – Did this role perpetuate comfort for men in the battlefield and physically engaging with the Holocaust? Not to excuse the horrific action of genocide through the Holocaust, but the ability to try and disseminate the individual intention versus the collective intention in the context of war, suffering and suppression was extremely insightful in how these multiple factors can feed into a problem.
<urn:uuid:8e516042-8999-4885-a830-e5e654dab846>
CC-MAIN-2018-09
https://hate2point0.com/2018/02/03/sweeper-furies-not-just-part-of-greek-mythology/
s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891811243.29/warc/CC-MAIN-20180218003946-20180218023946-00763.warc.gz
en
0.968816
315
3.421875
3
Introduction to Discrimination eLearning is a course covering issues that occur in the work environment as a result of the introduction of anti-discrimination legislation in Jersey. Introduction to Discrimination eLearning is a 30 minute online course that ensures employees are aware of the issues of equality and diversity in the workplace and what their role is in relation to these issues. Employers are responsible for the discriminatory actions of their employees. Complying with what is generally perceived as ‘reasonably practicable’ in this area is to communicate an Equality and Diversity Policy to all staff, and provide awareness training on the potential implications of discriminatory practices in the workplace. The course is deliberately aimed at an awareness level which allows for a reasonable completion time. In this training course employees will learn: - Definitions of ‘equality’ and ‘diversity’ - The benefits of being diversity aware - The role of an individual in relation to equality and diversity responsibilities in the workplace, and standards of behaviour expected - The different types of discrimination including bullying and harassment - How to make a complaint/seek help if they are either witness to, or a victim of, discriminatory practices BENEFITS TO YOUR ORGANISATION - Supports your organisation’s commitment to promoting equality and diversity in the workplace, and the prevention of discriminatory practices in your workplace - Helps to create a culture of respect for all employees by showing that diversity is a positive aspect of the workplace and not a threat - Allows your organisation to make clear what standards of behaviour are expected in relation to equality and diversity, what kinds of behaviour will not be tolerated and the consequences of breaking behaviour codes
<urn:uuid:125cabd5-9fb0-43bd-876a-90fad2b753e2>
CC-MAIN-2017-26
https://lawatworkci.com/product/introduction-discrimination-elearning/
s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128323870.46/warc/CC-MAIN-20170629051817-20170629071817-00327.warc.gz
en
0.940737
337
3.375
3
Students with Speech Impairments Participating in Recess Matthew D. Lucas, Ed.D., C.A.P.E. Carolyn R. Watson, Student The participation of students with speech impairments in recess can often be both challenging and rewarding for the students and teachers. This paper will address common characteristics of students with speech impairments and present basic solutions to improve the experience of these students in the recess setting. Initially the definition, symptoms, and prevalence of speech impairments will be presented. The paper will then address the special education classification of children with speech impairments, benefits of recess for children with speech impairments, and recommendations for these children in recess. Read or Download To Read this Article - or Download this Article (login require) To Download the Entire Winter 2013 Issue of JAASEP - (login required)
<urn:uuid:9862988a-7c2f-4e36-a645-9b15ba6108e4>
CC-MAIN-2015-22
http://www.naset.org/3697.0.html
s3://commoncrawl/crawl-data/CC-MAIN-2015-22/segments/1432207928015.28/warc/CC-MAIN-20150521113208-00119-ip-10-180-206-219.ec2.internal.warc.gz
en
0.923732
175
3.296875
3
There is a lot of confusion online about the safety of Borax in natural cleaning recipes, pest control applications and children's crafts recipes. Proponents claim that it is safe and natural, while others consider it a toxin to avoid. Part of the debate stems from people who are confused about the relationship between Borax and boric acid. The terms Borax and boric acid are often used interchangeably, though they are technically different. As you'll see, however, the differences are ultimately slight. Wikipedia defines Borax as: Borax, also known as sodium borate, sodium tetraborate, or disodium tetraborate, is an important boron compound, a mineral, and a salt of boric acid. Borax is commonly sold under the brand name "20 Mule Team" as a laundry aid, though it is also available from chemical companies for its uses as a pesticide, fungicide and cleaner, among other uses. Its chemical name is Sodium Tetraborate Pentahydrate and its Material Safety Data Sheet (MSDS) notes, "Sodium Tetraborate Pentahydrate is chemically and toxicologically related to Boric Acid" and "Sodium Tetraborate Pentahydrate is converted to Boric Acid in biological systems." Borax is often touted as safe and natural. While it is a naturally occurring mineral, that doesn't mean it is without dangers. It is officially classified as a poison. Government sites recommend that people who work with Borax use gloves and handle it with caution. Studies have linked it to reproductive problems in some lab animals, as well as a host of serious disorders (and death) at higher levels. Green Footsteps points out: Borax needs to be stored carefully because it is toxic if ingested. While this may be unlikely to happen, even by accident, anyone using it should be aware because of the dangers of very young children playing with it - as with any household compounds. Even as little as a teaspoonful could prove fatal if swallowed by a young child. For this reason, be very careful if using it anywhere near food and wipe up spills immediately. Many people online are quick to point out that Borax is technically different from boric acid, and they seem to believe that this somehow makes it safe. It is important to note that not only are the two substances extremely similar to begin with, but when Borax is exposed to an acid (such as stomach acid), it converts to boric acid. The 20 Mule Team's MSDS (Material Safety Data Sheet) for Borax specifically says: Nonetheless, the following effects have been reported for a component, sodium borate, and boric acid. Sodium borate upon entry into the body becomes boric acid. They further say: May cause gastrointestinal disturbances such as headache, nausea, vomiting, abdominal pain, and diarrhea, with delayed effects of skin redness and peeling. The National Library of Medicine's TOXNET, the national database on toxicology, hazardous chemicals, environmental health, and toxic releases, says of Borax: Ingestion of 5 to 10 g by young children can cause severe vomiting, diarrhea, shock and death. Note that this is the lethal dose for Borax, which is supposedly so much safer than boric acid. They further say: Effects from ingestion include abdominal pain, diarrhea, headache, nausea, vomiting, weakness, convulsions. Wikipedia lists more concerns: Sufficient exposure to borax dust can cause respiratory and skin irritation. Ingestion may cause gastrointestinal distress including nausea, persistent vomiting, abdominal pain, and diarrhea. Effects on the vascular system and brain include headaches and lethargy, but are less frequent. "In severe poisonings, a beefy red skin rash affecting palms, soles, buttocks and scrotum has been described. With severe poisoning, erythematous and exfoliative rash, unconsciousness, respiratory depression, and renal failure." Borax was added to the Substance of Very High Concern (SVHC) candidate list on 16 December 2010. The SVHC candidate list is part of the EU Regulations on the Registration, Evaluation, Authorisation and Restriction of Chemicals 2006 (REACH), and the addition was based on the revised classification of Borax as toxic for reproduction category 1B under the CLP Regulations. Substances and mixtures imported into the EU which contain Borax are now required to be labelled with the warnings "May damage fertility" and "May damage the unborn child". It is important to note that many of the products we regularly use around our homes are as dangerous or more. Even table salt and water are deadly in high enough doses. That said, Borax is not a harmless material and it's important to understand the dangers if you use it in your home -- especially around children. If you do use Borax in your home, follow these safety guidelines: - Keep it clearly labeled, and out of reach of children and pets. - Wear gloves when handling it, especially if you have damaged or broken skin. - Keep Borax away from eyes; use eye protection if necessary. - Take special precautions about inhaling it if you have asthma. - Do not use it in craft recipes if your children are likely to put the concoctions in their mouths or if they have broken or damaged skin on their hands. - Use special care when handling Borax if you are pregnant or planning to become pregnant. - Use caution using it in areas used by children and pets, such as carpets where babies crawl. - Wipe up spills carefully. - In case of accidental ingestion (one teaspoon or more for adults, much less for children), give water and seek medical attention. Want to stay in the loop? Be sure to subscribe to my column to be updated when I post articles. You can also find me on Pinterest and on examiner.com on the topics of homeschooling, attachment parenting and my national attachment parenting column, and on Facebook at All Natural Families and A Magical Homeschool.
<urn:uuid:0fdb9c80-e99d-475b-b12c-685ee32eb20e>
CC-MAIN-2015-22
http://www.examiner.com/article/just-how-safe-is-borax
s3://commoncrawl/crawl-data/CC-MAIN-2015-22/segments/1432207929096.44/warc/CC-MAIN-20150521113209-00132-ip-10-180-206-219.ec2.internal.warc.gz
en
0.940272
1,263
2.921875
3
Nazi is obviously a short form of National socialist, or Nationalsozialist to be precise, just as Sozi is a short form of Sozialist. But the word has a much more interesting story than that. Long before the rise of the NSDAP in the 1920s, people in at least southern Germany could be called Nazi if they were named Ignaz, or came from Austria or Bohemia (where they apparently had lots of Ignazes, perhaps after the well-known Loyola); it was supposedly also used as a generic name for soldiers of Austria-Hungary, like the German Fritz or Russian Ivan. It had to be used with caution between friends, though, since it could also mean "idiot" or "clumsy oaf". That's how it found it's way into politics; the fact that Adolf came from Austria (not Bohemia, though) could have made the pun even better. The Nazis supposedly made attempts to include the N-word in their own vocabulary in order to make it less derogatory, but unsuccessfully; since such a maneuver requires a sense of humor as well as irony, it was probably doomed to fail. "Nazi" was first used in English in 1930, as well as in Swedish, and probably some other languages as well; that's apparently the first year the Nazis made enough fuss to earn themselves the short form internationally. The double sense of the expression, as described above, was unfortunately lost in translation. An example of pre-hitlerian use of Nazi in southern Germany can be found in a "Bayerische Komödie in 4 Akten": Der Schusternazi, "the shoemaker nazi", by Ludwig Thoma in 1905. The title role is the shoemaker Ignaz Stanglmayer; without giving too much away, he receives a large inheritance, dumps his old friends in favor of some new ones, who turn out to have a piece of valuable russian land for sale, which they offer Ignaz for a friendship price... Und so weiter.
<urn:uuid:d337d82c-cbf4-40c7-9d5f-597a018dd544>
CC-MAIN-2015-22
http://www.faktoider.nu/nazi_eng.html
s3://commoncrawl/crawl-data/CC-MAIN-2015-22/segments/1432207928479.19/warc/CC-MAIN-20150521113208-00284-ip-10-180-206-219.ec2.internal.warc.gz
en
0.983698
420
2.734375
3
What's the Latest Development? An improved process for making large amounts of carbon nanotubes may allow a group of researchers at Rice University to make a new, highly efficient electricity grid this summer. Currently, the best production methods result in a mixture of different nanotubes, with varying dimensions and wildly different electrical properties. "Called 'amplification,' the new process should eventually allow the researchers to turn a nanogram of pure carbon nanotubes into a gram, then a kilogram, then a ton." Carbon nanotubes seen as crucial to developing a renewable energy grid. What's the Big Idea? Nanotechnology is poised to revolutionize nearly every engineering sector, enabling humans to control events at microscopic levels never before thought possible. "Andrew Barron is part of a group at Rice that wants to make something very large from these nanotubes: miles and miles of highly conductive electrical transmission lines for a more efficient energy grid, which will be important as the use of renewable energy grows." While Rice continues to research 'amplification', other methods are being developed which will require scaling production to become cost effective.
<urn:uuid:49b0a716-7113-48d9-854c-1e4670a5f2b7>
CC-MAIN-2016-40
http://bigthink.com/ideafeed/nanotechnology-the-future-of-electricity
s3://commoncrawl/crawl-data/CC-MAIN-2016-40/segments/1474738660709.12/warc/CC-MAIN-20160924173740-00251-ip-10-143-35-109.ec2.internal.warc.gz
en
0.92747
230
3.640625
4
Orfalea Center Thematic Research Cluster Global Futures: Uncertainty, Displacement, Security Pandemics Past Bibliography The works below are science journalism texts that address how different bodies of people (such as the general public, governments, NGOs, and scientific communities) came to know about pandemics previous to COVID-19. The lenses through which these works attempt to portray the moments of initial contact with emerging viruses are varied. While many authors point to the continued failure of governing bodies to detect and contain these highly contagious diseases, others also point to the role of social, cultural, and environmental factors that have either exacerbated mishandlings of pandemics or frustrated and complicated attempts to understand emerging viruses. These works are significant not only for how they depict the struggle to (epistemologically, biologically, and politically) “capture” emerging viruses, but also as epistemological and cultural objects themselves. This bibliography is a companion to the Mediations of COVID-19 Bibliography. Barry, John M. The Great Influenza: The Story of the Deadliest Pandemic in History. New York: Penguin Books, 2004. Branswell, Helen. “What the world learned in eradicating smallpox: Unity mattered.” STAT. May 8, 2020. Crosby, Molly Caldwell. The American Plague: The Untold Story of Yellow Fever, The Epidemic That Shaped Our History. New York: Berkley Books, 2006. DeSalle, Rob, ed. Epidemic! The World of Infectious Diseases. New York: The New Press, 1999. Donaldson, Ross. The Lassa Ward: One Man’s Fight Against One of the World’s Deadliest Diseases. New York: St. Martin’s Press, 2009. Garrett, Laurie. Ebola: Story of an Outbreak. New York and Boston: Hachette Books, 2014. Garrett, Laurie. Betrayal of Trust: The Collapse of Global Public Health. New York: Hyperion, 2011. Garrett, Laurie. The Coming Plague: Newly Emerging Diseases in a World Out of Balance. New York: Farrar, Straus, and Giroux, 1994. Greenfeld, Karl Taro. China Syndrome: The True Story of the 21st Century’s First Great Epidemic. New York and London: Harper Perennial, 2006. Johnson, Steven. The Ghost Map: The Story of London’s Most Terrifying Epidemic – and How it Changed Science, Cities and the Modern World. New York: Riverhead Books, 2006. Kelly, John. The Great Mortality. New York: HarperCollins, 2005. Khan, Ali S. and William Patrick. The Next Pandemic: On the Front Lines Against Humankind’s Gravest Dangers. New York: PublicAffairs, 2016. Kinsella, James. Covering the Plague: AIDS and the American Media. New Brunswick and London: Rutgers University Press, 1989. Kolata, Gina. Flu: The Story Of The Great Influenza Pandemic of 1918 and the Search for the Virus that Caused It. New York and London: Simon & Schuster, 1999. Kraut, Alan M.. Silent Travelers: Germs, Genes, and the Immigrant Menace. Baltimore and London: The Johns Hopkins University Press, 1994. Leavitt, Judith Walzer. Typhoid Mary: Captive to the Public’s Health. Boston: Beacon Press, 1996. Levitt, Alexandra M.. Deadly Outbreaks: How Medical Detectives Save Lives Threatened by Killer Pandemics, Exotic Viruses, and Drug-Resistant Parasites. New York: Skyhorse Publishing, 2013. McKenna, Maryn. Beating Back the Devil: On the Front Lines with the Disease Detectives of the Epidemic Intelligence Service. New York and London: Free Press, 2004. McNeill, William H.. Plagues and Peoples. New York: Anchor Books, 1976. Morse, Stephen S., ed. Emerging Viruses. New York and Oxford: Oxford University Press, 1993. Oshinsky, David M.. Polio: An American Story. New York and Oxford: Oxford University Press, 2005. Osterholm, Michael and Mark Olshaker. Deadliest Enemy: Our War Against Killer Germs. New York, Boston, and London: Little Brown Spark, 2017. Preston, Richard. Crisis in the Red Zone: The Story of the Deadliest Ebola Outbreak in History, and of the Outbreaks to Come. New York: Anchor Books, 2019. Quammen, David. Ebola: The Natural and Human History of a Deadly Virus. New York and London: W. W. Norton & Company, 2014. Shah, Sonia. Pandemic: Tracking Contagions, from Cholera to Ebola and Beyond. New York: Sarah Crichton Books, 2016. Shilts, Randy. And the Band Played On: Politics, People, and the AIDS Epidemic. New York: St. Martin’s Griffin, 1987. Yong, Ed. “The Next Plague Is Coming. Is America Ready?” The Atlantic. July 2018. Yong, Ed. “How A Pandemic Might Play Out Under Trump.” The Atlantic. December 2016. Zimmerman, Barry and David Zimmerman. Killer Germs: Microbes and Diseases That Threaten Humanity. New York and Chicago: McGraw-Hill, 2003.
<urn:uuid:e7ac23d5-f97a-4c4e-99e9-8984c697bd7f>
CC-MAIN-2022-05
https://www.orfaleacenter.ucsb.edu/pandemics-past-bibliography-global-futures-uncertainty-displacement-security/
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320304954.18/warc/CC-MAIN-20220126131707-20220126161707-00680.warc.gz
en
0.76446
1,182
2.59375
3
A child is nestled in your lap, and the picture book is opened wide. Maybe she’s a quiet observer while you read. Or maybe he points out every detail on the page. Whether your child likes to listen or likes to talk, how you engage them is important. In a recent study, College of Education and Human Ecology researchers found that teachers may be asking too many easy questions, or no questions at all, while reading to their preschoolers. Naturally, this leads to the question: What should families ask their preschoolers to help get them thinking as they read? “Book reading and the conversations embedded within book reading are really good ways to provide lots of language and stimulation to kids, which in turn strengthens the part of the brain that is responsible for language,” said EHE Distinguished Professor Laura Justice, the study’s author, in the podcast Positively Dad. It can be hard for families to know where to start when having these conversations during reading time, so the college’s experts offer the following advice. “It's most important that children are asked questions that they can expand upon – questions that ask them to describe something,” said Anneliese Johnson, principal of the college’s A. Sophie Rogers School for Early Learning. Families and teachers should be asking questions such as “What's more?” “How else?” or “Why do you think that?” rather than “Is that blue?” or “Who’s smaller?” “We want to ask questions that make kids infer things,” Justice said. “There’s good evidence that suggests that is good for kids’ development.” “What happens next” or “Why do you think the character is doing that” are types of inference questions to ask. Try not to get too caught up in the way you ask questions. “The best way to think about good questioning or good conversation is that it's authentic and honest,” Justice said in the podcast. “A good rule when you're reading with just one or two children is really to follow their lead. It's more important to talk with your child about the book than it is to finish the book in one sitting,” Johnson said. Keep in mind where your child is in his or her reading skills. It’s easy to recognize when a child is struggling with questions that are too complex. A three-year-old may not yet be developmentally ready to say what will happen next. “That's a complex executive thought process – to predict something in the future,” Johnson said. So if they can't answer “What happens next,” don’t worry. Change the question to “How do you think he will get out of that box?” “You're focusing the child's thinking and giving them a more attainable question. You’re still asking them an open-ended question, rather than just moving on in the book or not supporting that thinking skill,” Johnson said. If necessary, continue to ask simpler questions until the child can respond. Then the next time you read to your child, you’ll know at which level to start asking questions. Teachers also use a tactic called “serve and return.” A child asks a question or makes a statement; the adult expands upon it, offering more information and more words that help expand the child’s vocabulary and narrative skills. Then, the adult offers the conversation back to the child. The goal of serve and return interactions is to keep the back-and-forth conversation going as long as possible. The hardest questions for children to answer are ones for which they have little context. Questioning about space or dinosaurs might be difficult because children have never been around them. An Ohio child’s concepts of the ocean and snow might be much different than those of a child living in Florida. When reading books that don’t have many pictures, watch for signs that the book is too challenging. If the child asks lots of questions or makes comments about it, he likes it. If she’s disengaged or starts playing, you’ll know that the book is too advanced. Funny books can be very engaging, Johnson said, because they give children the rewarding experience of laughing with families or their classmates. She likes books by Mo Willems author of Don’t Let the Pigeon Drive the Bus! or The Pigeon Finds a Hot Dog! Children also love books with elaborate illustrations that they can search for hidden objects. The drawings give families and teachers a built-in way to ask questions. Another category of book younger children especially love: repetitive and rhyming books. Professor Justice also addresses how to support a child who is less willing to talk. “Something that can help is storybooks with manipulative features, like flaps you can lift,” she said, or books that make sounds. The best way to practice reading and asking your children compelling questions is to keep reading. “First and foremost, always incorporate reading into the daily life of your child. And the easy way to do that is bedtime, because it's also a relaxing activity where you can connect emotionally. It shows the child that reading is valued in the home,” Johnson said. Justice reinforced the importance of reading to children, especially at a younger age. “Birth to five is the most important period of brain development,” she said. “They’re building connections throughout their brain, and those connection pathways are going to serve them” for the rest of their lives. “Sitting down and reading a book, one on one, with a shared focus for 10-20 minutes a day is helping lay those pathways,” Justice said. “It is not a trivial activity. It's actually really important.” Reading to the child in your life is a privilege, and soon they’ll be reading on their own. Family life can be hectic, and reading sometimes slides off the schedule. “But if you're enjoying it, the child will enjoy it, so try not to make it stressful, but an enjoyable part of your daily routine,” Johnson said.
<urn:uuid:494c9c48-6a75-41b0-99bf-66f0189a9581>
CC-MAIN-2022-49
https://ehe.osu.edu/news/listing/reading-kids-questions-get-them-engaged
s3://commoncrawl/crawl-data/CC-MAIN-2022-49/segments/1669446711150.61/warc/CC-MAIN-20221207053157-20221207083157-00771.warc.gz
en
0.965868
1,339
3.46875
3
Urbes Mutantes: Latin American Photography 1944–2013 is a major survey of photographic movements in Argentina, Brazil, Chile, Colombia, Cuba, Mexico, Peru, and Venezuela. Taking the "mutant," morphing, and occasionally chaotic Latin American city as its focus, the exhibition draws particularly on street photography's depictions of the city during decades of political and social upheaval. It is divided into sections that explore public space as a platform for protest, popular street culture, the public face of poverty, and other characteristics of the city as described in photographs. Dispensing with arbitrary distinctions between genres of photographyart photography, photojournalism, documentary Urbes Mutantes points to the depth and richness of the extensive photographic history of the region. “As the 20th century progressed, amidst struggles for social justice and in defense of democracy and freedom, the city became a setting for uprisings and revolutions,” says Guest Curator Alexis Fabry. “Images became as important as the stories covering the events that shaped these Latin American nations. In certain cases, politics and art were inseparable.” Without attempting to provide an exhaustive account of Latin America’s photographic traditions, the exhibition spotlights alternative views of the region’s urban centers, bringing greater nuance to stock clichés or rigidly framed generic stereotypes. One of the challenges posed by this exhibition is to reveal how politically and socially committed visions can be paired with other types of urban documentation. These include formal experiments, such as abstract renderings of architecture and urbanization, or the more ironic and cynical works that emerged in the 1990s, which sought to question the region’s immutable truths and entrenched myths. Comprising more than 200 images, the exhibition is drawn from the collection of Leticia and Stanislas Poniatowski, the exhibition was first shown at the Museo de Arte del Banco de la República in Bogota in 2013. It was co-curated by Alexis Fabry and María Wills, and is accompanied by a bilingual catalogue published by Toluca Editions.
<urn:uuid:84fc384a-bbdc-489d-80de-0d3460b20a0c>
CC-MAIN-2017-09
https://www.icp.org/exhibitions/urbes-mutantes-latin-american-photography-1944%E2%80%932013
s3://commoncrawl/crawl-data/CC-MAIN-2017-09/segments/1487501172050.87/warc/CC-MAIN-20170219104612-00529-ip-10-171-10-108.ec2.internal.warc.gz
en
0.950549
426
2.625
3
Indonesia’s Fires Are Driving Climate, Public Health Crises Indonesia is in the grips of a double-headed climate and public health crisis as fires rage across the country. On Monday, the fires reached such a fever pitch that Indonesian Prime Minister Joko Widodo cut a trip to the U.S. short to return home and deal with the inferno that’s turning air toxic and putting more carbon dioxide in the atmosphere than the U.S. on a daily basis. Nearly 116,000 fires have been documented and air quality in Indonesia and its neighboring countries has suffered. The fires are a yearly occurrence, but this year is already the second-most prolific burn year on record and dry conditions driven by this year’s strong El Niño mean it still has a shot at the top spot. Air pollution and smoke from fires in Indonesia on Monday, October 26. Credit: Earth Wind Map “We are definitely looking at an extended fires season this year, perhaps rivaling 1997, with our current forecasts continuing the strong tilt toward dry (conditions),” Andrew Robertson, a climate scientist at the International Research Institute for Climate and Society (IRI), said. Robertson and others at IRI have developed fire forecasting tools for the Indonesian government to pinpoint outbreaks and plan a season ahead. The longer the fires burn, the harder they are to put out (and they could smolder through the rainy season). Deforestation Rate in Indonesia Surpasses Brazil El Niño Brings Floods, Risks — and Opportunities Why Do We Care So Much About El Niño? The current fires are burning in carbon-rich peatland and hardwood forests. That’s causing them to release as much greenhouse gas as the U.S. economy daily into the atmosphere. Fire emissions have already passed Japan’s annual greenhouse gas emissions and by the end of the year, the fires could account for up to 3 percent of the world’s greenhouse gas emissions from all human activities. That means the fires are contributing to the long-term climate crisis, but they’re also driving huge public health and economic woes for the region in the here and now. Schools, flights and outdoor events have been canceled in Indonesia, Singapore and Malaysia as the smoke permeates daily existence. As many as 500,000 people have reported haze-related illnesses, according to the Guardian. The Indonesian government is combating the fires with help from Australia and other countries in the region and costs are expected to run in the billions. However, fighting fires doesn’t address the recurring issue for a country where nearly 40 percent of its population works in agriculture. “Fighting Indonesia’s peatland fires is expected to exceed $14 billion this season and is only a short-term solution to a long term crisis,” Fred Stolle, a research at World Resources Institute, said. Smoke from fires in Borneo, Indonesia on October 19. Click image to enlarge. Credit: NASA Earth Observatory Farmers generally use fire — often illegally — to clear land for agricultural use ahead of the rainy season. In a regular year, they can cause widespread air pollution in Indonesia and across the region. But El Niño is helping enforce dry conditions and slow the arrival of the rainy season. This year’s efforts to control fires is a band-aid on the longer term issue of stopping fires before they start and restoring peatlands so that they keep their massive stores of carbon sequestered in the ground rather than in the atmosphere. Stolle and others at World Resources Institute have recommended improved land management techniques and providing farmers with tools and skills necessary to clear land without using fire. Miriam Marlier, a postdoctoral researcher at Columbia University who has studied this and past fire outbreaks in the region, said satellite data can also help scientists and policymakers target upwind and downwind locations to understand both the health impacts as well as the fire hot spots. “By identifying the locations that contribute most to exposure downwind of the fires, we can help to target the areas that should be the focus of fire prevention and fire fighting resources,” she said. While this year’s fires could dwindle when the rainy season finally arrives next month, it’s possible it could only be a respite and not a full reprieve. Robert Field, a research scientist at NASA Goddard Institute for Space Studies, warned that the region generally sees another short dry spell in February and with El Niño likely to still be in place then, it could mean a smoky start to the year. You May Also Like: Warming May Mean Stifling Future for Middle East Why Today Is a Critical Moment for the Clean Power Plan Patricia, Record-Setting Monster, Beelines For Mexico Pulp Fiction, The Series
<urn:uuid:2880d354-c36e-4f75-9efa-12efd546f6ac>
CC-MAIN-2019-47
https://www.climatecentral.org/news/indonesias-fires-climate-public-health-19601
s3://commoncrawl/crawl-data/CC-MAIN-2019-47/segments/1573496670821.55/warc/CC-MAIN-20191121125509-20191121153509-00214.warc.gz
en
0.942549
1,002
2.625
3
First, classification is an integral part of an organized and practical naming system, which Linnaeus recognized; he closely linked classification and naming. The official names for creatures prior to his work served two functions: to identify and describe. The names were in Latin (the universal "educated" language) and could be two or more words long. The first word was the generic name for the creature and the rest of the words formed a specific description. Sounds like a pretty good idea, but since you had no limit on the number of words in the specific description, you ended up with names like: Plantago foliis ovato-lanceolatus pubescentibus, spica cylindrica, scapo tereti. Understandably, Linnaeus recognized things were getting out of hand and that no one would be able to remember these ridiculously long names. He decided that each name should still be two parts, generic and specific, but that the specific name should be limited to only one word, while the lengthier descriptions should be listed elsewhere. And so, the lengthy Latin name listed above became Plantago media. Much better. This "binomial" (two name) system of naming is still used today, and we still use Latin since it is the classical universal scientific language; each Latin name is held in common around the world, so scientists in different countries can be on the same page when it comes to names and descriptions of life forms, which would otherwise be impossible because of language differences. |Plantago media (Photo credit: Sten Porse via Wikimedia Commons)| This binomial naming system inherently incorporates classification. Plants of the same species are able to reproduce with one another "naturally" (i.e. no human intervention) in the wild and produce viable offspring (seeds that will grow). Plants with the same generic name (aka "genus") are closely related, sharing many characteristics, but they do not reproduce with one another naturally in the wild. The next broader stage of relationship is at the "family" level, which usually includes a number of different related genera (plural of "genus") that share even broader characteristics. So, a plant not only has a name as an identity; its name is an indicator of relationships with the thousands of other plants and organisms in the web of life. |A simplified taxonomic family tree I drew to show how relationship and naming go hand in hand.| These relationships are important to know because plants that are related share similar characteristics. This one fact has a number of very useful practical implications like finding sources for medicinal compounds. If you found an especially effective medicinal compound in a rare plant, you would want to see if other more common species with this same compound exist to serve as sources for the medicine until a synthetic method for production is developed (which can take many years). It is often the case that closely related plants will produce similar compounds, so classification comes in handy here. A good example of this situation is Taxol, a powerful anticancer drug. The active compound in this drug, paclitaxel, was discovered in the bark of the Pacific yew (Taxus brevifolia), which was harvested from wild trees to produce the drug from 1967 to 1993, thereby killing each tree used for this purpose. Since the tree itself did not have a large range, the cancer-fighting drug was not widely available and the Pacific yew's future was in danger. So, scientists began to look elsewhere and discovered that the compound was also found in other species of yew trees and that needles from a common cultivated species (Taxus baccata) could be harvested and used in a semi-synthetic method of drug production, which made the medicine much more widely available and saved the Pacific yew from being harvested to extinction. Hooray for taxonomy! |A branch of Taxus brevifolia, the Pacific Yew (Photo credit: Jason Hollinger via Wikimedia Commons)| |Harvesting the bark for Taxol. (Photo credit: NCI via Wikimedia Commons)| |The finished product. (photo credit: drugdiscovery.com)| Another example regards avoiding poisonous plants. If you grew up with poison ivy (Toxicodendron radicans) in the north and know it is in the family Anicardiaceae, you could save yourself a lot of miserable itching by knowing which other plants are in the same family, like poisonwood (Metopium toxiferum) here in the Keys. Another plant family to learn to identify and avoid is Urticaceae, which contains many species of plants with stinging hairs. Lest you think "stinging hairs" don't sound so bad, check out the following video, which might motivate you to learn a little botany before your next trip through the woods: Classification also helps with conservation efforts, as it gives us an inventory of the world's plant species and tells us how related they are to one another. This information can, for example, help us prioritize which plants to conserve. While no species should be treated as expendable, classification comes in handy if we have to choose between spending our resources to preserve, say, one of the 1,200 species of orchids in the genus Dendrobium or the species Ginkgo biloba, which is not only the only species in the genus Ginkgo, it is the only genus in the family Ginkgoaceae (compared to 880 genera of orchids in the orchid family, Orchidaceae), and is even alone in its own DIVISION (to put that into perspective, another example of a plant division is "flowering plants," which includes about 250,000 species); in short, there is nothing else like it on earth. Conservation efforts don't have unlimited funds, so if we have a good classification of plants, we can try and conserve the most diversity with the limited resources available for these efforts. |A nice specimen of Ginkgo biloba. Photo credit (from Wikimedia Commons)| |The very distinctive leaves of Ginkgo biloba. (Photo credit: James Field via Wikimedia Commons)| Classification also helps when it comes to developing desired traits in plants such as higher fruit yield, disease resistance, or stress tolerance through breeding. While different species in the same genus do not reproduce naturally in the wild due to a number of factors, they will often produce viable seed if pollen from one is introduced to another by a human. To make a hybrid, you would find out which plants are closely related, then select from that group the species that have the qualities you want for breeding. For instance, if you wanted an especially tasty and disease-tolerant citrus tree, you would try to find a species of citrus with great fruit yield / taste and a related species of citrus noted for its disease resistance. Thanks to classification efforts, you can find a list of all the known citrus species in the world, and from there, you can find out which ones are tastiest and which ones have the least disease problems, then try and make some magic happen. This process has recently become very important for Florida farmers with the spread of citrus canker and citrus greening disease; hybridizations are made to produce trees with the best disease resistance and fruit quality. |Not exactly what you want to see on the grocery store shelves.| I hope this post has given you a better idea about some of the practical uses and benefits of plant classification in areas like drug discovery and development, food security, conservation, and the avoidance of painful and/or poisonous plants, which I hope will motivate you to take a closer look at those curious Latin names on our plant labels during your next visit!
<urn:uuid:76b4b3de-7fa0-45bd-a918-268348d747fa>
CC-MAIN-2017-13
http://thebotanicgardensatkonakai.blogspot.com/2014_02_01_archive.html
s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218189734.17/warc/CC-MAIN-20170322212949-00000-ip-10-233-31-227.ec2.internal.warc.gz
en
0.952675
1,583
4.03125
4
The Norway rat (Rattus norvegicus) is a stocky burrowing rodent, unintentionally introduced to North America by settlers who arrived on ships from Europe. into the United States about 1775, this rat has now spread throughout the contiguous 48 states. The Norway rat is found generally at lower elevations but may be found wherever Norway rats live in close association with people. They burrow to make nests under buildings and other structures, beneath concrete slabs, along stream banks, around ponds, in garbage dumps, and at other locations where suitable food, water and shelter are present. On farms they may inhabit barns, granaries, livestock buildings, silos, and kennels. In urban or suburban areas they live in and around residences, in cellars, warehouses, stores, slaughterhouses, docks, and in sewers. Although they can climb, Norway rats tend to inhabit the lower floors of multi-story buildings. Norway rats will eat nearly any type of food. When given a choice, they select a nutritionally balanced diet, choosing fresh, wholesome items over stale or contaminated foods. They prefer cereal grains, meats and fish, nut, and some types of fruit. General Biology, Reproduction and Behavior Norway rats are primarily nocturnal. They usually become active about dusk, when they begin to seek food and water. Some individuals may be active during daylight hours when the rat population is high, when disturbed (weather change, construction, etc.) or when their food source is Rats have poor eyesight beyond three or four feet, relying more on their hearing and their excellent senses of smell, taste and touch. Norway rats are very sensitive to motion up to 30-50 feet away. They are considered essentially colorblind. Rats use their keen sense of smell to locate food items and apparently to recognize other rats. Norway rats rely on their sense of smell to recognize the odors of pathways, members of the opposite sex who are ready to mate, differentiate between members of their own colonies and strangers, and to tell if a stranger is a strong or weak individual. Norway rats use hearing to locate objects to within a few inches. This highly developed sense (combined with their touch sensitivity) can pinpoint someone rolling over in bed to a six inch area. The frequency range of their hearing (50 kilohertz or more) is much higher than that of humans (about 20 kilohertz.) Norway rats have a highly developed sense of touch due to very sensitive body hairs and whiskers which they use to explore their environment. Much of a rodents movement in a familiar area relies heavily on the senses of touch and smell to direct it through time-tested movements learned by exploration and knowledge of its Their sense of taste is excellent, and they can detect some contaminants in their food at levels as low as 0.5 parts per million. This highly developed taste sensitivity may lead to bait rejection if the rodent baits are contaminated with insecticide odors or other chemicals. Norway rats usually construct nests in below-ground burrows or at ground level. Nests may be lined with shredded paper, cloth, or other fibrous material. Litters of 6 to 12 young are born 21 to 23 days after conception. Newborn rats are naked and their eyes are closed, but they grow rapidly, eating solid food at 2 1/2 to 3 weeks. They become completely independent at about 3 to 4 weeks and reach reproductive maturity at 3 months of age, sometimes as early as 8 weeks. Female Norway rats may come into heat every 4 or 5 days, and they may mate within a day after a litter is born. When eliminating Norway rats, remember that glue boards are not very effective on large rodents. Snap traps and live traps will work. The most effective control method for these rats is the use of weather proof bait blocks. For more information on control of Norway rats:
<urn:uuid:25d28b9a-b920-4fb0-a72f-678a5bf91f87>
CC-MAIN-2015-06
http://www.pestproducts.com/norway.htm
s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422120842874.46/warc/CC-MAIN-20150124173402-00179-ip-10-180-212-252.ec2.internal.warc.gz
en
0.957008
821
3.578125
4
Since 2006, U.S. honey bee populations have been in precipitous decline, with some estimates suggesting losses as high as 30% per year.1 While that’s terrible, the problem is far greater than just the destruction of a species. Without bees, a big piece of our food supply is in serious danger. Pollination by honey bees is key in cultivating the crops that produce a full one-third of our food. Scientists have been scrambling to understand the crisis — termed Colony Collapse Disorder — but have yet to find a single, definitive cause. There are likely multiple interacting causes, and mounting evidence suggests that one widely used class of pesticides may be a critical factor. Click here to find out more.
<urn:uuid:d82d68d2-f5d5-48db-a4a3-0ab1ebf605af>
CC-MAIN-2018-43
http://green.myninjaplease.com/?p=1875
s3://commoncrawl/crawl-data/CC-MAIN-2018-43/segments/1539583516071.83/warc/CC-MAIN-20181023044407-20181023065907-00081.warc.gz
en
0.9509
146
3.265625
3
ANSI B11-2008 General Safety Requirements Common to ANSI B11 Machines Risk assessment was introduced and specifically addressed by the American National Standards Institute in ANSI B11.TR3:2000 entitled Risk assessment and risk reduction–A guide to estimate, evaluate, and reduce risks associated with machine tools. The ANSI B11.TR3 explored three major premises of risks: - probability of occurrence—very likely, likely, unlikely, or remote - severity of harm—catastrophic, serious, moderate, or minor - exposure to hazard—frequency and duration, extent of exposure, or number of people exposed ANSI B11.TR3 explains that zero risk does not exist and is therefore unattainable. Some amount of residual risk remains even after the application of machine safeguarding. ANSI’s organization of the B11 series of documents and technical reports can be visualized as a pyramid structure (see Diagram 1). ANSI defines three categories of safety standards: - Type-A standards are BASIS standards. They give basic concepts, principles for design, and general aspects that can be applied to machinery. An example of a Type-A standard is the B11-2008 standard itself. - Type-B standards are GENERIC standards. They deal with one or more safety aspects of one or more types of safeguards that can be used across a wide range of machinery. ANSI B11.19 on safeguarding methods is an example of a Type-B standard. - Type-C standards are MACHINERY SAFETY standards. They deal with detailed safety requirements for a particular machine or group of machines. Examples of Type-C standards are ANSI B11.1 Mechanical Power Presses, ANSI B11.6 Lathes, ANSI B11.9 Grinders, and ANSI B11.15 Tube Benders. The recently published ANSI B11-2008 General Safety Requirements Common to ANSI Machines falls under both Type-A and Type-B standards. and goes further into the subject of risk assessment for machinery, even though it is not mentioned in its title. In the ANSI B11-2008 standard, “the expression ‘acceptable risk’ refers to the level at which further risk reduction will not result in significant reduction in risk; or additional expenditure of resources will not result in significant advantages of increased safety.” Achieving acceptable risk is the major point here. Risk reduction is considered complete when protective measures are applied and acceptable risk has been achieved for the identified hazards. At a minimum, this includes complying with local, regional, and national regulations. ANSI B11-2008 also states that the user and the supplier may have different levels of acceptable risk and that a collaborative effort is required between the two to achieve that goal. It is noted that acceptable risk is fundamentally a business decision made by each supplier or user in the context of its own unique circumstances. To minimize individual biases, a team approach is integral to assessing risk. This optimizes the combined judgment and expertise of individuals familiar with the tasks and hazards of a given machine tool. Four categories of risks are defined in ANSI B11-2008 as: - high risk—only acceptable when all reasonable alternatives/options (protective measures) have been reviewed and deemed impractical or infeasible - medium risk—undesirable but permissible when all reasonable alternatives/options (protective measures) have been formally reviewed and accepted by relevant stakeholders - low risk—usually acceptable - negligible risk—acceptable ANSI B11-2008 compares the classical approach to risk reduction using the hazard control hierarchy to a more progressive approach to risk reduction based on two stages (see Diagram 2). Stage one has two elements with a goal of eliminating the hazard(s) under analysis. Stage two has five elements that must be implemented in a balanced approach to reach the desired goal of reducing all remaining hazards or risks to an acceptable level. Clause 7 deals with specific risk reduction and safeguarding methods that contain the following 23 items: - access to machine controls - control systems - electromagnetic compatibility (EMC) - emergency stops - control of hazardous energy (lockout/tagout) - ergonomics/human factors - handling of machines, component parts, and materials - hydraulic and pneumatic (including vacuum) systems - ladders and platforms - machine tool systems - modified atmospheres - sanitation and hygiene - thermal systems - ventilation of airborne contaminants In essence, the common primary objective in both ANSI B11.TR3 and ANSI B11-2008 is that reasonably foreseeable hazards must be identified and dealt with. Both standards contain scoring/rating systems to establish various risk reduction categories. ANSI B11.TR3:2000 and ANSI B11-2008 are available by visiting the Association of Manufacturing Technology or calling 703-893-2900.
<urn:uuid:4e462275-f24f-409e-bf0f-6129e3751408>
CC-MAIN-2014-49
http://www.rockfordsystems.com/ansi-b11-2008-general-safety-requirements.html
s3://commoncrawl/crawl-data/CC-MAIN-2014-49/segments/1416400381177.56/warc/CC-MAIN-20141119123301-00025-ip-10-235-23-156.ec2.internal.warc.gz
en
0.900677
1,015
3.03125
3
Gibbon species have accumulated an unusually high number of chromosomal changes since diverging from the common hominoid ancestor 15–18 million years ago. The cause of this increased rate of chromosomal rearrangements is not known, nor is it known if genome architecture has a role. To address this question, we analyzed sequences spanning 57 breaks of synteny between northern white-cheeked gibbons (Nomascus l. leucogenys) and humans. We find that the breakpoint regions are enriched in segmental duplications and repeats, with Alu elements being the most abundant. Alus located near the gibbon breakpoints (<150 bp) have a higher CpG content than other Alus. Bisulphite allelic sequencing reveals that these gibbon Alus have a lower average density of methylated cytosine that their human orthologues. The finding of higher CpG content and lower average CpG methylation suggests that the gibbon Alu elements are epigenetically distinct from their human orthologues. The association between undermethylation and chromosomal rearrangement in gibbons suggests a correlation between epigenetic state and structural genome variation in evolution. Mammalian genomes are remarkably stable (with few exceptions). In humans, wrong recombination events occur quite rarely, manifesting themselves in genomic disorders or cancer. On exceptional occasions, the rate of genome evolution has been accelerated by genome-wide reshuffling events giving rise to some highly derivative karyotypes. The genomes of gibbon species (Hylobatidae) are an example of accelerated genome structural evolution; gibbons display a rate of chromosome evolution 10–20 fold higher than the default rate found in mammals (one chromosome change every 4 million years). As we are interested in investigating the possible genetic causes of this phenomenon, we sequenced a considerable number of chromosomal breakpoints in the northern white-cheeked gibbon genome and analyzed the genomic features of these sites. We observe that the gibbon breakpoints are mostly associated with endogenous retrotransposons called Alus, which are normally abundant in the genomes of primates. Furthermore, our analysis revealed that gibbon Alus have a lower content of methylated CpG when compared to the orthologous human Alus. In mammals, CpG methylation is known to be responsible for keeping retrotransposons in a repressed state and protect genome integrity. We therefore suggest that a glitch in the methylation apparatus might have driven the higher genome recombination in gibbons. Citation: Carbone L, Harris RA, Vessere GM, Mootnick AR, Humphray S, Rogers J, et al. (2009) Evolutionary Breakpoints in the Gibbon Suggest Association between Cytosine Methylation and Karyotype Evolution. PLoS Genet 5(6): e1000538. https://doi.org/10.1371/journal.pgen.1000538 Editor: Anne C. Ferguson-Smith, University of Cambridge, United Kingdom Received: February 27, 2009; Accepted: May 26, 2009; Published: June 26, 2009 Copyright: © 2009 Carbone et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Funding: No funding was available to support this work. Competing interests: The authors have declared that no competing interests exist. Gibbons (Hylobatidae) are small arboreal apes that inhabit the tropical and semi-deciduous forests of Southeast Asia and a portion of South- and East-Asia; their closest relatives are the great apes (human, chimpanzee, gorilla and orangutan). They are an excellent model in which to study mechanisms of chromosomal rearrangement during evolution, because their chromosomes have been accumulating changes at an accelerated rate in comparison to other apes –. As a result of this instability, the four genera of the gibbon family possess four different karyotypes (2n from 38 to 52). The genome shuffling observed in gibbons is in striking contrast to the high degree of karyotype conservation found in the other hominoids: there is only a single inter-chromosomal rearrangement separating humans from the great apes , but more than 40 such rearrangements have taken place on the gibbon lineage. Recent estimates based on the inferred karyotype of the common gibbon ancestor suggest that the rate of chromosomal rearrangements in these species is 20 times higher than in other primates . Given the great taxonomic diversity found within the family (four genera and fifteen species), it is tempting to speculate that segregating chromosomal changes mediated the speciation events in a relatively short time. The cause of this abundance of chromosomal changes is still undefined . Primate genomes harbor millions of interspersed repetitive elements , creating numerous opportunities for Non-Allelic Homologous Recombination (NAHR) events to produce deletions, duplications and chromosomal rearrangements. Chromosomal rearrangements caused by NAHR are nevertheless quite rare, and even on an evolutionary time scale mammalian chromosomes have proven to be very stable. Comparison of multiple mammalian karyotypes indicates that the average rate of gross chromosomal rearrangements is only approximately two events over 10 million years . Many repetitive DNA elements are rich in CpGs, which in mammalian cells are typically methylated. CpG methylation is an essential component of epigenetic mechanisms that maintain repetitive elements in a transcriptionally repressed state, thereby suppressing their proliferation ,. Cancer cells frequently exhibit a global decrease in genomic 5-methylcytosine, and it has been speculated that hypomethylation of repeat elements is an underlying factor in the high frequency of chromosomal rearrangements in cancer cells . In search of an explanation for the abundance of evolutionary chromosomal changes in gibbons, we have now characterized the sequence and molecular structure of 57 breakpoint sites in the northern white-cheeked gibbon (Nomascus leucogenys leucogenys, NLE). We had previously created a high-resolution physical map of the break of synteny regions for this species , using the human genome as a reference. This map allowed us to localize the breakpoints within an 80 Kbp range. We have identified an association between the breakpoints and Alu retroelements, and we find that Alu elements in the gibbon are undermethylated in comparison to their human orthologues. Our findings suggest that epigenetic activity of Alu sequences may have facilitated karyotypic evolution and disruption of the uniform rate of chromosomal changes in gibbon species. Identification and sequencing of 57 gibbon breakpoints To identify the breakpoints at the sequence level we selected 80 Bacterial Artificial Chromosome (BAC) from the Nomascus leucogenys leucogenys (NLE) genomic BAC library (CHORI-271) spanning the breakpoints of translocations and inversions. These BACs were selected from a high-resolution map that we constructed and from a complementary list of gibbon BACs identified as spanning breakpoints by BAC End Sequencing (BES). Out of these 80 BACs, 23 were sequenced using a shotgun approach and assembled to high quality sequence (Table S1). The final assembled sequences were individually aligned by BLAT to the most recent human genome assembly (hg18), and we identified the breakpoints between human and gibbon at the base pair level (Table S2). As we sequenced the BACs, we discovered multiple breakpoints inside the same clone in 8 of the 23 BACs including two cases previously reported by us. The complex structure of three of these BACs may be explained by their centromeric location in the gibbon (Table S2). In a few instances the presence of human segmental duplications did not allow for unambiguous mapping. To enrich our breakpoint dataset in a cost effective way, we pooled and shotgun sequenced at lower coverage the remaining 57 gibbon BACs (Protocol S1). This approach added 33 breakpoints to our dataset: 25 at the base pair level and 7 at the resolution of a small insert clone (about 6 Kbp) (Table S2). The remaining breakpoints could not be identified, due either to densely repeated regions or to lack of coverage. This brought the final number of breakpoints to 68 (57 at the base pair level). These results indicate that the frequency of breakpoints is higher than BES mapping alone can estimate. Hence assembly of the gibbon genome will be necessary to pinpoint all the breakpoints. Gibbon breakpoints show overlap with human- and gibbon-specific segmental duplications In a previous study we uncovered a significant association between gibbon break of synteny regions, identified at a resolution of 80 Kbp, and human segmental duplications (hSD) : 42% of the breakpoints were found to overlap with at least one hSD. In the current study we were able to identify the breakpoints at a higher resolution. This allowed us to further examine their relationship with SDs by measuring the correlation between a 1 Kbp window (−/+500 bp) including the breakpoint mapped on human and hSDs. We found that 15% of the breakpoints overlap with at least one hSD, which is significant (p = 0.0002) based on a random sampling simulation (performed as described in Materials and Methods) (Figure S1A). Recent studies have shown that a burst in duplication occurred in humans and chimpanzees after their divergence from other hominoids –. Thus we assume that the hSDs do not always correspond to gibbon SDs (gSD). As an assembled gibbon genome is not yet available, we used two methods to identify gSDs. First, we performed array-comparative genomic hybridization (array-CGH) of gibbon genomic DNA against human genomic DNA. This experiment allowed identification of large (>300 Kb) duplications/deletions that distinguish the two species. Second, following the method described by Bailey et al. , we mapped gibbon reads from the Trace Archives onto the human genome and identified putative gSD regions by detecting a higher depth of coverage by the reads (supporting online material). Of the gSDs identified by array-CGH, 37% were also identified as putative gSDs based on read coverage. Using the random sampling simulation approach mentioned above, we noticed that the overlap between the gibbon breakpoints and the gSDs is extremely large and statistically significant (Figure S1B), more than the overlap observed for the hSD. Examples of gibbon segmental duplications in breakpoints that could be detected by FISH are shown in Figure 1A. Even though the array-CGH and read-coverage-based gSD datasets do not show exact correspondence, we observed a significant correlation (p = 7.63e-9 by Fisher's Exact Test). We also validated, by Fluorescence in situ Hybridization (FISH), 11 duplications and 11 deletions identified by both methods (Figure 1B). Of note, the array-CGH results showed an excess of deletions in the gibbon relative to human (data not shown). We verified that 30% of the deletions are regions that are present in human at a higher copy number than in gibbon, confirming the occurrence of abundant human-specific duplication events. (A) Fluorescence in situ hybridization (FISH) experiments on NLE nuclei and metaphases using gibbon BACs spanning breakpoints which overlap with gSD. The fluorescent signals show a pattern typical of repeated sequences; (B) Images from the Array-CGH experiment using gibbon (test) versus human (reference) genomic DNA. Human chromosomes 2 and 3 are shown; duplications are represented in green and deletions in red. The duplications were validated by FISH on metaphases and nuclei of both human and gibbon using as probes the human BACs from the 32Kset. Duplicated regions present a higher depth of coverage of Trace Archives reads on the human genome as illustrated in the lateral panels. Gibbon breakpoints disrupt genes We looked at the relationship between breakpoints and genes. When mapped onto the human genome, 53% (36 out of 68) of the breakpoints occur within a gene and 19% occur within non-coding transcripts (Table S2). We hypothesize that when a breakpoint disrupts a gene, the selective pressure on the sequence should be reduced as a consequence of loss of function, unless the truncated protein is rescued and still functional. As a measure of relaxed selective constraint on these disrupted genes, we calculated the dN/dS ratio between non-synonymous (dN) and synonymous (dS) substitutions between human and gibbon (using macaque as the outgroup). This analysis was carried out only on the 23 fully sequenced BACs (Protocol S2). The same method was applied to an equal number of randomly selected gibbon BACs sequenced by the NIH intramural sequencing center (NISC) comparative vertebrate sequencing project (Table S3). This analysis showed a significant increase (p = 0.01, Mann-Whitney's U test) in the dN/dS ratio of gibbon genes when the breakpoint BACs are compared to the NISC BACs (Figure S2). It is worth noting that the p value becomes even smaller when the genes at <50 Kbp distance are considered, indicating a possible position effect. To confirm this trend, we sampled additional gibbon genes located at 500 Kbp and 1 M bp from the breakpoints, and found no differences when gibbon was compared to macaque (Table S4). Frequently, genes affected by the breakpoints are part of clusters: the ABCC family on HSA 16, the ABCA family on HSA 17, the growth hormone cluster on HSA 17, RFPL on HSA 22, MUC4 and MUC20 on HSA3, PLSCR (phospholipid scramblase) on HSA 3. The association between breakpoints and gene-clusters has at least two biological implications. First, gene clusters result from duplication events that may cause genome instability through NAHR. Second, the presence of other genes with redundant functions could mitigate natural selection against chromosomal rearrangements that disrupt genes. Gibbon breakpoints are enriched in interspersed and simple repeats The role of repeats in evolutionary or disease-causing chromosomal rearrangements is well documented –. We identified repeats within 150 bp of the 57 sequenced breakpoints with Repeat Masker. 81% of the breakpoints co-localized with at least one interspersed repeat. Alus and L1 LINEs are the most frequently represented, followed by simple repeats, as illustrated in Table 1. In 11 instances, one or more repeats span the breakpoint site in the gibbon. This could result either from an insertion after the breakage, or from a recombination event (Figure 2A). In the remaining cases, the repeats flank the breakpoint, and they are frequently truncated by the rearrangement event. Moreover, three breakpoints are next to blocks of repeats that were inserted sequentially in the gibbon genome, creating complex arrangements (Table S2). (A) Two examples of Alu–Alu mediated recombination events in the gibbon discovered by comparing the gibbon and great apes orthologous locations. In the example 1 (clone CH271-389E1) the AluY and the AluS on human chromosome 17 (HSA17) share high homology in two locations. In gibbon the AluS was broken as result of the inversion and the AluY was lost. A simple scenario is illustrated in example 2: two identical Alus located at the breakpoint boundaries on human chromosomes 2 and 17 (HSA2 and HSA17) recombined and most likely caused the translocation whose breakpoints was identified in clone CH271-262E11; (B) distance from breakpoints to Alus showing a decline in Alu content when moving from the breakpoint. (C) the proportion of CpGs per 100 bp is higher for Alus or Alu fragments closer to the breakpoints. Out of the 57 breakpoints, 11 co-localize with simple repeats of various types. Most of these breakpoints (6 out of 11) overlap with (AT)n-rich repeats which are either gibbon specific (CH271-254H12, CH271-171B20 and CH271-122E24) or shared by human (CH271-228C1, CH271-86M19, CH271-40A18). A different case is the breakpoint of a translocation HSA 3–5 that falls in the intra-genic tandemly repeated region (TR) of the mucin gene MUC4 (3q29). Analysis of Alu CpG content and methylation We were intrigued by the predominance of Alus at the breakpoint sites, as Alu-Alu recombination events have been reported as examples of Non Allelic Homologous Recombination (NAHR) ,. We verified that the proportion of Alus associated with breakpoints was significant when compared to other repeats by using a random sampling simulation (Figure S3A) (p = 0.001). At the same time this method showed that the association with LINE L1 in human was lower than expected by chance (Figure S3B). We then looked for features of Alus that may be distinctive in gibbon compared to other hominoids. To carry out this analysis we used the 23 assembled BACs to represent portions of the gibbon genome surrounding the breakpoints. First, we observed a decline in Alu density within the BACs with increasing distance from the breakpoints (Figure 2B). Furthermore, Alu fragments at or near (<150 bp) the breakpoints were almost twice as CpG-rich as the remaining Alu sequences in the same BAC (4.5 CpGs/100 bp compared to 2.4 CpGs/100 bp; t-test p<0.001). As shown in Figure 2C, the number of CpG doublets per 100 bp of Alu sequence declines rapidly as the distance from the breakpoint increases. Active Alus contain a relatively high number of CpG dinucleotides, which are linked to active retrotransposition . Normally, the epigenetic apparatus of the cell suppresses the activity of retrotransposons by adding methyl groups to cytosines in CpGs ,. Methyl-C tends to decay to T or A (therefore CpG become TpG/CpA) through a process known as CpG decay . Our data thus suggest a higher concentration of “active” Alus associated with breakpoints. We hypothesized that the higher rate of chromosomal breakage observed in gibbons is due to an active epigenetic state of these elements in the gibbon as compared to the common ancestor of the hominoids; the higher CpG content of these Alus suggests that they have been less methylated and consequently that they may have a different epigenetic state. The hypothesis predicts reduced CpG methylation of the gibbon breakpoint Alus in comparison to their human orthologues. We tested this prediction by performing bisulfite allelic sequencing of 14 orthologous Alus in human and gibbon, 8 of which were located near the breakpoint sites (<150 bp from the breakpoint) and 6 Alus outside of breakpoint regions but with similar CpG content to the breakpoint Alus (Materials and Methods and Table S5). As orthologous Alus are inserted in the genome of the common ancestor, we can safely assume that the CpG groups had the same amount of time to be methylated. Our results (Figure 3) demonstrate a significant reduction of CpG methylation in gibbon compared to human (p<0.001, Mann-Whitney U test). Alus at orthologous locations in human and gibbon would have been inserted into the genome of the common ancestor and would therefore be the same age in the two lineages. Even though the Alus are the same age, there is a difference in the methylation levels at the CpG sites skewing towards lower methylation in the gibbon. One exception is the Alu D (CH271-263C9) which shows lower methylation in human. Gibbon species carry an extraordinary number of chromosomal rearrangements, accumulated in a relatively short evolutionary time (15–18 mya). In order to uncover a possible genetic source for the genomic reshuffling observed in these species, we carried on a detailed analysis of 57 sequenced synteny breakpoints between the northern white cheeked gibbon (NLE) and human. Our molecular analysis revealed a scenario which, at a first glance, is similar to that described in other primates , where segmental duplications and repeats play a major role in chromosomal rearrangements (Figure 4). But a broader analysis, which took into account epigenetic modifications, uncovered a possible explanation for the high frequency of evolutionary chromosomal changes. The gibbon breakpoints are associated with Alu elements that have an unusually high CpG content, and in the gibbon these Alu elements are less methylated than their human orthologues. This may indicate that the epigenetic state of these Alus has predisposed them to recombination. This visualization was generated using Circos software (http://mkweb.bcgsc.ca/circos/). The lines in the inner circle represent inter-chromosomal (red) and intra-chromosomal (blue) rearrangements in gibbon relative to human. The outer circles provide genomic context. The outermost circle displays human chromosomes along with genomic coordinates and G-banding stains (NCBI Build 36.1). Purple lines represent human segmental duplications from the UCSC Segmental Dups Track. (http://genome.ucsc.edu/cgi-bin/hgTrackUi?g=genomicSuperDups). Orange lines represent gibbon segmental duplications we predicted based on read coverage. Green lines represent human genes from the UCSC RefSeq Genes Track (http://genome.ucsc.edu/cgi-bin/hgTrackUi?g=refGene. In this study we were able to confirm the correlation between breakpoints and human SD which we had reported previously . The higher resolution achieved in the present study, and the availability of gibbon sequences, allowed us to confirm association of the breakpoints with gibbon-specific SDs. As many breakpoints could not be mapped, due to the presence of these duplications (Table S2), the overlap is very likely to be more frequent than we have been able to demonstrate. It is noteworthy that we found only two breakpoints where SDs were present in both gibbon and human. As the intersection between gSD and hSD over the whole genome is much higher (32%), this observation suggests that the chromosomal rearrangements are mainly associated with “species-specific” duplications. The two cases of breakpoints in shared duplications may be explained by independent reuse of a breakpoint in regions susceptible to rearrangements . Nevertheless, we do not believe that SDs can be considered an underlying cause of the breakpoints, as we have only few examples of erroneous recombination events in these regions (Table 2). Very similar observations have very recently been reported . When studying evolutionary chromosomal rearrangement, it is tempting to search for sign of selection on genes that have been disrupted by the breakages. Recent work by Girirajan et al. found evidence that 3 of their 11 genes disrupted by breakpoints exhibited signatures of relaxed evolutionary constraint (average dN/dS = 1.09). Our approach was different, as we looked at all the genes within the sequenced BACs, and compared them with randomly selected regions of the gibbon genome. We did, however, identify 5 genes in our sample that are disrupted by breakpoints and for which we had adequate coverage. Although we found that their values of dN/dS (average dN/dS = 0.56) were not as high as those reported by Girirajan et al. , subsequent analysis on the remaining dataset of all non-disrupted genes located within 50 Kbp of a break point, revealed a significantly reduced difference of dN/dS between gibbons and macaques (from p = 0.001 to p = 0.06) (Table S4). We hypothesize that some of these genes may still be functional, perhaps producing a smaller transcript, and that some may have become non-functional recently enough that non-synonymous substitutions have not had a chance to accumulate. Nevertheless, it appears that there are position effects on genes near to but not interrupted by breakpoints, perhaps due cis effects of chromatin in the breakpoint region, leading to changes in expression. A genome-wide expression assay would be needed to define the major trend for the genes that have been disrupted but this approach may be complicated by the scarcity of tissues available from this endangered species. Breakage regions were found to co-localize with repeats. Whereas the known link between simple repeats and fragile genomic regions makes this observation intriguing, it is difficult to predict a cause-effect relationship between these repeats and the gibbon breakpoints. For many breakpoints we could readily observe that simple repeats were the result of gibbon-specific insertions by the repair mechanism after the break occurred. We therefore defined them as “filling” (Table 2) and we can assume that they followed the double-strand breaks. Our data point to a role for both Non-Allelic Homologous Recombination (NAHR) and non-homologous end joining (NHEJ) in double-strand break repair, with a prevalence of the latter. In 9 cases NAHR was driven by either Alu-Alu or SD mediated recombination (Table S2). In additional 15 cases, where long stretches of homology were not detected, we observed micro-homology or “filling” sequences which are both signs of NHEJ . In NHEJ the double-strand breaks are fused together without a requirement for extensive homology. For the remaining breakpoints it was not possible to pinpoint a mechanism, even though the absence of homology would lead us to speculate that NHEJ or some other complex mechanism occurred in most of them . While seeking a mechanism associated to the chromosomal reshuffling of gibbon species, our approach was to investigate Alu elements in more detail, given their higher concentration at the breakpoints. Independent evidence shows that this family of retrotransposons is particularly active in gibbons , strengthening our hypothesis. Our in silico and experimental data suggested that CpG cytosines in Alus are less methylated in gibbon than in human. CpG methylation has a major role in epigenetic suppression of endogenous retroelements in mammals. If this mechanism is attenuated, the repeated DNA sequences may threaten genome integrity: demethylation leading to an open chromatin structure at repeated sequences may cause structural and numerical variations . Multiple examples of correlations between methylation state and genome structural variation have recently come to light in cancer cells, where disrupted methylation patterns are common ,. Furthermore, it was recently observed that hypomethylated blocks in tumor cell lines correspond to fragile regions of the genome and synteny breakpoints in the mouse . This correlation suggests a common source of instability independent from genomic sequence and related to the epigenetic state of the DNA. O'Neill and colleagues showed that the genome of a hybrid between two species of Australian wallaby (marsupials) was hypomethylated when compared to the parental species . In these hybrids a hypomethylated retroviral element was abnormally replicated causing an evident centromeric expansion. The same group also reported double-minute chromosome formation in mouse interspecific hybrids (M. musculus×M. caroli) . Together with our findings this observation indicates that changes in methylation levels may explain perturbations of the uniform rate of genome evolution. Other mammal species (dog, mouse and rat), display very rearranged karyotypes and it will be important to investigate if the scenario we described in the gibbon is common to these species as well. Nevertheless, at the moment, the resolution of the synteny breakpoints for these species is still very far from the one needed to carry out an analysis comparable to the one we performed on the gibbon genome. We have presented here a scenario that may explain the genome reshuffling observed in gibbon species: hypomethylation of certain Alu elements may predispose them to recombination. We are currently investigating the magnitude of the genome hypomethylation in gibbon repeat elements, and whether repeats other than Alu are involved. At the moment we can only speculate about the possible causes of the difference in levels of methylation of Alus that we observed in the gibbon. One hypothesis is linked to the observation that CpG methylation is disrupted in hybrids ,. Population genetics theories propose that speciation may occur after hybrid recombination, followed by inbreeding and reproductive isolation due to the new genetic make-up. This idea is well accepted for plants, and it has recently been proposed for gibbon species . Hybridization may have gradually disturbed the apparatus responsible for the methylation of repeats in the hybrids, leading to higher numbers of chromosomal rearrangements . Very recently the implications of a specific class of small RNAs (piRNAs) in methylation of repeats have been discovered. A rapid divergence of these sequences during speciation could therefore explain the reduction in the cytosine methylation efficiency in cross-species hybrids . Materials and Methods Random sampling simulation The statistical analysis of the breakpoints repeat and duplication content was performed with the help of a C# application written in-house . Tracks of genome-wide repeat content for different subcategories of repeats and for segmental duplications content were prepared for input to the simulation software using data from http://genome.ucsc.edu human genome (hg18 release). The measure we used counted up the existence of at least one element of the corresponding track in each region of the set and returned a detailed report for the set. To attain the simulation, the program reallocates randomly al the regions maintaining the chromosome of origin and size as the initial counterpart. The same measurements were taken for each random set after a reiteration of 5,000 times. The resulting sampling distribution was then plotted to compare the original set of regions with the global genomic landscape. The track relative to gibbon specific segmental duplication was built as result of our in silico analysis of the trace archives. Subsequently the latter tracks were used in order to perform different overlap measurements with the set of 57 breakpoints. When mapped on the human genome and the regions with ambiguous mapping are removed, the dataset corresponds to 120 regions of about 500 bp size (on average). Another set which we called “stringent set” was also used to determine the overlap with hSD. In this set all the breakpoints form two BACs (CH271-298N13 and CH271-372B11) known to be centromeric in the gibbon and containing multiple breakpoints, were excluded. In silico segmental duplication detection Gibbon reads were downloaded from the NCBI Trace Archives and screened for quality. A total of 24,350,447 reads that passed quality screening were mapped to the human genome (build NCBI 36.1, UCSC hg18) using Pash ,. In order to remove highly ambiguous mappings, reads mapping to >500 locations with a score within 6% of its top mapping score were removed from consideration. Furthermore, reads that overlapped by >75% with repeats, as identified by RepeatMasker , were removed from consideration. A total of 15,518,707 mapped reads remained after filtering. Putative gibbon segmental duplications were identified following the method outlined in Bailey, et al., 2002 . The number of gibbon mapped reads was determined in 5000 bp windows across the human genome. The mean (31.11) and standard deviation (18.75) of mapped read counts was calculated across windows not overlapping with human segmental duplications. A read count cutoff of 3 standard deviations from the mean was applied meaning any 5000 bp region with >87 mapped reads was identified as a putative gibbon segmental duplication. This resulted in 1630 identified gibbon segmental duplications. 32,855 BACs, spanning 95% of the human euchromatic genome, have been assembled and re-arrayed into 384-well microtiter dishes ,. DNA was purified, amplified using the DOP-PCR method, and spotted on CMT-GAPS coated glass slides (Corning, UltraGaps). Genomic DNA from NLE was obtained from blood and anonymous human reference DNA was obtained from Children's Hospital Oakland Research Institute. Labeling and hybridization were performed essentially as described by . Hybridization images were generated by scanning the slides on a 4000B scanner (Axon). The images were first processed using GenePix Pro 5.1 (Axon Instruments). The primary experimental data (GenePix Results files) were subjected to fully standardized data-analysis (flagged spots removal, background subtraction and loess normalization) by uploading them to the BASE micro-array analysis software installation which performs standard normalization. The final output was human chromosome specific plots of Log2ratio values vs chromosome location as well as a whole genome view. Fluorescence in situ hybridization (FISH) Chromosome preparations were obtained from peripheral blood following standard procedures. Briefly, blood was incubated with cell culture media and phytohemagglutinin (GIBCO) for 72 hours (37°C, 5% CO2). Colcemid was then added (final concentration 0.05 ug/ml) and cells were harvested after a 1 hour incubation. Cells were spun down by centrifugation, the media was discarded and the pellet was resuspended in 8 ml of hypotonic solution. After incubating for 20 minutes, the standard fixative solution (1 part Acetic Acid, 3 parts Methanol) was added and cells were centrifuged at 2500 rpm for 5 minutes. The pellet was washed with fixative solution and cells were kept at 4°C overnight. DNA from BACs was extracted using PureLink Miniprep kit (Invitrogen, Cat#K2100-10). FISH experiments were performed essentially as described by Lichter et al. . BACs were labeled either with Cy3-dUTP or FITC-dUTP by standard nick-translation assay. Images were acquired using Nikon 80i microscope, equipped with CCD camera Cool Snap HQ2 (Photometrics) and software Nis Elements Br (NIKON). Elaboration of the images was done using Photoshop. Bisulfite allelic sequencing of Alu elements Primers for 14 Alus (Table S5) were designed using “MethPrimer” (http://www.urogene.org/methprimer/) making sure to target unique sequences flanking the Alu. Out of the 14 Alus, 8 were near the breakpoints (<150 bp); as our goal was to amplify Alus orthologous in human and gibbon, we had to take into account the synteny between human and gibbon and had to eliminate all the cases where the Alus were located across the breakpoint. The remaining Alus were located randomly in the gibbon genome but had to have a CpG content high enough to allow us to make a statistic. The genomic DNA from whole-blood from gibbon and human was bisulfite converted using EpiTect Bisulfite kit (Qiagen, Cat.# 59104PCR) and the amplification was performed using the FastStart Taq DNA Polymerase (Roche, Cat#12032929001). PCR products were purified and cloned using TA-cloning procedures (Qiagen PCR cloning Kit, Cat.# 231124). We sequenced 12 clones for each Alu in order to have fair representation of all the alleles. Random sampling simulations for human and gibbon segmental duplications. Random sampling simulations were carried on as described in Materials and Methods. Histograms were obtained for human SD (A) and the in silico set of gibbon SD (B). We also tested the overlap with a “stringent” sample (lighter color) where all the BP that in gibbon overlap with centromeres were removed. Even in this case it is evident that the overlap of the gibbon sample with both classes of SDs is significant. (1.42 MB TIF) dN/dS ratios for gibbon and macaque genes. The ratio of the average dN/dS compute for gibbons (vs. human) and macaques (vs. human) for all genes found within the fully sequenced BACs (Total), for genes found within the NISC database (NISC), for genes located within 50 kb (<50 kb) from the breakpoint found within the BAC sequences and genes located further than 50 kb (>50 kb) from the breakpoint found within the BAC sequences. p values were calculated using the nonparametric Mann-Whitney test. (8.62 MB TIF) Random sampling simulations for Alu and Line 1 elements. Random sampling simulations were carried on as described in Materials and Methods. The two charts show the histogram resulted from counting the overlap between random regions of the human genome and Alu (A) and Line (B). The random sampling was repeated 5,000 times in both cases. The corresponding value for the gibbon dataset is indicated by the blue arrow. (1.51 MB TIF) Sequenced Gibbon BACs. The fully sequenced and assembled 23 gibbon BACs from the genomic BAC library CHORI-271 (http://bacpac.chori.org/library.php?id=228) are reported here with the corresponding accession numbers. (0.04 MB DOC) Gibbon breakpoints mapping information and annotations. We report here the sequencing status and annotations for 80 BACs used in this study. All the BACs were End-sequenced and mapped on the human assembly (hg18). The breakpoint sequence was obtained for 46 BACs whereas for 7 clones we identify Trace Archives mate pairs whose mapping indicated the presence of the breakpoint but no breakpoint sequence was found. For 24 clones the breakpoints could not be narrowed down at a resolution higher than a BAC clone. For each breakpoint sequence repeat, segmental duplications and gene contents are annotated. Abbreviations: gSD = gibbon segmental duplication, hSD = human segmental duplication. (0.11 MB PDF) Estimates of dN/dS for each gene within the BAC sequences. The tables report the estimates of dN/dS for genes within the NISC BACs (left-most table) and for the fully sequenced gibbon BACs (center table). The small table on the right illustrates the averaged estimates of dN/dS for genes where human, gibbon and macaque sequences were available. (*) Hypothetical genes were removed to provide a more stringent analysis and genes found within BAC CH271-262e11 were omitted because they belong to a gene cluster family with high sequence identity. This sequence identity along with their position upstream and downstream of the BP leads to uncertainty in alignment data and gene coordinates. (0.04 MB XLS) Estimates of dN/dS for genes located within specified distances from the breakpoint tested in the study. Estimates of dN/dS for each gene found within 400–600 kb from the BP where human, gibbon and macaque sequences were available (left-most table). Estimates of dN/dS for each gene found within 0.9–1.1 Mb from the BP where human, gibbon, and macaque sequences were available (center table). The small table on the right side summarizes the average dN/dS values at various distances from the BP tested in the study. Distances were determined by calculating the minimum distance from either the start or end of the gene to the BP. (0.03 MB XLS) Primers used for Bisulfite Allelic sequencing. The table lists the primers that have been used to amplify the 14 Alus and carry on allelic bisulfite sequencing. (0.03 MB XLS) Breakpoints mapping strategy. (0.03 MB DOC) We would like to acknowledge Joseph Dhahbi and Dario Boffelli from Children's Hospital & Research Center Oakland for their valuable suggestions and Deepti Rokkam for her contribution. The whole genome shotgun sequencing of the northern white cheeked gibbon was performed at the Washington University Genome Sequencing Center and the Baylor College of Medicine Human Genome Sequencing Center. Conceived and designed the experiments: LC PJdJ. Performed the experiments: LC. Analyzed the data: LC RAH GMV SKK JDW JJ AM. Contributed reagents/materials/analysis tools: RAH ARM SH JR DM. Wrote the paper: LC DM. - 1. Jauch A, Wienberg J, Stanyon R, Arnold N, Tofanelli S, et al. (1992) Reconstruction of genomic rearrangements in great apes and gibbons by chromosome painting. Proc Natl Acad Sci U S A 89(18): 8611–8615.A. JauchJ. WienbergR. StanyonN. ArnoldS. Tofanelli1992Reconstruction of genomic rearrangements in great apes and gibbons by chromosome painting.Proc Natl Acad Sci U S A89(18)86118615 - 2. Muller S, Hollatz M, Wienberg J (2003) Chromosomal phylogeny and evolution of gibbons (Hylobatidae). Hum Genet 113(6): 493–501.S. MullerM. HollatzJ. Wienberg2003Chromosomal phylogeny and evolution of gibbons (Hylobatidae).Hum Genet113(6)493501 - 3. Misceo D, Capozzi O, Roberto R, Dell'oglio MP, Rocchi M, et al. (2008) Tracking the complex flow of chromosome rearrangements from the Hominoidea Ancestor to extant Hylobates and Nomascus Gibbons by high-resolution synteny mapping. Genome research 18(9): 1530–1537.D. MisceoO. CapozziR. RobertoMP Dell'oglioM. Rocchi2008Tracking the complex flow of chromosome rearrangements from the Hominoidea Ancestor to extant Hylobates and Nomascus Gibbons by high-resolution synteny mapping.Genome research18(9)15301537 - 4. Yunis JJ, Prakash O (1982) The origin of man: a chromosomal pictorial legacy. Science (New York, NY 215(4539): 1525–1530.JJ YunisO. Prakash1982The origin of man: a chromosomal pictorial legacy.Science (New York, NY215(4539)15251530 - 5. Girirajan S, Chen L, Graves T, Marques-Bonet T, Ventura M, et al. (2009) Sequencing human-gibbon breakpoints of synteny reveals mosaic new insertions at rearrangement sites. Genome research 19(2): 178–90.S. GirirajanL. ChenT. GravesT. Marques-BonetM. Ventura2009Sequencing human-gibbon breakpoints of synteny reveals mosaic new insertions at rearrangement sites.Genome research19(2)17890 - 6. Gibbs RA, Rogers J, Katze MG, Bumgarner R, Weinstock GM, et al. (2007) Evolutionary and biomedical insights from the rhesus macaque genome. Science 316(5822): 222–234.RA GibbsJ. RogersMG KatzeR. BumgarnerGM Weinstock2007Evolutionary and biomedical insights from the rhesus macaque genome.Science316(5822)222234 - 7. Fronicke L, Wienberg J, Stone G, Adams L, Stanyon R (2003) Towards the delineation of the ancestral eutherian genome organization: comparative genome maps of human and the African elephant (Loxodonta africana) generated by chromosome painting. Proceedings 270(1522): 1331–1340.L. FronickeJ. WienbergG. StoneL. AdamsR. Stanyon2003Towards the delineation of the ancestral eutherian genome organization: comparative genome maps of human and the African elephant (Loxodonta africana) generated by chromosome painting.Proceedings270(1522)13311340 - 8. Bourc'his D, Bestor TH (2004) Meiotic catastrophe and retrotransposon reactivation in male germ cells lacking Dnmt3L. Nature 431(7004): 96–99.D. Bourc'hisTH Bestor2004Meiotic catastrophe and retrotransposon reactivation in male germ cells lacking Dnmt3L.Nature431(7004)9699 - 9. Yoder JA, Walsh CP, Bestor TH (1997) Cytosine methylation and the ecology of intragenomic parasites. Trends Genet 13(8): 335–340.JA YoderCP WalshTH Bestor1997Cytosine methylation and the ecology of intragenomic parasites.Trends Genet13(8)335340 - 10. Wilson AS, Power BE, Molloy PL (2007) DNA hypomethylation and human diseases. Biochimica et biophysica acta 1775(1): 138–162.AS WilsonBE PowerPL Molloy2007DNA hypomethylation and human diseases.Biochimica et biophysica acta1775(1)138162 - 11. Carbone L, Vessere GM, ten Hallers BF, Zhu B, Osoegawa K, et al. (2006) A high-resolution map of synteny disruptions in gibbon and human genomes. PLoS Genet 2(12): e223.L. CarboneGM VessereBF ten HallersB. ZhuK. Osoegawa2006A high-resolution map of synteny disruptions in gibbon and human genomes.PLoS Genet2(12)e223 - 12. Kent WJ (2002) BLAT–the BLAST-like alignment tool. Genome research 12(4): 656–664.WJ Kent2002BLAT–the BLAST-like alignment tool.Genome research12(4)656664 - 13. Perry GH, Yang F, Marques-Bonet T, Murphy C, Fitzgerald T, et al. (2008) Copy number variation and evolution in humans and chimpanzees. Genome research 18(11): 1698–1710.GH PerryF. YangT. Marques-BonetC. MurphyT. Fitzgerald2008Copy number variation and evolution in humans and chimpanzees.Genome research18(11)16981710 - 14. She X, Liu G, Ventura M, Zhao S, Misceo D, et al. (2006) A preliminary comparative analysis of primate segmental duplications shows elevated substitution rates and a great-ape expansion of intrachromosomal duplications. Genome research 16(5): 576–583.X. SheG. LiuM. VenturaS. ZhaoD. Misceo2006A preliminary comparative analysis of primate segmental duplications shows elevated substitution rates and a great-ape expansion of intrachromosomal duplications.Genome research16(5)576583 - 15. Marques-Bonet T, Kidd J, Ventura M, et al. (2009) T. Marques-BonetJ. KiddM. Ventura2009 - 16. Bailey JA, Gu Z, Clark RA, Reinert K, Samonte RV, et al. (2002) Recent segmental duplications in the human genome. Science 297(5583): 1003–1007.JA BaileyZ. GuRA ClarkK. ReinertRV Samonte2002Recent segmental duplications in the human genome.Science297(5583)10031007 - 17. Kehrer-Sawatzki H, Sandig CA, Goidts V, Hameister H (2005) Breakpoint analysis of the pericentric inversion between chimpanzee chromosome 10 and the homologous chromosome 12 in humans. Cytogenet Genome Res 108(1–3): 91–97.H. Kehrer-SawatzkiCA SandigV. GoidtsH. Hameister2005Breakpoint analysis of the pericentric inversion between chimpanzee chromosome 10 and the homologous chromosome 12 in humans.Cytogenet Genome Res108(1–3)9197 - 18. Kehrer-Sawatzki H, Sandig C, Chuzhanova N, Goidts V, Szamalek JM, et al. (2005) Breakpoint analysis of the pericentric inversion distinguishing human chromosome 4 from the homologous chromosome in the chimpanzee (Pan troglodytes). Hum Mutat 25(1): 45–55.H. Kehrer-SawatzkiC. SandigN. ChuzhanovaV. GoidtsJM Szamalek2005Breakpoint analysis of the pericentric inversion distinguishing human chromosome 4 from the homologous chromosome in the chimpanzee (Pan troglodytes).Hum Mutat25(1)4555 - 19. Kehrer-Sawatzki H, Cooper DN (2008) Molecular mechanisms of chromosomal rearrangement during primate evolution. Chromosome Res 16(1): 41–56.H. Kehrer-SawatzkiDN Cooper2008Molecular mechanisms of chromosomal rearrangement during primate evolution.Chromosome Res16(1)4156 - 20. Lupski JR, Stankiewicz P (2005) Genomic disorders: molecular mechanisms for rearrangements and conveyed phenotypes. PLoS Genet 1(6): e49.JR LupskiP. Stankiewicz2005Genomic disorders: molecular mechanisms for rearrangements and conveyed phenotypes.PLoS Genet1(6)e49 - 21. Shaw CJ, Lupski JR (2005) Non-recurrent 17p11.2 deletions are generated by homologous and non-homologous mechanisms. Hum Genet 116: 1–7.CJ ShawJR Lupski2005Non-recurrent 17p11.2 deletions are generated by homologous and non-homologous mechanisms.Hum Genet11617 - 22. Jurka J (2004) Evolutionary impact of human Alu repetitive elements. Curr Opin Genet Dev 14: 603–608.J. Jurka2004Evolutionary impact of human Alu repetitive elements.Curr Opin Genet Dev14603608 - 23. Xing J, Hedges DJ, Han K, Wang H, Cordaux R, et al. (2004) Alu element mutation spectra: molecular clocks and the effect of DNA methylation. J Mol Biol 344: 675–682.J. XingDJ HedgesK. HanH. WangR. Cordaux2004Alu element mutation spectra: molecular clocks and the effect of DNA methylation.J Mol Biol344675682 - 24. Bailey JA, Baertsch R, Kent WJ, Haussler D, Eichler EE (2004) Hotspots of mammalian chromosomal evolution. Genome Biol 5: R23.JA BaileyR. BaertschWJ KentD. HausslerEE Eichler2004Hotspots of mammalian chromosomal evolution.Genome Biol5R23 - 25. Inoue K, Osaka H, Thurston VC, Clarke JT, Yoneyama A, et al. (2002) Genomic rearrangements resulting in PLP1 deletion occur by nonhomologous end joining and cause different dysmyelinating phenotypes in males and females. Am J Hum Genet 71: 838–853.K. InoueH. OsakaVC ThurstonJT ClarkeA. Yoneyama2002Genomic rearrangements resulting in PLP1 deletion occur by nonhomologous end joining and cause different dysmyelinating phenotypes in males and females.Am J Hum Genet71838853 - 26. Nakayama K, Ishida T (2006) Alu-mediated 100-kb deletion in the primate genome: the loss of the agouti signaling protein gene in the lesser apes. Genome Res 16: 485–490.K. NakayamaT. Ishida2006Alu-mediated 100-kb deletion in the primate genome: the loss of the agouti signaling protein gene in the lesser apes.Genome Res16485490 - 27. Roman-Gomez J, Jimenez-Velasco A, Agirre X, Castillejo JA, Navarro G, et al. (2008) Repetitive DNA hypomethylation in the advanced phase of chronic myeloid leukemia. Leuk Res 32: 487–490.J. Roman-GomezA. Jimenez-VelascoX. AgirreJA CastillejoG. Navarro2008Repetitive DNA hypomethylation in the advanced phase of chronic myeloid leukemia.Leuk Res32487490 - 28. Shann YJ, Cheng C, Chiao CH, Chen DT, Li PH, et al. (2008) Genome-wide mapping and characterization of hypomethylated sites in human tissues and breast cancer cell lines. Genome Res 18: 791–801.YJ ShannC. ChengCH ChiaoDT ChenPH Li2008Genome-wide mapping and characterization of hypomethylated sites in human tissues and breast cancer cell lines.Genome Res18791801 - 29. O'Neill RJ, O'Neill MJ, Graves JA (1998) Undermethylation associated with retroelement activation and chromosome remodelling in an interspecific mammalian hybrid. Nature 393: 68–72.RJ O'NeillMJ O'NeillJA Graves1998Undermethylation associated with retroelement activation and chromosome remodelling in an interspecific mammalian hybrid.Nature3936872 - 30. Brown JD, Golden D, O'Neill RJ (2008) Methylation perturbations in retroelements within the genome of a Mus interspecific hybrid correlate with double minute chromosome formation. Genomics 91(3): 267–273.JD BrownD. GoldenRJ O'Neill2008Methylation perturbations in retroelements within the genome of a Mus interspecific hybrid correlate with double minute chromosome formation.Genomics91(3)267273 - 31. Arnold ML, Meyer A (2006) Natural hybridization in primates: one evolutionary mechanism. Zoology (Jena) 109: 261–276.ML ArnoldA. Meyer2006Natural hybridization in primates: one evolutionary mechanism.Zoology (Jena)109261276 - 32. O'Donnell KA, Boeke JD (2007) Mighty Piwis defend the germline against genome intruders. Cell 129(1): 37–44.KA O'DonnellJD Boeke2007Mighty Piwis defend the germline against genome intruders.Cell129(1)3744 - 33. Coarfa C, Milosavljevic A (2008) Pash 2.0: scaleable sequence anchoring for next-generation sequencing technologies. Pac Symp Biocomput 102–113.C. CoarfaA. Milosavljevic2008Pash 2.0: scaleable sequence anchoring for next-generation sequencing technologies.Pac Symp Biocomput102113 - 34. Kalafus KJ, Jackson AR, Milosavljevic A (2004) Pash: Efficient Genome-Scale Sequence Anchoring by Positional Hashing. Genome Res 14: 672–678.KJ KalafusAR JacksonA. Milosavljevic2004Pash: Efficient Genome-Scale Sequence Anchoring by Positional Hashing.Genome Res14672678 - 35. Smit AFA, Hubley R, Green PRepeatMasker Open-3.0. AFA SmitR. HubleyP. GreenRepeatMasker Open-3.0.1996–2004 http://www.repeatmasker.org. 1996–2004 http://www.repeatmasker.org. - 36. Ishkanian AS, Malloff CA, Watson SK, DeLeeuw RJ, Chi B, et al. (2004) A tiling resolution DNA microarray with complete coverage of the human genome. Nat Genet 36: 299–303.AS IshkanianCA MalloffSK WatsonRJ DeLeeuwB. Chi2004A tiling resolution DNA microarray with complete coverage of the human genome.Nat Genet36299303 - 37. Krzywinski M, Bosdet I, Smailus D, Chiu R, Mathewson C, et al. (2004) A set of BAC clones spanning the human genome. Nucleic Acids Res 32: 3651–3660.M. KrzywinskiI. BosdetD. SmailusR. ChiuC. Mathewson2004A set of BAC clones spanning the human genome.Nucleic Acids Res3236513660 - 38. Veltman JA, Schoenmakers EF, Eussen BH, Janssen I, Merkx G, et al. (2002) High-throughput analysis of subtelomeric chromosome rearrangements by use of array-based comparative genomic hybridization. Am J Hum Genet 70: 1269–1276.JA VeltmanEF SchoenmakersBH EussenI. JanssenG. Merkx2002High-throughput analysis of subtelomeric chromosome rearrangements by use of array-based comparative genomic hybridization.Am J Hum Genet7012691276 - 39. Saal LH, Troein C, Vallon-Christersson J, Gruvberger S, Borg A, et al. (2002) BioArray Software Environment (BASE): a platform for comprehensive management and analysis of microarray data. Genome Biol 3: SOFTWARE0003.LH SaalC. TroeinJ. Vallon-ChristerssonS. GruvbergerA. Borg2002BioArray Software Environment (BASE): a platform for comprehensive management and analysis of microarray data.Genome Biol3SOFTWARE0003 - 40. Ventura M, Mudge JM, Palumbo V, Burn S, Blennow E, et al. (2003) Neocentromeres in 15q24-26 map to duplicons which flanked an ancestral centromere in 15q25. Genome Res 13: 2059–2068.M. VenturaJM MudgeV. PalumboS. BurnE. Blennow2003Neocentromeres in 15q24-26 map to duplicons which flanked an ancestral centromere in 15q25.Genome Res1320592068 - 41. Lichter P, Ledbetter SA, Ledbetter DH, Ward DC (1990) Fluorescence in situ hybridization with Alu and L1 polymerase chain reaction probes for rapid characterization of human chromosomes in hybrid cell lines. Proc Natl Acad Sci U S A 87: 6634–6638.P. LichterSA LedbetterDH LedbetterDC Ward1990Fluorescence in situ hybridization with Alu and L1 polymerase chain reaction probes for rapid characterization of human chromosomes in hybrid cell lines.Proc Natl Acad Sci U S A8766346638 - 42. Li LC, Dahiya R (2002) MethPrimer: designing primers for methylation PCRs. Bioinformatics 18: 1427–1431.LC LiR. Dahiya2002MethPrimer: designing primers for methylation PCRs.Bioinformatics1814271431 - 43. Yang Z (1997) PAML: a program package for phylogenetic analysis by maximum likelihood. Comput Appl Biosci 13: 555–556.Z. Yang1997PAML: a program package for phylogenetic analysis by maximum likelihood.Comput Appl Biosci13555556
<urn:uuid:f7797001-000b-4fe8-9bb0-e96c2c51c9c3>
CC-MAIN-2018-13
http://journals.plos.org/plosgenetics/article?id=10.1371/journal.pgen.1000538
s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257647146.41/warc/CC-MAIN-20180319194922-20180319214922-00164.warc.gz
en
0.898967
12,931
2.765625
3
Tempered glass/Reinforced glass is safety glass. Tempered glass is actually a kind of prestressed glass. In order to improve the strength of the glass, chemical or physical methods are usually used to form compressive stress on the surface of the glass. When the glass is subjected to external force, it first offsets the surface stress, thereby improving the bearing capacity and enhancing the resistance of the glass itself. Wind pressure, cold and heat, impact, etc. Pay attention to distinguish it from glass fiber reinforced plastic. Feature Editing Broadcast When the glass is damaged by external force, the fragments will be broken into small particles with obtuse angles similar to honeycombs, which are not easy to cause serious harm to the human body. The impact strength of tempered glass with the same thickness is 3 to 5 times that of ordinary glass, and the bending strength is 3 to 5 times that of ordinary glass. Tempered glass has good thermal stability, the temperature difference that can withstand is 3 times that of ordinary glass, and it can withstand a temperature difference of 300 °C. The first is that the strength is several times higher than that of ordinary glass, and it is resistant to bending. Second, it is safe to use. The increased bearing capacity improves the fragile nature. Even if the tempered glass is damaged, it will be small fragments without acute angles, which greatly reduces the damage to the human body. Compared with ordinary glass, the resistance to rapid cooling and rapid heating of tempered glass is improved by 3 to 5 times. It is a type of safety glass. To ensure the safety of qualified materials for high-rise buildings. Disadvantages of tempered glass: 1. The tempered glass can no longer be cut and processed. The glass can only be processed to the required shape before tempering, and then tempered. 2. Although the strength of tempered glass is stronger than that of ordinary glass, tempered glass has the possibility of self-explosion (self-breaking), while ordinary glass does not have the possibility of self-explosion. 3. The surface of the tempered glass will be uneven (wind spots), with a slight thickness thinning. The reason for the thinning is that after the glass is softened by hot melt, it is rapidly cooled by strong wind, so that the crystal gap inside the glass becomes smaller and the pressure becomes larger, so the glass is thinner after tempering than before tempering. Under normal circumstances, 4~6mm glass will be thinned by 0.2~0.8mm after tempering, and 8~20mm glass will be thinned by 0.9~1.8mm after tempering. The specific degree depends on the equipment, which is why tempered glass cannot be used as a mirror. 4. The flat glass used for construction after passing through the tempering furnace (physical tempering) will generally be deformed, and the degree of deformation is determined by the equipment and technicians. To a certain extent, the decorative effect is affected (except for special needs). Preparing Editing Broadcast Tempered glass is obtained by first cutting ordinary annealed glass to the required size, then heating it to about 700 degrees close to the softening point, and then rapidly and uniformly cooling it (usually 5-6MM glass is heated at 700 degrees for about 240 seconds, Cool down for about 150 seconds. 8-10MM glass is heated for about 500 seconds at a high temperature of 700 degrees, and cooled for about 300 seconds. In short, according to the thickness of the glass, the time for heating and cooling is also different). After tempering, uniform compressive stress is formed on the surface of the glass, while tensile stress is formed inside, which improves the bending and impact strength of the glass, and its strength is about four times that of ordinary annealed glass. The tempered glass that has been tempered cannot be cut, ground, etc. or damaged, otherwise it will be "shattered" due to the destruction of the uniform compressive stress balance. Category Edit Broadcast 1. Tempered glass is divided into flat tempered glass and curved tempered glass according to shape. Generally, the thickness of flat tempered glass is 11, 12, 15, 19mm, etc.; the thickness of curved tempered glass is 11, 15, 19mm, etc., and the specific thickness after processing depends on the equipment and technology of each manufacturer. But curved (ie curved tempered) tempered glass has a maximum arc limit for each thickness. That is to say, R R is the radius. 2. Tempered glass is divided into flat tempered and curved tempered according to its appearance. 3. Tempered glass is divided into: superior products and qualified products according to its flatness. High-quality tempered glass is used for car windshields; qualified products are used for architectural decoration. ⒈ Physical tempered glass is also called quenched tempered glass. It heats ordinary flat glass in a heating furnace to a temperature close to the softening temperature of glass (600°C), eliminates the internal stress through its own deformation, then removes the glass from the heating furnace, and then uses a multi-head nozzle to blow high-pressure cold air to the glass. Tempered glass can be obtained by cooling it quickly and evenly to room temperature on both sides. This kind of glass is in a stress state of internal tension and external compression. Once local damage occurs, the stress will be released, and the glass will be broken into countless small pieces. These small pieces have no sharp edges and corners and are not easy to hurt people. ⒉Chemical tempered glass is to improve the strength of glass by changing the chemical composition of the surface of the glass, generally by applying ion exchange method for tempering. The method is to immerse the silicate glass containing alkali metal ions into the molten lithium (Li+) salt, so that the Na+ or K+ ions on the surface of the glass are exchanged with Li+ ions, and a Li+ ion exchange layer is formed on the surface. The expansion coefficient is smaller than the Na+ and K+ ions, so that the outer layer shrinks less and the inner layer shrinks more during the cooling process. When cooled to room temperature, the glass is also in a state where the inner layer is pulled and the outer layer is compressed. Similar to physical tempered glass. According to the degree of toughness ⒈Tempered glass: degree of tempering=2~4N/cm, surface stress of tempered glass of glass curtain wall α≥95Mpa; ⒉Semi-tempered glass: degree of tempering=2N/cm, surface stress of semi-tempered glass of glass curtain wall 24Mpa≤α≤69Mpa; ⒊Super-strength tempered glass: degree of tempering>4N/cm. Product application editor broadcast Flat tempered and curved tempered glass are safety glass. Widely used in high-rise building doors and windows, glass curtain walls, indoor partition glass, lighting ceilings, sightseeing elevator passages, furniture, glass guardrails, etc. Usually tempered glass can be used in the following industries: ⒈Architecture, building formwork, decoration industry (eg: doors and windows, curtain walls, interior decoration, etc.) ⒉Furniture manufacturing industry (glass coffee table, furniture matching, etc.) ⒊Home appliance manufacturing industry (TVs, ovens, air conditioners, refrigerators and other products) ⒋Electronics, instrument industry (mobile phones, MP3, MP4, clocks and other digital products) ⒌Automotive manufacturing industry (automobile window glass, etc.) ⒍Daily products industry (glass cutting board, etc.) ⒎Special industry (military glass) Since tempered glass is broken, the fragments will be broken into small uniform particles and there is no sharp corners like glass knives, so it is called safety glass and is widely used in automobiles, interior decoration, and high-rise windows that open to the outside. Emergency methods edit broadcast Tempered glass is obtained by cutting ordinary annealed glass to the required size, heating it to a near softening point, and then rapidly and uniformly cooling it. After tempering, uniform compressive stress is formed on the surface of the glass, while tensile stress is formed inside, which greatly improves the performance of the glass. The tensile strength is more than 3 times that of the latter, and the impact resistance is more than 5 times that of the latter. It is also this feature that the stress characteristic has become an important symbol for identifying real and fake tempered glass, that is, tempered glass can see colored stripes on the edge of the glass through the polarizer, and when observed on the surface of the glass, you can see black and white. Alternating spots. Polarizers can be found in camera lenses or glasses. Pay attention to the adjustment of the light source when observing, so that it is easier to observe. The automatic explosion of tempered glass without direct mechanical external force is called self-explosion of tempered glass. According to industry experience, the self-explosion rate of ordinary tempered glass is about 1~3‰. Self-explosion is one of the inherent characteristics of tempered glass. There are many reasons for the expansion of self-explosion, which can be briefly summarized as follows: ①Influence of glass quality defects A. There are stones, impurities and bubbles in the glass: impurities in the glass are the weak point of the tempered glass and the place where the stress is concentrated. In particular, if the stone is in the tensile stress area of the tempered glass, it is an important factor leading to the explosion. Stones exist in glass and have a different coefficient of expansion than the vitreous body. The stress concentration in the crack area around the stone increases exponentially after glass tempering. When the expansion coefficient of the stone is smaller than that of the glass, the tangential stress around the stone is in tension. Crack propagation that accompanies stones is highly susceptible. B. The glass contains nickel sulfide crystals Nickel sulfide inclusions generally exist as small crystalline spheres with a diameter of 0.1-2mm. The appearance is metallic, and these miscellaneous inclusions are Ni3S2, Ni7S6 and Ni-XS, where X=0-0.07. Only Ni1-XS phase is the main reason for the spontaneous explosion of tempered glass. The theoretical NIS is known at 379. There is a phase transition process at C, from the α-NiS hexagonal system in the high temperature state to the β-NiS trigonal system in the low temperature state, accompanied by a volume expansion of 2.38%. This structure is preserved at room temperature. If the glass is heated later, the α-β state transition may occur rapidly. If these impurities are inside the tempered glass under tensile stress, the volume expansion will cause spontaneous bursting. If a-NIS exists at room temperature, it will slowly transform to the β state after several years and months, and the slow increase in volume during this phase transition does not necessarily cause internal rupture. C. Due to improper processing or operation, the glass surface has defects such as scratches, cracks, deep cracks, etc., which are easy to cause stress concentration or self-explosion of tempered glass. ②The stress distribution in tempered glass is uneven and offset When the glass is heated or cooled, the temperature gradient along the thickness of the glass is not uniform and asymmetric. The tempered products have a tendency to self-explode, and some produce "wind explosion" when they are chilled. If the tensile stress area is shifted to one side of the product or shifted to the surface, the tempered glass will self-explode. ③Influence of the degree of tempering, experiments have shown that when the degree of tempering increases to level 1/cm, the number of self-explosions reaches 20% to 25%. It can be seen that the greater the stress, the higher the degree of tempering and the greater the amount of self-explosion. Development history editing broadcast The development of tempered glass can be traced back to the middle of the 17th century. A Prince of Rhine named Robert once did an interesting experiment. He put a drop of molten glass in icy cold water and made a kind of glass. Very hard glass. The high-strength, granular glass resembles water droplets with long, curved tails called "Prince Robert Grains." However, when the tail of the granule was bent and broken, it was strange that the whole granule suddenly collapsed violently and even became a fine powder. The above approach is very similar to the quenching of metal, but this is the quenching of glass. This quenching does not change the composition of the glass, so it is also called physical tempered, so tempered glass is called tempered glass. The first patent for glass tempering was obtained by the French in 1874. The tempering method is to heat the glass to a temperature close to the softening temperature, and then immediately put it into a liquid tank with a relatively low temperature to increase the surface stress. This method is the early liquid tempering method. Frederick Siemens of Germany received a patent in 1875, and Geovge E. Rogens of Massachusetts, USA, applied the tempering method to wine glasses and lampposts in 1876. That same year, Hugh O'heill of New Jersey was granted a patent. In the 1930s, Saint-Gobain in France, Triplex in the United States, and Pilkington in the United Kingdom all began to produce large-area flat tempered glass for automobile windshields. In the 1930s, Japan also successively carried out the industrial production of tempered glass. Since then, the world has begun the era of mass production of tempered glass. After 1970, the British company Triplex succeeded in tempering glass with a thickness of 0.75~1.5mm in liquid medium, ending the history that physical tempering could not temper thin glass, which was a major breakthrough in tempered glass technology. The history of tempered glass in China first started in 1955, when Shanghai Yaohua Glass Factory started trial production, and in 1958 Qinhuangdao City tempered glass factory successfully started trial production. In 1965, Qinhuangdao Yaohua Glass Factory began to produce tempered glass for military industry. In the 1970s, Luoyang Glass Factory was the first to introduce Belgian tempering equipment. During the same period, the chemical tempered glass of Shenyang Glass Factory was put into production. Since the 1970s, tempered glass technology has been comprehensively promoted and popularized around the world, and tempered glass has been used in the fields of automobiles, construction, aviation, electronics, etc., especially in construction and automobiles. If you want to get more information of our products,pls visit Ningbo Sunshinelux Lighting Co., Ltd. is a manufacturer specializing in the production ofremote control flood lights, ufo led high bay light 150w, led worklight tripod, flexible work light with magnetic base, motion activated solar flood light, street lamp solar, led streetlights. Sunshinelux as a professional chinese outdoor lights, best lighting manufacturers, rechargeable led work light manufacturers, china led high bay light 200w suppliers, solar street lights manufacturers china, road lighting manufacturer in China. please believe that we will provide the best quality, best service, best price, please contact us.
<urn:uuid:c986d4a8-36fc-4dc5-9d34-696f84b81368>
CC-MAIN-2023-06
https://www.nbsunlux.com/industial-news/led-flood-light-suppliers.html
s3://commoncrawl/crawl-data/CC-MAIN-2023-06/segments/1674764499790.41/warc/CC-MAIN-20230130003215-20230130033215-00248.warc.gz
en
0.932539
3,289
3.84375
4
Electricity’s share in transport energy is projected to increase sevenfold from its 2011 level by 2050. Yet while great strides are being made in the electrification of passenger vehicles, commercial vehicles and other heavy vehicles are slower on the uptake. Here, Roger Brereton, head of sales at steering systems specialist Pailton Engineering, explains some of the unique challenges on the road to electrification in different sectors. If we look at progress toward the electrification of vehicles, we find a lack of uniformity. Electrification of passenger vehicles is accelerating, but different sectors present unique challenges. The opportunities presented by electric vehicle technology also varies depending on the type of vehicle we are looking at. For this reason, the uptake will be slower and more complicated in some sectors and we should not expect the success of passenger vehicles to be easily replicated in every other sector. According to many predictions, electrification of bus fleets will proceed at a rapid pace in the coming years. Currently, approximately 17% of the world’s buses are electrified. According to Bloomberg New Energy Finance, that figure is likely to rise to 60% by 2040. The increase is being driven by improvements in the technology but also by active government intervention. In California, for example, it is mandated that by the end of this decade, any new purchases of bus fleets by mass transit agencies must be of electric buses. When looking at these figures we need to consider the impact of the Chinese market. Of the 425,000 electric buses in the world last year, 99% of them were in China. The largest bus fleet in Europe was in London, which recently celebrated a deal to add 78 electric double deck buses to its existing fleet of approximately 200 buses. For bus electrification to become more widespread outside of China, substantial infrastructural challenges will need to be overcome. Building expensive charging infrastructure in densely populated urban areas is difficult. However, charging is an equally significant problem for buses outside of cities, which must travel longer distances and therefore suffer from even greater range anxiety. In the wake of the coronavirus pandemic, the sector will need significant government support and far-sighted policies to achieve the optimistic growth forecasts. There is also the issue of standardization. Engineers and manufacturers have not coalesced around a standard for where in the bus the battery should be located. Competitive innovation will help drive the technology forward, but it also means that the conventional parts of the vehicle, such as the steering system for example, must often be custom built. Solaris Bus & Coach Similar issues with range anxiety, lack of infrastructure and excessive battery weight hold back the uptake of electrification in the commercial vehicle sector. These vehicles face different demands in comparison to passenger cars. For example, long haul vehicles are heavier and so require larger and heavier batteries to get them moving. Research by Benchmark Mineral Intelligence has estimated that a 1,000 L (264.2 gal.) fuel tank weighing 800 kg (1,763.7 lbs.) would provide the equivalent level of energy to a 20,000 kg (44,092.5 lbs.) lithium ion battery. Electric vehicles cannot travel as far on a single charge, when compared with vehicles that used an internal combustion engine. They also take longer to charge. For the reasons above, the uptake of electrification in this sector has been and will continue to be slower when compared to other vehicle types. Nor is it clear whether electric battery powered vehicles will be better suited to this sector than other low carbon alternatives like the hydrogen fuel cell. In the meantime, there are alternative ways of reducing vehicle emissions. For example, platooning is one option being explored. This involves the linking of two or more vehicles in a convoy, often employing automated technologies. Another example is greater emphasis on high quality parts. Improved conventional components that can reduce both the overall weight of the vehicle and the friction generated while driving can also improve fuel efficiency. The military sector is a major contributor to greenhouse gas (GHG) emissions. One prominent academic study reached the conclusion that if the U.S. military were a country, it would be the 47th largest GHG emitter in the world. The sector faces strong political pressure to reduce its carbon footprint. There are exciting reports of developments in hybrid technology. In August, for example, the Ministry of Defence announced its latest round of investments in electric and hybrid drivetrains. As well as the environmental benefits, advocates of alternative fuel types for military vehicles claim operational benefits, such as quieter and therefore stealthier vehicles. However, there is good reason to expect the uptake of this technology in the military sector will be at a very slow pace. Diesel vehicles will continue to dominate in the coming years and the reasons behind this fact have more to do with the inherent limitations of the electric battery technology than with politics. The key obstacle to electrifying military vehicles is the weight of the battery. If the vehicles are too heavy, they simply cannot perform or compete in the way a vehicle powered by an internal combustion engine would. This factor is particularly important in the context of rising demand for light-weight, nimble vehicles that are better suited to the asymmetric warfare and complex battlegrounds of the 21st century. Different vehicles face different demands. The challenges and opportunities from electrification are not uniform across each sector. In comparison to passenger vehicles, many other sectors face a steeper uptake of the technology because of issues like greater range anxiety and inadequate infrastructure. In some sectors, it may be the case that electrification is still a long way off or is less viable than alternative means of cutting emissions. Design flexibility and innovation will therefore be key. ©cherezoff – stock.adobe.com This article was contributed by Pailton Engineering which provides custom-engineered steering systems and steering parts for heavy goods vehicles.
<urn:uuid:836cd9f6-94a8-4a15-900a-4f432c344e07>
CC-MAIN-2021-04
https://www.oemoffhighway.com/trends/electrification/article/21203728/heavy-vehicle-electrification-uptake-to-be-slower-as-design-challenges-remain
s3://commoncrawl/crawl-data/CC-MAIN-2021-04/segments/1610704833804.93/warc/CC-MAIN-20210127214413-20210128004413-00432.warc.gz
en
0.952437
1,191
2.53125
3
W3C stands for the "World Wide Web Consortium". The W3C was founded and is chaired by the original pioneer of the Internet - Tim Berners Lee. The W3C sets the standards by which websites should be built. Here is their mission statement: "To lead the World Wide Web to its full potential by developing protocols and guidelines that ensure long-term growth for the Web." Building websites that adhere to the W3C guidelines is often much harder than building websites that don't. It is therefore usually more costly to do. The advantage is that you can be sure that your website will be more accessible to more people amongst many other benefits.
<urn:uuid:7362ee42-17d1-4497-8b0b-b266e81194d7>
CC-MAIN-2014-10
http://www.digiflip.tv/
s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1394009885896/warc/CC-MAIN-20140305085805-00023-ip-10-183-142-35.ec2.internal.warc.gz
en
0.972451
136
2.6875
3
Also known as low blood sugar. It occurs when blood glucose levels drop to below 4 mmol/l. This can happen if a person with diabetes has taken too much insulin or has exercised more than usual. Symptoms of hypoglycaemia include dizziness, shakiness, rapid heartbeat, sudden hunger, cold or clammy skin, fuzzy vision, confusion, mood changes, and tingling or numbness in the hands, arms, tongue, or lips. Hypoglycaemia can occur even while using an insulin pump and, if left untreated, can lead to unconsciousness and diabetic coma. Pre-Diabetes – blood glucose (sugar) levels higher than normal but not high enough to be called diabetes. People with pre-diabetes may or may not develop diabetes. Other names for pre-diabetes are impaired glucose tolerance, impaired fasting glucose, borderline, subclinical, chemical, or latent diabetes.
<urn:uuid:41b01106-9296-4451-bc4e-60fb93580c44>
CC-MAIN-2020-10
https://www.diabetesqualified.com.au/explandict/hypoglycaemia/
s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875146004.9/warc/CC-MAIN-20200225014941-20200225044941-00040.warc.gz
en
0.874452
185
3.265625
3
School and district leadership isn’t about a position or title–it’s about improving practices around digital learning If educators want to see results in student engagement and achievement, they must adapt their leadership practices to an increasingly digitally-focused learning environment. This was the focus of a CoSN 2016 spotlight session by Eric Sheninger, a senior fellow at the International Center for Leadership in Education and a former high school principal. “Leadership is not about position, titles, or power. Leadership is about the actions you take,” he said. During the session, Sheninger highlighted seven pillars of effective digital school leaders and talked about how, during his time as a high school principal, he and his staff modeled each pillar. Those pillars focus on how school leaders can ensure that their policies and practices highlight the best examples of digital learning success in schools. “The environment in which kids learn is dramatically different,” he said. “We fault our kids for being so engaged with technology…how can we prepare kids for the future if we, those tasked with educating kids, are stuck in the past? If teaching, learning, and leadership don’t change, we’re never going to get results.” Pillar 1: Student learning and engagement “Technology is a tool–it’s not a learning outcome,” Sheninger said. “What do you want in your vision? What do you want your kids to be able to do with technology that will allow them to demonstrate conceptual mastery?” Engagement can begin with creating projects and learning opportunities that mean something. “If you don’t get instructional design right, technology is just going to speed up the rate of failure. It’s about building a foundation.” Pillar 2: School environment When he was principal of New Jersey’s New Milford High School, Sheninger said a change in learning spaces changed student engagement for the better. In fact, he said, data indicates the learning environment design can impact student engagement and achievement by up to 25 percent. Pillar 3: Professional learning and growth An unlimited number of professional learning opportunities are available on social media and through professional learning networks (PLNs), Sheninger said. Modeling the practices educators want to see from students is the first step. Pillar 4: Communications Communication has changed drastically because of technology, Sheninger said, and now educators are in an era of mass dialogue. “Don’t we want to take advantage of that in our own leadership capacity?” he asked. School stakeholders want news about school events, staff and student accomplishments, and district successes. “It’s about being proactive, not reactive,” he said. “Digital leadership is not just about information; it’s about meeting your stakeholders where they are.” Sheninger also advised using multiple methods of news distribution, because people use various social media channels and communication methods. “You can’t put all your eggs in the Twitter basket or the Facebook basket,” he said. Pillar 5: Public relations “If you don’t tell your story, someone else will,” he said. “You need to tell stakeholders what actually happens in your schools and in your district. There are great stories to share–leverage the media.” Digital leadership is about becoming a storyteller-in-chief and sharing school accomplishments. Sheninger advised pushing out good news and accomplishments in various forms–news stories, photos, etc.–across various social media channels. Pillar 6: Branding Branding is a combination of your vision, mission, and values. “In education, a brand is not about selling. It’s about sharing, telling, and building relationships,” he said. A school brand should convey student achievement, teacher and administrator quality, extracurriculars, innovation, and partnerships. Pillar 7: Opportunity Digital leaders should consistently look for opportunities to improve existing programs, strategies, and resources.
<urn:uuid:c86b3d80-0315-4c5e-b375-b54b787023a1>
CC-MAIN-2019-04
https://www.eschoolnews.com/2016/04/08/7-pillars-of-todays-digital-leadership/
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547584518983.95/warc/CC-MAIN-20190124035411-20190124061411-00517.warc.gz
en
0.951837
868
2.59375
3
TECHNOLOGY CAN ENHANCE LEARNING ACHIEVEMENT I recently attended the H. E. McCracken Middle School and Bluffton High School technology night showcase and was impressed on several levels. As I went through the various exhibits, it was evident the students were very proud of their projects and their enthusiasm was clearly demonstrated in their presentations. Computer technology and how it relates to science, business, math, English, social studies, related arts, SPED/ESOL and JROTC was on display. The JROTC was especially impressive with their well tailored uniforms and inspiring deportment. When asked, many of the student presenters knew the career path they wanted to follow and had even selected a college they felt would help them achieve success. However, many felt learning would be greatly enhanced if more technology was integrated into the classroom. One student was already using an iPad to complete a portion of his homework. Some suggested the availability of nooks or kindles in the school library would greatly improve interest and access to reading materials. Even at the middle school level, the students showed a keen awareness of the need for more technology integrated curriculum. The technology on display was clearly a useful tool and learning facilitator. The presentations were a fun experience that could have been enjoyed by all parents and students alike. However, parents and other student attendance was sparse. If students are to reach their full learning potential, technology infused curriculum is important, along with parent, teacher, and community involvement.
<urn:uuid:9b9b2877-9a0e-4fd9-88b2-297ae8dd6ae6>
CC-MAIN-2015-11
http://www.blufftontoday.com/bluffton-opinion/2012-02-08/technology-can-enhance-learning-achievement
s3://commoncrawl/crawl-data/CC-MAIN-2015-11/segments/1424936461944.75/warc/CC-MAIN-20150226074101-00238-ip-10-28-5-156.ec2.internal.warc.gz
en
0.98865
306
2.875
3
Preppers Staple Food Storage Quantities "Experts" have said that the average adult will consume the following amounts of fresh food per year. - Meat - 150 to 200 pounds - Flour - 200 to 300 pounds - Sugar or honey - 60 pounds - Fats or Oils - 60 pounds - Salt - 5 pounds - Powdered Milk - 75 pounds - Vegetables and Fruits - 600 to 700 pounds - Water - 375 gallons These figures are just basic guidelines, however considerations should be made from the aspect of preserved foods, rather than fresh foods. Meat: In severe conditions, people could easily get by with less protein than 150 pounds of fresh meat per year, as that averages to almost a half pound per day. A canned, cooked one pound ham, for example, would be a real treat once a week, and easily feed a family of four. For weekday meals for a family of four, a 5 ounce can of tuna, canned chicken, 12 ounce can of luncheon meat, or 12 ounce can of corned beef can be used in a casserole (or whatever) and provide the required protein. Flour: The listed amount of 200 to 300 pounds of flour per year is fairly realistic, in a SHTF scenario people would be making their own bread and pasta, for example. Using a hand cranked mill to produce flour from whole wheat is a sure way to limit the amount of flour required. Sugar or honey: The recommended 60 pounds is the absolute minimum needed, in reality far below the actual amount desired, as sweeteners are the carbohydrates needed for energy, and survival is hard work. The 60 pounds listed does not take into account home canning, for example, and people will need to make jellies and jams and can fruits, all of which require a considerable amount of sugar or honey. Fats or oils: Again, this is an absolute minimum amount needed, as 60 pounds of fats or oils does not go far when used in baking, frying, and other uses. In hard times, people actually require fat in their diet in order to do hard work. In every country in which food is rationed, cooking oils are one of the first items of scarcity. Corn oil stores for years, and so does plain, inexpensive hydrogenated lard. Salt: Five pounds of iodized table salt would be the recommended minimum per person per year, but what about making kraut, salt preserving meat, or preserving fish in a barrel of salt? For those needs, a family should have at least 50 pounds of fine grade, non iodized salt, available for less then $5.00 from a feed and seed store. Powdered Milk: The 75 pounds recommended per person is fine, but for cooking needs a couple of cases (48 cans) of canned, condensed milk is an absolute necessity. Vegetables and Fruits: In hard times, greens and fruits can indeed be a vital food item, as they provide the vitamins and minerals our bodies require to remain healthy. Storing vegetables and fruits is where a food dehydrator really shines. Combine the dried veggies with fresh greens from a garden and canned fruit juices and sauces, and the 600 pound per year amount becomes far more attainable.
<urn:uuid:fecb65ea-ffa8-4da6-844c-da7a50e0f764>
CC-MAIN-2015-06
http://isurvivalskills.blogspot.com/2012/11/preppers-staple-food-storage-quantities.html?utm_source=feedburner&utm_medium=feed&utm_campaign=Feed%3A+SurvivalSkills+%28Survival+Skills%29
s3://commoncrawl/crawl-data/CC-MAIN-2015-06/segments/1422115855561.4/warc/CC-MAIN-20150124161055-00113-ip-10-180-212-252.ec2.internal.warc.gz
en
0.943248
674
2.734375
3
Torch Lily or Red Hot Poker of the genus Kniphofia 228/365 Torch Lily, is yet another example of a plant that is not really a lily of the genus, lilium, but is of the genus Kniphofia. It's an exotic plant native to South Africa that produce various shades of red, orange, and yellow. They grow from 2-5 feet high and are extremely hardy, drought and heat tolerant, tolerating just about any type of well-draining soil. Named after 18th century German physician and botanist, Johann H. Kniphof, the plant is closely related to the aloe and produces an abundance of nectar. The plant attracts both bees and hummingbirds. So a lily by any other name is either an Agapanthus or a Kniphofia. Note to readers: yesterday's post, I was wrong! Mildred is not a bumble bee but a robber fly. They can be deceptively similar to bumble bees. I had no idea. The educational info about bumblebees still stands. Robber flies are tricky, but still pretty cool. No cash for the treat jar but you'd like to show the love? No problem! Connect with me on LinkedIn and endorse my creative writing skills. Let me know how the pups and I can love ya back! "Your project is guaranteed to meet superior Siberian standards or I will fatally masticate it. You have my "woo" on that!" "I keep your project safe from crows, coyotes, and flies. I prefer to be paid in salmon treats and tennis balls." "I manage the treat jar & the staff's daily payroll of cookies and bones. The staff is excellent at math and let me know when I come up short."
<urn:uuid:e6fe973c-eeb8-46b0-a7c1-0ad5d7753cfa>
CC-MAIN-2020-16
https://wilddingo.com/2015/10/torch-lily/
s3://commoncrawl/crawl-data/CC-MAIN-2020-16/segments/1585370499280.44/warc/CC-MAIN-20200331003537-20200331033537-00460.warc.gz
en
0.947698
375
2.5625
3
Research & Tools Research, Reports, and Data This report by McKinsey & Company has been widely acknowledged as providing a "common, neutral fact base" on the gaps that exist between students who are White or Black, rich or poor, and looks at the dramatic difference in student performance in students from similar backgrounds across school systems and classrooms. The report also highlights the impact of the gaps on our economy and on individual life outcomes, and the information will be useful in discussions on addressing the achievement gaps. At this web site view and download a number of briefs developed by Thedocumenting the connection between poverty and the achievement gaps, looking at the impact of the gaps on our national economy. - The Economic Benefits of Reducing High School Dropout Rates in America's Fifty Largest Cities - The High Cost of High School Dropouts: What the Nation Pays for Inadequate High Schools - Facts For Education Advocates: The Economic Impacts of Education (co-published with the College Board) - Dropouts, Diplomas, and Dollars: U.S. High Schools and the Nation’s Economy - Potential Economic Impacts of Improved Education - Hidden Benefits: The Impact of High School Graduation on Household Wealth - Healthier and Wealthier: Decreasing Health Care Costs by Increasing Educational Attainment - Saving Futures, Saving Dollars: The Impact of Education on Crime Reduction and Earnings This report was prepared for America's Promise Alliance one of NEA’s partners in addressing the dropout crisis. It examines the economic and employment picture for individuals with varied educational levels, including dropouts, looking at income levels, employment rates, and poverty rates. - Focus On What Works: What Educators Need to Know to Move from Issues to Action - Are They Really Ready to Work? Employers' Perspectives - Poverty and Potential: Out-of-School Factors and School Success - The Crisis in the Education of Latino Students - Resources for Breaking Down Barriers To Learning - Students Affected by Achievement Gaps - NEA Research, Tools, & Talking Points on Achievement Gaps - Effective Practices in Closing Achievement Gaps
<urn:uuid:7113fc45-b9e8-43a4-a99a-3197e641e49e>
CC-MAIN-2016-50
http://www.nea.org/home/1051.htm
s3://commoncrawl/crawl-data/CC-MAIN-2016-50/segments/1480698544678.42/warc/CC-MAIN-20161202170904-00103-ip-10-31-129-80.ec2.internal.warc.gz
en
0.921758
446
3.71875
4
Hello everyone, thought I'd share some bits about the processes I’ve been slowly refining over the last few months. As some of you may already be aware, my photographic 'heroes' are pictorialist photographers such as Julia Margaret Cameron, and pioneers such as Carleton Watkins. This has led me down the path of looking at all kinds of different processes - coating my own glass plates, chemical paper reversal, cross processing with very basic disposable cameras, and experimenting with DIY lenses and cameras. For today though, I'm going to waffle on about my experiments using silver-gelatin paper in large format cameras as an alternative to the usual sheet film, a technique which I have been hoping would give my work a flavour of the pictorialist aesthetic. I began experimenting with a chemical reversal process for photographic paper earlier in the year which used relatively innocuous chemistry - regular old Thiourea based sepia toner. The process does take some time, but the result is a one-off (unique) sepia toned silver gelatin print. I have been working on this with a photo-buddy of mine Mr Jonny Sutton, and as we are unaware of a proper name for this process, though I’m sure someone must have thought of this before, we have been (jokingly) calling it the Cox-Sutton Papyrotype when we've been introducing it to some of our students. The method we've worked out may seem somewhat counterintuitive if you already have some experience using sepia toners, but basically it involves toning the silver halides after development, then bleaching away any metallic silver created by the developer before fixing. Resin coated paper reversed with sepia toner. This process is very exciting, but I found the paper had very little latitude - it's a bit too contrasty for my tastes, and the amount of overexposure needed seemed to change depending on the amount of UV light around. Fibre based paper also seems to look mottled with this process, which I think is to do with the higher silver content. Fibre based paper reversed with sepia toner – note the ‘mottled’ effect. So - back to the drawing board, I decided to look into how to reduce contrast in paper negatives made with variable contrast material which is the most readily available paper around these days. Thinking along the lines of darkroom printing, where a magenta filter is used to increase contrast in a print, and yellow filters are used to reduce contrast, I decided to give it a go with a regular yellow Y2 / K2 filter on the lens of my camera. I also decided to develop them in approximately half-strength paper developer (Ilford Multigrade 1+19) which should be less aggressive with the highlights than full strength. I have tried Rodinal in the past, but I had difficulty finding a developing time, dilution and temperature which both developed the negative and didn't stain the paper, and it doesn’t last all that long in a tray. As I understand it (please correct me if I am wrong!) variable contrast paper has two layers, one which is sensitive to blue and UV giving higher contrast, and another more sensitive to green light which is lower in contrast. Adding the yellow filter should reduce the exposure of the blue / UV layer, and as a result decrease the contrast. ISO Tests in negative and digitally inverted. I shot an ISO test, starting at 12 ISO and increasing the exposure by a stop each time until I ended up at 0.375 ISO. I shot one set with the yellow filter, and one without, then developed them by inspection in the darkroom to determine my developing time. I found that on an overcast day, 0.375 ISO with the yellow filter developed for 2 and a half minutes gave me the best tonal range - a good amount of shadow detail without sacrificing density. Overexposing the negative and pulling it in development without a yellow filter gave results which may well be good enough for scanning at around 0.75 ISO, but probably wouldn’t be dense enough to make a good contact print the old fashioned way. An easy way to meter for ISO 0.375 is to set your meter for ISO 3 then multiply the given exposure time by 8. If your meter doesn’t read low enough, you can always take a reading at ISO 6 and multiply by 16 for example. I have yet to test it on a sunny day, but I believe it may change the ISO rating as there would be more UV around, depending how much the yellow filter lets through. Closeups of the best results with and without the filter. Exposing and developing the paper in this way still gives me the 'antiquarian' look - the paper gives a certain softness and texture to the photograph, and most early processes were also only blue / UV sensitive – but it also gives me a reasonable tonal range which you don't normally achieve with multigrade paper. It does however mean exposures can be very long, some examples in this article needed an 8 minute exposure at f/32, end even at f/5.6 you may be looking at an exposure of 30 seconds to 1 minute on a cloudy day. Maybe I need to invest in a Victorian style neck-brace for my portrait sitters! Thankfully, photographic paper doesn’t seem to suffer the same failure of reciprocity as film – I’m guessing it must have been designed to cope with long exposures pretty well as this is more common in printing. Scanned paper negative – this was an 8 minute exposure at f/32 with a yellow filter. Another interesting observation I have made is that you can make a remarkably sharp scan of a paper negative, whereas the traditional contact print method gives you a much softer, more painterly aesthetic. I don’t think there is a clear winner here – I like the softness and tonality of the contact prints, but those used to sharper photographs may cringe at the thought of carrying a large format camera to make such a soft photograph. Scanned contact print from 5x4 paper negative – nice and soft! Scanned straight from the same negative – nice and sharp! As for the reversal, I am hoping to test it out again soon with my new method for exposing paper - the experiments continue :) In the meantime I'm planning on starting a new body of work using the processes I’ve figured out with my 5x4 and 10x8 cameras in the coming months. I may also make some more dry plates for this project but perhaps this is a topic for another time! Keep your eyes peeled at the following places for further updates, and don’t hesitate to get in touch if you would like to discuss this stuff in finer detail. Prints and zines are available from my website. Scanned 5x4 paper negative.
<urn:uuid:24ff9db1-9d1d-4d05-a3c1-46edd2383e9f>
CC-MAIN-2019-30
https://yourexhibition.net/latest/2018/12/3/what-im-working-on-steven-cox
s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195526408.59/warc/CC-MAIN-20190720024812-20190720050812-00530.warc.gz
en
0.95911
1,423
2.546875
3
Poor air quality can exacerbate underlying lung problems such as bronchial asthma. The Department of Respiratory and Critical Care Medicine at SGH explains the health effects of air pollution. In Singapore, the air quality is generally good due to the strict pollution standards. Yet doctors still see many patients whose lives and health are directly affected by the global phenomenon of air pollution. The common effects of air pollution include irritation to the eyes, nose and throat, and upper respiratory infections such as bronchitis and pneumonia. "Poor air quality can also exacerbate underlying lung problems, such as bronchial asthma, chronic obstructive pulmonary disease (COPD), or allergic rhinitis," say doctors from the Department of Respiratory and Critical Care Medicine at Singapore General Hospital (SGH), a member of the SingHealth group. Some patients may even need to be hospitalised. Fine air particles can accumulate in the lungs We can divide pollutants into two types: - gas pollutants (e.g. carbon monoxide, sulphur dioxide, etc.) and - particles in the air Broadly speaking, gas pollutants can be detrimental to our health as they take part in gas exchange during each breath we take. However, they are not retained in the body and are excreted. Particles in the air are another story. While the bigger particles are filtered by the respiratory tract (which includes our nostrils, airways and lungs), the finer particles are deposited in our lungs and may build up over a prolonged period. Such prolonged exposure to pollution can lead to the onset of chronic lung diseases. In Singapore, the most preoccupying source of air pollution is the haze attributable to forest fires in Indonesia. This typically happens around October, when farmers carry out their slash and burn method of cultivation. The haze can be exacerbated depending on the prevailing wind direction which carries the smoke particles. Air pollution may then reach the moderate to unhealthy range. During the smoke haze pollution of 1997, there was an immediate 30 per cent increase in outpatient attendance and increased accident and emergency attendance for haze-related conditions. Measurements showed that 94 per cent of haze particles were smaller than 2.5 microns in diametre, which is the size that bypasses our filtering system and gets deposited in our lung tissues. Ways to protect yourself from air pollution Should you want a face mask to protect yourself against the haze and air pollution, make sure you get one that can catch the smallest particles. Look for filters that contain an electro-static charge, which can trap fine dust and particles below 0.3 microns. As for inside your home, using an air conditioner may help remove air pollutants. Why? Air conditioning makes the air drier, effectively removing many water-soluble pollutants. Stand-alone HEPA air cleaners can also be considered. The best ones are those equipped with True HEPA filters, which can capture high proportions of very small particles.
<urn:uuid:ee1e3834-43f1-4c8d-8d07-fbe60498d247>
CC-MAIN-2017-47
https://www.healthxchange.sg/head-neck/ear-nose-throat/health-hazards-air-pollution
s3://commoncrawl/crawl-data/CC-MAIN-2017-47/segments/1510934807344.89/warc/CC-MAIN-20171124085059-20171124105059-00295.warc.gz
en
0.925351
607
3.59375
4
Safely Ever After: Educating Children About Child Predators The point of educating children about child predators is to make them more aware of their surroundings, not to make them afraid. The "doom and gloom" method can have a negative impact, since the goal is to inform children and teach them how to look out for themselves if they feel threatened, not to make them afraid of the world. One effective technique Fitzgerald talks about using with kids is the thumbs up/thumbs down method. When talking about safe people, give them a thumbs up. This includes people and situations that are safe, fun, and do not harm a child in any way. A thumbs down includes any person or situation that is harmful or breaks the rules. "Parents tend to go overboard when talking to their children about a topic like this, because it is a topic that scares them, and rightfully so," explains Fitzgerald. She points out that news and media coverage have sensationalized these situations, making the problem with child predators seem more widespread than it used to be. Although people talk about it more, child predators are no more prevalent today then they were 15 or 20 years ago. It is important to inform children, but being overprotective and approaching this topic too seriously will not help a child in the long run. Pattie Fitzgerald is the founder and creator of Safely Ever After, Inc. and has been teaching child safety awareness since 2001. Parents, teachers, and school administrators specifically request her Safely Ever After programs because they present important and accurate information in a calm and humorous way. Being a mom herself, Fitzgerald knows this take is important, especially when dealing with kids and when talking about a topic that can scare them. For more information on Pattie Fitzgerald's Safely Ever After Program, visit her website at www.safelyeverafter.com.
<urn:uuid:98cae289-a436-4fec-95f4-2e425b79d94f>
CC-MAIN-2015-48
http://life.familyeducation.com/safety/stranger-safety/59178.html?page=2
s3://commoncrawl/crawl-data/CC-MAIN-2015-48/segments/1448398447729.93/warc/CC-MAIN-20151124205407-00020-ip-10-71-132-137.ec2.internal.warc.gz
en
0.963052
374
3.03125
3
Summary: Students participate in a posted offer market experiment that includes an exogenous demand shift and an exogenous supply shift sequentially. Time paths of observed asked prices, contract prices, competitive equilibrium prices, generated efficiency levels and distribution of gains are displayed graphically. Motivation: Posted-Offer Markets are frequently used in society - retail markets are a familiar example. Teaching Posted-Offer Markets may involve following questions: Participation in a posted offer market experiment provides students with experience of the market, trains their intuition, and helps them better understand the properties of this institution. - Why such an institution exists? - Which type of markets favor it? - How does it perform in comparison with one-sided auctions and double auctions? This posted offer market configuration supports up to 18 subjects - 9 sellers and 9 buyers. Buyer and seller role assignments alternate as subjects connect (e.g., if only 12 subjects join the game, then there will be 6 sellers and 6 buyers). After running the experiment, an informal discussion can be carefully guided to cover issues such as: - how posted prices were chosen, - what drove their adjustments, - which prices were more likely to lead to sales, - what were gains from trade, - how were gains distributed between the seller and the buyer involved in the trade, - how these gains relate to market efficiency, - how the efficiency changed with repetition, - what were the competitive market predictions, - which elements were involved in transmitting structural (demand, supply shocks) changes and so on. For a formal discussion, see Davis D. D. and C.A. Holt, 1993. Experimental Economics, Princeton University Press. In the presence of thick markets, like buyers in retail markets, one-sided and double auctions are characterized by high negotiation costs. Posted-offer markets present a market institution with negotiation costs that are significantly lower. In this institution, agents on one side of the market, normally sellers, simultaneously post :take-it-or-leave-it" prices and buyers simply shop for goods at the posted prices. A simple (computerized) trading rule goes as follows. A number of sellers and buyers interact in the following order: Observed behavior in the laboratory reveals that this institution, in the absence of market power: - Sellers decide their prices simultaneously and privately. Each seller chooses a price and quantity she would like to sell at that price. - After all sellers have made their decisions, prices but are revealed to everyone in the market. - Buyers make their decisions sequentially and according to a random order. During his turn, a buyer decides from which available sellers he will purchase, and how many available units he will buy. - After all buyers have the chance to make their purchases, the market closes. Each seller earns for each sold unit the difference between the posted price and its production cost whereas each buyer earns for each bought unit the difference between the unit's value for the buyer and the paid price. - The market re-opens for the next trading period, and the interactive process is a repetition of the previous steps. However, in the presence of market power competitive equilibria are poor predictors. - drives prices to competitive equilibrium level; - slows down the speed of convergence compared with one-sided auctions and double auctions; - generates prices that are higher than the competitive equilibrium price; - generates a surplus that (i) increases with repetition, (ii) favors the active side of the market, and (iii) is smaller than the one observed in one-sided and double auctions; - is rigid with respect to changes in demand but it is sensitive to changes in supply.
<urn:uuid:8fd95b81-af33-4f2b-94f7-16e93b433052>
CC-MAIN-2018-26
http://www.econport.org/econport/request?page=web_experiments_modules_po
s3://commoncrawl/crawl-data/CC-MAIN-2018-26/segments/1529267861163.5/warc/CC-MAIN-20180618203134-20180618223134-00395.warc.gz
en
0.959175
763
2.828125
3
Vitamins are very important in human vitamin however for probably the most part can’t be manufactured by your body. This crowd of nutritional vitamins contains vitamin C and the massive group of B nutritional vitamins — B1 (thiamin), B2 (riboflavin), niacin, B6 (pyridoxine), folic acid, B12 (cobalamine), biotin, and pantothenic acid. Nutritional vitamins ought to come firstly from a balanced and various weight loss plan with loads of fruit and greens. This vitamin additionally helps your body resist infection Which means although you can’t all the time avoid getting sick, vitamin C makes it somewhat harder in your physique to turn into contaminated with an sickness. When somebody reduces meals intake in an effort to drop body fat, they’re almost assured a nutrient deficiency. Supplemental nicotinic acid could trigger flushing of skin, itching, impaired glucose tolerance and gastrointestinal upset. Scientists additionally observed that these diseases have been absent amongst people …
<urn:uuid:8d6e9d78-161a-452b-be88-f20088874024>
CC-MAIN-2018-39
http://www.monroehornetsfootball.com/tag/encyclopedia
s3://commoncrawl/crawl-data/CC-MAIN-2018-39/segments/1537267157574.27/warc/CC-MAIN-20180921225800-20180922010200-00327.warc.gz
en
0.914688
213
2.625
3
The Anglican Use Gradual of the Sarum Office Carolinum. The Devotions of His Sacred Majestie in His Solitudes and Sufferings, Rendred in Verse. Set to Musick for 3 Voices and an Organ, or Theorbo. A Sermon concerning Church-musick An Essay on the Cultivation of Church Music by Edward Hodges, Mus. Doc. Sermon Preached at St. Michael's Church, Sittingbourne, on the Occasion of the First District Festival of Parish Choirs, on Wednesday, April 20, 1870. Surpliced Boy Choirs The Importance of Musical Knowledge for the Priesthood of the Church, by the Reverend James Nevett Steele. New York: James Pott, 1894. Tucker, Priest-Musician: A Sketch which Concerns the Doings and Thinkings of the Rev. John Ireland Tucker, S.T.D. Two Addresses. Music, Considered in its Effect upon, and Connection with, the Worship of the Church, by Sir John Stainer. The Essential Elements of Churchly Music. A Linguistic Bridge to Orthodoxy: In Memoriam Isabel Florence Hapgood Choir School in the American Church: A study of the choir school and other current chorister training models in Episcopal and
<urn:uuid:c2fb6673-2c1a-4fba-82e0-3d9d38babdd7>
CC-MAIN-2014-41
http://anglicanhistory.org/music/index.html
s3://commoncrawl/crawl-data/CC-MAIN-2014-41/segments/1410657137108.99/warc/CC-MAIN-20140914011217-00307-ip-10-234-18-248.ec2.internal.warc.gz
en
0.816177
291
2.640625
3
Josheen Oberoi looks at the history and artistic practice of Mithila (popularly known as Madhubani) paintings New York: Mithila paintings, sometimes referred to as Madhubani, originate in the region of Mithila from where they also derive their name (as is often the case in artistic traditions in India). This nomenclature is fitting since it is the geographical origin and medium of the artwork that unites this genre. Beyond that this tradition encompasses a diversity of aesthetic styles and content. As a living tradition, Mithila painting has had a dramatic trajectory over the last fifty years, and its evolution is crucial to understanding its value. Mithila is the birthplace of Sita, located on the plains of Bihar. This mineral rich area plays into the aesthetics of Mithila paintings – the primary mediums for this tradition have always been natural vegetable dyes and mineral oxides. This accessibility of medium also tangentially relates to the original praxis of Mithila painting since it is the women of Mithila who had been the sole practitioners for generations creating ceremonial and devotional floor paintings (aripana) and wall murals (kohabar) for festivals and auspicious occasions like weddings using simple brushes made of bamboo and raw cotton. These paintings created sacred spaces for their domestic rituals. The first known recording of Mithila painting occurred in the 1930s. After an earthquake in Bihar in 1934, W.G. Archer, a British civil servant, found these wall and floor paintings in his survey of the area. He and his wife Mildred photographed and published their subsequent research over the next fifteen years. The 1950s and 60s saw a greater interest from Indian scholars. However, it was a 1966 drought in the area that led to its transformation from a localized and domestic art form to a national one. In the 1970s, the All India Handicrafts Board in an effort to provide economic assistance for the drought affected encouraged women of the region to transfer their wall paintings on to paper, for sale. The works in “Colours of the Earth” from the Story by Saffronart primarily come from this early period, marking this critical chapter in the history of Mithila painting. An art form independent of stylistic influences, Mithila painting was practiced by women from every caste. However, art historian Neel Rekha notes in her dissertation “Art and Assertion of Identity: Women and Madhubani Paintings” that one outcome of this change to paper as a medium in the 1970s was the emergence of different styles of painting. These can be broadly categorised as Geru, Bharni, Kachni, Tantric, Gobar (cow dung), and Godana (tattoo). In the early 70s, most paintings were close to the Geru style, the folk art tradition, with an absence of ornamentation and thick black lines. While the Bharni (filled) style uses strong colours and mostly eschews lines, the Kachni (lined) style is marked by the intricate use of line to create dense, beautiful patterns. In content, Maithil artists also retain their autonomy – subjects range from mythological epics and celebrations of rituals and important events to snapshots of their daily life. Theirs is picture writing, and their ideas and experiences remain key to this art, as they were when this tradition was private to these women’s homes. Importantly, the names of these women artists from the 1970s onwards, like Indrakala Devi, Annapurna Devi, amongst others are documented and attributed, creating a significant canon of Mithila art. Ganga Devi, a Karn Kayastha, Sita Devi, a Mahapatra Brahmin, and Jumna Devi, a Harijan, were important early painters in this tradition, each with a distinct aesthetic. Both Ganga Devi and Sita Devi have represented India at major cultural exhibitions in Japan, Russia, Europe and the USA. Strong scholarship on Mithila paintings, from the 1970s onwards has been vital to building knowledge around the art form. Yves Vequad, a French novelist and journalist, produced a pioneering book and a film, The Women Painters of Mithila,in the 1970s. Another researcher, Raymond Lee Owens, set up the Master Craftsmen Association of Mithila in 1977 and the Ethnic Arts Foundation in 1980. This association, which is still active, provides the artists of the region with a regular source of income through exhibitions and sales. The Mithila Art Institute (MAI) was set up in 2003 at the bequest of the late Owens, and is an important source for the transmitting of techniques and the specific Maithil painting culture. In Tokamachi, Japan, the Mithila Museum exhibits about 850 Madhubani paintings at any given time. Works by Maithil artists are also in the permanent collections of Crafts Museum, New Delhi, Syracuse University, New York, and the Philadelphia Museum of Art. At the latter two institutes, important exhibitions of Mithila paintings were held in 2002-03. Another significant exhibition was held at the Janakpur Women’s Development Center; their website has an incredible diversity of work available for viewing, while a documentary made on five Mithila artists by the University of Wisconsin-Madison documents the diversity of practices within this oeuvre. Interestingly, the predominance of international exhibitions vis-a vis domestic suggests an appreciation for this tradition abroad that has been somewhat lacking in India. However, the government of Bihar stated early this year that plans are underfoot to open a school for teaching Mithila. The shared history of Mithila paintings and the intuitive understanding of pictorial language is a constant within the diversity of the visual aesthetics that this tradition represents. Key to the economic independence of women from the region from the 70s onwards, it continues to be dominated by women artists, although now not exclusively so. The early works from the 1970s began this chapter of contemporary Mithila painting and have shown a consistent demand, even as newer artists emerge. In a recent Saffronart auction in August 2012, works by artists like Bachi Devi, Sashikala Devi (the image above) and Kali Devi, all from the 1970s, received a strong response. “Mithila painting is part decoration, part social commentary, recording the lives of rural women in a society where reading and writing are reserved for high-caste men” (Arminton, Bindloss & Mayhew, 2006, p. 315). The suggestion that these paintings empowered women by providing them tools of communication and documenting their lives has remained central to its narrative. The aesthetics and vibrancy of this artistic tradition has persevered in the independent voice that each Mithila artist represents.
<urn:uuid:33dc1ba6-d822-4821-9c51-c0c63393dc95>
CC-MAIN-2017-26
https://blog.saffronart.com/2012/11/06/the-living-tradition-of-mithila-paintings/
s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128320823.40/warc/CC-MAIN-20170626152050-20170626172050-00502.warc.gz
en
0.964071
1,399
3.5625
4
Despite the fact that our solar system is large, and full of planets, it’s hard to wrap our minds around the fact that the entire universe makes our solar system look laughably small. We continue to search the depths of the universe for clues about our existence, and in the middle of doing so, we occasionally find new stars, planets, and moons. Scientists have recently discovered a planet 100 light years away that they’re saying is very similar to Jupiter, a planet right here in our solar system. It was spotted with the Gemini Planet Imager, which is located in Chile and was created purely for searching for other planets in our universe; this is the first time that the Gemini Planet Imager has discovered a new planet. This is of course big news for the people who have created the new imaging telescope, as James Graham, a GPI project scientist will tell you: "This is exactly the kind of planet we envisioned discovering when we designed GPI," Graham said. "We wanted to find planets when they're young so we can figure out the formation process." They’ve named the planet 51 Eridani b, but the fact that it’s like Jupiter isn’t the only reason that it’s getting so much attention. In fact, what scientists are interested in is the fact that the planet appears to be very young – just 20 million years old, which means the planet just came to be about 40 million years after the dinosaurs are thought to have died off right here on Earth. To put into perspective just how much younger this planet really is than some of the planets here on our solar system, it’s worth noting that our sun is thought to be 4.5 billion years old. "Only a handful of planets have been imaged like this so far, and they were each at least five times the mass of Jupiter, so it’s incredibly hard to do," Professor Bruce Macintosh of Stanford University notes. "Planets like Jupiter in our solar system are a billion times fainter than the sun, so planets are usually lost in the bright light from their star." The high levels of methane gas in the newly-discovered planet’s atmosphere is intriguing to scientists, because it’s reportedly the highest amount of methane gas ever found in an exoplanet. 51 Eridani b is also said to be about twice the size of Jupiter, despite being much younger, and have a temperature of around 800 degrees Fahrenheit, which doesn’t exactly sound like the type of planet life like ours would like to live on. Astronomers are hoping that because the planet is so young, they’ll be able to gather more clues about how it was formed that could help us to understand more about Jupiter and how it came to be. Perhaps the Gemini Planet Imager will soon discover additional interesting planets somewhere in our universe. Source: Discovery News
<urn:uuid:2f7fef6d-a482-4b7d-b867-0aac6dc2cefe>
CC-MAIN-2019-04
https://www.labroots.com/trending/space/1507/jupiter-like-exoplanet-discovered-100-light-years-away
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583684033.26/warc/CC-MAIN-20190119221320-20190120003320-00544.warc.gz
en
0.960479
605
3.65625
4
Babies - Modern vs. Historical Having a baby in modern times has become less risky for mother and child since ancient and historical times. Advancements in scientific research and medicine have provided care options for conditions that would have ended in death for the baby, or the mother, or both. Breech Birth Risks In ancient times a baby that was in breech position when labor began would often caused death to the mother. Sometimes the child could be saved, but blood loss and the strain of a difficult labor increased the possibility of death in the new mother. Women died in childbirth if the infant was trapped in the birth canal or if there was unstoppable bleeding after the delivery. Nowadays, baby images in the form of ultrasounds let medical professionals know for certainty the position of the baby. Often the position can be determined by a skilled practitioner by feel, but an ultrasound can give a better idea as to whether or not the baby can be manipulated into the head down position. In the 21st century, a c-section is often scheduled if a baby is in the breech position. In the 1800s there was no choice but to deliver the baby feet first. The risks were high including irreparable tearing in the mother which could lead to life-threatening infections or excessive bleeding. It could also end in the infant's since there was the chance that the child's oxygen supply would be cut off during the delivery and the baby would die. Baby Information: Modern Infant Mortality Rates Infant mortality rates are at an all time low especially in the developed world. There's such a high rate of confidence in the likelihood of survival of mother and child that it's common for women to have baby showers at some point in the third trimester of pregnancy instead of waiting for the infant to be born. Statistics from 2009 from the World Bank showed that infant mortality in the United States was 6.8 per 1,000 live births. Canada had 5.3 and the United Kingdom had 4.6. Sierra Leon had a one of the world's highest rates of infant death at 112.8 deaths for every 1,000 live births. In the late 1800s and early 1900s infant mortality rates were on average 100 per 1,000 live births with some European countries like Russia and Germany reaching rates as high as 250 deaths per 1,000 live births. This is according to the report, The Decline of Infant Mortality in Europe 1800 to 1850 edited by Carlo A. Corsini and Pier Paalo Viazzo. Wealth did not make any family immune to infant mortality, according to the report. Agricultural laborers tended to have the lowest infant mortality rate up until the early 1900s. Parenting Attitudes 1800s vs. Today Today children tend to be the focus of many couples lives. Babies pictures are taken from a young age and often done repeatedly throughout the first year of a child's life. This wasn't the case in the 1800s since the process of getting a picture taken was so complicated and expensive. Old style photography required the individual to sit still for an extended period of time which could be a challenge for a parent to do with an active baby. The focus and purpose of babies names has changed too. In the 1800s it was more common for names with family ties or religious significance to be chosen. Today some modern parents choose the most unique names possible to reduce the chance of their children having the same name as dozens of their peers when they grow older.
<urn:uuid:2cc932d9-6dc1-40eb-9d91-d8fa266dcc39>
CC-MAIN-2022-05
https://www.pregnancy-info.net/childbirth-history/babies.html
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320300722.91/warc/CC-MAIN-20220118032342-20220118062342-00699.warc.gz
en
0.973251
709
3.390625
3
Live action role play games or Larps share many traits to tabletop RPGs and some video games. Through character development and playing those characters through game situations, players can build empathy for others in addition to their own character. The added benefit of the Larp is to cultivate empathy in situations that are less personal, less risky. For example, a group may learn to deal with bullying in a fantasy setting where an ogre is the bully and is played by the kindest individual in the room or racism is studied through playing a first contact situation with aliens set in the future. At Inquiry Hub, students have designed and played two Larps aimed at grades 9 to 12 with 60 participants. The first Larp was fantasy-based as seen in the video. It emphasized collaboration among individuals while also developing a storyline that put two teams in opposition. In the end, the players had to leave the Island because their actions lead to its destruction. However, through self-sacrifice there was a way to be saved. Recently, our students entered the Cloud 7 space station with a dystopian sci-fi theme. The players had to learn to work together to reboot the station's AI that had been infected by a malevolent virus. In addition, players had to understand their roles in the hierarchy and make alliances with other groups, including androids who wished for more rights. This proved to be a significant challenge for the students, but after the Larp we were able to talk more about how negotiations would have been much more successful in comparison to using force to subdue presumed threats. Our experiences have highlighted the potential of Larps to develop character, identity, and problem solving skills.
<urn:uuid:0cc22e34-3c48-4554-b06a-05ad28a93ccd>
CC-MAIN-2019-30
https://collaborate.teachersguild.org/challenge/empathy/select/educational-live-action-role-play-edularp
s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195527204.71/warc/CC-MAIN-20190721205413-20190721231413-00278.warc.gz
en
0.983635
336
3.546875
4
The last round of king tides, the highest high tides of the year, is coming up January 2-4, 2018. Residents and visitors to Oregon’s coast are invited to capture these high water events in photos. The Oregon King Tides Photo Project is part of a worldwide initiative; anyone with a camera can help document the extent of the extreme high tides, and help us catch a glimpse of what sea level rise will look like in our region. Everyone is welcome to participate, just pick a place, snap a photo and share it online. What are King Tides? The term ‘king tide’ describes a high tide event when the sun, moon, and earth are in alignment, causing greater than usual gravitational pull on the tides. When king tides occur during intense rain or storm events, the water level rise can cause flooding, erosion, and other impacts to infrastructure and property. Why is it important to document them? King Tide events give us the opportunity to peek into the future and see what the impacts of sea level rise could look like in our coastal communities. Even a small increase in sea level could escalate the impacts of winter storms along the Oregon coast, intensify chronic hazards like erosion and flooding, and reduce the width of the public beach. Photographing king tides is an effective way to help coastal communities identify areas prone to flooding, expose potential impacts of sea level rise, and start planning for the future. How can I take a helpful photo? Helpful king tide photos show water levels adjacent to a fixed feature like a piling, seawall or bridge abutment. Including fixed features allows actual water levels to be documented and tracked over time. Good photos also must include the location, the date and time of the photo, and the viewer’s direction for each picture. Two photos taken from the same spot, one during the king tide and the other at a typical high tide are also very effective in highlighting these high water events. Find tide tables for your area and instructions for how to take and upload photos on the King Tides website: www.oregonkingtides.net. Check out some photos already taken during the 2017-2018 season!
<urn:uuid:db53c0ed-b1d5-4509-aaa9-0d373d1a9456>
CC-MAIN-2018-05
https://oregon.surfrider.org/become-a-citizen-scientist-by-photographing-the-last-round-of-king-tides/
s3://commoncrawl/crawl-data/CC-MAIN-2018-05/segments/1516084891543.65/warc/CC-MAIN-20180122213051-20180122233051-00524.warc.gz
en
0.911941
447
2.90625
3
replication is easy, sharding is very easy too. MongoDB provides high availability with replica sets. A replica set consists of two or more copies of the data. Each replica set member may act in the role of primary or secondary replica at any time. All writes and reads are done on the primary replica by default. Secondary replicas maintain a copy of the data of the primary using built-in replication. When a primary replica fails, the replica set automatically conducts an election process to determine which secondary should become the primary. Secondaries can optionally serve read operations, but that data is only eventually consistent by default. MongoDB scales horizontally using sharding. The user chooses a shard key, which determines how the data in a collection will be distributed. The data is split into ranges (based on the shard key) and distributed across multiple shards. (A shard is a master with one or more slaves.). Alternatively, the shard key can be hashed to map to a shard – enabling an even data distribution. MongoDB can run over multiple servers, balancing the load or duplicating data to keep the system up and running in case of hardware failure. MongoDB can be used as a file system, called GridFS, with load balancing and data replication features over multiple machines for storing files. This function, called grid file system, is included with MongoDB drivers. MongoDB exposes functions for file manipulation and content to developers. GridFS can be accessed using mongofiles utility or plugins for Nginx and lighttpd. GridFS divides a file into parts, or chunks, and stores each of those chunks as a separate document. MapReduce can be used for batch processing of data and aggregation operations. The aggregation framework enables users to obtain the kind of results for which the SQL GROUP BY clause is used. Aggregation operators can be strung together to form a pipeline – analogous to Unix pipes. The aggregation framework includes the $lookup operator which can join documents from multiple documents, as well as statistical operators such as standard deviation. MongoDB is not relational so you lose all the capabilities of relational databases. No joins. You design as if they were never an option. It scales well in a narrow range but other NoSQL solutions are better at scaling. Sharding is annoying, complex and seems like it was bolted on. It uses eventual consistency, which is good and bad. No transactions except at the record level. If you want a transaction you have to fit the whole transaction into a single record. Hard to secure properly without going with an Enterprise license. There is no patching, you have to do full upgrades and full upgrades are issued several times a year. So you have to build a patch schedule around those. There is no guarantee that your upgrade will work with your given driver so tests must be scheduled before upgrading.
<urn:uuid:7c62afd5-2258-461f-b541-2f8f80acf577>
CC-MAIN-2018-34
https://www.g2crowd.com/products/mongodb/reviews
s3://commoncrawl/crawl-data/CC-MAIN-2018-34/segments/1534221211935.42/warc/CC-MAIN-20180817084620-20180817104620-00452.warc.gz
en
0.939574
594
2.9375
3
English language Stylistics ENGLISH LANGUAGE STYLISTICS G. Misikova – Linguistic stylistics I. Ur. Galperin- Stylistics Crystal D. and Davy D. Looking into English Style / pertaining to non-literary designs Style and Styleme Štylistika je náuka to výbere a spôsobe využívania jazykových a mimojazykových alebo umeleckých prostriedkov a postupov uplatňovaných sixth is v procese použitia komunikácie. Jazykový štýl u spôsob prejavu, ktorý vzniká cieľavedomým výberom, zákonitým usporiadaním a využitím jazykových a mimojazykových prostriedkov so zreteľom na tematiku, situáciu, funkciu, autorov zámer a obsahové zložky prejavu. Štyléma prvok, pomocou ktorého sa štylizuje, napr. slovo, veta alebo časť celku/ textu, ktorý má takú výrazovú podobu, že pri štylistickej typológii textu signalizuje jeho zameranie, napr. vulgarizmus v humornom texte. Style is a coherent, repeating set of anticipated features, which usually identify design of an individual or maybe a larger, recognizable group. Style can be a variation in a person's speech or perhaps writing and normally varies from formal to casual in line with the type of situation, the person or perhaps person's addresses, the location, the subject discussed etc . Some language specialists use the term variety various other register. Local speaker has intuitive know-how about linguistics appropriateness, foreign scholars have no intuitive sense regarding linguistic appropriateness or less strong or no awareness of conformity, mainly because they haven't grown up in a relevant linguistics environment. Purpose of stylistics is usually to study the varieties of a language in as much details as possible, in order that we can point to the formal linguistics features, which define them and understand the constraints on their employ. 3 practical aims in the stylistics: 1 . To spot and analyse language habits and those popular features of language, which are restricted to selected kinds of cultural contexts. installment payments on your To explain, so why these features have been used as opposed to various other alternatives three or more. To classify these types of features to categories based on the view of their function. Stylistics is a kind of roofing of all numbers of language, it can be closely linked to graphitic and graphology. Graphitic is the research of created or published shape and graphology study regarding the system. Shade, type, size of text can be stylistically relevant, even punctuation. Stylistics and phonetics – the sculpt voice of the sermon gentleman is different of TV advertisement, and different coming from businessmen. Stylistics and phonology - the suprasegmental features can as well report a degree of stylistic significance, like falling expression for distress or reluctance. Stylistics and morphology – the unique use of plural for exaggeration for example or the use of demonstratives in a negative function. Stylistics and syntax – in change of word purchase which doesn't mean error, but stylistic significance: He was angry. Versus Angry he was. intensified Duplication is always combined with rhythm. Stylistics and lexicology- in texts there are a lot of phrases, which may carry only pure that means, but another added as well: Weakling – slaboch: -ling as endsilbe bears expressiveness; is carefully related to stylistic significance. Various other examples: Neutral thin, positive willowy and negative lanky. Stylistics and semantics – jokes depending on homonymy or perhaps homophony, the usage of pun is usually stylistic significant. Our tasks throughout the session: 1 . The investigation from the inventory of special language media, which usually secure the desirable a result of the utterance. These media are called expressive means NA and stylistic devices SECURE DIGITAL. When dealing with them, we will give attention to the artistic function of language on the synonymous means of rendering one and the same idea, emotional colouring in language or perhaps the interrelation between language and thought. NA s and SD h we shall examine on phonetic, syntactic and lexical level. 2 . - study selected types of texts, which will...
<urn:uuid:e036b2c6-39ed-491a-83fb-9c3413677db3>
CC-MAIN-2020-05
https://thehomeschoolofmoney.com/english-stylistics/09320-essay-about-english-stylistics.html
s3://commoncrawl/crawl-data/CC-MAIN-2020-05/segments/1579251694071.63/warc/CC-MAIN-20200126230255-20200127020255-00476.warc.gz
en
0.772773
1,057
2.734375
3
You cannot compare apples with oranges, but of course, you can compare an orange and a նարինջ (narinch), because both refer to the juicy fruit used to make your everyday breakfast beverage and both have the same origin. And, as we will see, apples were somehow part of the origins of the English orange. Oranges probably originated in Southeast Asia, and were already cultivated in China around 2500 B.C. However, the ultimate origin of both English orange and Armenian narinch (pronounced narinj in Classical Armenian) is India. The fruit was called naranga, which means “orange tree,” in the Vedas, the sacred books of Indian religion written in Sanskrit, although the origin of the word is unknown. It seems to come from a non-Indo-European language of the Indian peninsula, such as Telugu, Malalayam, or Tamil. The fruit went from India to Western Asia with the Arabs as intermediaries. Along went the name: it remained as narang in Persian, turned naranj in Arabic (Arabic does not have a g), and became narinj in Kurdish and Armenian. According to German linguist Heinrich Hubschmann and his disciple Hrachia Adjarian, the Armenian word originated from Persian narang. The fruit went to Europe through Portuguese travelers to China, and through Arabs. It received the name of “Chinese apple” in some languages: they are called sinaasappel in Dutch and appelsin in Low German, literally “China’s apple” (hence Russian апельсин apelsin “orange,” which you may hear sometimes in Eastern Armenian). Interestingly, “Chinese apple” is the name of the pomegranate in British English. Interestingly, Spanish and Portuguese most probably adopted the word through Arabic influence in the Iberian Peninsula (Spanish naranja and Portuguese laranja), but Portugal helped spread the word to Southern Europe and the Middle East: Greek πορτοκάλι (portokáli) and Turkish portakal. Believe it or not, Arabs today call the fruit burtuqāl برتقال; the word nerinj is used for a different citric. The name and the fruit reached England through a more indirect path. Old Italian borrowed the word from Arabic and turned it into melarancia (mela + (n)arancia “apple of orange”). The French calqued the word from the Italians and turned it into pome orenge (“apple of orange”). Finally, Old English borrowed orenge/orange from Old French, without the “apple” part. (In the end, the French dropped the word “apple” too.) And this is how English orange sounds quite close to Armenian narinch, only with the vowel o at the beginning. There is one difference, though: orange in English means both the fruit and the tree. In Armenian, we have two different words, although close enough: narinch for the fruit and նարնջենի (narncheni) for the tree. The suffix eni is equivalent to the English “-tree,” as in khntzoreni “apple tree.”
<urn:uuid:3f1a0aeb-04a3-4a5d-9130-a9bb69f39cca>
CC-MAIN-2019-04
http://armenianlanguagecorner.blogspot.com/2013/09/the-apple-that-came-from-china-or-india.html
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583831770.96/warc/CC-MAIN-20190122074945-20190122100945-00291.warc.gz
en
0.950551
728
3.265625
3
In the 2004 reauthorization of the Individuals with Disabilities Education Act (IDEA) the process for determining which children have a specific learning disability (SLD) was altered to include response to intervention (RTI). This change stemmed from criticism of how children were tested for an SLD, which was primarily looking at the discrepancy between a child’s intellect (IQ) and ability. If a severe discrepancy was evident the child was determined to have an SLD. Since an SLD may manifest itself in the imperfect ability to listen, think, speak, read, write, spell, or do mathematical calculations using only the discrepancy model can’t accurately identify all children with an SLD. This is where RTI can be used to fill in the gap where the discrepancy model isn’t working. There is no one definition of RTI but it’s most commonly used as, 1) an early predictor of at-risk students, 2) intervention for general education students with academic or behavioral difficulties, and 3) as a component of SLD determination. Usually, RTI incorporates a tiered system of support that increases the amount and intensity of intervention as you move to a higher tier. Most children in RTI will stay in the lowest tier and after a period of intervention will be back on track. For those children that need more intensive and longer support they will be moved to higher tiers. If a child isn’t showing any progress in RTI, this can be used as evidence that the child has an SLD. Thus, a child whose testing didn’t show a severe discrepancy between intellect and ability can still be determined to be a child with an SLD. RTI can be a very useful tool in the SLD determination process but it should not be used as a stall tactic by the school district to delay assessments for special education. Some School Districts, after receiving written requests from parents to assess their child for special education, have used RTI as an excuse to stop the process. While IDEA does require general education interventions to be tried prior to special education there is nothing in the law that allows the school to stop assessment timelines after a parent requests an assessment. When a parent requests a special education assessment for their child make sure: 1) It’s in writing, 2) The request includes assessing in all areas of suspected disability, and 3) The request specifically requires any general education interventions, such as RTI, to be started during the assessment period and that the intervention doesn’t affect the legally required assessment timelines required by IDEA. The general rule of thumb is there will be approximately 90 days between a parent’s request for assessment and the first IEP meeting. These 90 days include 15 days for the School to put together an assessment plan, at least 15 days for the parents to approve the assessment plan, and 60 days for the assessments. This is plenty of time for the School to try RTI and see how the results are working prior to the first IEP meeting. If used properly, RTI is a powerful general education intervention tool and SLD determination method. It is a positive step forward in fixing the current education system and I look forward to seeing it implemented, in the proper manner, by all School Districts. January 21, 2011 Update: The United States Department of Education Office of Special Education and Rehabilitative Services issued a memo reminding School Districts that Response to Intervention cannot be used to delay-deny an evaluation for eligibility under the Individuals with Disabilities Education Act. You can download the memo here: OSERS Memo on RTI
<urn:uuid:a5bfcac3-6171-4900-82d7-b06135231faf>
CC-MAIN-2016-07
http://www.specialeducationadvisor.com/response-to-intervention-rti-specific-learning-disability-sld/
s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701156448.92/warc/CC-MAIN-20160205193916-00084-ip-10-236-182-209.ec2.internal.warc.gz
en
0.946323
742
3.265625
3
Hypothyroidism is really a condition indexed by an under-active thyroid gland—when a thyroid problem doesn’t produce enough hormones. There are numerous treatments available, however the fundamental concept may be the same—and it’s name is thyroid hormone substitute therapy. To best understand the objective of thyroid hormone substitute therapy, you must know the interaction of T3 and T4—the two thyroid hormones. T3 and T4 The entire name of T3 is triiodothyronine, and T4’s complete name is tetraiodothyronine or thyroxine. T3 and T4 take control of your body’s metabolic process. Without having an adequate amount of them, your metabolic process slows lower. Your metabolism dictates how rapidly you process food, how quickly your heart beats, just how much heat the body creates—and even how rapidly you are able to think. Essentially, T3 and T4 are responsible for the way your body uses energy. T3 and T4 aren’t equal in strength T3 may be the more active hormone of these two. While T3 is more powerful, taking synthetic T4 hormone is the standard strategy to hypothyroidism. The reason behind it is because the majority of the T3 within our physiques really was once T4. When T4 hormones touch other cells within the blood stream, they provide up an iodine atom to have interaction with individuals cells. When T4 loses an iodine atom, it might be T3. If this T4 into T3 conversion occurs, T3 then conveys the metabolic "message" to another cells through the body. The advantage of taking only T4 treatments are that you are allowing the body to do a few of the actions it is supposed to do, that is taking T4 and altering it into T3. The half existence of T4 can also be longer when compared with T3 (seven days versus 24 hrs), this means that it’ll stay for a longer period within your body after ingestion. The objective of Thyroid Hormone Substitute Therapy If you’re prescribed a kind of thyroid hormone substitute therapy, the aim would be to make amends for the possible lack of hormone secreted from your thyroid. Generally, you’ll have a daily dose of T4 inside a pill taken orally. But you need to realize that every patient’s therapy will change. There is no cookie-cutter dosage or plan for treatment with regards to thyroid hormone substitute therapy. The way the body absorbs the endocrine system, along considering the variety of hormones it requires, is really varied. Your plan for treatment is going to be individualistic. As a result, you are very likely some extent of experimentation with regards to locating the dosage and type of therapy that works well with you. Though synthetic T4 supplements would be the most prescribed type of thyroid hormone substitute therapy, there are a number of forms, including animal thyroid supplements. Synthetic T3 can also be from time to time given included in treatment in a few instances, most frequently after thyroid surgery, when awaiting the radioiodine ablation in situation of cancer. Thyroid hormone substitute therapy is an extremely individualized course of treatment, which is impressive when prescribed correctly. The aim of thyroid hormone substitute therapy, generally, would be to normalize your thyroid stimulating hormone (TSH) levels. Both you and your physician will talk about what treatment option will best alleviate your hypothyroid signs and symptoms, enabling you to live a proper, normal existence.
<urn:uuid:121c9616-4901-4d89-8807-e9a83ff9766a>
CC-MAIN-2019-26
https://suamaytinhhaiphong.com/what-s-thyroid-hormone-substitute-therapy-treating/
s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627999000.76/warc/CC-MAIN-20190619143832-20190619165832-00373.warc.gz
en
0.922193
734
2.578125
3
High above the surface, Earth’s magnetic field constantly deflects incoming supersonic particles from the sun. These particles are disturbed in regions just outside of Earth’s magnetic field – and some are reflected into a turbulent region called the foreshock. New observations from NASA’s THEMIS – short for Time History of Events and Macroscale Interactions during Substorms – mission show that this turbulent region can accelerate electrons up to speeds approaching the speed of light. Such extremely fast particles have been observed in near-Earth space and many other places in the universe, but the mechanisms that accelerate them have not yet been concretely understood. The new results provide the first steps towards an answer, while opening up more questions. The research finds electrons can be accelerated to extremely high speeds in a near-Earth region farther from Earth than previously thought possible – leading to new inquiries about what causes the acceleration. These findings may change the accepted theories on how electrons can be accelerated not only in shocks near Earth, but also throughout the universe. Having a better understanding of how particles are energized will help scientists and engineers better equip spacecraft and astronauts to deal with these particles, which can cause equipment to malfunction and affect space travelers. “This affects pretty much every field that deals with high-energy particles, from studies of cosmic rays to solar flares and coronal mass ejections, which have the potential to damage satellites and affect astronauts on expeditions to Mars,” said Lynn Wilson, lead author of the paper on these results at NASA’s Goddard Space Flight Center in Greenbelt, Maryland. The results, published in Physical Review Letters, on Nov. 14, 2016, describe how such particles may get accelerated in specific regions just beyond Earth’s magnetic field. Typically, a particle streaming toward Earth first encounters a boundary region known as the bow shock, which forms a protective barrier between the solar wind, the continuous and varying stream of charged particles flowing from the sun, and Earth. The magnetic field in the bow shock slows the particles, causing most to be deflected away from Earth, though some are reflected back towards the sun. These reflected particles form a region of electrons and ions called the foreshock region. This image represents one of the traditional proposed mechanisms for accelerating particles across a shock, called a shock drift acceleration. The electrons (yellow) and protons (blue) can be seen moving in the collision area where two hot plasma bubbles collide (red vertical line). The cyan arrows represent the magnetic field and the light green arrows, the electric field. Credits: NASA Goddard’s Scientific Visualization Studio/Tom Bridgman, data visualizer Some of those particles in the foreshock region are highly energetic, fast moving electrons and ions. Historically, scientists have thought one way these particles get to such high energies is by bouncing back and forth across the bow shock, gaining a little extra energy from each collision. However, the new observations suggest the particles can also gain energy through electromagnetic activity in the foreshock region itself. The observations that led to this discovery were taken from one of the THEMIS – short for Time History of Events and Macroscale Interactions during Substorms – mission satellites. The five THEMIS satellites circled Earth to study how the planet’s magnetosphere captured and released solar wind energy, in order to understand what initiates the geomagnetic substorms that cause aurora. The THEMIS orbits took the spacecraft across the foreshock boundary regions. The primary THEMIS mission concluded successfully in 2010 and now two of the satellites collect data in orbit around the moon. Operating between the sun and Earth, the spacecraft found electrons accelerated to extremely high energies. The accelerated observations lasted less than a minute, but were much higher than the average energy of particles in the region, and much higher than can be explained by collisions alone. Simultaneous observations from the additional Heliophysics spacecraft, Wind and STEREO, showed no solar radio bursts or interplanetary shocks, so the high-energy electrons did not originate from solar activity. “This is a puzzling case because we’re seeing energetic electrons where we don’t think they should be, and no model fits them,” said David Sibeck, co-author and THEMIS project scientist at NASA Goddard. “There is a gap in our knowledge, something basic is missing.” The electrons also could not have originated from the bow shock, as had been previously thought. If the electrons were accelerated in the bow shock, they would have a preferred movement direction and location – in line with the magnetic field and moving away from the bow shock in a small, specific region. However, the observed electrons were moving in all directions, not just along magnetic field lines. Additionally, the bow shock can only produce energies at roughly one tenth of the observed electrons’ energies. Instead, the cause of the electrons’ acceleration was found to be within the foreshock region itself. “It seems to suggest that incredibly small scale things are doing this because the large scale stuff can’t explain it,” Wilson said. High-energy particles have been observed in the foreshock region for more than 50 years, but until now, no one had seen the high-energy electrons originate from within the foreshock region. This is partially due to the short timescale on which the electrons are accelerated, as previous observations had averaged over several minutes, which may have hidden any event. THEMIS gathers observations much more quickly, making it uniquely able to see the particles. Next, the researchers intend to gather more observations from THEMIS to determine the specific mechanism behind the electrons’ acceleration. In a remarkable demonstration of the extreme limits of nanoscale engineering, researchers from the US and China have used the tip of a scanning tunnelling microscope (STM) to cleave and form selected chemical bonds on a complex molecule. The work marks another step along the road of ever-increasing control of single-molecule manipulations. The team, led by veteran molecule manipulator Wilson Ho of the University of California, Irvine, designed a compound consisting of a chain of three aromatic rings with a tail at each end, comprising an acetyl group attached to the ring system by a sulfur. This molecule was then adsorbed to a nickel-aluminium surface and the tip of the STM was positioned over one of the ‘tail ends’ of the molecule. Electrons from the tip were injected into the sulfur–acetyl bond. At a particular threshold energy of the electrons, the bond snaps. ‘This occurs with surprisingly high efficiency,’ Ho says, with the energy being localised at the bond and with relatively little dissipation across the rest of the molecule. Similarly, the bond between the sulfur and the aromatic ring can also be selectively cleaved at a given end of the molecule. By this method the researchers sequentially broke all four bonds, leaving the aromatic chain remaining. In a second series of experiments the team cleaved the acetyl groups, exposing the sulfurs. They deposited gold atoms onto the surface and gradually nudged them closer to the sulfurs with the STM tip. ‘Sometimes a bond forms, but sometimes it doesn’t,’ Ho says. By ‘energising’ the sulfur and the gold with electrons from the STM, a much higher rate of bond formation was achieved. ‘By exciting the gold and the sulfur, we can form a bond through a kind of “nano-welding” process,’ says Ho. The microscope was also able to image the patterns of electronic distribution as the various bonds formed and broke. The team used 1,4-bis[4’-(acetylthio)styryl]benzene to test its ‘nano-welding’ technique Other experts in the field recognise the significance of the new work. Geoff Thornton, of UCL and the London Centre for Nanotechnology, UK, says: ‘It’s a lovely piece of work.’ Thornton adds: ‘The modification of the molecular electronic structure on attaching a gold atom at both ends of the molecule is also fascinating, but I was left wondering what role the alloy substrate might play in this process.’ ‘Ho’s team shows that the techniques that work on small and well-studied molecules also work beautifully on large extended chemical structures,’ says Peter Sloan of the University of Bath, UK. ‘Atomically engineered molecules probably won’t be appearing be appearing in the high-street soon, but these researchers have taken a small but important step.’ Y Jiang et al, Nat. Chem., 2012, DOI: 10.1038/nchem.1488 Watch this and other space videos at http://SpaceRip.com In high-res 1080p. Explores one of the deepest mysteries about the origin of our universe. According to standard theory, the early moments of the universe were marked by the explosive contact between subatomic particles of opposite charge. Featuring short interviews with Masaki Hori, Tokyo University and Jeffrey Hangst, Aarhus University. Scientists are now focusing their most powerful technologies on an effort to figure out exactly what happened. Our understanding of cosmic history hangs on the question: how did matter as we know it survive? And what happened to its birth twin, its opposite, a mysterious substance known as antimatter? A crew of astronauts is making its way to a launch pad at the Kennedy Space Center in Florida. Little noticed in the publicity surrounding the close of this storied program is the cargo bolted into Endeavor’s hold. It’s a science instrument that some hope will become one of the most important scientific contributions of human space flight. It’s a kind of telescope, though it will not return dazzling images of cosmic realms long hidden from view, the distant corners of the universe, or the hidden structure of black holes and exploding stars. Unlike the great observatories that were launched aboard the shuttle, it was not named for a famous astronomer, like Hubble, or the Chandra X-ray observatory. The instrument, called the Alpha Magnetic Spectrometer, or AMS. The promise surrounding this device is that it will enable scientists to look at the universe in a completely new way. Most telescopes are designed to capture photons, so-called neutral particles reflected or emitted by objects such as stars or galaxies. AMS will capture something different: exotic particles and atoms that are endowed with an electrical charge. The instrument is tuned to capture “cosmic rays” at high energy hurled out by supernova explosions or the turbulent regions surrounding black holes. And there are high hopes that it will capture particles of antimatter from a very early time that remains shrouded in mystery. The chain of events that gave rise to the universe is described by what’s known as the Standard model. It’s a theory in the scientific sense, in that it combines a body of observations, experimental evidence, and mathematical models into a consistent overall picture. But this picture is not necessarily complete. The universe began hot. After about a billionth of a second, it had cooled down enough for fundamental particles to emerge in pairs of opposite charge, known as quarks and antiquarks. After that came leptons and antileptons, such as electrons and positrons. These pairs began annihilating each other. Most quark pairs were gone by the time the universe was a second old, with most leptons gone a few seconds later. When the dust settled, so to speak, a tiny amount of matter, about one particle in a billion, managed to survive the mass annihilation. That tiny amount went on to form the universe we can know – all the light emitting gas, dust, stars, galaxies, and planets. To be sure, antimatter does exist in our universe today. The Fermi Gamma Ray Space Telescope spotted a giant plume of antimatter extending out from the center of our galaxy, most likely created by the acceleration of particles around a supermassive black hole. The same telescope picked up signs of antimatter created by lightning strikes in giant thunderstorms in Earth’s atmosphere. Scientists have long known how to create antimatter artificially in physics labs – in the superhot environments created by crashing atoms together at nearly the speed of light. Here is one of the biggest and most enduring mysteries in science: why do we live in a matter-dominated universe? What process caused matter to survive and antimatter to all but disappear? One possibility: that large amounts of antimatter have survived down the eons alongside matter. In 1928, a young physicist, Paul Dirac, wrote equations that predicted the existence of antimatter. Dirac showed that every type of particle has a twin, exactly identical but of opposite charge. As Dirac saw it, the electron and the positron are mirror images of each other. With all the same properties, they would behave in exactly the same way whether in realms of matter or antimatter. It became clear, though, that ours is a matter universe. The Apollo astronauts went to the moon and back, never once getting annihilated. Solar cosmic rays proved to be matter, not antimatter. It stands to reason that when the universe was more tightly packed, that it would have experienced an “annihilation catastrophe” that cleared the universe of large chunks of the stuff. Unless antimatter somehow became separated from its twin at birth and exists beyond our field of view, scientists are left to wonder: why do we live in a matter-dominated universe?
<urn:uuid:7760f66e-32bb-4cdf-8942-18290719f38a>
CC-MAIN-2017-47
http://myscienceacademy.org/tag/electrons/
s3://commoncrawl/crawl-data/CC-MAIN-2017-47/segments/1510934806317.75/warc/CC-MAIN-20171121055145-20171121075145-00263.warc.gz
en
0.94211
2,817
4.1875
4
Page 1 of 4 Please PAUSE the "How to Draw a Dragonfly" video after each step to draw at your own pace. For the first few steps, don't press down too hard with your pencil. Use light, smooth strokes to begin. Step 1: Draw a small circle as a guide for the dragonfly's head. The circle doesn't have to be perfect. It's just a guide. Step 2: Draw an oval under the head for dragonfly's midsection or thorax. The thorax should be almost twice as tall as the head and a bit thinner than the head. Step 3: Under the thorax, draw a very long oval-like shape as a guide for the dragonfly's abdomen. The abdomen should be about three times as tall as the thorax and should be skinnier than the thorax. Step 4: Draw the dragonfly's leg guides using a series of lines on the sides of the thorax. The first pair of legs are near the top of the thorax pointing up. Bend the lines to indicate the joints. The second pair of legs are right below the first, and they point to the sides more. They are similar to wide letter V's. The third and final pair of legs are closer to the bottom of the thorax, and they point downward. Step 5: Draw two long oval-like shapes on the sides of the thorax as guides for the first pair of dragonfly wings. Sketch lightly at first to get the shape of the guide right. Notice that the dragonfly wing guide has a diagonal orientation and is not strictly horizontal. Draw the guide for the other front wing on the right side the same way. Try to draw both sides similarly so your drawing has symmetry.
<urn:uuid:300accd0-4adc-4a38-83df-bcc96516b44e>
CC-MAIN-2022-05
https://how2drawanimals.com/8-animals/108-draw-dragonfly.html
s3://commoncrawl/crawl-data/CC-MAIN-2022-05/segments/1642320302622.39/warc/CC-MAIN-20220120190514-20220120220514-00582.warc.gz
en
0.935578
362
3.96875
4
Simple Machines Enrichment Program Introduce students to mechanical engineering with step-by-step builds and open-ended experimentation in this kit! This kit teaches students about simple machines and how engineering is important to our society. Included step-by-step instructions apply real-world application as students build visual models of levers, pulleys, and inclined planes that proliferate structures all over the world. As the course progresses, students advance to discover wedges, worm gears, and even complex machines. Critical thinking is encouraged as students experiment with the models and build upon existing knowledge. In a grand finale, students are challenged to design the ultimate machine of their imaginations! Kit includes 3,000+ pieces: 8 Universal II Fischertechnik kits (400+ pieces each) plus 8 student guides, 1 instructor guide, and 4 convenient storage bins to keep it all organized. See full contents list below. Recommended for Upper Elementary and up. - Fischertechnik Building Pieces, Set of 3,000+ - Rubber Bands, 1/4 lb - Gratnell Tubs with Lids and Inserts, 4 Ea - Instructor Manual, 1 Ea - Printed Student Guides, 8 Ea - Digital Curriculum Download Click here to download a sample of the NGSS-aligned curriculum manual.
<urn:uuid:57a386fa-3a25-4836-86d8-afff9eb29945>
CC-MAIN-2023-14
https://stem-supplies.com/simple-machines-enrichment-program
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296949355.52/warc/CC-MAIN-20230330163823-20230330193823-00059.warc.gz
en
0.877944
300
3.53125
4
Wednesday, December 29, 2010 Romania - Maramureş Romania - Maramureş - Rural traditions in Hoteni. Sent by Tatiana from Romania. This is from Wikipedia : Maramureş (in Romanian; Rusyn: Марамуреш / Maramuresh; Hungarian: Máramaros; Latin: Marmatia; Ukrainian: Мармарощина / Marmaroshchyna, Мараморщина / Maramorshchyna)) is a geographical, historical and ethno-cultural region in northeastern Carpathians, along the upper Tisza River; it covers the Maramureş Depression and the surrounding Carpathian mountains. Before the Treaty of Trianon, the whole territory belonged to Hungary, and formed Máramaros County. The borders then divided the region between Romania and Czechoslovakia. The southern section is now part of Maramureş County in northern Romania; the northern section is part of Zakarpattia Oblast of western Ukraine. Alternatively, the name Maramureş is used for the Maramureş County of Romania, which contains the southern section of the former historical region and also small parts of the interbellic Satu Mare County and Sălaj County (former pre-World War I Szatmár and Szilágy counties). Maramureş is a valley totally enclosed by mountains, Oaş, Gutâi, Ţibleş and Rodnei to the west and south, Maramureş Mountains and Ukrainian Carpathians (Wooded Carpathians) to the east and north, with a thin opening at Khust, with several dozen small mountain rivers and creeks flowing into the river Tisza (Tisa). It is forested and not easily accessible. The limits of the region is between the parallels of 47o33' N and 40o02' N and the meridians of 23o15' E and 25o03' E. Maramureş represents one of the largest depressions in the Carpathians, covering an area of about 10000 km². Its length from Khust to Prislop Pass is about 150 km and the width from North to South is up to 80 km. The main mountain passes linking Maramureş with the neighboring regions are high and in the past were hardly accessible in the winter. Prislop Pass is 1414m high to the East towards Moldavia, Dealul Ştefăniţei, 1254m in the south towards Transylvania, Pintea Pass 987 m towards Baia Mare and Fărgău 587 in the west towards Oaş Country. In the north Frasini Pass (Yasinia) is 931 m high, linking the region to Galicia. The mountains surrounding this region occupy more than a half of area, and reach 2000 m by several peaks like Pietrosul 2303m in Rodnei Mountains to the south and Hovîrla 2061 in Muntele Negru (Cernahora) to the north. The heavy forested mountains sustain many protected species of plants like yew (Taxus baccata), larch (Larix decidua), Swiss pine (Pinus cembra), edelweiss (Leontopodium alpinum); and animals like lynx (Lynx lynx), chamois (Rupicapra rupicapra), alpine marmot (Marmota marmota), golden eagle (Aquila chrysaetos), and capercaillie (Tetrao urogallus). Here lived the last Romanian wisent, hunted to extinction in 1852. Lostriţa, a local variety of salmon, still live in the mountain rivers. On this ground protected areas have been created here, Rodnei Mountains National Park in 1930, Maramureş Mountains Natural Park in 2004. Many other caves, gorges, cliff formations and lakes are officially protected areas.
<urn:uuid:f5ece025-b138-4509-a98a-53863829f963>
CC-MAIN-2017-51
http://projeksatudunia.blogspot.com/2010/12/romania-maramures.html
s3://commoncrawl/crawl-data/CC-MAIN-2017-51/segments/1512948513866.9/warc/CC-MAIN-20171211183649-20171211203649-00500.warc.gz
en
0.875304
884
2.75
3
It is easy to think of metadata as a few information fields that you tack onto an article when you submit it to a CMS. In this vision, metadata is something fairly small, a triviality in comparison to the content itself. The reality is just the opposite; for any piece of content or data, the metadata is bigger than the data. Metadata is data that explains data. And since data needs to be explained to be useful, metadata also needs metadata to explain it. This means that behind every piece of data there is a metadata cascade that eventually leads back to unstructured stories. Therefore the metadata for any given piece of data or content is all of the data that explains the entirety of that data/content. That’s a lot. Consider a fairly simple example. When you create a table in an HTML document, you are adding metadata to the content in the form of table markup. The table markup allows a browser to display the content as a table. But how does the browser know what the HTML tag <table> means or how to display it? The table tag is metadata to your content, but it is also data, and as such it needs metadata to explain what it means. That metadata is the HTML specification. Take a look. It’s pretty big. Bigger than your table, probably. The metadata is bigger than the data. Of course, the HTML specification is part of the metadata of millions of documents. The total size of all the HTML documents in the world is vastly greater than the size of the HTML specification. This is as it should be. Content or data that has common metadata can be processed by common algorithms. If we did not have a huge amount of content that shared the metadata of the HTML specification, we would not have the Web. Useful content always has a combination of common metadata and unique metadata. Without common metadata, it could not share any common processing, and would be very hard to manage. Without unique metadata, it would be redundant. Some of its common metadata it will share with a very large body of content, while other common metadata may only be shared with a limited range of content. As an author, your responsibility for metadata does not end with the few pieces of data you type into the metadata record when you submit your content to the CMS. Your responsibility goes much deeper than that. You are responsible for making sure that your content conforms to its metadata — all of it. Whatever format you are writing in, be in HTML, DITA, DOCBook, MarkDown or whatever, you have to tag your content with the appropriate tags. That is applying metadata. The process is fairly simple: you pick from a list of available tags. But the implications are significant. An entire processing infrastructure relies on the metadata attached to your content, and the entire metadata cascade of that explains that data. You may never have read the DITA specification, but if you are writing in DITA, your content is being processed by algorithms written by someone who did (or should have). And they wrote those algorithms on the good faith assumption that when you wrapped some piece of content in a particular DITA tag, that that tag was the correct tag to use for that piece of content: that your content conformed to the metadata you applied to it. Which means, in effect, that the algorithm is written on the tacit assumption that you have read the DITA spec, fully understood it, completely remembered it, and consistently applied it to every piece of content that you wrote. Which, of course, you haven’t. Just, as you haven’t read the data dictionary of the bank’s database tables when you type information into an ATM or an online banking app. To a very large extent, we could characterize usability in digital systems as avoiding the need for users to understand the full metadata cascade behind the data they create. (The rest of usability is about avoiding mechanical awkwardness — like typing angle brackets for XML tags.) How is this done? - By providing natural language prompts that guide the writer (an interpretation of their true metadata). - By designing systems such that you only ask writers for data in a domain they understand well (meaning that they actually do understand most of the relevant metadata). - By designing systems that do lots of validation of data input and give good feedback when the data provided is inconsistent with the metadata. Unfortunately, all of these things are difficult to do when it comes to content. - Not only is it hard to fit natural language prompting onto the screen without obscuring the document itself, the author is creating so much metadata so quickly as they type that a high level of guidance would kill productivity. And in many cases, the markup schema provides so many choices at every juncture that the full guidance to selecting between them is overwhelming. This forces the author to learn the schema and hold it in their head when writing, a cognitive load that is made worse by the size and complexity of many schemas. - Much of the metadata that authors are asked for relates to the publishing domain or the content management domain, which the average author is not familiar with. Even the abstract document structures that typically make up a good portion of many schemas are not part of the native way that authors think about documents. Authors are therefore forced to learn a good deal about these foreign domains in order to create content that is valid for the metadata they are applying to it. - It is easy enough to validate if a person is attempting to withdraw more money than they have in their bank account. It is much more difficult to validate that a string of sentences validly fulfil the requirements of the markup attached to them. And this is made significantly worse by the very loose structures found in many schemas. By creating schemas that are largely free of constraints, we put all the burden on authors to follow appropriate constraints themselves, which again means they have to know and understand the metadata cascade behind those constraints. That the metadata is bigger than the data is true for all systems. The successful systems are the ones that have successfully managed this fact by encapsulating the metadata in a way that does not require extensive metadata knowledge of users. The content management industry has not done a good job of this. Many content management projects succeed, but many fail — far more than we should expect at this stage in the development of the industry. Some projects that were judged an initial success break down over time. Such failures can plausibly be traced back either to a loosening of metadata discipline, or a loss of understanding of the complex metadata system among users, or to the accumulation of content that does not fully conform to the metadata requirements, but which was not detected at the time it was created. (Such non-conformance can actually become self-perpetuating, as non-compliant workarounds are used to compensate for existing non-compliant content.) A large part of this failure lies in its failure to design systems with this problem in mind. Complex metadata requirements accumulate during system design without much thought as to how they may be successfully hidden from users. Attempts to solve the problem often take the form of a gloss over the system — putting WYSIWYG interfaces over XML tags. But these approaches really only address the mechanical awkwardness aspects of usability. They do not encapsulate metadata requirements successfully. As noted above, the encapsulation of metadata is an inherently difficult proposition for content — moreso perhaps than for any other type of data. So we should not expect that content management will ever be as easy as using an ATM. But we could benefit greatly from a ground up rethink of the content management model that had this problem at the heart of it.
<urn:uuid:1e195944-23a0-4c8d-946a-3beb4ac14f32>
CC-MAIN-2019-51
https://everypageispageone.com/2015/10/14/the-metadata-is-bigger-than-the-data/
s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575540487789.39/warc/CC-MAIN-20191206095914-20191206123914-00533.warc.gz
en
0.947372
1,581
2.890625
3
Christmas could have increased the risk of diabetes for millions of people, experts have warned. Charity Diabetes UK said that eating and drinking too much over the festive period could have increased many people's risk of developing prediabetes - an under-diagnosed and symptomless condition that puts you up to 15 times more likely to develop Type 2 diabetes. The charity said that high-calorie foods enjoyed by most people over Christmas could boost their weight over a very short time. With mince pies containing about 200 calories each, Christmas cake with 250 calories per slice and mulled wine at 250 calories a glass, the recommended calorie intake of 2,000 a day for women and 2,500 for men can be quickly exceeded. The charity said that if people were already overweight, had high blood pressure or a family history of Type 2 diabetes, they were at serious risk of prediabetes. Those with prediabetes have blood sugar levels higher than normal, but not high enough to be diagnosed as Type 2 diabetes. However, Diabetes UK said that prediabetes could often be reversed through losing even just a moderate amount of weight, adopting a healthy, balanced diet and increasing physical activity levels. "Prediabetes is often a precursor to Type 2 diabetes which can lead to a shortened life expectancy and devastating complications such as heart disease, stroke, blindness, kidney failure and amputation," the Scotsman quoted Diabetes UK chief executive Douglas Smallwood as saying. "Make it your New Year's resolution to eat more healthily and take regular physical activity," Smallwood added.
<urn:uuid:efba26d5-76f2-4b67-bb42-5e770d82299c>
CC-MAIN-2014-10
http://www.dnaindia.com/health/report-christmas-could-have-increased-diabetes-risk-for-millions-say-experts-1328330
s3://commoncrawl/crawl-data/CC-MAIN-2014-10/segments/1393999679121/warc/CC-MAIN-20140305060759-00035-ip-10-183-142-35.ec2.internal.warc.gz
en
0.965662
320
2.890625
3
Students of the Karlsruhe University of Applied Sciences recently designed an EV called the E-Quickie whose claim to fame is that it gets its energy from electric conducting paths on the ground. Receivers underneath the car take energy from the tracks through electric induction, directing it onward to the car’s electrical hub drive.The principle is not unlike that used with AGVs, except that the E-Quicky gets its power, not just its directional instructions, via the tracks below it. Individual vehicle components (e.g. steering and braking system, chassis), use high-tech materials. The outer skin of the vehicle’s body is carbon. The entire vehicle weighs 60 kg and it looks as though 40 kg might be possible through further optimization. In May, students took part in a race wherein the car completed 40 laps on the 222-m conductor track. The car also carries small batteries for power when the car is not over its power-carrying tracks. “We went to the start with half-filled batteries and returned with full ones,” says the project director. Batteries power the car when it leaves the track and to go, say, to the garage. The motor only has a horsepower of 2 kW and still the vehicle reaches a speed of 50 km/hr.
<urn:uuid:fc9a691f-537f-42b0-ab61-be9930d670e2>
CC-MAIN-2018-13
http://www.powerelectronics.com/markets/electric-vehicle-wireless-twist
s3://commoncrawl/crawl-data/CC-MAIN-2018-13/segments/1521257647649.70/warc/CC-MAIN-20180321121805-20180321141805-00559.warc.gz
en
0.960608
269
3.109375
3
Tarentum, Magna Graecia |Died||347 BC (aged 80–81)| Archytas (//; Greek: Ἀρχύτας; 428–347 BC) was an Ancient Greek philosopher, mathematician, astronomer, statesman, and strategist. He was a scientist of the Pythagorean school and famous for being the reputed founder of mathematical mechanics, as well as a good friend of Plato. Life and work Archytas was born in Tarentum, Magna Graecia and was the son of Mnesagoras or Histiaeus. For a while, he was taught by Philolaus, and was a teacher of mathematics to Eudoxus of Cnidus. Archytas and Eudoxus' student was Menaechmus. As a Pythagorean, Archytas believed that only arithmetic, not geometry, could provide a basis for satisfactory proofs. Archytas is believed to be the founder of mathematical mechanics. As only described in the writings of Aulus Gellius five centuries after him, he was reputed to have designed and built the first artificial, self-propelled flying device, a bird-shaped model propelled by a jet of what was probably steam, said to have actually flown some 200 meters. This machine, which its inventor called The pigeon, may have been suspended on a wire or pivot for its flight. Archytas also wrote some lost works, as he was included by Vitruvius in the list of the twelve authors of works of mechanics. Thomas Winter has suggested that the pseudo-Aristotelian Mechanical Problems is an important mechanical work by Archytas, not lost after all, but misattributed. Archytas named the harmonic mean, important much later in projective geometry and number theory, though he did not invent it. According to Eutocius, Archytas solved the problem of doubling the cube in his manner with a geometric construction. Hippocrates of Chios before, reduced this problem to finding mean proportionals. Archytas' theory of proportions is treated in book VIII of Euclid's Elements, where is the construction for two proportional means, equivalent to the extraction of the cube root. According to Diogenes Laertius, this demonstration, which uses lines generated by moving figures to construct the two proportionals between magnitudes, was the first in which geometry was studied with concepts of mechanics. The Archytas curve, which he used in his solution of the doubling the cube problem, is named after him. Politically and militarily, Archytas appears to have been the dominant figure in Tarentum in his generation, somewhat comparable to Pericles in Athens a half-century earlier. The Tarentines elected him strategos, 'general', seven years in a row – a step that required them to violate their own rule against successive appointments. He was allegedly undefeated as a general, in Tarentine campaigns against their southern Italian neighbors. The Seventh Letter of Plato asserts that Archytas attempted to rescue Plato during his difficulties with Dionysius II of Syracuse. In his public career, Archytas had a reputation for virtue as well as efficacy. Some scholars have argued that Archytas may have served as one model for Plato's philosopher king, and that he influenced Plato's political philosophy as expressed in The Republic and other works (i.e., how does a society obtain good rulers like Archytas, instead of bad ones like Dionysus II?). Archytas may have drowned in a shipwreck in the shore of Mattinata, where his body lay unburied on the shore until a sailor humanely cast a handful of sand on it. Otherwise, he would have had to wander on this side of the Styx for a hundred years, such the virtue of a little dust, munera pulveris, as Horace calls it in Ode 1.28 on which this information on his death is based. The poem, however, is difficult to interpret and it is not certain that the shipwrecked and Archytas are in fact the same person. The Archytas curve is created by placing a semicircle (with a diameter of d) on the diameter of one of the two circles of a cylinder (which also has a diameter of d) such that the plane of the semicircle is at right angles to the plane of the circle and then rotating the semicircle about one of its ends in the plane of the cylinder's diameter. This rotation will cut out a portion of the cylinder forming the Archytas curve. Another way of thinking of this construction is that the Archytas curve is basically the result of cutting out a torus formed by rotating a hemisphere of diameter d out of a cylinder also of diameter d. A cone can go through the same procedures also producing the Archytas curve. Archytas used his curve to determine the construction of a cube with a volume of half of that of a given cube. - Archita; Pitagora, Sito ufficiale del Museo Archeologico Nazionale di Napoli, retrieved 25 September 2012 - Morris Kline, Mathematical Thought from Ancient to Modern Times Oxford University Press, 1972 p. 49 - Diogenes Laertius, Vitae philosophorum, viii.83. - Aulus Gellius, "Attic Nights", Book X, 12.9 at LacusCurtius - ARCHYTAS OF TARENTUM, Technology Museum of Thessaloniki, Macedonia, Greece - Modern rocketry - Automata history - Vitruvius, De architectura, vii.14. - Thomas Nelson Winter, "The Mechanical Problems in the Corpus of Aristotle," DigitalCommons@University of Nebraska - Lincoln, 2007. - J. J. O'Connor and E. F. Robertson. Archytas of Tarentum. The MacTutor History of Mathematics archive. Visited 11 August 2011. - Eutocius, commentary on Archimedes' On the sphere and cylinder. - Plato blamed Archytas for his contamination of geometry with mechanics (Plutarch, Symposiacs, Book VIII, Question 2): And therefore Plato himself dislikes Eudoxus, Archytas, and Menaechmus for endeavoring to bring down the doubling the cube to mechanical operations; for by this means all that was good in geometry would be lost and corrupted, it falling back again to sensible things, and not rising upward and considering immaterial and immortal images, in which God being versed is always God. - von Fritz, Kurt (1970). "Archytas of Tarentum". Dictionary of Scientific Biography 1. New York: Charles Scribner's Sons. pp. 231–233. ISBN 0-684-10114-9. - Huffman, Carl A. Archytas of Tarentum, Cambridge University Press, 2005, ISBN 0-521-83746-4 |Wikiquote has quotations related to: Archytas| |Wikisource has original works written by or about: |Wikimedia Commons has media related to Archytas.| - Archytas entry by Carl Huffman in the Stanford Encyclopedia of Philosophy - O'Connor, John J.; Robertson, Edmund F., "Archytas", MacTutor History of Mathematics archive, University of St Andrews. - Diogenes Laërtius, Life of Archytas, translated by Robert Drew Hicks (1925). - Pseudo-Aristotle, Mechanica – Greek text and English translation - Complete fragments (Greek–Spanish bilingual edition) - Fragments and Life of Archytas - FARAH CALDERÓN, Walter: Entre la mentira y la verdad pitagórica: el caso de Arquitas de Tarento.
<urn:uuid:fc85df75-87d0-40b9-8620-1f2190824760>
CC-MAIN-2016-07
https://en.wikipedia.org/wiki/Archytas
s3://commoncrawl/crawl-data/CC-MAIN-2016-07/segments/1454701155060.45/warc/CC-MAIN-20160205193915-00095-ip-10-236-182-209.ec2.internal.warc.gz
en
0.941351
1,653
3.46875
3
Key word: plan 1 Verb + plan Use the correct form of these verbs: change go ahead with - It’s important that you all look ahead and ____________ plans for the future. - We had to____________ our holiday plans at the last minute, when my mother was taken into hospital. - The school was forced to____________ its plans for a new library due to a lack of financial support. - I’m afraid the picnic had to be cancelled. The bad weather ____________ all our plans. - The committee unanimously ____________the plan. Not one person opposed it. Everybody supported it. - Yesterday the government ____________ plans to create thousands of new jobs. - If I ____________ my plan to open a restaurant, can I count on your support? I’ll need good advice and some financial backing to____________ my plans off the ground. 2 Common adjective collocations Use the following adjectives in these sentences: audacious contingency crazy detailed future original - Although____________ plans were drawn up for a new housing development, it didn’t get planning permission. - What a____________ plan! It’ll never work. Who on earth thought this one up? - Managers have drawn up a ____________ plan to keep the factory running if there is a power failure. - We looked at a number of different plans for a bridge over the river, but after much discussion we decided to stick with the ____________ plan of building a tunnel under it. - The interviewer asked me what my ____________ plans were. I could only tell him that they were dependent on getting this job! - The prisoners put together an ____________ plan to escape from prison using a helicopter. 3 Noun + preposition + plan Complete the sentences with these nouns: beauty change details favour opposition success - There’s been a _____________ of plan. We’re leaving next week instead of this week. - We expect to come up against a lot of _____________ to the plan. A number of people have voiced their objections to it already. - Remember that the_____________ or failure of the plan depends on you. - Everybody wants to know how the government is going to pay for its ambitious proposals, but the PM wouldn’t give away any_____________ of their plan. - The government pledged its support for the plan to build a new runway at Gatwick airport. However, none of the local residents are in _____________ of the plan. - The_____________ of this plan, what makes it so good, is that it won’t cost the members of the club a penny! In 1-3 ‘cancel’ and ‘drop’ are also possible. In 1-4 ‘spoiled’ and ‘scuppered’ are also possible. 2. Note these ways of expressing opposition to a plan: We remain strongly opposed to these new plans. Parents are fiercely opposed to the plan to close the local primary school. I am totally opposed to the plan to legalise the use of cannabis for personal use. We are all fundamentally opposed to this plan to change the voting system. The workers are bitterly opposed to the plan to increase the length of the working day. 3. Note these expressions: There was an airline strike, so all our plans fell through at the last minute. I don’t have any plans for the weekend. Your best plan would be to take the train. If everything goes according to plan, we should arrive around midday. It soon became obvious that plan A wasn’t going to work, so we had to resort to plan B. 1 make 2. change 3. scrap 4. wrecked 5 approved 6. announced 7. go ahead with, get - detailed 2. crazy 3. contingency 4. original 5. future 6. audacious - change 2. opposition 3. success 4. details 5. favour 6. beauty Main IELTS Pages: This website is to develop your IELTS skills with tips, model answers, lessons, free books, and more. Each section (Listening, Speaking, Writing, Reading) has a complete collection of lessons to help you improve your IELTS skills. Subscribe for free IELTS lessons/Books/Tips/Sample Answers/Advice from our IELTS experts. We help millions of IELTS learners maximize their IELTS scores! Latest posts by IELTS Editor (see all) - Improve Your Vocabulary in IELTS Speaking & Writing – Key Word: Thing - April 1, 2018 - Boost Your Lexical Resource – Key Word: Rule - March 26, 2018 - Using Collocation to Boost Your IELTS Score – Key Word: promise - November 20, 2017
<urn:uuid:6ea9a368-eff9-4525-aaf7-871acf7652d4>
CC-MAIN-2018-30
https://ieltsmaterial.com/using-collocation-to-boost-your-ielts-score-key-word-plan/
s3://commoncrawl/crawl-data/CC-MAIN-2018-30/segments/1531676589029.26/warc/CC-MAIN-20180716002413-20180716022413-00095.warc.gz
en
0.919255
1,034
2.65625
3
Inflation has once again gone outside the bounds of the Bank of Canada’s 1- to 3-per-cent target, falling to an annualized rate of 0.8 per cent. This comes on the heels of Mark Carney’s admission that economic growth and inflation are much slower than they had anticipated a few months ago. Economic forecasting is incredibly difficult, with so many factors influencing the rate at which an economy grows and the rate of inflation. There are a number of economic indicators that a forecaster can use to predict the movement of the economy. The most notable of these is the yield curve, which is the difference in interest rates on bonds of different maturity lengths. For the past nine months, the yield curve has been screaming that the economy was slowing down. Yet the Bank of Canada chose not to listen. The relationship between the yield curve and inflation expectations is well known in economics. A 1987 paper from the Kansas City Fed defines the relationship as follows: “The view that the yield curve is an indicator of inflation expectations has a basis in economic theory. Inflation expectations influence the shape of the yield curve by affecting expected short term interest rates. When investors revise their expectations about long-term future inflation rates upward, theory predicts the yield curve will steepen.” In other words, when the gap between a long-term and short-term interest rate rises, inflation is on the way. If the gap narrows, expect inflation to fall. This view is shared by the Bank of Canada’s own research. A 2010 Bank of Canada study by Fung and Remolona concluded: “In this paper, we construct a … model to estimate inflation expectations and inflation-risk premiums in Canada and the United States using bond yields of 2-, 5-, and 10-year maturities. The results suggest that there is useful and substantial information that can be extracted from the yield curve.” Following the Bank of Canada’s advice, I examined the difference between the yields on 2-year and 5-year government bonds, using data provided by the Bank of Canada. The data used is from the period where the target for the overnight rate was 1 per cent (from October, 2010, to today); I limit the data to this period so the yield curve is not being affected by monetary policy. The results are striking (click here to view the chart). The gap between the 2- and 5-year bond yields rises until March, 2011, where the gap peaks at 94 points. The gap steadily falls to a 35 point difference by the end of 2011. After holding steady for a few months, it drops sharply between March, and May, 2012, losing 18 of 38 points. By the Bank of Canada’s own research, such a quick, substantial drop in the yield curve suggests the market is expecting a slowdown in inflation and economic growth. Yet after considering national and international factors, the bank’s July 2012 Monetary Policy Report ends with the sentence: “Overall, the bank judges that the risks to the inflation outlook in Canada are roughly balanced over the projection period.” I wish the Bank of Canada would explain why it held a neutral stance in mid-2012 when the yield curve, as defined by the bank’s own research, was forecasting disinflation.Report Typo/Error
<urn:uuid:2ed6f563-4c58-40f4-8874-3e8ac996bea4>
CC-MAIN-2017-30
https://www.theglobeandmail.com/report-on-business/economy/economy-lab/why-hasnt-the-bank-of-canada-been-watching-yield-curve-data/article7912941/
s3://commoncrawl/crawl-data/CC-MAIN-2017-30/segments/1500549423903.35/warc/CC-MAIN-20170722062617-20170722082617-00341.warc.gz
en
0.957914
695
2.828125
3
History of the Court Courts in Rupert’s Land and the North West Territories Indigenous peoples in what is now Alberta had a long tradition of resolving disputes through a system of community based restorative initiatives, storytelling and elder advice. The British Parliament in 1803 permitted magistrates and Justices of the Peace to be appointed by the Governor of Lower Canada beyond the borders of Upper and Lower Canada. Further legislation in 1821 permitted Justices of the Peace “to sit and hold Courts of Record for the Trial of Criminal Offences” in the territories then being administered by the Hudson’s Bay Company, known as Rupert’s Land. Following the North West Territories becoming part of Canada in 1869, in 1873 the Parliament of Canada permitted the Governor-General to appoint magistrates and created the North West Mounted Police. In 1874 and 1879, the commissioner of the police and then assistant commissioners were given the powers of a magistrate. In 1880, legal qualifications were required for magistrates. Magistrates and Provincial Court Judges In 1906, the newly created province of Alberta established An Act respecting Police Magistrates and Justices of the Peace, SA 1906 c13. The legislation provided for magistrates to be appointed that have practiced law for at least three years. In 1922, the requirement of legal experience was eliminated for police magistrates. In 1955, the designation “police magistrate” was changed to “magistrate.” The new Province also established District Courts (An Act respecting the District Courts, SA 1907 c4) and a Supreme Court for Alberta (An Act respecting the Supreme Court, SA 1907 c3). These courts would eventually become the Court of Queen’s Bench and the Court of Appeal of Alberta. A Juvenile Court was created in 1913, and in 1918 magistrates were given powers to hear “small debts”. The first Family Court was created in 1952. Magistrates had jurisdiction over the Small Debts Court and sat in the Juvenile Court and the Family Court along with judges from the Supreme and the District courts. An amendment to the Police Magistrates and Justices Act in 1970, An Act to Amend the Police Magistrates and Justices Act, SA 1970 c 71, changed the names of magistrates to Provincial Judge. The Provincial Court Act of 1971, SA 1971 c 86, created the Provincial Court of Alberta. In 1975, Justice Cam Kirby conducted a review of the Court, which resulted in the “Administration of Justice in the Provincial Courts of Alberta” report, or the “Kirby Report”. Most of the recommendations of the report were accepted and implemented in the following years. The Provincial Court of Alberta as it exists today was established in 1978, by the Provincial Court Act 1978, SA 1978 c 70 (in effect on May 15, 1980). The Act amalgamated what used to be the Magistrates Court (Provincial Court since 1973), the Juvenile Court, the Small Claims Court and the Family Court to form the modern Provincial Court, and instituted four key service areas: Family, Youth, Small Claims (currently called Civil) and Criminal. Justices of the Peace The Justices of the Peace played an important role in the criminal justice system since ancient times. In the English legal system this office was created to perform local government functions, including that of magistrate. English justices of the peace, although often not legally trained and unpaid, had wide powers to commence criminal proceedings, conduct trials and give sentence in the form of substantial fines or imprisonment up to six months. The office was considered vital to the criminal justice system partly because of the rural nature of the population and the impracticality of placing judges at each location. Additionally, the local Justice could bring to bear upon a case his knowledge of local affairs and of the accused himself. In Canada, the office of Justice of the Peace was preserved when in 1892 the Criminal Code of Canada was adopted. Pursuant to the Criminal Code, Justices of the Peace are empowered to receive and swear informations, issue warrants for arrest and issue search warrants. Initially, Canada followed the English example in that Justices of the Peace were not salaried, but paid on a case by case basis. Usually they were untrained local citizens who were not full time government employees. As the volume of cases increased, and as the provincial courts were expanded to cover rural areas, it became customary to appoint a senior Court clerk as a Justice of the Peace to receive informations, as well as grant adjournments and hear bail applications when a provincial Judge was not available to do so. The Justice Statutes Amendment Act 1998, S.A. 1998, c. 18 (in force February 1, 1999), increased the judicial functions of Justices of the Peace to the current practice. As a practical development, Justices of the Peace in urban centres started working on a full time basis. As a consequence of this Act, a law degree became a requirement for Justices of the Peace. Education was always a priority for the Magistrates Court and the Provincial Court. Starting in 1956, the newly created Alberta Magistrates Association commenced annual education conferences for its members. The first conference was held on December 28, 1956 at the Legislative Building in Edmonton. The current Alberta Provincial Judges Association and the Society of Justices of the Peace for Alberta adhere to the tradition of continuing education with two annual education conferences. The current Provincial Court is the first point of contact between the justice system and an individual. In recent years, the Alberta Provincial Court has become the deciding court for the majority of the legal disputes litigated in the Province. Almost 225,000 adult and youth criminal charges were concluded in Provincial Court in 2012-13, and almost 290,000 adult and youth criminal charges were concluded in 2016-17. Civil Court commenced close to 18,000 claims in 2017 and Family Court commenced more than 16,000 family claims and child applications in 2016-17. The past 30 years have brought a lot of change in the Court with the recognition that the traditional adversarial process is not necessarily the most appropriate process for every crime, or for every population. As a result, different special courts have been set to deal with cases in a therapeutic and culturally appropriate manner. Alberta Board of Review, Provincial Courts, Administration of Justice in the Provincial Courts of Alberta (Edmonton: The Board, ) Attorney General for Alberta, Courts in Alberta (Edmonton: Attorney General for Alberta, 1979) Ell v Alberta, 1999 ABQB 45 Peter Bowal and Tyler Ivie, “The expanding role of the Justice of the Peace” (April/May 2004) Law Now 19.
<urn:uuid:8bd0442a-f8e2-499d-bcb2-f2448e872470>
CC-MAIN-2019-51
https://www.albertacourts.ca/pc/about-the-court/history-of-the-court
s3://commoncrawl/crawl-data/CC-MAIN-2019-51/segments/1575541307813.73/warc/CC-MAIN-20191215094447-20191215122447-00073.warc.gz
en
0.960526
1,372
3.671875
4
A new international study, led by researchers from the University of Cambridge in the United Kingdom, confirms evidence that dairy fats may decrease the risk of type 2 diabetes. For the study, a team of the researchers analyzed the data received from 16 prospective cohorts that included 63,682 participants from 12 countries. Having analyzed the data, the researchers discovered that people who consumed more dairy products had the lower risk of type 2 diabetes. The lead researcher Dr. Fumiaki Iamamura says: “We hope that our findings and existing evidence about dairy fat will help inform future dietary recommendations for the prevention of lifestyle-related diseases.” According to a new research, full-fat dairy products did not raise the possibility of death from any form of heart disease. The research was carried out by an international team of scientists from Reading University in the UK and Copenhagen University in Denmark. The team analyzed twenty-nine studies and discovered that dairy products, including high-fat ones, do not increase the risk of heart attacks and strokes. Ian Givens, a professor of nutrition at Reading University, explains: “There’s quite a widespread but mistaken belief among the public that dairy foods, in general, can be bad for you, but that’s a misconception. While it is a wide belief, our research shows that that’s wrong.” Fat (and full-fat dairy in particular) doesn’t make you fat. In fact, according to a number of studies it can even make you leaner. For example, a review published in the European Journal of Nutrition in 2013 found that people who eat full-fat dairy tend to be leaner than those who chose low-fat versions. And this is not the only study proving that full-fat products are good for your health and weight. Here are five major benefits of eating full-fat milk, yogurt, cheese and ice-cream: Easier weight loss. A lower risk of diabetes. A healthier heart. A calmer digestive tract. Lower sugar intake. More information about eating dairy and studies proving these benefits are here. All original content on these pages is fingerprinted and certified by Digiprove
<urn:uuid:4377987d-5550-4cd3-8283-6eb47867c630>
CC-MAIN-2019-04
https://www.medigoo.com/tag/dairy/
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583656897.10/warc/CC-MAIN-20190116052151-20190116074151-00074.warc.gz
en
0.935171
456
3.0625
3
The skill of sustainable architecture is really a practice of contracting structures while considering relevant issues like ecological growth and sustainable progress. The fundamental motto of this sort of architecture would be to lessen the impact of wide scale construction on atmosphere by sticking with a minimalist attitude regarding use of energy sources, building materials and development space. The next points pointed out below might help in further promoting this noble practice: Using Small Space Is Crucial To Eco-friendly Architecture Building smaller sized houses is just about the order during the day. We can’t manage to waste natural sources and earth space to satisfy our need for constructing and residing in extravagant houses. Smaller sized houses means more and more people get accommodation without wasting an excessive amount of natural sources. Solar Power And Eco-friendly Architecture Coincide Alongside All of our energy sources are near getting depleted. Hence use of solar power has been encouraged in just about all quarters from the globe. As well as that, the inclusion of solar heating techniques make these structures even more comfortable to reside in. A properly contracted passive solar power ought to be sufficient for letting in enough sunlight within the rooms. Water Conservation Permitted Only Through Eco-friendly Architectural Practices Among the primary standpoints of eco-friendly architecture would be to encourage conservation water. Minimum use of water will be adopted while building eco-friendly homes. Ideally the toilets, faucet aerators, showerheads ad flow limitations ought to be selected meticulously to reduce water usage during these houses. Residents should also be asked to plant drought-tolerant plant within their gardens. Alternative Energy, The Primary Perspective of Eco-friendly Architecture An very effective approach to producing usable energy while saving fossil fuel is possible by sang nature agents like hydel power, wind power, solar energy etc. Conserving Local And Natural Sources Via Eco-friendly Architectural Practices Nature continues to be kind to all of us people for countless years. Nature continues to be offering all of the sources we have to build our homes. Therefore we must rely on them wisely while constructing newer structures to leave a few of these sources for the generations to come. You may also plant increasingly more trees around your house to boost natural atmosphere. Lastly stay with using local building sources as that may save fuel and which otherwise could have been utilized in transporting these to the development site. Eco-friendly architecture when adopted honestly helps you save money, but more to the point saves the atmosphere. So act smart and show your ex and support to make Eco-friendly architecture increasingly popular in the future. Have you got any question regarding Sustainable Architecture? You’ll find the solution to your query at Archinomy. Archinomy is really a website produced by students of architecture. At Archinomy, you’ll find situation studies and much more specifics of all topics regarding architecture. For those who have any question, you are able to publish them within the Architecture Forum and they’ll surely be clarified by other people of forum.
<urn:uuid:bdc0c4e7-52b0-4129-86aa-e8b97288f9de>
CC-MAIN-2021-31
https://mylittlehomeblog.com/5-details-that-need-considering-for-eco-friendly-architecture/
s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046153980.55/warc/CC-MAIN-20210730185206-20210730215206-00443.warc.gz
en
0.941539
610
3.171875
3
Your child's eyes needs several essential nutrients so that they can function properly. As they are in their growing stage, a healthy diet is required for a sharp eyesight. For a health eyesight, you should ask your kids to exercise regularly and take regular while reading or working on laptop/mobile and take adequate rest by keeping their eyes closed. Eat Right for the Sharp Sight All these measures will strengthen their eyesight but the key to healthy eye sight is nutritious and balanced diet.Diet including such essential elements such as Vitamins, Minerals, Zin, etc. can help to avoid many eye conditions, such as. - Age related muscular regeneration - Dry eyes - Poor low light vision Here are seven type of food that carry such essential elements and helps to boost your child’s eyesight. There are various types of Citrus fruits available in the market at any time of the year; however, orange is the best in this case. Citrus fruits are rich in vitamin C that provides many health advantages like improving blood vessels; reduce the risk of cataracts and problems related to muscular regeneration. Oranges can be easily consumed as Juice, direct after peeling off, etc. You can also add orange in pudding or can make home made Popsicle for your kids. Deep-water fish such as Salomon, Tuna, and Markel is a great source of omega 3 fatty acids. Omega 3 fatty acid helps in visual development in children’s and also maintain the health of retina by regaining DHA. DHA helps in keeping your eyes moisturized. It maintains proper lubrication in the child’s eye and prevents dry eyes syndrome. Consider adding fish in your meal plan at least 3 or 4 times in a week, this could be the easy way to bring the essential elements in their routine. Eggs are all-time best and preferred food to eat for healthy eyesight. Eggs are not only a great source of protein and contains lutein, an antioxidant that prevent blindness and fight Macular Degenration. They are also loaded with Vitamin A, Zinc and Zeaxanthin. Vitamin A from egg helps as a safeguard to the cornea; it prevents conditions like dry eyes and weak night visions. Similarly, zinc is very beneficial for the health of the retina and prevents problems like low light vision. Zeaxanthin reduce the chance of serious eye conditions like age-related muscle regeneration. Leafy green vegetables and caots Green vegetables like kale, spinach, and collard greens are known for their antioxidant properties. Many terms these vegetables as superfoods as they are rich source of vitamins, and minerals. These vegetables are a great source of antioxidants like Lutein and Zeaxanthin. It helps in preventing serious eye conditions like cataracts, muscular degeneration, and any small defects emerging in eyes. It acts like a defense mechanism to eliminate such conditions. Carrots can be an alternative for egg in getting vitamin A. They are also packed with the benefit of Beta carotene that helps in preventing eye infection and other conditions. Almonds, walnuts, and pistachios are rich in vitamin E which acts as an antioxidant to preserve children's eyesight from conditions like cataracts and muscular regeneration. Vitamin E helps to guard against the unstable molecule which attacks healthy tissues in eyes. Dry fruits can easily be consumed as a snack or in any recipe. Legumes like beans, lentils, peas, chickpeas, lupins are rich in zinc and also provide bioflavonoids to the body. Consumption of such seed grain in a diet plan can lower the risk of cataracts, muscular degeneration, and retinal infections. The seed grains can be consumed in the form of many delicious recipes. They are good source of Vitamin E and Zinc. Help keep your kid’s eyes healthy and disease-free by incluidng sunflowers seeds in their diet. Published on: 02nd July 2020
<urn:uuid:96338b98-0154-45e0-a047-cceb356821ca>
CC-MAIN-2020-34
https://www.mommunity.in/article/eat-right-for-the-sharp-sight/
s3://commoncrawl/crawl-data/CC-MAIN-2020-34/segments/1596439738944.95/warc/CC-MAIN-20200812200445-20200812230445-00364.warc.gz
en
0.93889
833
3.296875
3
Emma Donahue, Staff Writer Columbine. Virginia Tech. Sandy Hook. From the years 1996 to 2012, there have been ninety mass shootings in the United States. The runner up is the Philippines, with only eighteen. Recently, we have backtracked rather than progressed with regards to gun policy. In February, President Trump rolled back an Obama-era regulation that was aimed at preventing the mentally ill from purchasing guns. Despite being home to less than five percent of the world’s population, we own thirty-five to fifty percent of its civilian owned guns, making us the number one firearm per capita nation. We also have the highest gun homicide and suicide rate, although a Pew study showed that the majority of Americans own a gun for personal protection. Legislation to ban semiautomatic assault weapons was recently defeated in the senate, in spite of the bill’s popular support in the wake of the Las Vegas and San Antonio shootings. Currently, we have bans on concealed and specific categories of weapons, as well as restrictions on sales to certain groups of people. The Gun Control Act of 1986 prohibited under eighteen year olds, convicted criminals, the mentally disabled, and dishonorably discharged military personnel from buying firearms. In 1993, The Brady Handgun Violence Prevention Act required those without a gun license to go through a background check before purchasing a gun from a federally authorized dealer. There have also been setbacks, like when the Supreme Court retracted the law that banned handguns in Washington D.C., or when Idaho, Alaska, and Kansas attempted to nullify federal gun legislation. Many analysts believe that the United States could benefit by modeling our gun policy after certain countries with lower gun-related crime. In Canada, major gun reforms were passed following a school shooting in 1989 where the perpetrator used a semiautomatic rifle. Now, there is a twenty-eight day waiting period for gun purchases, mandatory safety training, more thorough background checks, bans on large capacity magazines, and increased restrictions on military grade weapons and ammunition. Their three categories of firearms include non-restricted (rifles and shotguns, which don’t need to be registered), restricted (handguns, semi automatic rifles and shotguns), and prohibited (automatic weapons). Australia is another prime example of how restrictive policies can decrease violence: since their recent implementation of new gun control laws, there have been declining gun deaths and no mass shootings. Their murder rate due to guns has fallen to one per 100,000, compared to our five per 100,000. Additionally, armed robberies occur half as frequently there as they do here. In the United Kingdom, gun control reform was spurred by the Hungerford massacre in 1987. The direct result of this tragic event was the Firearms Amendment Act, which expanded the list of banned weapons and increased registration requirements. The Scotland Dumblane shooting in 1996 lead to the Snowdrop Petition, which was instrumental in pushing legislation to ban handguns and implement a temporary gun buyback. Japan is known for having among the most strict laws, and have a very low gun homicide rate as a result. Most guns are illegal there; under the Firearm and Sword law, the only guns permitted are shotguns, air guns, and specific, situational exceptions which require a series background, drug, and mental health tests. In Germany, any gun purchaser under the age of twenty-five are subject to a psychiatric evaluation that they must pass in order to obtain a firearm. License applicants in Finland can only purchase guns if they can prove that they are an active member in shooting or hunting clubs, pass an aptitude test, a police interview and be in possession of a safe storage unit. Similarly, Italian laws also require purchasers to establish a legitimate reason for their need of a firearm. French applicants for guns must have no record and also pass a background check which takes into account reason for the purchase. Perhaps the reason we have failed to progress as much as these countries is due to a certain mindset created by forces like the NRA. Supporters of increased gun-rights tend to argue that high rates of ownership don’t directly correlate with the strictness of gun laws. The NRA has been continuously opposing safe storage laws, saying it is pointless to own a gun if you can’t reach it in time to defend yourself. However, only a small amount of victims are able to actually use a gun in their defense. A national crime victimization survey showed that 99.2% of 6 million victims (from the years 2007-2011) involved in non-fatal violent crimes did not protect themselves with a gun. Furthermore, in a study of 198 cases of unwanted entry into family homes in Atlanta, it was found that the invader was twice as likely to obtain the homeowners’ gun than to have it used against him/her. There is also a debate surrounding right to carry laws (RTC), as the NRA has been pushing for a Supreme Court decision that would make this right a matter of the Constitution. But, research conducted at Stanford University found that the thirty three states that have adopted RTC laws between 1979-2014 experience gun-related crime rates fourteen percent higher than if these laws had not been adopted. So, despite the significant evidence that more restrictive policies (both internationally and domestically) have resulted in less gun violence, it seems that the American mindset towards what we views as a right must shift before any real progress can be made.
<urn:uuid:c4880f65-8ed9-4ece-b74e-3eb0ad55b1dd>
CC-MAIN-2019-26
http://globalvu.org/2018/02/the-debate-on-global-gun-policy/
s3://commoncrawl/crawl-data/CC-MAIN-2019-26/segments/1560627999261.43/warc/CC-MAIN-20190620145650-20190620171650-00221.warc.gz
en
0.970658
1,097
2.71875
3
What is a basic Database? Well, a database is a collection of data stored. This data is accessed electronically using a computer. Types of databases Storage location of this database is shared,among individuals or organizations Data is not found/stored in different locations, for that organization or group. Data is stored in a personal computer. e.g laptop or desktop. Data is stored over virtual environment. e.g iCloud for Apple Why you need a database. Information is easier to access when it’s organized in a database as it is sorted accordingly, Allows you to know the amount of stock, what’s on the shelves and what’s been sold then gives you a notice when stock is low or a lot. This allows you to check your sale’s metrics, sale’s performance and how customer’s come into contact with you. Allows you to create budget’s on prices of products, this can now be be used to create, Monthly based statements. Quick in turn can be used to create a better and more lean expenditure at the end of the day.
<urn:uuid:c0737ddb-2aab-4662-9ba7-cc9ab71bd063>
CC-MAIN-2020-45
https://igroup24.com/database/
s3://commoncrawl/crawl-data/CC-MAIN-2020-45/segments/1603107888931.67/warc/CC-MAIN-20201025100059-20201025130059-00592.warc.gz
en
0.87846
235
3.09375
3
2 She told me. She is cooking. I say to Jim. We’ll go surfing.Miss Gao said. She likes oranges.Mr Smith told his wife:“I bought a dress for you.”Do you know? When will the meeting begin?I knew. How does David go to school?Can you tell me? Where did you buy the book?Jim’s aunt asked him:“How often do you clean your bedroom?”I want to know. Are you from America?She told me (that) she was cooking.I say to Jim (that) we’ll go surfing.Miss Gao said (that) she liked oranges.Mr Smith told his wife (that) he bought a dress for her.Do you know when the meeting will begin?I knew how David went to school.Can you tell me where you bought the book?Jim’s aunt asked him how often he cleaned his bedroom.I want to know if / whether you are from America. 3 If / Whether clause as objects: 1.从句是一般疑问句的宾语从句 ;2. 从句的引导词用 if 或 whether , 不能省略 ;3. 从句的句式要由疑问句的句式改为陈述句的句式.4. 时态,标点应随着主句的变化而变化. 4 1)“ Has Sally Maxwell arrived?” Do you know? Model:1)“ Has Sally Maxwell arrived?” Do you know?--- Do you know if Sally Maxwell has arrived?2)“Do you miss the UK, or your relations?”Can I ask ?--- Can I ask if you miss the UK, or your relations?从句主句陈述语序引导词 5 He asked Jane if / whether she could borrow her pen to him. What did Ted ask Jane ?Can you borrowyour pen to me?He asked Jane if / whether she could borrowher pen to him. 6 What does the baby cat ask? Have you got any fish?What does the baby cat ask?He asks if his father has got any fish. 7 Am I cool?What does the babyask?He asks if / whether he is cool. 8 What did the girl want to know? Can you hold me up?What did the girl want to know?The girl wanted to know if the man could hold her up? 9 What did the girl ask the young man? “Do youlove me?”What did the girl ask the young man?She asked the young man if / whether he loved her. 10 What does the baby wonder ? Can I fly?He wonders if / whether he can fly. 11 What does the girl ask the Spiderman? “Will youtake me tothe sky?”What does the girl ask the Spiderman?She asks the Spiderman if / whether hewill take her to the sky. 12 Do you know if / whether Mary likes apples? Do you know? Does Marry like apples?I want to know. Was he at home yesterday?Can you tell me? Have they finished the work ?I didn’t know. Are you Tom?He knew. Does the girl come here by bus?Mr. Smith asked Miss Gao:“ Do you work in a school?”The farmer asked his son:“ Can you grow fruit?”8. The teacher asked :“Does the earth turn around the sun?”Do you know if / whether Mary likes apples?I want to know if / whether he was at home yesterday.Can you tell me if / whether they have finished the work?I didn’t know if / whether you were Tom.He knew if / whether the girl came here by bus.Mr. Smith asked Miss Gaoif / whether she worked in a school.The farmer asked his son if / whether he could grow fruit.The teacher asked if / whether the earth turns around the sun. 13 宾语从句三要素:"语序"、"时态"、"引导词” 宾语从句的语序 1). 宾语从句的语序是陈述语序,即“连接词+主语+谓语+其它成分”。而不是疑问句的倒装结构。Can you tell me where will we go?(╳)Can you tell me where we will go?( √)2). 陈述句变为宾语从句时,要注意人称和时态的变化,语序不变。eg. She said, “I will leave a message on the desk.” → She said she would leave a message on the desk.3)一般疑问句和特殊疑问句变为宾语从句时,也要注意人称和时态的变化,后面接陈述语序。eg. “Where are the tickets?” I asked him. → I asked him where the tickets 14 宾语从句的时态宾语从句中谓语动词的时态,常常受到主句谓语时态的制约,此为时态呼应。1)如果主句谓语是一般现在时或将来时,从句谓语的时态不受限制;2)如果主句谓语的时态是一般过去时,从句一般要随着改为相应的过去时态(一般过去时、过去进行时、过去将来时、过去完成时)。eg. I thought (that) you are free today. (╳)I thought (that) you would be free today. ( √)3)【注意】当宾语从句叙述的是客观真理时,不管主句谓语的时态如何,从句都用一般现在时。eg. The teacher told us (that) the earth moves around the sun. 15 eg. He knew (that) he should work hard. 宾语从句的连接词1.连词that,只起连接作用,在从句中不作句子成分,也无词汇意义,在口语中常被省略。eg. He knew (that) he should work hard.2.连词if 、whether,它们起连接作用,在从句中不作句子成分,作“是否”解,在口语中多用if。eg. Tom don't know if/whether his grandpa liked the present.He asked me whether or not I was coming.一般情况下,if 和whether可以互换,但以下3种情况只能用whether:①在不定式前:Whether to go there or not hasn't been decided.②在介词前:It depends on(依靠) whether it is going to rain.③与or not连用:They are talking about whether to go there or not.3.连接代词who, whom, whose, what, which,连接副词when, where, why, how, 它们起连接作用,作句子成分,各有自己的意义。eg. The teacher asked the new students which class he was in.I wonder where he got so much money. 16 【注意】1.标点由主句决定,如主句是陈述句、祈使句,则用句号;是疑问句则用问号。eg. I heard she had been to the Great Wall. Can you tell me which bus I should take?2.要注意个别句子中主从句人称的一致。 17 Mr. Smith thought. Does Mrs. Smith get back? Mr. Smith thought ______ Mrs. Smith ______ ______.Miss Green says:“I like singing.”Miss Green says ______ ______ ______ singing.Do you want to know ? Why didn't I see the film ?Do you want to know ______ ______ ______ see the film ?Nobody knew. Who is knocking at the door?Nobody knew ______ ______ knocking at the door.We could see. There are many animals in the zoo.We could see _______ _______ many animals in the zoo.The boy asked his mother :“ Do you like me ?”The boy asked his mother ________ ______ _______ ______.The girl asked:“ Is the moon far away from the earth?”The girl asked _____ the moon ______ far away from the earth.Uncle Wang told his daughter:“I bought a computer for you.”Uncle Wang told his daughter ______ _______ a computer for______.if got backthat she likeswhy I didn'twho wasthere werewhether she liked himif ishe boughther 18 Activity 2 Choose the correct words to complete the sentences. --- Please tell me if / why you want to come to my party.--- I’d love to come. Thanks!2. I don’t know how / why he did it.--- I’ll show you.I’m not sure when / whether Tony will come or not.--- No, he isn’t. 19 4. --- I want to know how long / old he is. --- He is ten years old.Do you know if / where the orchestra will play?--- Yes, in the concert hall in town.Can you tell me that / who sings this song?--- Yes, it’s Robbie Williams.Can I ask you if / why you like living here?--- Because everyone is so friendly. 20 8. --- Do you know when / what the next English club is? --- Yes, it’s on Thursday, at five o’clock.Can you tell me how / when Sally is playing?--- Yes, at ten o’clock, I think.I’ve heard that / who our team won the match. Is it true?--- Yes, it is. Isn’t it wonderful!
<urn:uuid:9938a3b9-12bf-43c7-9b82-0fde28e6d508>
CC-MAIN-2021-10
https://slideplayer.com/slide/245466/
s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178362899.14/warc/CC-MAIN-20210301182445-20210301212445-00534.warc.gz
en
0.909509
2,822
3.109375
3
San Antonio in Bexar County, Texas — The American South (West South Central) John W. Smith (November 4, 1792 - January 12, 1845) Smith moved to San Antonio in 1828 and was soon in banking-mercantile, brokerage-contracting business. Loyal to democracy and opposed to dictatorship, he was active in defending Texas against Mexico, 1835-42. He participated in "Affair at Gonzales" (Oct. 2, 1835), "Storming of Bexar" (Dec. 1835), siege of the Alamo (March 1836), and the Battle of the Salado (Oct. 1842). Divorced on Jan. 15, 1831, in Missouri from Harriet Stone, he married Maria Curbelo, a descendant of the Canary Islanders who came to San Antonio in 1731. Attorney for many pioneer Texans, he was dominant political figure in Bexar County during 1836-45 era. Elected first county clerk on May 1, 1837, and first Mayor of San Antonio Sept. 18, 1837, by Jan. 1838 he had been appointed postmaster as well. From 1842 to 1845 he served the Republic as senator from Bexar. During regular session of the 9th Congress, he died at Washington-on-the-Brazos. His grave is located in a State Park there. This marker stands at the site of his early home in La Villita De San Fernando. Erected 1973 by Location. 29° 25.269′ N, 98° 29.327′ W. Marker is in San Antonio, Texas, in Bexar County. Marker is at the intersection of King Philip Alley and East Nueva Street, on the right when traveling north on King Philip Alley. Touch for map. Marker is in this post office area: San Antonio TX 78205, United States of America. Other nearby markers. At least 8 other markers are within walking distance of this marker. St. Philip's College (within shouting distance of this marker); Kitchen (within shouting distance of this marker); Herrera House (within shouting distance of this marker); Bolivar Hall (within shouting distance of this marker); Original Site of St. Philip's College (within shouting distance of this marker); Little Church of La Villita (within shouting distance of this marker); Losana House (within shouting distance of this marker); The Weaving Building And Kiln (within shouting distance of this marker). Touch for a list and map of all markers in San Antonio. Categories. • Politics • Settlements & Settlers • War, Texas Independence • Credits. This page was last revised on June 16, 2016. This page originally submitted on May 3, 2015, by Bernard Fisher of Mechanicsville, Virginia. This page has been viewed 234 times since then and 66 times this year. Photos: 1, 2. submitted on May 3, 2015, by Bernard Fisher of Mechanicsville, Virginia.
<urn:uuid:b5d0cbfb-4d0f-4ca5-9abb-6d411b48f32d>
CC-MAIN-2017-51
https://www.hmdb.org/marker.asp?marker=82880
s3://commoncrawl/crawl-data/CC-MAIN-2017-51/segments/1512948521292.23/warc/CC-MAIN-20171213045921-20171213065921-00057.warc.gz
en
0.944345
623
2.765625
3
“The United Nations Convention on the Rights of the child. Everyone in the U.K, including children and young people, have rights that are recognised and protected. These rights are granted by legislation, for example, Human right legislation, The Children Act, etc, which have evolved from a variety of sources. One of these sources is The United Nations Convention on the Rights of the child. Before this was adopted any laws relating to the care and protection of children and young people gave all rights to the child’s parents to do as they felt fit. The Convention on the Rights of the child is an international treaty which has been adopted by the U.K. It was approved by the general assembly of the United Nations on 20 November 1989 and it set out principles relating to children (this includes all children and young people under 18). The U.K government agreed to be bound by the convention in 1991 and it underpins all childcare practice. The convention is made up of a series of articles on Rights. The wording of these articles is at times dated and so the terminology, in some cases, is now not appropriate. However, the principles are still important. The articles are divided into 3 groups – *basic principles which apply to all rights: these cover non-discrimination on grounds of gender, religion, disability, language, other opinion, ethnic or social origin. The best interests of the child are always to be considered by adults and organisations when making decisions about children. The child’s view is to be heard and taken seriously. *Civil and political rights: these include rights to a name and nationality, access to information and protection from abuse, neglect, torture or the deprivation of liberty. *Economic, social, cultural and protective rights: these include rights to life and opportunities, a decent standard of living, day to day care, health care and a healthy environment, education and protection from exploitation. The declaration of the ‘rights of the child’ (article 31) from the UN Convention states that children are especially vulnerable and need protection, and also that their human rights should be respected. Play workers need to work in a manner which protects these rights and empowers children and young people to assert them. Within this comprehensive set of rights it was identified that a child has a right to play, rest and have leisure. It was internationally agreed that these were of vital importance and even essential to the development and well-being of a child. The requirements of national legislation on the Rights of children and young people. The legal aspects of your role as a play worker are important. To be an efficient and effective play worker you need to be aware of relevant legislation, ensure implantation and understand that legislation is constantly being updated to reflect current issues. You need to keep abreast of the changes and adapt as necessary to include these in your practice. The Children Act. The Children Act (1989) covered a wide range of topics and set out the duties of Local Authorities. It was established to form a balance between the rights and responsibilities of the parents and the rights and well being of the child. The Act covered all aspects relating to the well-being and protection of the child. The Children Act 2004 provides a framework for the wider strategy for improving children’s lives. It covers the services which every child accesses as well as those services which children and young people with additional needs access. Every child Matters (2003) called for radical improvements in opportunities and outcomes for children. The 5 outcomes identified are 1) Be Healthy 2) Stay safe 3) Enjoy and achieve 4) Make a positive contribution 5) Achieve economic well-being” (Level 3 Play work Book pages 4,5 and 6) Good practice suggests that the need for clear policies and procedures...
<urn:uuid:cac72e3a-6dd1-4852-981c-e807a687f5f7>
CC-MAIN-2017-13
http://www.studymode.com/essays/Understand-The-Organisational-Framework-For-Play-1003617.html
s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218189032.76/warc/CC-MAIN-20170322212949-00469-ip-10-233-31-227.ec2.internal.warc.gz
en
0.963882
771
3.875
4
In Memory of Rabbi Abraham Leibtag Just a short ‘vort’ on Parshat Shekalim (Shmot 30:11-16) which we read for maftir this week. A ‘BASE’ AND A ‘CROWN’ What did Bnei Yisrael do with the money collected for the “machazit ha’shekel”? From the commandment at the beginning of Parshat Ki-tisa it is not very clear, for it simply states that it should be given: “l’avodat OHEL MOED” – for the needs of the Tabernacle (see 30:16) Ramban enters into a lengthy discussion in regard to whether this “parshia” discusses the yearly obligation that each jew give a “machazit ha’shekel” for the “korbanot tzibur”, or if it is only a specific mitzvah for Bnei Yisrael in the desert who now prepare to build the Mishkan. However, no matter how we understand this parshia in Parshat Ki-tisa, from the opening psukim in Parshat Pekudei it is clear that there was indeed a special donation of the “machazit ha’shekel” for the specific purpose of BUILDING the Mishkan. Even though those details appear to be a bit technical, if we read ‘between the lines’, we can find a very significant message. [I suggest that you read 38:21-31, i.e. the opening psukim of Parshat Pekudei, before continuing.] In Parshat Pekudei we learn that Bnei Yisrael were permitted to donate as much gold and copper [“zahav” & “nechoshet”] as they chose. The primary use of the gold was to make the keilim of the Mishkan (aron, shulchan, menorah etc.), while the copper was used to make the “keilim” of the “chatzer” (i.e. the mizbach n’choshet and kiyor). [This donation totalled approx. 30 “kikar” of gold, and 70 “kikar” of copper / note that the approx. weight of a “kikar” is 30 kilo (or 65 lbs.).] Silver, was a different story. The only silver collected was from the “machazit ha’shekel”, where each member of the male population age 20 (and up) gave a half a shekel (see 38:26). Since the census totalled 603,550 -> a total silver came to 301,775 shekel. Since every 3,000 shekel is a “kikar”, this resulted with 100 “kikar” and a remainder of 1,775 shekel. What did they do with this silver? Actually, it was quite simple. The primary use of the silver was to make “adanim” [weighted base sockets] to support the “kerashim” [wooden planks] of the Mishkan. Since there were a total of 48 “kerashim” which formed the support walls of the Mishkan [20+20+8], and each “keresh” required TWO “adanim”, a total of 96 “adanim” were needed. Plus, four additional “kerashim” were needed to support the “parochet”, but each of these “kerashim” stood on only one “eden”, so a grand total of 100 [96+4] “adanim” were needed to form the base support of the “kerashim” of the Mishkan. Therefore, from this 100 “kikar” of silver they made 100″adanim”, each one weighing 1 “kikar” (about 30 kilograms), and thus forming a very solid base for the Mishkan. With the leftover 1775 shekels of silver, they made clips to connect the curtains to the poles of the outer courtyard [“chatzer”], and a sliver plating for the heads of those poles as well. So why is this important? It is significant that the very BASE of the Mishkan, the silver “adanim” that SUPPORT the entire structure is made from the material which was donated EQUALLY by every member of Am Yisrael. On top of that base, we find “keilim” and materials which were made from the wide range of the donated gold and copper, but the base itself was different. Similarly, from the outside of the Mishkan, the silver coating on the very top of each of the poles of the courtyard [“amudei ha’chatzer”] formed a silver like crown surrounding the Mishkan. This too, was made from the ‘equal from all’ donation of the silver of the “machazit ha’shekel”. Therefore, the very base at the bottom, as well as the shiny crown at the top, reflect the joint donation where each member of Am Yisrael is equal. However, in between them, we find the “keilim” made from the extra donations of gold and copper from private individuals. This may reflect the proper balance between the need on the one hand for everyone to be equal and work together at the very base level of Judiasm, yet at the same time allowing for each individual to make his own personal contribution in any additional realm that he may choose. Yet, all said and done, when one looks from the outside, the ‘finishing touches’ [the silver crown surrounding the Mishkan] must still reflect the very same unity which forms its base. In later generations, after the Mishkan was built, the “machazit ha’shekel” was used to buy the “korbanot tzibur”. Again we find that the very basic “korbanot TAMID” offered daily on the MIZBAYACH reflected the unity and collective nature of Am Yisrael. In addition to that korban TAMID, each individual was permitted to offer an add additional “korban nedava” [voluntary offering], but once again, at the base level, we all stand in front of God as “knesset Yisrael” – one collective unit.
<urn:uuid:95ba1e17-1115-4c10-b9a9-7016d8f4725a>
CC-MAIN-2015-48
https://www.ou.org/holidays/four-shabbatot/shabbat-shekalim-2/
s3://commoncrawl/crawl-data/CC-MAIN-2015-48/segments/1448398468396.75/warc/CC-MAIN-20151124205428-00293-ip-10-71-132-137.ec2.internal.warc.gz
en
0.955883
1,432
2.578125
3
Forest fires raging in Malaysian and Indonesian Borneo in this image, taken on the 8th of August, send thick plumes of smoke arching north-northeast offshore. These fires threaten to intensify a “brown haze” across southern Asia. Smoke from the fires blew over major towns on Borneo island, causing air quality to plunge to its worst level this year in parts of Malaysia’s eastern Sarawak state on Borneo. In this region, more than 2,471 acres (1,000 hectares) of wildfires – roughly the size of 1,500 soccer fields – are ablaze in several forests, according to the AP. Many fires can also be seen in Indonesia’s province of West Kalimantan, both near and far from the Sarawak border. The fires are believed to be caused by plantation operators who set brush fires to clear land during the dry season. State authorities have said they are considering tighter restrictions to ban setting fires on peat soil areas, which are harder to extinguish. The UN has identified this “brown haze”, a near-permanent cloud across southern Asia, as one of the world’s worst environmental hazards. The haze is caused by the smoke from forest fires and agricultural burning, as well as industrial emissions and inefficient wood and dung burning stoves. It can reduce the solar energy reaching the Earth’s surface by up to 15 per cent, reports the New Scientist, altering the Asian monsoon, reducing harvests and killing as many as a million people a year from respiratory diseases.
<urn:uuid:68d43a02-ad92-47b3-b8f5-15fcfa9c90b8>
CC-MAIN-2013-48
http://www.eosnap.com/tag/sarawak/
s3://commoncrawl/crawl-data/CC-MAIN-2013-48/segments/1386164645800/warc/CC-MAIN-20131204134405-00080-ip-10-33-133-15.ec2.internal.warc.gz
en
0.955139
331
3.296875
3
The resulting map of Staten Island was ordered by General Howe shortly after the British landed at Staten Island in July of 1776. According to the Staten Island Historian, the map was found in the Duke of Northumberland's collection in England, and is claimed to have been surveyed without any surveying instruments; hence the inaccurate shape of the Island. It also shows a conflict as to where General Howe's headquarters were located. The Howe map marks the Bancker house, located near the Decker Ferry in Port Richmond as his headquarters. Legend has it that General Howe was staying at the Rose & Crowne at Newe Dorpe. He may have been using both the Bancker residence and the Tavern at different intervals during his stay on Staten Island. The Taylor and Skinner map, another British map commissioned in 1783 is an extremely accurate depiction of the Island! I have yet to see this map, as it is in the U.K. archives. A British box sextant (pictured below), late 18th century. This was used on shore by military and naval officers busily engaged in surveying the colony of British Columbia. Light, portable and easier to transport, it was one of essential the tools for Colonial expansion and mapping. Like a nautical sextant, the box sextant measures angles, but it is smaller, enclosed version placed inside a cylindrical box. Its graduated arc is at half degrees from 0° to 120°. The readings on its vernier were read to single minutes using a magnifying glass. The small removable telescope was used to take long sights, but ordinary observations were usually made through a peephole in the glass. The top (lid) of the cylindrical brass box has been removed to show the knobs that work the sextant. The large knob operates a pinion that engages the toothed, circular segment inside the sextant that mounts the index glass and index arm. The smaller knobs are adjusting screws to eliminate index error and adjust the horizon glass.
<urn:uuid:d6aea36e-316f-4bf9-89ff-82311e5110d0>
CC-MAIN-2015-35
http://revolutionaryrichmondnewyork.blogspot.com/2010/08/general-howe-map-of-staten-island-1776.html
s3://commoncrawl/crawl-data/CC-MAIN-2015-35/segments/1440646249598.96/warc/CC-MAIN-20150827033049-00315-ip-10-171-96-226.ec2.internal.warc.gz
en
0.975179
416
3.421875
3
A new book on the origins of Australia’s refugee policies idealises the approach of the Fraser government. But it has plenty of evidence on why it’s no model argues Ian Rintoul Malcolm Fraser was the Liberal politician who became Prime Minister when Gough Whitlam was sacked by Governor-General Sir John Kerr in 1975. In more recent years, Fraser became known as an outspoken critic of the harsh policies of successive Liberal and Labor governments towards refugees. He resigned from the Liberals in 2010, because he thought the party had moved “too far to the right”. He even became a patron of the Asylum Seeker Resource Centre in Melbourne. His visage is now immortalised as a huge mural on the side of the ASRC building in Footscray. Compared to the years of mandatory detention since its introduction under Labor in 1992, and the vicious detention regime established in the Howard years, the Fraser government’s record on paper looks creditable. Even so, while around 70,000 refugees were resettled between 1976 and 1982, only about 2000 were boat arrivals. Another 80,000 came through resettlement channels in later years. Clare Higgins’ work unearthing Immigration Department archives has added invaluable information for understanding the Fraser government and the dark role of the Immigration Department in shaping immigration policy. It deserves to be read. Who knew, for example, that Christmas Island was first suggested as a detention island in 1978 by NT Chief Minister, Paul Everingham? But her central thesis, that the present inhumane refugee policy has its origins in the mid-1980s, with the Liberal Party breaking with the approach adopted by the Fraser government, is questionable. There is plenty of material in Higgins’ book that shows present refugee policy is a continuation of the approach of the Fraser government rather than any break. In fact the political underpinnings of the present policy were established under Fraser. We will decide At the beginning of 1977, the Fraser government prepared a brief for a UNHCR conference which read in part that the Australian government, “will wish to retain its discretion to determine ultimately who can enter Australian territory and under what conditions newcomers may remain.” Sound familiar? Fraser admits in his autobiography that the government’s willingness to accept more refugees was a product of increasing boat arrivals, not the scale of the refugee crisis itself. Fraser didn’t use Abbott’s notorious slogan, but he did want to stop the boats. Fraser’s willingness to resettle Vietnamese refugees from camps in Malaysia, Indonesia and Thailand was fundamentally based on the assessment that resettlement could help prevent boats travelling to Australia. Fraser went to extraordinary lengths to prevent boats arriving. Higgins herself recounts the stories of Immigration officers disabling and sinking asylum boats by drilling holes in their hulls to disable the boats and prevent them being used to travel on to Australia. This was the sharp end of Fraser’s determination to “stop the boats” and ensure that the Vietnamese stayed in countries of first asylum. But the “border protection” politics that we are now so familiar with were also on display when asylum boats first arrived. On 22 November 1977, Fraser’s Immigration Minister Michael Mackellar spoke to the NSW branch of Institute of International Affairs, saying, “No country can afford the impression that any group of people who arrive on its shores will be allowed to enter and remain.” On 26 November 1977, the Sydney Morning Herald ran a front page headline, “Fraser Warns Refugees,” that they were liable to be returned. Similarly, in November 1977, Fraser’s Foreign Minister told Radio Australia that, “Australia could not be regarded as a dumping ground.” The international commitment from Canada, the US and France to resettle the Vietnamese from the camps in south-east Asia was crucial to “stopping the boats” getting to Australia. In April 1978, an Australian government delegation went to the US to urge it to use its influence to ensure that Malaysia, Thailand and Indonesia prevented the Vietnamese from leaving the camps to sail on to Australia. Space limitations prevent a discussion here about who first coined the “queue jumpers” phrase, although Whitlam certainly used the term in the 1977 federal election. For our present purposes, it is enough to know that then Labor shadow Immigration Minister, Moss Cass, recalls an immigration department official using the term at a bi-partisan briefing in April or May 1978. That meeting decided that processing refugees in offshore camps would be prioritised over those who arrived by boat. In April 1979, Mackellar told The Australian that he, “no longer considered refugees arriving by boat to be ‘queue jumpers’.” The camps in south-east Asia were full and more boats were thought to be on the way to Australia. But just seven months later, when a boat of 50 people arrived off the West Australian coast, MacKellar issued a press statement expressing concern at the resumption of boat arrivals. “The whole Indo-Chinese refugee program could be placed in jeopardy if refugees continue to arrive in such a manner”, he said. “Apart from the enormous risks involved in such a perilous journey, it is unfair to those in refugee camps who are prepared to wait for orderly processing”. Mackellar may not have been using the term “queue jumpers”. But his public comments, just as surely, spelled out the queue jumper argument in terms that are still used by politicians today. People smuggling legislation Higgins tells the story of the threatened arrival of the Hai Hong, a large steel-hulled vessel carrying 2500 Vietnamese asylum seekers, in November 1978. Indonesia denied it entry. Malaysia refused to allow the passengers to disembark unless there were guarantees of resettling countries. The Hai Hong was left in limbo off the coast of Malaysia as diplomatic cables flew around the world. Australia made it clear that it would not resettle any asylum seekers on the Hai Hong. But the Fraser government was very worried that the Hai Hong might head for Australia. Advice prepared in 1978 by the Immigration Department canvassed the whole range of possible mistreatments of boat arrivals that have been progressively imposed over the years. They included, to try and intercept the ship and turn it around at the then three mile ocean boundary; to hold the ship in a remote location and try to force it to leave; to establish a transit camp in a remote locality; and, if compelled to land the passengers, they could be given refugee status but denied resettlement in Australia. By managing to control the number of boat arrivals, the Fraser government avoided implementing these options. But the proposals were not rejected for all time. In May 1979, Mackellar was still telling Cabinet that a, “detention centre in Darwin” could be necessary. The government, however, did respond to the possibility of large numbers arriving with the Immigration (Unauthorised Arrivals) Act 1980. It was Australia’s first anti-people smuggling legislation, introduced into Parliament by Mackellar’s successor, Ian Macphee. Macphee’s introduction repeats the now familiar refrain, that, “Australia would not always be in a position to accept without question large numbers of refugees who push their claims for resettlement ahead of those of their compatriates (sic) who wait patiently in the camps.” Fraser in context Any consideration of the Fraser government’s approach as a compassionate alternative also has to consider the overall political context. Foreign policy considerations played a major role in determining refugee policy. In October 1976 dozens of students at the Thammasat University demonstrating against military dictatorship were killed by the Thai military. Unlike Whitlam, who had allowed all Vietnamese students in Australia to stay after Saigon fell in 1975, Fraser did not allow all Thai students to remain in Australia. Fraser had little sympathy for the students who were considered left-wing. He also did not want students from other regional dictatorships, like Indonesia, getting any ideas. Many Thai students were subject to deportation orders. Fraser’s flexibility on human rights can also be seen in regard to East Timor. In 1978, Fraser followed Whitlam’s shameful de facto recognition of Indonesia’s 1975 annexation, and gave de jure recognition in 1979. It is understandable that refugee supporters look back positively on a time when there was no mandatory detention and when the government organised for around 150,000 Vietnamese to be brought into the country. As Higgins puts it, “Fraser used his own record to be a vocal critic of policies under later governments… he painted a picture of his own government’s decision making in which [there was] ‘a moral and ethical’ obligation to accept Vietnamese boat arrivals…” But in fact, ideas like stopping the boats, queue jumpers, making a distinction between good, deserving refugees who stayed in camps to be selected and bad refugees who arrived in Australia by boat, were all approaches that Fraser’s policies pre-figured. If the refugee campaign had been around under Fraser, it would have had to fight against “stopping the boats” and against vilifying asylum seekers as queue jumpers, just as we do today. When refugee supporters hanker for the Fraser government, it represents a hope for a genuinely humanitarian refugee policy, where refugees who arrive in transit countries are processed and resettled in Australia, where asylum boats desperately trying to gain protection are welcomed and assisted, rather then being turned back; and where asylum seekers arriving in Australia are not subjected to punitive detention, and certainly not expelled to offshore prisons. That is what the refugee movement is fighting for. But to end mandatory detention, to close Manus and Nauru, we need to understand the significant ways that means breaking from the politics of the Fraser government. Influential commentators like Robert Manne, Tim Costello, John Menadue and Frank Brennan have argued that the refugee movement should accept turnbacks in return for the closure of Nauru and Manus Island. Others argue that the movement should put forward proposals that would be more acceptable to the Labor and Liberal parties—ideas like “regional processing” and “burden sharing” (note the name!) are drawn from the Fraser experience. The refugee movement needs to fight for a welcome refugee policy that does not compromise human rights. Asylum by Boat—Origins of Australia’s Refugee Policy By Claire Higgins, UNSW Press, $29.99
<urn:uuid:646ca45c-2439-41f4-8e72-f44e16abd59c>
CC-MAIN-2019-09
https://www.solidarity.net.au/refugees/fraser-wanted-stop-boats/
s3://commoncrawl/crawl-data/CC-MAIN-2019-09/segments/1550247482478.14/warc/CC-MAIN-20190217192932-20190217214932-00607.warc.gz
en
0.965961
2,195
2.640625
3
Fiction is a beautiful art Fiction is a remarkable art. It is thrilling and fascinating. Life would not be the same without imaginative stories unravelling across pages in books and eBooks, or through fiction finding expression through audio storytelling, or as cinema in theaters. Most individuals regard fiction as a means of reaching out to artistic creativity; for entertainment, engagement, and information, or simply as a work of masterly workmanship to admire and enjoy. This can be said for the majority of people— “They value fiction as the creative work it truly is”. Fiction and creativity, in their many alluring casts in the material world, are engraved on the same stone. Imparting life lessons Other than providing creative entertainment, fiction also allows one to understand the difference between right and wrong. Fiction has the capacity to disperse messages of morality, motivation, strength, courage, peace, and other virtues. On the other hand, fiction also dwells upon matters of misadventures, crimes, tragedies, and the dark side of life—this is necessary, for these also act as sources of valuable life lessons—that one must remember and use, wherever applicable. Fiction can have varied other benefits in addition to imparting life lessons. Therefore, fiction is irreplaceable and precious for innumerable reasons. Everyone has access to fiction—in the form of literature, cinema, and other media. Therefore, one can say that fiction does have the ability to influence. It could affect someone intellectually, creatively, and in other ways—perhaps immensely or minutiae. Given the vast reach of books of fiction and cinema, it is however obvious, that fiction has a powerful charm that draws readers and audiences in. Fiction and personal life A majority, watches these creations on silver screens, or read their favorite fictive story, immerse themselves in it, enjoys it and then continues pursuing their love for fiction, in its many forms. It does not influence their personal lives. Positively, the lessons learnt, the creativity engendered, the entertainment value they offer, also become an important byproduct of reading good quality fiction or seeing great cinema, for them. However, fiction could affect a few people, not so positively, in their personal lives, too—an individual who is sensitive to what they see on the screen or how they absorb stories they read—someone to whom, fiction is magnified to a point, where it may overwhelm them on how they perceive their own real life situation and its outcome. Perhaps, the number of such individuals is very less; almost negligible, but imaginary fiction could shape their thoughts and actions in subtle ways—not always, but in certain cases. In-depth research would however be needed to enable some representative figures to come through. For those whose, personal lives are deeply or even moderately influenced by fiction—fiction could have a different connotation. It could change the way one thinks about or perceives situations—the way outcomes or solutions arising in the lives of the protagonists, also flash in front of one’s eyes, when faced with a similar or closely related situation. Stories in books or tales on celluloid could linger on in such an adult’s mind—kind of like in a repetitive mode—perhaps focusing on how ‘the truth shall prevail’, ‘how the enemy shall be annihilated’, ‘how long lost friends would meet’, and ‘how good shall overcome the bad’, or perhaps even how ‘things could go terribly wrong’ and so on. Regardless of what age they might be running at present, fiction could percolate into their lives, sometimes in a trickle and sometimes, in copious measures. Plots and palettes Exemplifying this further, one can take a look at the general shape some fictional creations take. In some cases, the protagonist finds themselves in a difficult situation, and someone far stronger than them or braver rescues them. In other instances, the protagonist falls in love with someone and their story has a happy ending or coincidences happen, which ultimately bring them together, leading up to a lovely denouement. Different stories have different plots and palettes: some are awash with emotions, some with adventure and so on. Then there are romances and love filled chronicles; narrations dealing with hate, anger, revenge, wars, battles—or there are science fiction accounts too. Before the individual knows it, somewhere in their memories, the pieces of fiction have etched their plot. If the fragments left behind are illuminating, peaceful, motivating, empowering—there is no cause for concern. However, if they are associated with destruction, violence, and other negatives—then some caution could be practiced. Illustrating further, here is a look at some representative pieces of fiction: An incredible film called Apocalypto directed by Mel Gibson, for example, had an intense storyline, and was based on a real ancient culture—that of the Mayans. The film had sadness, fear, cruelty. It also had bravery in the face of all of this—a truly heart rendering tale of hope and courage, where there was possibility of none. The peaceful protagonist, Jaguar Paw, loses his family members to a terrifying massacre carried out by a very cruel enemy tribe. He, along with some other members of his tribe are captured by an evil clan who drag them to their village, as slaves—where the caught tribes-people are readied for a ritual sacrifice. However, in a very unbelievable escape, where he is outnumbered, and where no possible rescue could have been attempted—Jaguar Paw rescues himself. He has also hidden his pregnant wife in an earthen well in the village, and before it fills with rain, he has to return and save her too. In spite of the worst possible situation, the protagonist was caught in—he escapes and how—and also saves his wife and child. This movie is a brilliant piece of fiction. Although it may have been inspired partly by real life events and happenings, it still is historical fiction. However, if a particular sensitive viewer is deeply influenced by such a film, it would be a good idea, to remember from the movie—its strengths—not the intensity of its cruelty or sadness. Learn from the lessons they teach, but do not let the tragedies fester in the lanes of one’s memory. Focus on the courage and strength Jaguar Paw showed, and how he fought an almost inevitable death. Gone with the Wind, a famous American novel and later on a film, showed the feminine lead in a contradictory light; Scarlett O’Hara—a woman loved by many, loses a lot, and is seemingly left alone toward the end. Scarlett O’ Hara is initially shown as a woman in love with Ashley Wilkes. However, he marries Melanie Hamilton, and Scarlett marries Melanie’s younger brother, Charlie. Scarlett marries thrice in the novel. Her first two husbands die, due to unfortunate circumstances. Her third husband is Rhett Butler, whom she marries much later, although having met him somewhere in the early stages of the novel. Scarlett has a volatile temper and an inconsistent manner, and her love story with Rhett, someone who loved her quite unconditionally, almost up to a point, leaves her—probably because he finally felt that there was no way she was ever going to love him the way he had always wanted her to. They also lose a child causing a rift in their relationship. Various problems befall Scarlett and her family during the course of the narrative. Scarlett O’Hara is also a good woman, but her personality had flaws and the story does not end the way it should have. It is a beautiful piece of creative fiction. Her mistakes are something to learn from, but the end is sad—she loses her near and dear. Although one does expect that she will win Rhett back, but the reader is left to speculate. A very famous short story by Anton Chekhov called The Grasshopper/The Fidget/The Butterfly—is another marvellous piece of fiction; unusual for its profound insight into the contradictions and irony of life. It is the story of Olga Ivanovna and her husband Osip Stepanitch Dymov—it dwells upon understanding the importance of a treasure one does not know one possesses—until one has lost it. Olga is an artist and a socialite, who marries a doctor. In the initial stages of the story, she partially recognizes his true worth, she thinks of him as someone very special she had always wanted by her side, and appreciates him for it. Nevertheless, this does not last long. Influenced by her friends, and her own insecurities in having a husband so lacking in artistic inclination, she finds a new lover who she feels is more worthy of her—she finds a talented artist and neglects her simple husband. Tragically, this new person in her life eventually abandons her after having an affair with her. Toward the end of the story, Olga realizes that the great and rare man she had always wanted by her side was already by her side—Osip, her husband—someone who always took great care of her. But this realization comes to her, rather late—Osip dies at the time; a death, partly caused by her indifference to him. There are moral messages given here, those of understanding the value of contentment, of one’s responsibilities and the worth of love. Some losses are permanent and cannot be revoked—these are valuable lessons imparted by fiction. However, would anyone want such a loss for themselves, in real life? Many other such pieces of fiction fetch and portray with intricacy the smithereens left behind by the catastrophes, unpredictability and atrocities of life. Even if fiction has elements of true life stories, it still is fiction. Cinema or fiction in print was plotted and expressed—the director made it—the author wrote it. They had control over all the events—the events happened as was planned by them. Such stories are compelling and spell genius. They are sad epic stories. They make great fiction—great reads and incredible cinema. In real life, however, one wishes the opposite of such tragedies. On the contrary, if works of fiction are to a high degree, based mainly on true stories, then fiction takes on an entirely different colour and should be viewed differently. Fiction and the future For those few, who find cinematic or fictional stories in literature influencing their personal choices and decisions for the future, should just respect fiction as an art, and allow one’s personal life to take its own course. Although fiction could in some occurrences become reality, or reality could find a cast of fiction, but a natural unfolding of events should be sought, instead of imprinting fictive stories on one’s life. Unlike a perfectly fictitious environment, real life is natural; it does not have a bound script. Fiction, although guided and shaped formally, and howsoever redacted it may be, is not entirely real. For those who at times, align their current personal lives or futurity with fiction—should borrow from fiction consciously, as long as it is positive. Fictional plots should also not become a parameter for predicting future outcomes either—the future should be allowed to manifest naturally, without putting imaginary fictive outcomes along the road. Making predictions that are congruent with hope and optimism could be a good practice though. Every individual’s life story is their own—Original. Quality works of fiction are meant to shine, as creative masterworks, in an exceptional panorama of artful and sophisticated effort. They are sublime pieces of art, and not necessarily real and therefore may not essentially have something to do with an individual’s actual life, now or in the future. Author Website: www.trishabhattacharya.com Picture Credit: www.pixabay.com
<urn:uuid:ad138c1d-e0b8-481b-811f-bc8aa825d386>
CC-MAIN-2018-47
http://contributoria.com/issue/2015-01/546deede4357a0e474000075.html
s3://commoncrawl/crawl-data/CC-MAIN-2018-47/segments/1542039742963.17/warc/CC-MAIN-20181115223739-20181116005739-00446.warc.gz
en
0.973627
2,478
2.53125
3
Controversy over the federal budget deficit : a theoretical perspective Recently, the Office of Management and Budget projected that the fiscal 1985 federal budget deficit would exceed $200 billion and that out-year reductions would be gradual at best. These prospects have engendered a debate concerning the economic effects of government deficits and the attendant high level of government borrowing. In this article, Michael Dotsey investigates this topic theoretically using alternative macroeconomic models. Dotsey first examines a standard Keynesian model and finds that its prediction of a depressing future effect of government borrowing stems from its unrealistic view of the determinants of consumption. He then discusses the Ricardian equivalence idea that the economic effects are the same whether the government finances its expenditures by taxation or by borrowing. His discussion includes an evaluation of several objections that have been raised against Ricardian equivalence. Dotsey also suggests that one must estimate the optimal size of the government deficit as a necessary first step in evaluating current deficit levels. Volume (Year): (1985) Issue (Month): Sep () |Contact details of provider:|| Web page: http://www.richmondfed.org/| More information through EDIRC |Order Information:|| Web: http://www.richmondfed.org/publications/ Email: | Please report citation or reference errors to , or , if you are the registered author of the cited work, log in to your RePEc Author Service profile, click on "citations" and make appropriate adjustments.: - Barro, Robert J, 1974. "Are Government Bonds Net Wealth?," Journal of Political Economy, University of Chicago Press, vol. 82(6), pages 1095-1117, Nov.-Dec.. - Robert E. Lucas Jr. & Nancy L. Stokey, 1982. "Optimal Fiscal and Monetary Policy in an Economy Without Capital," 532, Northwestern University, Center for Mathematical Studies in Economics and Management Science. - Lucas, Robert Jr. & Stokey, Nancy L., 1983. "Optimal fiscal and monetary policy in an economy without capital," Journal of Monetary Economics, Elsevier, vol. 12(1), pages 55-93. - Robert B. Barsky & N. Gregory Mankiw & Stephen P. Zeldes, 1987. "Ricardian Consumers With Keynesian Propensities," NBER Working Papers 1400, National Bureau of Economic Research, Inc. - Drazen, Allan, 1978. "Government Debt, Human Capital, and Bequests in a Life-Cycle Model," Journal of Political Economy, University of Chicago Press, vol. 86(3), pages 505-16, June. - Aschauer, David Alan & Greenwood, Jeremy, 1985. "Macroeconomic effects of fiscal policy," Carnegie-Rochester Conference Series on Public Policy, Elsevier, vol. 23(1), pages 91-138, January. - Chan, Louis Kuo Chi, 1983. "Uncertainty and the neutrality of government financing policy," Journal of Monetary Economics, Elsevier, vol. 11(3), pages 351-372. - Bryant, John, 1983. "Government Irrelevance Results: A Simple Exposition," American Economic Review, American Economic Association, vol. 73(4), pages 758-61, September. When requesting a correction, please mention this item's handle: RePEc:fip:fedrer:y:1985:i:sep:p:3-16:n:v.71no.5. See general information about how to correct material in RePEc. For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: (William Perkins) If references are entirely missing, you can add them using this form.
<urn:uuid:42030b81-0b0d-4073-a004-5662ea752950>
CC-MAIN-2015-14
https://ideas.repec.org/a/fip/fedrer/y1985isepp3-16nv.71no.5.html
s3://commoncrawl/crawl-data/CC-MAIN-2015-14/segments/1427131300905.36/warc/CC-MAIN-20150323172140-00138-ip-10-168-14-71.ec2.internal.warc.gz
en
0.800706
798
2.546875
3
For decades, researchers have debated about how to classify Pluto. Is it a planet? Is it a dwarf planet? Or is it something else entirely? Now, scientists from the Southwest Research Institute suggest it may fall into the third category. As Neel V. Patel reports for Popular Science, Pluto could be made up of billions of comets all mashed together. The researchers present their ideas in a study published in the journal Icarus. Scientists had long believed the dwarf planet Pluto was formed the way planets come to be: they start as swirling dust that's gradually pulled together by gravity. But with the realization that Pluto was a Kuiper belt dwarf planet, researchers began speculating about the origins of the icy world. In recent decades, scientists have tossed around the idea that Pluto could be a giant comet. But they had no way to test these speculations. That is, until the summer of 2015, when New Horizons zipped by the tiny world. The historic flyby yielded breathtaking images, spectacular data—and the possibility of testing out the wild comet proposal. The researchers turned to Sputnik Planitia—the western lobe of the massive heart-shaped icy expanse stamped on Pluto’s side—for the task. As Christopher Glein, lead author of the paper and researcher at the Southwest Research Institute, explains to Patel, the researchers used the data from New Horizons on this icy expanse to estimate the amount of nitrogen on Pluto and the amount that’s escaped from its atmosphere. The researchers then pulled together composition data gathered by the European Space Agency’s Rosetta mission. The craft orbited Comet 67P/Churyumov-Gerasimenko for two years before purposefully crash-landing into its surface in 2016. “[W]e used the nitrogen abundance from Rosetta, and scaled that up to the mass of Pluto,” Glein says. Both analyses gave surprisingly similar estimates. Glein explains the conclusions in a statement: “We found an intriguing consistency between the estimated amount of nitrogen inside the [Sputnik Planitia] glacier and the amount that would be expected if Pluto was formed by the agglomeration of roughly a billion comets or other Kuiper Belt objects similar in chemical composition to 67P, the comet explored by Rosetta.” The conclusions are far from definitive but hint that the comet idea is an intriguing possibility. However, there are still a few caveats. For one, researchers aren’t sure that comet 67P has an average comet composition, Patel reports. For another, New Horizons only captured information about Pluto at a specific point in time, which means nitrogen rates could have changed over the last billions of years. As Mike Walls writes for Space.com, there’s also still the possibility Pluto formed “from cold ices with a chemical composition closer to that of the sun.” One big challenge to the theory is the low amounts of carbon monoxide on the dwarf planet—a find that runs counter to the situation of most comets. But that doesn't preclude the comet idea: Carbon monoxide may be buried deep under the glacier, or even trapped in a below-surface ocean. Despite these uncertainties, Caltech planetary scientist James Tuttle Keane, who was not involved in the study, tells Gizmodo's George Dvorsky the study still adds to the important conversation about how the solar system formed. “This paper is an exciting example of the science that can be achieved when combining data from different, international, planetary science missions,” he says. “There’s been long debate about the role and significance of comets in the construction of planets… This study represents a new piece to this long-standing puzzle.” As Patel reports, there’s only one way to confirm the new theory: Land on Pluto to gather more data.
<urn:uuid:d692988e-7c2b-4fc6-b86b-878bd3a82250>
CC-MAIN-2020-40
https://www.smithsonianmag.com/smart-news/researchers-think-pluto-made-billions-comets-180969183/
s3://commoncrawl/crawl-data/CC-MAIN-2020-40/segments/1600400244353.70/warc/CC-MAIN-20200926165308-20200926195308-00792.warc.gz
en
0.930755
804
3.90625
4
« back to encyclopedia search results Blackberry (Rubus Villosus) The plant belongs to the rose family - so is also closely related to the raspberry. In spite of its name, it's not a true berry, but a collection of drupelets on a cone-shaped receptacle to which they adhere - unlike the raspberry, which comes away from its 'hull'. Blackberries are nourishing fruit, containing carbohydrates and a high proportion of vitamin B1. When cooked, they go well with apples, but quince and pears suit them too. They make good jam, but some people find this too pippy - to avoid this sieved jelly (generally known as 'Bramble Jelly' is an option. Blackberry; Medicinal uses: A 'tea' made from the dried root can help relieve water retention If you enjoy cooking take a minute to look at ‘Simon Scrutton French Cookery Classes’ on Google – and learn how to make top class bistro-style dishes. Suitable for beginners upwards.
<urn:uuid:315f7fd3-9749-4f47-91eb-7272c7abe651>
CC-MAIN-2018-22
http://www.gourmetbritain.com/food-encyclopedia/5786/blackberry-rubus-villosus/
s3://commoncrawl/crawl-data/CC-MAIN-2018-22/segments/1526794864648.30/warc/CC-MAIN-20180522073245-20180522093245-00515.warc.gz
en
0.933003
222
2.84375
3
What are the long-term effects of living in a technological world: Since the beginning of the 21st century, technology has been developing at a rapid pace. From the early days of the Internet to today’s smartphones and smart devices, technology has become an integral part of our lives. But as we rely on technology more and more, it’s important to ask: What are the long-term effects of living in a technological world? In this blog, we’ll explore the long-term effects of technology on our lives, both positive and negative. We’ll also discuss how Americans feel about digital privacy issues in the era of the COVID-19 pandemic. So if you’re wondering about the long-term effects of technology on our world, read on! What are the long-term effects of living in a technological world? It’s hard to say what the long-term effects of living in a technological world will be. Some people believe that we are becoming too dependent on technology and this can have negative consequences. Others believe that technology is making our lives easier and more efficient in the long run. Only time will tell which side is right. In the meantime, be sure to keep an open mind and use technology responsibly. This way, you can enjoy the benefits of the tech world while avoiding any negative long-term effects. Learn More About What Is Technology. Long-Term Advantages of Living in a Technological world We can’t help being constantly surrounded by technology, but what are its long-term effects? For the most part, the tech world has been a great boon to us. Firstly, it has made our life easier in many ways. For example, we can now work remotely, which gives us more flexibility in our schedules. We also have a wealth of information at our fingertips, which can help us make better decisions. However, living in a technological world also has some disadvantages, such as increased screen time and dependence on technology. Therefore, while the technological world has its merits, we must be aware of the consequences of living in a world where technology is central to our lives. Long-Term Disadvantages of Living in a Technological world As the world becomes more and more technological, it’s no surprise that this has many long-term effects. One of the most important is the growing dependence on gadgets and machines. This can lead to the loss of essential life skills such as cooking or gardening. It can also lead to feelings of isolation and loneliness, as people increasingly rely on technology for social interaction. Furthermore, too much screen time has been linked to various health problems like obesity and sleep disorders. So, what are the best ways to reduce the effects of living in a technological world? Well, one thing is for sure – keeping up with the latest technology and gadgets is important, but don’t let them dominate your life. Also, make sure to get plenty of exercise, socialize with friends and family, and take breaks now and then. All of these things are important for long-term health and well-being. Long-term Positive Effects of Technology We are constantly surrounded by technology. From our phones to the internet, it’s hard to go a day without encountering some form of it. While technology has some negative effects, such as addiction and social media addiction, there are also many long-term positive effects. For example, technology helps people stay connected. It helps people access information and learns new things more easily than ever before. Additionally, technology can help improve efficiency and productivity, both in the workplace and at home. The key is to make positive use of technology and enjoy its benefits. Long-term Negative Effects of Technology Technology has undeniably revolutionized the way we live and work. However, there are also long-term negative effects of technology that we need to be aware of. These negative effects can range from health problems such as neck pain and eye strain to social isolation and anxiety. Additionally, constant exposure to blue light from screens can disrupt your sleep patterns and cause fatigue. As we become more dependent on technology, it is important to be mindful of the negative effects it can have on our physical and mental health. If you’re experiencing any of these effects, it’s important to take a break and spend time offline to reboot. As the world becomes increasingly technology-driven, we are seeing an increase in psychological effects. These effects can include anxiety, depression, social isolation, and addiction. It is important to be aware of these risks and take steps to protect yourself. Some simple tips include setting limits on screen time, taking breaks from social media, and staying active in the real world. By doing so, you will help reduce the negative psychological effects of living in a technological world. Physical health effects The world we live in is constantly changing and along with it, our technology is also changing. Continuous exposure to blue light from screens can cause eye fatigue and strain. This can lead to problems such as tired eyes and eye strain, which can have long-term physical health effects. Poor posture and repetitive movements can lead to musculoskeletal problems such as neck pain and carpal tunnel syndrome. Being constantly connected can lead to feelings of social isolation and anxiety. Taking a break from technology can help you recuperate and restore your physical and mental health. So, the next time you find yourself overwhelmed with technology, remember to take a break! You’ll be glad you did. Eyestrain is a relatively common problem that can be traced back to the excessive use of screens. When you’re constantly looking at a screen, your eyes begin to feel tired and strained. This can lead to headaches and difficulty focusing – in short, everything from mild eyestrain to full-blown eye pain! To avoid this headache-inducing predicament, take regular breaks from screens by spending time outdoors or doing something else active. If eyestrain does happen, see an optometrist as soon as possible for relief. Poor posture is one of the most common physical health effects of living in a technological world. We often hunched over our devices for long periods, which can lead to back pain, headaches, and neck pain. Some simple things you can do to improve your posture and reduce the risk of these problems include taking regular breaks and stretching your muscles. In addition, make sure you use devices that allow you to stand up straight while working on them – this will help reduce the risk of poor posture altogether! Technology is addictive Technology is addictive, and this addiction can have long-term effects on our mental health and social lives. For example, addiction to technology can lead to social isolation, lack of sleep, and even depression. However, there are many positive aspects to living in a technological world, such as increased efficiency and connectivity. Therefore, it is important to take breaks from technology and limit screen time. In addition, it is helpful to have conversations with friends and family about the effects of technology on our lives. By doing so, we can help each other fight addiction and reap the benefits of living in a technological world. It’s no secret that the long-term effects of living in a technological world are largely unknown. However, some health experts believe that excessive screen time can cause problems like neck pain, headache, and eye strain. Other potential problems include sleep disorders, anxiety, and depression. To avoid any of these problems, it’s important to take a break from technology and spend time outside in nature. This will help you restore balance and optimize your health. How Americans see digital privacy issues amid the COVID-19 outbreak The COVID-19 pandemic has had a profound effect on the way Americans view technology. Before the pandemic, many people were living comfortably in a world of digital convenience. However, the outbreak of COVID-19 has forced many people to re-evaluate their relationship with technology. For some, it means reconsidering the long-term effects of living in a world of constant digital communication. For example, some are concerned about the impact of technology on our social media habits and our sense of privacy. Others are taking steps to protect their digital privacy by using a VPN and encrypting their data. Overall, the COVID-19 pandemic has sparked a debate about the benefits and drawbacks of living in a technology-driven world. It is important to be aware of the long-term effects that technology can have on individuals and society as a whole. Although technology has many advantages, it also has some disadvantages. The most important thing is to use technology in moderation and to be aw
<urn:uuid:ad96774e-b7e6-4092-8dc7-fc98545d1ff1>
CC-MAIN-2023-14
https://www.techow99.com/what-are-the-long-term-effects-of-living-in-a-technological-world/
s3://commoncrawl/crawl-data/CC-MAIN-2023-14/segments/1679296949181.44/warc/CC-MAIN-20230330101355-20230330131355-00636.warc.gz
en
0.951986
1,811
2.671875
3
When I was skating, we skated exclusively on outdoor ice, and there was ice for only a fraction of the year. As a result, we had to do a lot of cross-training to stay in shape: lifting weights, running — which I discovered through Allan Claremont, an exercise physiologist at the Madison Speed Skating Club — and cycling. The physical requirements of skating and cycling are similar. In fact, it was cycling as a means of cross-training for skating that led me to start racing a bit, and I found I really enjoyed it. Cross-training is essentially performing any type of exercise outside of your primary sport to maintain or improve your level of fitness. The classic goal of cross-training is to employ a variety of exercise so that the fatigue of a single muscle group doesn't prevent you from working your cardiovascular system and other muscle groups. When you're first starting an exercise program, cross-training can allow you to go longer than you might otherwise, which more effectively boosts your body's basic systems: heart, lungs, circulation and muscles. That, in turn, creates a very strong, wide base to support the pyramid of skills specific to your sport that you will develop over time. Repetitive-use injuries can be avoided using cross-training. An avid athlete in any sport runs the risk of impact-related or repetitive-use injuries. To increase your exercise load without piling on more sessions of the same sport, add a cross-training workout instead. Alternative forms of exercise develop whole-body conditioning by recruiting muscles your primary activity may not activate. When a bicyclist does gym training, for example, she can specifically focus on muscle groups that cycling neglects. It can also accelerate weight loss. One study showed that runners who maintained their usual running schedules and added just one 30-minute cross-training workout (in this case, cycling) per week lost an extra pound of fat every 10 weeks, provided they didn't increase their eating. Once you have developed the habit of exercise, like an elite athlete you will very rarely want to go long periods without it, even when you need to stay away from your favorite sport when it's offseason or due to injury. Here are a few activities I recommend to my patients as appropriate cross-training modalities when they are injured or in other circumstances. Runners can cross-train using cycling, cross-country skiing, hiking, skating, even martial arts. Those who develop knee pain during running, soccer, basketball or aerobic dance can cross-train without pain riding a bike, in most cases. People with Achilles tendinitis can often do aqua jogging or swimming without kicking (doing the breaststroke, for example). Rowers or tennis players with elbow or wrist injuries can run or ride a bike. If you have shoulder pain, use your legs to get your heart and lungs working by running or using a StairMaster. When it's too hot or too cold outdoors, the gym is ideal. You can use any of the ergometers (treadmill, cyclette, rowing machine or StairMaster) or weights. "Active rest," such as stretching, works for a day or two off from any other activity. Eric Heiden, M.D., a five-time Olympic gold medalist speed skater, is an orthopedic surgeon in Utah. He co-wrote "Faster, Better, Stronger: Your Fitness Bible" (HarperCollins) with exercise performance physician Max Testa, M.D., and DeAnne Musolf. Visit fasterbetterstronger.com.
<urn:uuid:3721b904-597c-48be-b4b6-77e7b6cc89ab>
CC-MAIN-2018-09
http://www.ctnow.com/sc-health-0707-heiden-20100707-story.html
s3://commoncrawl/crawl-data/CC-MAIN-2018-09/segments/1518891816094.78/warc/CC-MAIN-20180225031153-20180225051153-00774.warc.gz
en
0.957371
740
2.71875
3
Trait theories centre on the belief that the individual is more important than the situation, identifying the personality characteristics of successful leaders is of paramount importance. Traits, such as above average intelligence and high levels of self assurance were identified as being synonymous with successful leaders. Most studies eluding to this point were committed in the 1950’s or earlier, so suggest notions of imperialism or elitism. Possession of all of these traits is an unreasonable ideal.The fading of the prominence of trait theories can be attributed to an increase in democratic culture stating that anyone has the potential to be a successful leader if they display the appropriate attributes. There are three main leadership styles that could be implemented. The main differences between them reside with the focus of power. An autocratic system involves the manager setting the objectives and insisting on obedience. Poor cohesion and poor worker motivation often result. Democratic type leaders encourage participation in decision making, and can result in increased motivation and worker efficiency.A Laissez-faire style allows subordinates to carry out activities freely, but can bring about unstructured and confounding worker efforts. Fiedler (1976), in attempting to put forward a theory to explain the most effective leadership style argued that it is easier to change someone’s role or power, or to modify the job he has to do, than to change his leadership style” (Hall, et al. 1994). Fiedler is, therefore, suggesting that an individual’s leadership style is more a function of a specific personality trait, therefore an autocrat will always remain so; unable to adapt for the situation in question.More than 800 individual studies were committed during this research, so despite no methodology could be found for possible criticism of his findings, Fiedler’s perspective has to be strongly considered by the prospective manager or scholar. Participate styles such as democratic or laissez-faire can be argued to benefit both employer and employee through satisfying subordinate urges for responsibility, self-actualisation and esteem. Employing appropriate management style alone is not enough to ensure overall effectiveness, thus the advent of contingency theories. Contingency theories focus on a broader spectrum of variables involved in a leadership situation than the oversimplified situational approach. No one style of leadership is appropriate to all situations. Fiedler, as previously mentioned was a particularly prominent theorist, stating that the favourability of the leadership situation, relationship between the leader and the group and the structure of the task are the determinants to be used when choosing appropriate leadership characteristics. He largely concentrated on the relationship between leadership and organisational performance, and devised the ‘least preferred co-worker scale’ (LPC). This scale demonstrates a rating of which subordinate a leader would work least well with. However, interpretations of the LPC scale have proved inconsistent, so although it provides a useful tool to the potential leader or manager, reactions must be guarded. Vroom and Yetton provide an alternative contingency model. Their analysis is based on three aspects, decision quality, decision acceptance and the time constraints placed on the decision making process.’Decision rules’ were then outlined to help the manager adopt the most appropriate leadership style. Another alternative contingency model is that of the path-goal theory, championed by theorists such as House, House and Dessler. The model is based on the understanding that an individual’s motivation is governed by the expectations of increased reward, in return for increased effort and in the confidence of success if effort is boosted. This model relies heavily on sociological perspectives of motivational expectancy theory.All contingency approaches suffer if the manager appears to have an inconsistent leadership style. Communication patterns within a group can determine effectiveness. However, when a status discrepancy exists, the higher status members speak more and have more influence, while those of lower status speak less and are likely to defer to superiors’ (adapted from: Hurwitz, Zander and Hymovitch, 1953). This shows how adopting a flatter, hierarchical structure can promote freer contributions, and group effectiveness.The environment is another element that has to be considered, as leader, subordinates and the task in question does not operate in a vacuum, but as a part of a wider external field. Nearly all organisational tasks are not isolated, they impact upon other areas, internally and externally of the organisation. Pressures can come from areas such as national or international, social, economic or political. An effective leader has to try to shape the environment and be comfortable with being shaped by it.
<urn:uuid:50687aaa-a9b4-45cc-ac64-d27911147c05>
CC-MAIN-2021-10
https://woodstock-online.com/13913-management-style-alone/
s3://commoncrawl/crawl-data/CC-MAIN-2021-10/segments/1614178376006.87/warc/CC-MAIN-20210307013626-20210307043626-00433.warc.gz
en
0.953165
920
3.46875
3
Popular Science Monthly/Volume 74/April 1909/The Origin of the Theory of Natural Selection |THE ORIGIN OF THE THEORY OF NATURAL SELECTION| I BEG to thank the council of the Linnean Society for the very great honor they have done me, in coupling my name with that of Charles Darwin on the celebration of this anniversary, and for the still greater and more exceptional honor, of perpetuating my features with those of my illustrious forerunner, upon the medal you have now awarded me. With your permission I propose to make a few remarks both as to the actual relations between Darwin and myself prior to July, 1858, and also to some peculiarities of our respective life-histories which brought about those relations, and which will, I hope, be both novel and of some general interest. Since the death of Darwin in 1882, 1 have found myself in the somewhat unusual position of receiving credit and praise from popular writers under a complete misapprehension of what my share in Darwin's work really amounted to. It has been stated (not unfrequently) in the daily and weekly press, that Darwin and myself discovered "natural selection" simultaneously, while a more daring few have declared that I was the first to discover it, and that I gave way to Darwin! In order to avoid further errors of this kind (which this celebration may possibly encourage), I think it will be well to give the actual facts as simply and clearly as possible. The one fact that connects me with Darwin, and which, I am happy to say, has never been doubted, is that the idea of what is now termed "natural selection" or "survival of the fittest," together with its far reaching consequences, occurred to us independently, and was first jointly announced before this society fifty years ago. But, what is often forgotten by the press and the public, is, that the idea occurred to Darwin in October, 1838, nearly twenty years earlier than to myself (in February, 1858); and that during the whole of that twenty years he had been laboriously collecting evidence from the vast mass of literature of biology, of horticulture and of agriculture: as well as himself carrying out ingenious experiments and original observations, the extent of which is indicated by the range of subjects discussed in his "Origin of Species," and especially in that wonderful store-house of knowledge—his "Animals and Plants under Domestication}}," almost the whole materials for which works had been collected, and to a large extent systematized, during that twenty years. So far back as 1844, at a time when I had hardly thought of any serious study of nature, Darwin had written an outline of his views, which he communicated to his friends Sir Charles Lyell and Dr. (now Sir Joseph) Hooker. The former strongly urged him to publish an abstract of his theory as soon as possible, lest some other person might precede him—but he always refused till he had got together the whole of the materials for his intended great work. Then, at last, Lyell's prediction was fulfilled, and, without any apparent warning, my letter, with the enclosed essay, came upon him, like a thunderbolt from a cloudless sky! This forced him to what he considered a premature publicity, and his two friends undertook to have our two papers read before this society. How different from this long study and preparation—this philosophic caution—this determination not to make known his fruitful conception till he could back it up by Overwhelming proofs—was my own conduct. The idea came to me, as it had come to Darwin, in a sudden flash of insight: it was thought out in a few hours—was written down with such a sketch of its various applications and developments as occurred to me at the moment,—then copied on thin letter-paper and sent off to Darwin-—all within one week. I was then (as often since) the "young man in a hurry"; he, the painstaking and patient student, seeking ever the full demonstration of the truth that he had discovered, rather than to achieve immediate personal fame. Such being the actual facts of the case, I should have had no cause for complaint if the respective shares of Darwin and myself in regard to the elucidation of nature's method of organic development had been thenceforth estimated as lieing, roughly, proportional to the time we had each bestowed upon it when it was thus first given to the world—• that is to say, as twenty years is to one week. For, he had already made it his own. If the persuasion of his friends had prevailed with him, and he had published his theory, after ten years'—fifteen years'—or even eighteen years' elaboration of it—I should have had no part in it whatever, and he would have been at once recognized, and should be ever recognized, as the sole and undisputed discoverer and patient investigator of the great law of "natural selection" in all its far-reaching consequences. It was really a singular piece of good luck that gave to me any share whatever in the discovery. During the first half of the nineteenth century (and even earlier) many great biological thinkers and workers had been pondering over the problem and had even suggested ingenious but inadequate solutions. Some of these men were among the greatest intellects of our time, yet, till Darwin, all had failed; and it was only Darwin's extreme desire to perfect his work that allowed me to come in, as a very bad second, in the truly Olympian race in which all philosophical biologists, from Buffon and Erasmus Darwin to Richard Owen find Robert Chambers, were more or less actively engaged. And this brings me to the very interesting question: Why did so many of the greatest intellects fail, while Darwin and myself hit upon the solution of this problem-—a solution which this celebration proves to have been (and still to be) a satisfying one to a large number of those best able to form a judgment on its merits? As I have found what seems to me a good and precise answer to this question, and one which is of some psychological interest, I will, with your permission, briefly state what it is. On a careful consideration, we find a curious series of correspondences, both in mind and in environment, which led Darwin and myself, alone among our contemporaries, to reach identically the same theory. First (and most important, as I believe), in early life both Darwin and myself became ardent beetle-hunters. Now there is certainly no group of organisms that so impresses the collector by the almost infinite number of its specific forms, the endless modifications of structure, shape, color and surface-markings that distinguish them from each other, and their innumerable adaptations to diverse environments. These interesting features are exhibited almost as strikingly in temperature as in tropical regions, our own comparatively limited island fauna possessing more than 3,000 species of this one order of insects. Again, both Darwin and myself had, what he terms "the mere passion of collecting,"—not that of studying the minutiae of structure either internal or external. I should describe it rather as an intense interest in the mere variety of living things-the variety that catches the eye of the observer even among those which are very much alike, but which are soon found to differ in several distinct characters. Now it is this superficial and almost child-like interest in the outward forms of living things, which, though often despised as unscientific, happened to be the only one which would lead us towards a solution of the problem of species. For nature herself distinguishes her species by just such characters—often exclusively so, always in some degree—very small changes in outline, or in the proportions of appendages, as give a quite distinct and recognizable facies to each, often aided by slight peculiarities in motions or habits; while in a large number of cases differences of surface-texture, of color, or in the details of the same general scheme of color-pattern or of shading, give an unmistakable individuality to closely allied species. It is the constant search for and detection of these often unexpected differences between very similar creatures, that gives such an intellectual charm and fascination to the mere collection of these insects; and when, as in the case of Darwin and myself, the collectors were of a speculative turn of mind, they were constantly led to think upon the "why" and the "how" of all this wonderful variety in nature—this overwhelming, and, at first sight, purposeless wealth of specific forms among the very humblest forms of life. Then, a little later (and with both of us almost accidentally) we became travellers, collectors and observers, in some of the richest and most interesting portions of the earth; and we thus had forced upon our attention all the strange phenomena of local and geographical distribution, with the numerous problems to which they give rise. Thenceforward our interest in the great mystery of how species came into existence was intensified, and—again to use Darwin's expression—"haunted" us. Finally, both Darwin and myself, at the critical period when our minds were freshly stored with a considerable body of personal observation and reflection bearing upon the problem to be solved, had our attention directed to the system of positive checks as expounded by Malthus in his "Principles of Population." The effect of this was analogous to that of friction upon the specially-prepared match, producing that flash of insight which led us immediately to the simple but universal law of the "survival of the fittest," as the long-sought effective cause of the continuous modification and adaptation of living things. It is an unimportant detail that Darwin read this book two years after his return from his voyage, while I had read it before I went abroad, and it was a sudden recollection of its teachings that caused the solution to flash upon me. I attach much importance, however, to the large amount of solitude we both enjoyed during our travels, which, at the most impressionable period of our lives, gave us ample time for reflection on the phenomena we were daily observing. This view, of the combination of certain mental faculties and external conditions that led Darwin and myself to an identical conception, also serves to explain why none of our precursors or contemporaries hit upon what is really so very simple a solution of the great problem. Such evolutionists as Robert Chambers, Herbert Spencer and Huxley, though of great intellect, wide knowledge, and immense power of work, had none of them the special turn of mind that makes the collector and the species-man, while they all—as well as the equally great thinker on similar lines, Sir Charles Lyell—became in early life immersed in different lines of research which engaged their chief attention. Neither did the actual precursors of Darwin in the statement of the principle—Wells, Matthews or Prichard—possess any adequate knowledge of the class of facts above referred to, or sufficient antecedent interest in the problem itself, which were both needed in order to percieve the application of the principle to the mode of development of the varied forms of life. And now, to recur to my own position, I may be allowed to make a final remark, I have long since come to see that no one deserves either praise or blame for the ideas that come to him, but only for the actions resulting therefrom. Ideas and beliefs are certainly not voluntary acts. They come to us—we hardly know how or whence, and once they have got possession of us we can not reject or change them at will. It is for the common good that the promulgation of ideas should be free—uninfluenced by either praise or blame, reward or punishment. But the actions which result from our ideas may properly be so treated, because it is only by patient thought and work, that new ideas, if good and true, become adopted and utilized; while, if untrue or if not adequately presented to the world, they are rejected or forgotten. I therefore accept the crowning honor you have conferred on me today, not for the happy chance through which I became an independent originator of the doctrine of " survival of the fittest," but, as a too liberal recognition by you of the moderate amount of time and work I have given to explain and elucidate the theory, to point out some novel applications of it, and (I hope I may add) for my attempts to extend those applications, even in directions which somewhat diverged from those accepted by my honored friend and teacher—Charles Darwin. - Reply on receiving the Darwin-Wallace medal of the Linnean Society of London on July 1, 1908.
<urn:uuid:c7677093-8bc6-4471-b349-4ae2c6856721>
CC-MAIN-2019-18
https://en.wikisource.org/wiki/Popular_Science_Monthly/Volume_74/April_1909/The_Origin_of_the_Theory_of_Natural_Selection
s3://commoncrawl/crawl-data/CC-MAIN-2019-18/segments/1555578643556.86/warc/CC-MAIN-20190424134457-20190424160457-00238.warc.gz
en
0.97797
2,616
3.078125
3
Dr. Paul Auerbach is the world's leading outdoor health expert. His blog offers tips on outdoor safety and advice on how to handle wilderness emergencies.See all posts » What to Do in the Case of Serious Food Allergies An EpiPen (an epinephrine auto-injector)An allergic reaction in an outdoor setting can rapidly become a life-threatening emergency. While most of us think of food allergies as annoyances, they can be quite serious or even life threatening. Itchy skin rashes can progress to breathing difficulty, swollen soft tissues (e.g., lips, tongue, throat) that compromise the airway, and low blood pressure or even shock. Therefore, it’s important to be familiar with the signs and symptoms of severe allergy and to be prepared to respond rapidly in the event of an emergency. The National Institute of Allergy and Infectious Diseases has released Food Allergy Guidelines for healthcare professionals to help guide the care of patients with life-threatening food allergies. The full guidelines can be found at http://www.niaid.nih.gov/topics/foodallergy/clinical/Pages/default.aspx. Here are some key points: The first-line treatment for a very severe allergic reaction (anaphylaxis) is an injection of intramuscular (IM) epinephrine (adrenalin). Failure to administer epinephrine early in the course of treatment has been repeatedly implicated in fatalities from anaphylaxis. IM epinephrine (given in the front or side of a big muscle of the thigh, e.g., the vastus lateralis muscle) is recommended over injection into subcutaneous tissues because this method of injection provides a more rapid increase in bloodstream and tissue concentrations of epinephrine. When a 1:1,000 epinephrine solution (this is a standard preparation) is used, the recommended dose is 0.01 mg/kg with a maximum dose of 0.5 mg. Autoinjectors, such as the EpiPen, typically are preloaded to deliver 0.3 mg for adults and 0.15 mg for children. Epinephrine has an onset of action within minutes but is rapidly broken down in the body. Therefore, the effect is often short-lived—repeated doses may be necessary. If a patient responds poorly to the initial dose or has ongoing or progressive symptoms despite initial dosing, repeated dosing may be required after five to 15 minutes. 10 percent to 20 percent of persons who receive epinephrine will require more than one dose before recovery from symptoms. These medications are often given at the same time as epinephrine. They are not meant to be given sequentially as listed, with the exception of epinephrine as first-line treatment. “H1” antihistamines are useful only for relieving itching and the raised red rash (wheals) of an allergic reaction. They do not relieve airway obstruction caused by tissue swelling, shortness of breath, wheezing, gastroenteric symptoms, or low blood pressure. Therefore, they should be considered additional therapy and should not be substituted for epinephrine. For the treatment of bronchial spasm that is not responsive to IM epinephrine, inhaled bronchodilators, such as albuterol, should be used as needed and should be considered to be additional therapy to the administration of epinephrine. Very little information is available to support or refute the use of corticosteroids (“steroids’) for the treatment of acute anaphylaxis. However, their use is prevalent and supported by many health care professionals. Corticosteroids are not helpful in the treatment of acute anaphylaxis due to their slow onset of action (four to six hours). These agents often are given because of their anti-inflammatory properties that benefit recovery from allergic and inflammatory disease, and also because they may help prevent waxing and waning, or prolonged, reactions, which occur in up to 20 percent of afflicted persons. Many patients with severe allergic reactions require infusion of intravenous (IV) fluids. Massive fluid shifts can occur rapidly in anaphylaxis due to leaky blood vessels. Any patient whose low blood pressure does not respond promptly and completely to injected epinephrine should receive large-volume fluid resuscitation, with normal saline being the preferred treatment. Large-volume fluid resuscitation should be initiated immediately in patients who present with a severe drop in blood pressure when moving from the lying-down to the upright (standing) position, low blood pressure in any position, or incomplete response (with respect to blood pressure) to IM epinephrine. Supplemental Oxygen Therapy Oxygen should be administered initially to all patients experiencing anaphylaxis. The patient should be placed in a lying-down position (if tolerated) with the lower limbs raised. Persons with breathing difficulty or vomiting may not tolerate a recumbent position. Anaphylaxis has many different symptoms, and can even look like a severe asthma attack (even for those not diagnosed with asthma). Symptoms include: - Heart: drop in blood pressure (anaphylactic shock), fainting, chest pain, paleness - Respiratory: swelling of the throat, hoarseness, choking, difficulty breathing, coughing, wheezing - Ear, nose and throat: mouth tingling, itching, swelling; watery, red, itchy eyes; runny itchy nose, nasal congestion - Abdomen: nausea, vomiting, diarrhea, abdominal cramps, uterine cramps - Skin: hives, flushing, itching, swelling - Neurological: a sense of impending doom, anxiety, headache Allergic reactions do not always present with skin symptoms; most often they are multi-system, but can also present with only a single organ system being affected (e.g., with only severe airway constriction). Recent Blog Posts Jul 01, 2013 In Advance of a Wildfire Feb 11, 2013 Topical Ivermectin Lotion for Treating Head Lice Feb 04, 2013 Public Health Interventions and Snowmobile Fatality Rates
<urn:uuid:71acd02e-f420-4c8d-a142-e287262cc04a>
CC-MAIN-2016-50
http://www.healthline.com/health-blogs/outdoor-medicine/serious-food-allergies
s3://commoncrawl/crawl-data/CC-MAIN-2016-50/segments/1480698541214.23/warc/CC-MAIN-20161202170901-00369-ip-10-31-129-80.ec2.internal.warc.gz
en
0.912098
1,282
3.015625
3
Sunday, April 19, 2015 April 19th, 1915 - Today Joffre circulates a memorandum to his army commanders outlining the proper conduct of offensive operations. The objective of a major attack, writes Joffre, is to achieve a break through, and that once an assault has begin it is to be continued until the the German line has been decisively ruptured. This is to be done via continuous attacks, whereby pressure on the Germans is to be maintained by constantly sending more units into the attack until the strain becomes so great that the line breaks. The memorandum also compares the proper conduct of an offensive to a symphony, in which a wide range of parts have to work together harmoniously under the direction of the commanding officer if success is to be achieved.
<urn:uuid:a4c9c6ab-4529-46b8-96c2-ac6a9f5f9abf>
CC-MAIN-2022-40
http://diarygreatwar.blogspot.com/2015/04/april-19th-1915.html
s3://commoncrawl/crawl-data/CC-MAIN-2022-40/segments/1664030335304.71/warc/CC-MAIN-20220929034214-20220929064214-00024.warc.gz
en
0.975876
155
3
3
The previous post in this series is: The home page in this series on building raised bed gardens is Raised Beds Jim’s Way. For my example, throughout these pages, I’ll be “building” two side-by-side beds 3′ 6″ (3 feet 6 inches) by 12′ (inside dimensions). I’ll have a single 3′ walking path between them. There will be a 4′ walking path around the outside as well. Finally, we will make the beds out of 2″x12″ treated lumber1. So, the length of the garden will be the length of the planks plus 8′. The garden’s width will be the total width of the beds (7′) plus 4′ for each of the two outside walkways, and another 3′ between each bed. That adds up to 7′ + 8′ + 3′ = 18′ for our width. The garden’s length is 12′ plus the two 4′ paths around the outside, or 20′ total. For setting up precise locations in your chosen area, we’re going to have to use a little trigonometry. But, I’ll show how to do it without requiring any actual math to be performed. The first step is to choose an object to which you want to align your beds. This can be a fence, the side of a building, or any number of other possibilities. I’ll refer to it as the “yard object” in these steps. If you’re using a building as your guide to setting these up in your yard, add some clearance between the side of the garden and that building, to prevent the building’s shadow from falling on the garden. In my location, that’s about 7′. In my example, the long side of the beds is closest to the building. So, add 7′ to the other direction (the 16′ side), making the total 23′ by 18′. However, the gap between the bed and the yard object can include the walking path around the outside, so this can be 20′ instead of 23′. I’ll use 20′. We will be marking six locations using my example dimensions. Two stakes will be placed at the edge of the yard object we’re aligning with. Two optional stakes are placed the width of the gap between the wall of the yard object and the start of the garden (7′). Two final stakes will be placed the width of the garden from the second pair of stakes (16′). Pick a spot on your yard object closest to one corner of the garden area. Put a stake #1 in the ground. With the tape measure, measure off the distance along the yard object at the length of the garden (18′). Put stake #2 here. If the building or fence is not long enough, put stake #2 in the ground along a line extended from the object. Stakes #3 and #4 are optional, used only if you have a gap between your yard object and the garden itself. In our case, they would go at 7′ from the first two stakes. Stakes #5 and #6 go the width of the bed from the last two stakes you placed (#1/2 or #3/4). In our case, these would be 16′. Also, measure to the yard object as well as to the other stakes. This can let you know if there’s a problem with the other measurements. Unless you are lucky or really good at eyeballing straight lines, the corners of the box you just made with stakes are not going to be square. To fix that, we will use a math trick to get the plot corners square. Measure from stake #1 to stake #6 (diagonally across the bed area). Then measure from stake #2 to stake #5. (The measuring lines will make an X across the garden area.) To make the garden actually have square corners, move stakes #5 and #6 until these two diagonal distances are the same. Always maintain the same distance from the yard object (20′) and the distance between the two stakes you are moving (18′). So, you basically have to move both stakes #5 and #6 at the same time, in the same direction, and the same amount. The amount you should move them is about half of the difference between the two diagonal measurements. Measure again, and move the stakes accordingly. Once the diagonals are the same size, your garden area is now measured. Leave the stakes in the ground. If you used all six stakes, repeat the above process for stakes 3, 4, 5, and 6. This time measure from statke #3 to stake #6, and again between stake #4 and stake #5. Move stakes #3 and #4, leaving #5 and #6 in place. When done, you should be able to eyeball a line through stakes #1 to #2 and #3 (and 4, 5, & 6). The next post in this series is:
<urn:uuid:455fcde1-bcc0-4aff-8bb2-5362bde16f34>
CC-MAIN-2021-31
https://god-isgod.com/gardening/align-raised-beds-to-yard/
s3://commoncrawl/crawl-data/CC-MAIN-2021-31/segments/1627046154127.53/warc/CC-MAIN-20210731234924-20210801024924-00638.warc.gz
en
0.943901
1,064
2.65625
3
In 2002 Ben Cohen (of Ben & Jerry’s fame) got together with the core players in the anti-sweatshop movement and started SweatX. SweatX was planned to be structured as a worker owned cooperative sewing facility in Los Angeles, CA. Early in the first year of operations, with full support from company management, the workers of SweatX voted to be represented by the Union of Needletrades, Industrial and Textile Employees (UNITE). A positive labor-management relationship was forged and production commenced. But after 14 months of operation — and before worker ownership had been established — SweatX closed in 2004. SweatX was a business answer to a social problem. The problem was, and still is, the lack of corporate accountability in the global sweatshop industry that drives the manufacture of almost all of the clothing we wear. Clothing brands are companies with a mission to make money for their shareholders. These companies typically do not care about the individual worker and seek out manufacturing centers that oppress workers and poison the environment.? “Between 1865 and 1890, in the aftermath of the Civil War, virtually every important American labor reform organization advocated ‘cooperation’ over ‘competitive’ capitalism and several thousand cooperatives opened for business during this era. The men and women who built cooperatives were practical reformers and they established businesses to stabilize their work lives, families, and communities. Yet they were also utopians—envisioning a world free from conflict where workers would receive the full value of their labor and freely exercise democratic citizenship in the political and economic realms.” The Practical Utopians – Steve Leikin The majority of workers that gave up the cooperation model and flocked to wage labor factories suffered deeply for that decision. It was not easy for the early cooperative movement to prosper when competing with conventional, capitalist employers as they were better capitalized with European fortunes. For the most part, these conventional firms became sweatshops and launched the garment industry on a difficult path for workers. The Triangle Shirtwaist Factory Fire, in New York City (1911), galvanized the citizens of the United States of America and helped produce a social contract with the people who sewed clothing for a living. The years following the fire and the New Deal brought reforms that guaranteed worker safety and the ability of a professional sewer to make a living in a safe and just work environment. Many of these reforms were pioneered by Worcester native, Francis Perkins, the first woman appointed to the US Cabinet.? This system of clothing manufacturing thrived for 50 years in the United States. In exchange for great clothing, laborers from all backgrounds had the opportunity to provide for their families, enjoyed their careers, and thrived financially. Americans fought and earned these opportunities, in large part through the creation of strong labor unions that gave them a collective voice. We have lost many of these opportunities due to the proliferation of international free trade agreements and union busting. American workers suffered a mass exodus of jobs and the cycle of clothing factory worker oppression began anew throughout the 1980’s and 90’s. The broken social contract of this industry and others lead to the street revolts at the 1999 WTO meeting in Seattle and the 2000 Democratic National Convention. The SweatX effort entered the scene at this juncture and attempted to reset history in favor of the worker. They failed. WorX Printing Cooperative is our next attempt to right this wrong.
<urn:uuid:5ab71f83-9e0b-48d0-ae2d-33d3096d9f99>
CC-MAIN-2019-39
https://worxprinting.coop/creating-a-movement/
s3://commoncrawl/crawl-data/CC-MAIN-2019-39/segments/1568514572517.50/warc/CC-MAIN-20190916100041-20190916122041-00215.warc.gz
en
0.97245
705
3.125
3
Mrs L. Stewart-Young – CL Mr W. Gray (Biology and PT Sustainability) Mrs C. Kellagher (Biology) Mr C. Craig (Biology) Advanced Higher – the course consists of 3 taught units and a research investigation. The content covered includes cells and proteins, organisms and evolution and investigative biology. Higher – the course includes 3 taught units. These are DNA and the genome, metabolism and survival and sustainability and interdependence. Pupils must pass an internal unit assessment for each of the units and complete a practical assessment. 20% of the final grade is awarded from a research assignment, which is externally marked. Higher Human – the course includes 4 taught units. These are human cells, physiology and health, neurobiology and communication and public health. 20% of the final grade is awarded from a research investigation, which is externally marked. National 3/4/5 – the courses have 3 taught units and a research assignment. The courses covers cell biology, multicellular organisms and life on earth. S3 – the S3 course provides a valuable introduction to the knowledge and skills required in Biology. Pupils will study cell structure, movement of molecules across membranes, enzymes, DNA, respiration, photosynthesis, reproduction and the brain.
<urn:uuid:4510ab8d-f9ec-4e24-91c1-19dd58f60a33>
CC-MAIN-2020-24
http://dalkeith.mgfl.net/biology-department-2/
s3://commoncrawl/crawl-data/CC-MAIN-2020-24/segments/1590347390437.8/warc/CC-MAIN-20200525223929-20200526013929-00417.warc.gz
en
0.928763
267
2.890625
3
Basic HTML | Site management | Assets | Library of Page Elements| Find and Replace | Templates | Style Sheets | Forms | Imagemaps | Mouseovers | Flash buttons & text | Frames | Tables: Formatting & Sorting | Preview Browsers | Font Lists | Colors | Automating Tasks | Pasting in a Script Cascading Style sheets Cascading Style Sheets let you specify the appearance of some text element in HTML. Basically, you define a 'class' of that element (or a generic class of any element) that has the settings you want, and then apply it to the elements you want. This allows you to, for instance, have two kinds of headings, citations, etc. in the same document. It also allows you to make global changes to the settings for all the elements of that type. So, instead of laboriously changing a bunch of different sets of text to one color of green, then changing your mind and making them yellow one at a time instead, you can give all of the elements the same class, and experiment with the colors for the whole class. If you save the styles to a separate style sheet, you can use them in any other web page too. To set up a Cascading Style sheet in your document: Once you've set the settings to your liking, click OK, then Done in the Edit Style Sheet window. Now, when you want to apply that style, select the area you want to apply the style to, and select CSS Styles from the Text menu. Choose the style you want to apply. To remove the style from something, select it, go to CSS Styles, and choose 'none.' Suppose you want to change the attributes for one kind of element throughout the entire document? When you set up a 'custom style' in Dreamweaver, and set it on something, Dreamweaver will either add the class as an attribute of the element (if you have selected the whole element), or create a 'span' element around the text you have selected and add the class to that. If the settings you have created in CSS don't appear to be applied properly, be sure the class is applied to the whole element. Editing an existing style If you want to change the attributes of a style you've already created, you will need to edit the Style Sheet. Special Types of Styles-- Advanced To change the style of links (unvisited, visited, active, or when they are hovered over), create a new style as above, but choose the Advanced selector type, then choose a:link, a:visited, a: active, or a:hover. Then create the style and set the options as above. Adding a Stylesheet an existing page To use an existing stylesheet on a page, go to the Text menu, choose CSS Styles, Manage Styles. Then click on the style you want to use, and click the "Attach" button. The next page has some sample stylesheet applications: StyleSheet examples
<urn:uuid:0f9b09d0-fcf8-4761-b631-38e4218511eb>
CC-MAIN-2015-14
http://www.lehigh.edu/computing/web/IntDreamweaver/css.html
s3://commoncrawl/crawl-data/CC-MAIN-2015-14/segments/1427131298538.29/warc/CC-MAIN-20150323172138-00287-ip-10-168-14-71.ec2.internal.warc.gz
en
0.817635
632
3.3125
3
Trying to learn a new language? The library has some resources to help. Learn a new language through interactive self-paced lessons either on a computer or through the app. Over 70 languages to study! Access Transparent Language for free on the A-Z List of Learning Databases. Learn more: Transparent Language Help Video. Pimsleur Language Systems Founded by Dr. Paul Pimsleur, Pimsleur Language Programs have been offered since since 1963. Pimsleur is primarily an audio-based language learning program. Find downloadable titles through MELSA: Twin Cities Metro eLibrary or CDs and workbooks at the library. Mango Languages employs organic language acquisition and emphasizes learning grammatical principles through realistic conversations. Download the audiobook learning through MELSA: Twin Cities Metro eLibrary, Living Language was originally developed in 1946 by foreign language education experts to teach overseas-bound service personnel and diplomats. Download the audiobook version at MELSA: Twin Cities Metro eLibrary.
<urn:uuid:8d2a28f1-de6c-4887-82c0-993386c5d336>
CC-MAIN-2023-40
https://library.stillwatermn.gov/research-and-learn/language-learning/
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233510284.49/warc/CC-MAIN-20230927071345-20230927101345-00639.warc.gz
en
0.873693
219
2.75
3
Another theory involves the subduction of the African plate under the Eurasian plate. Population Sixty one people died due to mudflows, being crushed to death and poisonous gases , while had to be rescued. Mount Etna is a volcano. When a volcano is close to erupting it starts to release gases. However, even before the eruption of Mount St Helens, scientists thought that the it might still be a few weeks away. Lava flows from the southern flank vents seriously threatened the tourist facilities around the Rifugio Sapienza between 23 and 25 November, and a few days later destroyed a section of forest on the southwestern flank. Schools in the town have been shut down, although the church has remained open for people to pray. Global change Unit 1: The reasons why Mount Etna is located where it is are complex. Report Mon 22nd May, A detailed plan is needed for dealing with a possible eruption. Heavy tephra falls caused by the activity on the southern flank occurred mostly in areas to the south of the volcano and nearly paralyzed public life in Catania and nearby towns. Impact on landscape and population Landscape The mountain was reduced from a height of m to m as the eruption created the largest landslide ever recorded. Preparing for volcanic eruptions A detailed plan is needed for dealing with a possible eruption. This is a easy website to access for all ages especially primary school. Sign up to Comment. Hundreds of small earthquakes are caused as magma rises up through cracks in the Earth’s crust. Why do people live near volcanoes? One theory envisages a hot-spot or mantle-plume origin for this volcano, like those that produce the volcanoes in Hawaii. Villagers also continued their tradition of parading their patron saint through the streets to the railway station, to try to ward off the lava flow. These warning signs are picked up by volcanologists experts who study volcanoes and the volcano is monitored. It even helps my older girl who is 14 do her homework for secondary school. The higher the sulfur content of these gases, the closer the volcano is to erupting. Disadvantages Ash fall- towns and farms are covered by tonnes of ash, destroying crops Poisonous gases are produced, choking and killing people Lava flows, with temperatures between c and c Mud flows are formed from ash and rain mixing, creating a catastrophic river of mud. Scientists can also check for gas emissions sulphur dioxide and increased thermal activity at the crater. Friction and heat cause the plate to melt and, as it melts, molten rocks are formed. Advantages Volcanic ash produces very fertile soil, ideal for farming e. Emergency services were also on hand to rescue those people needing help. Report Thu 7th April, This has been a fantastic website for my children. National 5 Subjects National 5 Subjects up. Volcanoes have a wide range of effects on humans. Mount Etna Case Study Worksheet. Impact on landscape and population Landscape The mountain was reduced from a height of m to m as the eruption created the largest landslide ever recorded. Volcanoes – Revision 2 – KS3 Geography – BBC Bitesize Sicilian city ztudy in ash [28 October ]. Japan and Iceland Valuable minerals and metals can be found near volcanic sites such as tin, copper, silver and gold. Everyone who could be affected needs to know the plan and what they should do if it needs to be put into action. Predicting volcanic eruptions As a volcano becomes active, it gives off a number of warning signs. Global environmental hazards – Revision 7 – National 5 Geography – BBC Bitesize Helens, Washington State, began its most recent series of eruptions in when a massive landslide and powerful explosive eruption created a large craterand ended six years later after more than a dozen eruptions of lava built a dome in the crater. This is because they have been active in recent years and people who live nearby would benefit from early-warning signs of an eruption. For a named area, describe the causes of the example which you have chosen and its impacts on the people living there. It is usually the destructive nature of volcanoes which is more bitesixe documented. I find this website very good for all types stufy things. However, even before the eruption of Mount St Helens, scientists thought that the it might still be a few weeks away.
<urn:uuid:2afeadfd-0b7d-492d-bd70-db65b6962a95>
CC-MAIN-2020-10
https://feather2s.us/
s3://commoncrawl/crawl-data/CC-MAIN-2020-10/segments/1581875145438.12/warc/CC-MAIN-20200221014826-20200221044826-00551.warc.gz
en
0.967491
907
3.578125
4
Most people know how augmented reality (AR) technology works thanks to revolutionary games like Pokémon Go and social media platforms like Snapchat. However, AR is more than that. Today, it’s now being used in several fields, including education and training. Here are some use cases of AR in education and training professionals, helping you see what it can bring to the learning experience. Refining Professionals’ Skills The average lifespan of professional skills is five years. Fortunately, thanks to AR, companies, and organizations can help their employees refine their skills by re-educating or retraining them. It’s a more convenient and cost-effective way of polishing their skills. For instance, AR can help pilots hone their skills without the hassle of dealing with immigration policies from different countries and hiring an immigration lawyer to help them out. It saves the company and the professional more time and money. Medical Education and Training Medical professions, no matter the niche, require a high level of proficiency and accuracy. That’s because any mistake may leave a negative impact on a patient’s health and well-being. AR in medical education can drastically reduce this by giving students an immersive experience and familiarizing themselves with the human body more. Through interactive and visual representations, these can create simulations that can train medical students by practicing medical procedures on virtual patients. All these can help medical students improve both their knowledge and skills. AR allows them to practice surgery and do other medical procedures to virtual patients more, honing their techniques and abilities without harming anyone. Augmented Reality Classroom The simplest and widespread AR use in education is in the traditional classroom. It can support textbook materials by adding a different dimension to the learning process, making it simpler and more engaging thanks to its innovative, practical illustrations of complex concepts. For instance, AR-enabled scanners that can scan covers of textbooks can help students receive a short description of what they’re about, allowing them to make better-informed decisions regarding selecting learning material most suitable for them. AR can make the learning experience for students who have gotten so used to visual stimulation and interactivity. Plenty of schools and other educational facilities have implemented AR in their classrooms. AR-Enabled Worksheets at Home The AR application where it allows teachers to start giving their students AR-enabled school worksheets isn’t far. It aims to encourage students to explore more educational content at their own account, helping them handle homework and projects better when not interacting with an educator. Several educational institutions are testing out ‘printable’ AR worksheets, making information and graphics “pop-out” literally. It emphasizes crucial concepts while breaking the boundaries of classic textbook learning despised by many students. AR lets you interact with 3D models, allowing you to set rotation, transparency, color, styles, dimensions, and more. All these allow users to have a more in-depth insight of particular things, such as human body parts used by medical students or building structures utilized by architects. It can help you learn better and explore different topics deeper. Thanks to the technological advancements today, you’re bound to see excellent AR applications in education and training worldwide — and those mentioned are some of the best ones, steadily improving and opening more options for students and professionals alike.
<urn:uuid:37205cd2-6db8-46e9-b4fb-3373c4a23ee1>
CC-MAIN-2023-40
https://www.factoryschool.com/uses-of-augmented-reality-in-education-and-career/
s3://commoncrawl/crawl-data/CC-MAIN-2023-40/segments/1695233510734.55/warc/CC-MAIN-20231001005750-20231001035750-00378.warc.gz
en
0.941241
684
3.3125
3
A road into Berners Bay is an old, outdated, dead-end idea. According to an article in the April 24 Juneau Empire, "25 years later, conservation act drives boom," conservation of Alaska lands in parks and preserves has been an economic bonanza for the state. The article details the economic benefits enjoyed by the community of Seward due to Kenai Fjords National Park, a $52 million a year business. Unlike the proposed Kensington Mine, Kenai Fjords National Park did not require a dead-end road (paid for by tax dollars). The park also did not operate for 10 or 15 years, close down and leave behind an environmental mess to be cleaned up by the state. Over the planned life of the mine, the Kensington will dump more than 4 million tons of mining waste into Lower Slate Lake, this toxic slurry will have to be held in check by a dam and will likely be left for the state to monitor for many years. When the Kensington Mine closes down after 10 or 15 years of operation, it will leave behind a legacy of environmental degradation. The proposed Kensington Mine's industrial facilities will sprawl across the bay, two port facilities will handle barges carrying supplies, fuel containers, and thousands of tons of ore. Ferries will shuttle employees back and forth across the bay up to five times daily. This industrial boat and barge traffic will likely disturb the bay's spectacular marine life, including Steller sea lions and whales. The fuel spills and impacts from the construction of docks will harm the bay's rich eulachon and herring schools. The bay's rich fisheries and marine life will continue to suffer even after the mine is closed down. Instead of living in the past and promoting a dead-end road into Berners Bay for the benefit of a socially irresponsible, multinational mining company with terrible environmental and labor records, we should be looking to the future and promoting the protection of Berners Bay. Juneau's elected officials should be taking a leadership role and planning for the future, calling for the creation of a state marine park at Berners Bay, not a dead-end mining road. © 2017. All Rights Reserved. | Contact Us
<urn:uuid:b2cecab0-0eb4-48d3-97aa-4a487a3fa6db>
CC-MAIN-2017-13
http://juneauempire.com/stories/070705/let_20050707013.shtml
s3://commoncrawl/crawl-data/CC-MAIN-2017-13/segments/1490218188785.81/warc/CC-MAIN-20170322212948-00518-ip-10-233-31-227.ec2.internal.warc.gz
en
0.944875
451
2.75
3
We hear a lot about the need to stay physically active in order to support both our physical and mental health. You probably already know that exercise is important for heart health and your weight, but what do studies show about its impact on your mood — does exercise actually help with depression? There’s good reason to believe that there are, in fact, many uses of exercise for depression, including because physical activity fights fatigue and can reduce pain. Exercise can relieve anxiety too — plus it often leads to improved sleep, focus, self-esteem and productivity. What type of exercise is best for mental health? As explained more below, it’s important to do the types you enjoy and that you can stick with long term, such as walking outside, lifting weights or commuting by bike. Study Findings: Exercise Combats Depression A 2022 systematic review and meta-analysis published in the journal JAMA Psychiatry investigated the association between physical activity and depression in adults. The review, which included 15 studies and over 191,000 participants, found an inverse curvilinear association between physical activity and incident depression, meaning exercise decreased depression up until a certain point. Adults who reported meeting physical activity recommendations — which is equivalent to about 2.5 hours of moderate exercise per week, such as brisk walking or cycling — had lower risks of developing depression compared to adults who reported not being physically active at all. Even relatively small doses of physical activity were associated with substantially lower risks of depression. In fact, the biggest difference in depression symptoms was seen between people who were sedentary and those who did at least some exercise every week. In the study, adults who did half the recommended amount of physical activity each week had about 18% less risk of depression compared to sedentary adults. Those meeting the recommended amount of exercise had 25% lower risk. Beyond this point, there were diminishing additional benefits observed. In other words, doing higher volumes of physical activity above the recommended amount (2.5 hours per week) won’t necessarily protect your mental health any more. In conclusion, just like many studies have demonstrated over the past several decades, making an effort to stay active and fit can serve as a natural remedy for depression. Why It Matters: Researchers estimate that if less active adults achieved the current physical activity recommendations each week, then about 12% of depression cases could be prevented. According to the World Health Organization, depression is the leading cause of mental health–related disabilities worldwide. What does depression do to a person? Symptoms and signs of depression can include: - Persistent feelings of sadness, worthlessness and hopelessness - Low motivation for physical activities - Social withdrawal and isolation, along with loneliness - Aches and pains - Sleep disturbances - Brain fog and trouble concentrating - Irritability and restlessness - Changes in appetite - Digestive issues - Sexual dysfunction - In severe cases, suicidal thoughts and attempts How to Fit More Exercise Into Your Day Which exercise is good for depression? Just about all types are beneficial, including strength training, aerobic exercises, and both moderate and high-intensity exercise (or HIIT workouts). Moderate exercise is the most widely recommended types for adults of all ages, including older adults. It’s defined as “activity that gets you moving fast enough or strenuously enough to burn off three to six times as much energy per minute as you do when you are sitting quietly.” Here are some examples of moderate exercise: - Brisk walking outside or on a treadmill - Cycling indoors or outdoors at a moderate pace - Gentle hiking - Playing tennis - Lifting weights - Doing bodyweight exercises like yoga or pilates - Heavy cleaning in your home or even mowing your lawn or shoveling show More vigorous exercise can boost your mindset too, such as running or jogging, playing most sports like soccer, and cycling at a fast pace. How long do you need to exercise to help depression? In the study mentioned above, meeting exercise recommendations offered the most mental health perks. The current recommendation is 2.5 hours of moderate exercise per week, or about 30 minutes most days of the week. If you enjoy doing more, then by all means do so, but don’t feel obligated if it feels like too much. How can you add more movement and exercise to your day? Below are some ideas for sneaking more physical activity into your day: - Try commuting by bike or foot whenever possible. - Park farther away from destinations and walk. - Take the stairs whenever possible rather than elevators. - Do yoga or pilates at home when you have free time, such as by following a YouTube video. - Do light calisthenics, such as squats or lunges, when watching TV. - Take your pet for daily walks, or walk together as a family. - Use a standing or treadmill desk instead of sitting for hours. - Try having “walking meetings” or walking while on the phone. Other Ways to Fight Depression - Work on getting enough sleep. Aim for seven to nine hours per night. - Eat a healthy diet. Increase your consumption of fruits, vegetables, legumes, whole grains, nuts and seeds, and consume plenty of omega-3 foods and healthy fats like fish and olive oil. Limit your intake of added sugar, processed foods, fast foods, commercial bakery goods and sweets. - Reduce your alcohol intake, and quit using drugs and/or smoking. - Get some fresh air and sunlight every day. - Consider taking supplements that may help your mood, such as fish oil, vitamin D, B vitamins, adaptogenic herbs and St. John’s wort. - Studies show there are many benefits of exercise for depression and other mood issues. A 2022 meta-analysis found that meeting exercise recommendations can slash your risk of depression by about 25%. - The review, which included 15 studies, uncovered significant mental health benefits from being physically active, even at low levels. - Moderate exercise for about 2.5 hours per week, or about 30 minutes most days of the week, is thought to be enough to support mental health. - Try including more walking, swimming, cycling, hiking and even cleaning or mowing into your routine to boost your activity level.
<urn:uuid:d8e0d46d-86eb-49fd-a7b0-d2ca96c7337a>
CC-MAIN-2024-10
https://draxe.com/health/study-shows-exercise-fights-depression/
s3://commoncrawl/crawl-data/CC-MAIN-2024-10/segments/1707947474659.73/warc/CC-MAIN-20240226094435-20240226124435-00346.warc.gz
en
0.949898
1,308
3.296875
3
Now that your child’s a preschooler, you can expect longer, more complex conversations about all sorts of things. His vocabulary will continue to grow, and he might show that he understands the basic rules of grammar and you can look forward to some entertaining stories. Here are some of the things your preschooler might do from 4-5 years. At four, she’ll use around 1500 different words and understands even more. By five, she’ll have an even wider range of words that she understands and uses. Your child will begin to learn and use more: - connecting words (‘when’, ‘but’) - words that explain complicated emotions (‘confused’, ‘upset’, ‘delighted’) - words that explain things going on in his brain (‘don’t know’, ‘remember’). He’s also learning more and more adjectives that let him explain things better. Sentences and grammar Your child will speak in increasingly complex sentences by joining small sentences together, and will use sentences in different ways. For example, she’ll be able to say both ‘The dog was chasing the cat’ and ‘The cat was chased by the dog’ to mean the same thing. By five, your child will be able to use long sentences of up to nine words. Your child will develop the ability to talk about things that have happened in the past, rather than just things that are currently happening. He’ll also get better at using past tense and plurals. Did you know? Reading to your child is an important activity in supporting speech and language development. Reading regularly to your child from an early age helps build a stronger relationship, supports speech development, reading and school performance. By five, your child will understand and use words that explain when things occur, such as ‘before’, ‘after’ and ‘next week’. She might still have trouble understanding complicated ideas, such as ‘at the same time’. He’ll begin to understand figures of speech, such as ‘You’re pulling my leg’ and ‘He’s a couch potato’. Your child will follow directions with more than two steps, even if the situation is a new one. For example, ‘Give your ticket to the man over there, and he’ll tear it, and then we can go to the movie’. But your child might do what she hears first and ignore words that tell her the order she should do things in. For example, she might ignore the word ‘before’ in the sentence ‘Before we go into the movie, give your ticket to the man’. By the age of 4½-5 years, almost every word your child says can be understood by strangers. He might still have difficulty using some speech sounds – for example, saying ‘fing’ for ‘thing’ or ‘den’ for ‘then’ – and might mispronounce some complex words, such as ‘ambulance’ and ‘spaghetti’. Conversation and storytelling Your child will continue to improve her storytelling, although she might still give too much or not enough information. She might also have trouble telling things in order and making it clear who’s being spoken about. But he’ll have greater appreciation of others’ perspectives, so he might add more useful background information in conversation. For example, ‘I went to Mark’s and we had cake and Mark is from my preschool’. Your child will be getting better at taking turns in conversations with a group of people. And she’ll start talking at the right volume for the situation. She might make requests more politely, using words such as ‘can’, ‘would’ and ‘could’. He’ll begin to use language to tease and tell jokes. Tip: Children grow and develop at different rates. The information in this article is offered as a guide only. If you’re at all concerned about your child’s language development, check with your public health nurse or doctor. © Raising Children Network Limited, reproduced with permission. Resources & Links: HealthLink BC: Speech and Language Milestones, Ages 3 to 5 Years HealthLink BC: Encouraging Language Development in Your Preschooler HealthLink BC: Speech and Language Development: Red Flags HealthLink BC: Speech and Language Delays: Common Misconceptions HealthLink BC: Speech Problems: Normal Disfluency (stuttering)
<urn:uuid:1da9d20a-1cbb-461c-80ac-97f9abe177bd>
CC-MAIN-2019-04
https://www.healthyfamiliesbc.ca/home/articles/language-development-4-5-years-preschool
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583823140.78/warc/CC-MAIN-20190122013923-20190122035923-00358.warc.gz
en
0.948668
998
3.796875
4
Griscelli syndrome is an inherited condition characterized by unusually light (hypopigmented) skin and light silvery-gray hair starting in infancy. Researchers have identified three types of this disorder, which are distinguished by their genetic cause and pattern of signs and symptoms. Griscelli syndrome type 1 involves severe problems with brain function in addition to the distinctive skin and hair coloring. Affected individuals typically have delayed development, intellectual disability, seizures, weak muscle tone (hypotonia), and eye and vision abnormalities. Another condition called Elejalde disease has many of the same signs and symptoms, and some researchers have proposed that Griscelli syndrome type 1 and Elejalde disease are actually the same disorder. People with Griscelli syndrome type 2 have immune system abnormalities in addition to having hypopigmented skin and hair. Affected individuals are prone to recurrent infections. They also develop an immune condition called hemophagocytic lymphohistiocytosis (HLH), in which the immune system produces too many activated immune cells called T-lymphocytes and macrophages (histiocytes). Overactivity of these cells can damage organs and tissues throughout the body, causing life-threatening complications if the condition is untreated. People with Griscelli syndrome type 2 do not have the neurological abnormalities of type 1. Unusually light skin and hair coloring are the only features of Griscelli syndrome type 3. People with this form of the disorder do not have neurological abnormalities or immune system problems. Griscelli syndrome is a rare condition; its prevalence is unknown. Type 2 appears to be the most common of the three known types. The three types of Griscelli syndrome are caused by mutations in different genes: Type 1 results from mutations in the MYO5A gene, type 2 is caused by mutations in the RAB27A gene, and type 3 results from mutations in the MLPH gene. The proteins produced from these genes are found in pigment-producing cells called melanocytes. Within these cells, the proteins work together to transport structures called melanosomes. These structures produce a pigment called melanin, which is the substance that gives skin, hair, and eyes their color (pigmentation). Melanosomes are formed near the center of melanocytes, but they must be transported to the outer edge of these cells and then transferred into other types of cells to provide normal pigmentation. Mutations in any of the three genes, MYO5A, RAB27A, or MLPH, impair the normal transport of melanosomes within melanocytes. As a result, these structures clump near the center of melanocytes, trapping melanin within these cells and preventing normal pigmentation of skin and hair. The clumps of pigment, which can be seen in hair shafts when viewed under a microscope, are a hallmark feature of the condition. In addition to their roles in melanosome transport, the MYO5A and RAB27A genes have functions elsewhere in the body. Specifically, the protein produced from the MYO5A gene transports materials within nerve cells (neurons) that appear to be critical for cell function. The protein produced from the RAB27A gene is found in immune system cells, where it is involved in the release of certain compounds that kill foreign invaders (such as viruses and bacteria). Mutations in these genes impair these critical cell activities, leading to the neurological problems and immune system abnormalities found in Griscelli syndrome types 1 and 2, respectively. This condition is inherited in an autosomal recessive pattern, which means both copies of the gene in each cell have mutations. The parents of an individual with an autosomal recessive condition each carry one copy of the mutated gene, but they typically do not show signs and symptoms of the condition. - hypopigmentation immunodeficiency disease - partial albinism with immunodeficiency - Centers for Disease Control and Prevention: Developmental Disabilities - MalaCards: griscelli syndrome, type 1 - MalaCards: griscelli syndrome, type 2 - MalaCards: griscelli syndrome, type 3 - Orphanet: Griscelli syndrome - The Merck Manual for Healthcare Professionals: Hemophagocytic Lymphohistiocytosis (HLH) - University of Bristol: Melanin, Melanocytes, and Melanosomes
<urn:uuid:01e2a188-7391-44ec-b854-809694caa60b>
CC-MAIN-2019-04
https://ghr.nlm.nih.gov/condition/griscelli-syndrome
s3://commoncrawl/crawl-data/CC-MAIN-2019-04/segments/1547583658928.22/warc/CC-MAIN-20190117102635-20190117124635-00430.warc.gz
en
0.936467
923
3.953125
4
The RMS Titanic was once the largest floating ship in the world. It had a massive hull, towering smoke stacks, an opulent staircase, gargantuan propellers, enormous boilers and plenty of deck space. It was to sail on a wide, expansive sea. Everything about the Titanic and its journey was big. Its biggest promise — one few stopped to question — was that it was an unsinkable ship. On April 10, 1912, it left Southampton, England, to travel west across the Atlantic Ocean and reach New York City on the other side. It carried 2,228 passengers and crew. But hubris on the part of the White Star Line, and lax maritime safety regulations of the time, ensured most on board never reached the U.S. To add deck space, designers of the ship had a second row of lifeboats removed. That left only 20 lifeboats — not nearly enough for all who were on board. When the ship struck an iceberg 600 kilometres off the coast of Newfoundland and sank, its once awe-inspiring size and capacity became a hindrance. Women, children and some men were ushered into lifeboats that had no hope of holding every person on the ship. More than 1,700 people were still on the Titanic when it broke apart and slipped into the frigid North Atlantic. Anyone who fell into the water would have died within minutes because of hypothermia. When the news reached Europe and North America — that the mammoth unsinkable ship was had indeed sunk — the headlines in newspaper were big too. Seven hundred and five people survived and more than 1,5oo died.
<urn:uuid:48b12690-99c7-45ee-b5e1-8d04cf99213f>
CC-MAIN-2017-26
http://nationalpost.com/news/titanic-pictures-shipwreck-anniversiary/wcm/e071c02c-1cd6-45f5-9e9b-ebdeea55a318
s3://commoncrawl/crawl-data/CC-MAIN-2017-26/segments/1498128320438.53/warc/CC-MAIN-20170625050430-20170625070430-00638.warc.gz
en
0.985782
335
3.65625
4
This March saw 650 top dementia researchers attend the Alzheimer’s Research UK Research Conference. Delegates travelled from across the country and beyond to London to present their findings and gain insight into the cutting-edge science paving the way for new dementia treatments. A highlight of the conference, in which so much was covered, was a presentation from Dr John Skidmore, Chief Scientific Officer at our Cambridge Drug Discovery Institute (DDI). He spoke with fascinating insight about the progress his team is making towards drugs that could tackle the processes driving damage to the brain in the diseases that cause dementia. Delving into the Drug Discovery Institute Back in 2015 we announced our £30m Drug Discovery Alliance and launched three flagship Drug Discovery Institutes at the Universities of Cambridge, Oxford and University College London (UCL). The DDIs now employ scientists in state-of-the-art facilities to fast-track the development of new treatments for Alzheimer’s disease and other dementias. In his presentation, Dr Skidmore talked about the work of the Cambridge Drug Discovery Institute and updated attendees on significant progress his team is making towards drugs that could limit damage to nerve cells in diseases like Alzheimer’s. Maintaining a balance Proteins are the building blocks of life. There are an enormous variety of proteins in our bodies, and the activity of these complex molecules is what drives the processes that keep our cells functioning. A lot of medical research is about understanding these cellular processes — finding out how they go wrong and how we can put them right. The Cambridge DDI has a particular focus on mechanisms that keep the levels of all the body’s proteins in check, a process known as ‘proteostasis’. There is a continuous flux in protein levels in our bodies, through the production of new proteins and the activity of waste disposal mechanisms. Our cells are usually able to maintain a careful balance between these processes, but in a number of brain diseases this balance starts to go awry and certain proteins build up inside or around nerve cells. Proteins are made from long chains of smaller molecules called amino acids. These long chains are folded into particular shapes. When proteins are folded incorrectly they can form sticky tangles, which clump together in larger and larger units, interfere with the way cells work, and lead to damage. Misfolded proteins play a central role in diseases including Alzheimer’s disease, frontotemporal dementia, and dementia with Lewy bodies. A waste disposal process known as autophagy can bag up misfolded proteins ready for destruction and recycling. Normally this goes on behind the scenes without any glitches, however, when there is too much protein building up, the cell’s waste disposal system can become overwhelmed. Dr Skidmore and his team are working on drugs that boost autophagy and help clear away these clumps of protein. The Cambridge DDI is positioned to seize upon new findings that emerge from academic research at the University of Cambridge and take forward work that could lead to new dementia drugs. Dr Skidmore works particularly closely with the Institute’s Lead Academic Scientist, Prof David Rubinsztein, who is a leading expert in the role of autophagy in degenerative brain diseases. A key part of the drug discovery process is picking the right target. Autophagy involves a complex chain of molecular events, with each link being an interaction between specific proteins. Scientists at the DDI are working on drugs that latch on to these proteins and alter their activity in a way that boosts autophagy. Following leads thrown up by the work of Prof Rubinsztein and his colleagues, the Cambridge DDI has zeroed in on proteins called PI5P4 kinases. These proteins limit autophagy in nerve cells, so by blocking them with targeted drugs, the DDI is hoping to increase autophagy and boost the cell’s ability to clear away toxic proteins. Using a combination of computer modelling and ‘test tube’ experiments DDI team has looked at many thousands of chemical compounds to see if they are able to block the activity of PI5P4 kinases. This led to the identification of a number of compounds which the team improved further through rounds of design, chemical synthesis and testing, building in the properties needed for a drug. Through these experiments, the scientists at the Cambridge Drug Discovery Institute have developed a number of lead compounds that are able to increase autophagy and clear the clumps of misfolded proteins in experiments carried out in cells. These compounds also have many of the properties required of drugs, such as being able to get into the brain when administered orally. There is a legal requirement for all drugs to be tested in animals before they can progress into trials involving people. This important part of the drug discovery process allows researchers to move from experiments in a dish in the lab, to being able to see the effect of a drug in a living brain that has many important similarities to those of people. The DDI is in close contact with pharmaceutical companies and if these compounds are successful in animal studies, there is a clear path to develop compounds ready for human clinical trials. The Alzheimer’s Research UK Drug Discovery Alliance was set up to bridge the gap between university research and the pharmaceutical companies. It is fantastic to see how the Cambridge DDI took what was an interesting finding about a cellular process, and thanks to your support, turned it into potential drugs that are now entering a key stage of preclinical testing. This is crucial progress, but there is still a long distance between this stage of research and something that makes a difference to people’s lives. There are promising dementia drugs that are already in clinical trials, and while we are optimistic that one or more of these will be effective, to have the best chance of success it is crucial that we maintain a healthy pipeline of potential drugs. The picture below shows the progress of different projects in the Drug Discovery Alliance last year. Protein housekeeping is just one of the disease processes being explored, and the Institutes are making progress on a number of different fronts, with a growing portfolio of projects. We look forward to bringing you further updates about these projects as more make it through to preclinical testing and clinical trials. It has been over 15 years since the last dementia treatment became available, but thanks to your support Alzheimer’s Research UK is investing in initiatives like the Drug Discovery Alliance, which are dedicated to bringing about new treatments sooner. Research has the power to change lives. Donate today to support pioneering science that can defeat dementia.
<urn:uuid:7b6f3a20-6e50-45f3-9cde-bdc27feb28d7>
CC-MAIN-2019-30
https://www.dementiablog.org/drug-discovery-alliance-an-update-on-progress/
s3://commoncrawl/crawl-data/CC-MAIN-2019-30/segments/1563195525500.21/warc/CC-MAIN-20190718042531-20190718064531-00445.warc.gz
en
0.951244
1,351
2.734375
3
People with eczema are especially vulnerable to allergic reactions to certain preservatives in lotions and creams, according to a new study. With eczema, the skin doesn't function as a barrier the way it normally would, said senior author Dr. Donald V. Belsito, a dermatologist at Columbia University Medical Center in New York City. People with eczema often apply lots of moisturizers and topical medications to combat dryness and itchiness, Belsito said. "The more exposure, the more likely you are to develop an allergic reaction to a chemical," he said. In his study published in the Journal of the American Academy of Dermatology, patients with eczema were more likely to have allergic reactions to several cosmetic preservatives, called "formaldehyde releasers," than people without eczema. Of 2,500 people tested for allergic reactions in the study, 342 had eczema. After a battery of individual allergy tests, the eczema group was more likely to have a reaction to preservative chemicals quaternium-15, imidazolidinyl urea, DMDM hydantoin and 2-bromo-2-nitropropane-1,3-diol. Many things might cause these chemicals to be harmful, Belsito said. One possibility is that this group can penetrate the outer layer of skin and bind to immune cells, activating them. All of the chemicals in this group also release formaldehyde. Eczema patients in the study did not seem to be predisposed to allergies to parabens, formaldehyde or diazolidinyl urea, which are not formaldehyde releasers. It is unclear what makes the "formaldehyde releasers" problematic, the authors write. It could be that those products are more commonly used so people with eczema have more exposure to them, and more opportunity to develop a reaction. About eight percent of all people have irritation or allergic reaction to formaldehyde itself, according to a previous study. These are not dangerous allergic reactions, just very uncomfortable, like poison ivy, Belsito said. Skin reactions usually go away if you stop using the product and see your doctor, who may give you a steroid cream, said Michael Dyrgaard Lundov, a senior researcher at the National Allergy Research Centre in Copenhagen, Denmark. He was not involved in the new study. Eczema - which is an excessive immune response similar to those in asthma and allergic reactions - causes painful, red scaly patches on the skin. It's becoming more common, Belsito said, and in some areas up to 30 percent of kids have it. Some people with mild eczema haven't been diagnosed - they just think they have sensitive skin, he said. Eczema patients "should be advised to treat their skin with ointments, which are unlikely to contain antimicrobial preservatives," he said. Microorganisms need water to proliferate, Lundov said. Products without water, like ointments, therefore don't need to include antimicrobials. Eczema patients should also avoid products with fragrance, which can cause irritation as well, he said. If eczema patients must use creams or lotions containing antimicrobials, they should choose products preserved with parabens, Belsito and Lundov agreed. Parabens are one of the most common cosmetic preservatives, in use for more than 70 years, according to the U.S. Food and Drug Administration. Less than one percent of people had a reaction to parabens in a previous study. Allergic reactions depend on how potent a chemical is, how often you are exposed, how much of the chemical is in the product you use and the unique tolerance level of your own skin, Lundov said. "Despite the massive use of parabens in cosmetic products the parabens are rarely a cause of allergic contact dermatitis," Lundov said. "This is because it is not a very potent allergen." "Quaternium-15 and the other formaldehyde releasers are used in lower concentrations and fewer products, but is still a bigger problem because the releaser itself and the released formaldehyde are more potent allergens than the parabens," he said.
<urn:uuid:2fa240d1-9028-4e9a-888b-0983603b3021>
CC-MAIN-2016-22
http://www.newsmax.com/Health-News/eczema-lotion-preservatives-allergy/2013/12/12/id/541549/
s3://commoncrawl/crawl-data/CC-MAIN-2016-22/segments/1464049274059.91/warc/CC-MAIN-20160524002114-00221-ip-10-185-217-139.ec2.internal.warc.gz
en
0.959457
895
2.75
3