text
stringlengths 307
28.2k
| __index_level_0__
int64 0
904
|
|---|---|
Mon 20 Jul 2009
As far as scary-looking fish go, the adult sea lamprey is right up there with the sharks. It’s about 3 feet long, with concentric rings of teeth inside a jawless, suction-cup mouth. After trapping its prey by sucking, the lamprey scrapes off the skin using a tooth at the end of its tongue, and then sucks out the blood. Since the lamprey has good taste in fish – preferring trout, salmon, bass –it’s the target of a federal lampricide program in the Great Lakes region, where it has nearly devoured some of these stocks.
But no creature is all bad. “The lamprey has a really beautiful spinal cord,” says Ona Bloom, an MBL Research Fellow from The Feinstein Institute for Medical Research. Together with Jennifer Morgan (Univ. of Texas at Austin) and David Parker (Univ. of Cambridge), who are recipients of an MBL sponsored Albert and Ellen Grass Faculty Research Grant, Bloom and her colleagues are studying the lamprey’s remarkable ability to recover from severe spinal cord injury. “We have been talking for years about collaborating on these studies,” says Bloom. “ The MBL gives us a wonderful opportunity to do that.”
One of the most primitive of living vertebrates, the sea lamprey is a good animal in which to analyze the acute response to spinal cord injury, which is Bloom’s focus, and how the nervous system functions are restored, which interests Morgan and Parker. “We want to understand the differences between the robust recovery in lampreys and how that contrasts with what happens in humans with spinal cord injury,” says Morgan. “If we can understand processes in the lamprey that contribute to recovery, then we can see if those processes can promote better recovery in higher vertebrates.”
| 582
|
exactly located (exactlyLocated)
The actual, minimal location of an
Object. This is a subrelation of the more general Predicate
SUMO / BASE-ONTOLOGY
Related WordNet synsets
- the precise location of something; a spatially limited location; "she walked to a point where she could survey the whole street"
Agar obj is partly located in region, to yah kuch subobj nahin, ki subobj is a part of obj aur subobj is exactly located in region.
(partlyLocated ?OBJ ?REGION)
(part ?SUBOBJ ?OBJ)
(exactlyLocated ?SUBOBJ ?REGION))))
Agar obj is exactly located in region, to yah kuch otherobj nahin, ki otherobj is exactly located in region aur otherobj is not equal to obj.
(exactlyLocated ?OBJ ?REGION)
(exactlyLocated ?OTHEROBJ ?REGION)
(equal ?OTHEROBJ ?OBJ))))))
"thing ki jagah time tha" is equal to region agar hai thing is exactly located in region during time.
(WhereFn ?THING ?TIME)
(exactlyLocated ?THING ?REGION)))
| 261
|
By Margaret Southern
For many people in the United States, food comes from the grocery store. Most Americans would be hard pressed to determine the real origins of much of the food we eat. But residents of the Cayos Cochinos islands off the northern coast of Honduras know exactly where their food comes from: the ocean outside their doors.
Here, it’s not just their stomachs that rely on the bounty of the ocean—it’s almost their entire economy. But overfishing and industrial fishing continues to put pressure on their marine resources. The Nature Conservancy is working with partners and locals to protect their reefs through education, conservation and income-generating projects.
In 2007, the Conservancy and the Cayos Cochinos Foundation identified almost 20 income-generating projects with the goal of increasing income for households that depend heavily on the reef.
One of these projects was the construction of a small tourist complex managed by the East End community. The community hopes to attract tourists with two eco-friendly cabins, which were built with outside funding, and an ocean-view restaurant serving traditional Garifuna meals, funded by the Conservancy. The Conservancy is connecting the community with tourism service providers to promote best practices and ensure the protection of conservation targets.
The Cayos Cochinos Foundation, along with the Conservancy and other groups, has also worked with the communities to establish management plans that encourage no-take zones and the construction of artificial reefs that serve as nurseries for small fish. And through other partnerships, laws have been passed that will eventually prohibit industrial fishing around the marine protected area.
All of these initiatives are important in creating sustainability in the reefs and communities of Cayos Cochinos. According to Francisco Velasquez, a school teacher in East End, if it weren’t for the Conservancy, they wouldn’t still be living on that island.
“Our livelihood depends on fishing, but in the past large ships would come in and sweep the whole ocean clear of fish. There was nothing left for us to eat,” Velasquez said. “But now with the conservation of this area the fish are returning, and through the tourism project we are creating new solutions to increase our income and improve our lives.”
In Cayos Cochinos, these projects would not have been possible without Anthony Ives. A former San Francisco stock trader who at one point was managing 50 billion dollars for the state of California, Ives found his way to Honduras after Sept. 11 through the Peace Corps. He created the Foundation Heart Ventures-Grupo de Apoyo al Desarollo (IHV-GAD) to have a focus on education, conservation and job creation, with the idea that it is not possible to separate the three objectives.
“We have found that an integrated approach is much more sustainable, and we’re very happy that the Conservancy shares our long-term vision to sustainable conservation,” Ives said. “We hope to convince other organizations that through an integrated approach that involves education, we can achieve sustainable development while focusing on the environment.”January 03, 2011
| 195
|
Portrait of Katherine Parr, Queen of England (1512-1548)
English School, Late Sixteenth Century
“As the famous rhyme suggests, Katherine Parr’s record as the last of Henry VIII’s six wives was unique. She survived.”
Oil on panel
21¼ x 15¾ inches, 54 x 40 cm
Possibly William Bryant, sale at Edward Foster’s, London, 15th May 1823 as ‘Holbein’
Archibald Alexander Leslie Melville, 13th Earl of Leven and 12th Earl of Melville
by whom sold, Robinson and Fisher, 29th November 1934 lot 107, as ‘Holbein’
Anon sale Christies London 9th June 1944, Lot 14
1967 J Gold Coll.
Anon Sale Sothebys London 11th July 1983 (lot 37)
English Private Collection
Katherine Parr, by Susan E. James (London 1999) p 421 fig 30, as showing the Queen c. 1546;
Ian Tyers, Dendrochronoligical Survey, 2005;
Inscribed at top, Catharina Regina Uxor Henrici VIII
As the famous rhyme suggests, Katherine Parr’s record as the last of Henry VIII’s six wives was unique. She survived. Though Anne of Cleves, the sad ‘Flanders Mare’ unable to arouse England’s most insatiable monarch, lived on until 1557 it is only Katherine who was neither divorced, beheaded, or died. She was by any standards a remarkable woman: beautiful enough to marry the King of England, despite having neither royal nor court background; shrewd enough to remain his Queen, despite court plots and an attempt on her life; and courageous enough to sustain the Protestant cause, despite Henry’s latent sympathies for the Roman faith. She was Regent of England during Henry’s invasion of France in 1544. And with her publication of religious works such as Prayers or Meditations in 1545, she became not only the first English Queen to publish a work of prose, but the first woman to do so in the sixteenth century.
Katherine became Queen of England in July 1543. Henry was her third husband, but, on this occasion, not her first choice. She had instead fallen in love with the dashing courtier Thomas Seymour, and was understandably wary of Henry’s past form when it came to marital relations. Five wives had failed – what chance did a sixth have of success? Nonetheless, to turn down the King’s offer of marriage was unthinkable. Katherine, a deeply devout woman, determined that if she was to be Queen, she would be Queen with a purpose. That purpose was to further the cause of the Protestant Reformation.
In doing so Katherine, literally, risked her life. Never afraid to exercise her sharp mind, Katherine had become accustomed to discussing religion with Henry VIII. Though this was at first welcomed by the King, the conservative factions of court and church were terrified of any radical words whispered into the Royal ear - that after all was how Anne Boleyn had first led Henry towards Lutheranism. To conservatives like Bishop Gardiner and Chancellor Wriothesley the answer seemed obvious – Katherine should meet the same fate as Anne. At first, Henry, increasingly irascible from ulcerated legs, indicated that Katherine’s days were numbered. An arrest warrant was drawn up, and, amid rumours of ‘a new queen’, arrest could only have been followed by death. But Katherine succeeded in persuading Henry of her good faith and innocent naivety. “Is it even so, Sweetheart?”, said the King, “Then perfect friends we are now again…”[i] Thus did Tudor Royalty kiss and make up.
Katherine’s victory checked any conservative renaissance in the final years of the King’s reign. From now all eyes turned to the future (Protestant) reign of Edward VI. Here, Katherine appears to have been less successful, and for once followed her heart rather than her head. With ill-considered haste, she took Thomas Seymour as her lover within weeks of Henry’s death in 1547, and married him just months later. In doing so she lost any chance she may have had in exercising power during Edward’s minority. And yet, perhaps her final and most enduring success was yet to come, for in helping to restore the Princesses Mary and Elizabeth to the line of succession she had extended the Tudor dynasty by half a century. Katherine died after giving birth to a daughter in 1548.
The iconography of Katherine Parr is of particular interest. It is ironic that so few portraits of the Queen appear to survive, given that she was the foremost patron of portraiture in mid-Tudor England. There are several reasons why the Queen liked portraiture, not least because she evidently liked art. But perhaps the most intriguing reason may lie in Henry VIII’s habit (undoubtedly annoying to Catherine) of repeatedly portraying himself with Jane Seymour. Was Katherine’s jealousy manifested in art? Was her decision to commission the first full-length portraits of Elizabeth and Mary as Princesses, part of her desire to elevate them from illegitimate bastards to heirs of the English crown? Whatever the reasons, her legacy to the advancement of English portraiture cannot be doubted.
There are five recorded certainly known portraits of Katherine Parr that survive. The first is a miniature formerly in the collection of Horace Walpole (now at Sudeley Castle), which is probably by Lucas Hornebolt. The second and third, in the National Portrait Gallery, are a full-length (once erroneously called Lady Jane Grey) by Master John, and a half-length by an unknown artist. A fourth (Lambeth Palace) shows a young Katherine in the 1530s. And now the present example represents a fifth, and shows the Queen towards the end of her life.
And yet, Katherine’s own records show that she commissioned at least more than a dozen portraits of herself; “give me one of your small pictures”, her fourth husband Thomas Seymour wrote, “if ye have any left…”[ii] The contrast between Katherine’s commissions and those extant portraits gives a useful indication of how little survives from the sixteenth century – in this case less than a third. The Queen’s chamber accounts show that John Bettes the Elder painted up to seven miniatures – none survive – and nor apparently do any other miniatures by Hornebolt, aside from the possible Sudeley example.
Records also show that Katherine was painted by Hans Eworth, the Dutch artist considered the closest thing to Holbein’s heir[iii]. Such patronage was an indication of Katherine’s desire to support the new, for Eworth had only arrived in England c.1543. His earliest known work is dated 1549. The almost enamel-like flesh tones and bright colouring of the cheeks in this portrait, together with the distinctive modeling of the eyes, may suggest that the artist of this picture was influenced in some way by Eworth’s now lost original. The accomplished handling of the detail in Katherine’s out-turned collar, and the delicate portrayal of her hair, is also reminiscent of Eworth’s Mary Neville, Lady Dacre (National Gallery of Canada). That the jewelry Katherine wears in this portrait is similar to that recorded in her inventories, not to mention the intelligent depiction of Katherine’s slight physique, further suggests that it is based on a contemporary ad vivum example.
| 904
|
Eating and drinking
My partner is so restless now that he can’t seem to sit and finish a meal. But he’s happy to take a sandwich and eat it as he goes.
I used to get so upset when my sister ate with her fingers in company. I used to try to make her use her fork and for a while I didn’t invite anyone round. But then I thought about it and I decided that it really didn’t do anyone any harm. She wasn’t worried and my friends all said they didn’t mind – it was just me being embarrassed. So I decided not to worry about it any more.
You may have found that mealtimes are becoming more and more difficult. Perhaps the person with dementia makes a mess eating and you need to help them handle knives and forks. This might irritate you or make you feel embarrassed in front of guests. You may find that they eat more often because they forget having eaten or that they forget to eat altogether. They might start to eat everything or just one thing. They might also be perturbed by the difficulties they are experiencing. They might feel embarrassed or frustrated at not being able to eat correctly and suffer from other problems such as constipation, forgetting to chew food and loss of or disturbance of their sense of taste. Eating and drinking can clearly be problematic for you both. However, there are a number of ways you can help the person to maintain a healthy diet and enjoy meals, whilst at the same time contributing to maintaining a certain level of independence.
How to cope with eating and drinking
Provide assistance, whilst encouraging independence
So as not to make the person with dementia feel like a child or embarrassed, it is best to encourage them to do as much as possible themselves. For example, if they cannot handle a knife, instead of spoon-feeding, you could prepare bite-sized pieces of food or cut it up before serving it. By changing the way to eat or serve food, you can leave the person with the chance to remain independent longer. Sometimes, a simple reminder (e.g. to use a spoon instead of a fork) may be all that is needed. It might help to have your own meal at the same time. However, at some stage, you may find that spoon-feeding is necessary. If this occurs, you should be careful not to try when the person is drowsy, agitated or lying down, as otherwise they could choke. Certain medication can lead to problems swallowing as food becomes lodged in the person’s mouth. If they have difficulty swallowing, the following guidelines might be useful to bear in mind.
Make sure that the person sits upright, keeps their chin down and takes small sips of drink. Avoid vessels which force the head back too much (specially designed cups and mugs exist). Alternate food and drink to help clear the mouth. Encourage the person to try to swallow each mouthful of food twice (e.g. to swallow a second time even if the mouth is empty). If they have not swallowed, it can be useful to present an empty spoon. Encourage frequent swallows to counteract dribbling. Avoid mixtures of liquid and solid food (e.g. certain soups or cornflakes), stringy textures (e.g. cabbage, runner beans); floppy textures (e.g. lettuce, cucumber) and small, hard textures (e.g. peanuts and sweet corn).
Making mealtimes a pleasant experience
As a result of these difficulties, the person with dementia may start to dislike mealtimes, seeing them as an ordeal rather than a pleasant experience. However, giving the person enough time to try to manage alone or a little assistance may help them to feel more relaxed. Background music might also help contribute to a relaxed atmosphere. They might not manage to eat without being slightly messy or untidy, but in this case, it is perhaps best not to draw attention to it, as it could make them less inclined to try. Finally, they might want to be more involved, e.g. help to prepare food or clear the table. Mealtimes provide an ideal opportunity to give the person with dementia the chance to remain active and feel useful.
How to prevent problems with eating and drinking
You may be able to prevent some problems from occurring, by taking certain precautions or slightly adapting the way you normally prepare food and serve it.
- The routine : Give more time; serve plates of food instead of having self-service; remove condiments once used; have smaller meals more often if the person forgets having eaten; serve the main meal at midday to prevent night-time indigestion, etc.
- The utensils and layout : Try to find out about specially designed utensils, non-breakable dishes, non-slip mats and drinking containers that cannot be knocked over; use mugs if possible for certain food as the ability to hold a mug tends to be retained longer than the ability to use a knife and fork; provide a sharp contrast between plates and the table and avoid strong patterns on table cloths as this can cause confusion.
- The kind of food served : Prepare finger foods if the person has difficulty handling a knife and fork or is restless and cannot stay long at table; bite sized pieces avoid the need to use a knife; prepare food in a liquidiser rather than serving baby food if they cannot manage solid food so as to avoid embarrassment; avoid milk or citrus juice which tends to produce more mucus and can aggravate respiratory problems and drooling, if this is a problem.
Ensure a sufficient amount of liquid per day
It is important that you try to give the person with dementia enough to drink to avoid dehydration, constipation and more confusion. If possible, try to encourage them to drink 1½ litres (8 cups) of liquid per day. However, there are certain cases where even more attention needs to be taken, e.g. with people who have vomiting, diarrhoea or diabetes and people who are taking diuretics (water pills) or heart medication. On the more practical side, you will need to make sure that drinks are not too hot, as otherwise the person might scald themselves without realising. If they tend to become agitated, it is best to limit drinks which are stimulants (e.g. coffee, tea, hot chocolate, etc.). Possible alternatives include fruit and milk drinks, as well as decaffeinated drinks. Otherwise let them drink what they are used to.
Consult a doctor in case of loss of appetite
You should not worry too much if the person with dementia loses weight, as this can be a natural consequence of the disease. They might even be eating normally. However, it is nevertheless a good idea to contact your doctor just to check that there is not a physical cause and in case nutritional supplements should be given.
Last Updated: jeudi 06 août 2009
| 754
|
RWJF Priority: Use pricing strategies—both incentives and disincentives—to promote the purchase of healthier foods
Prices can significantly affect family food choices and are emerging as an important strategy in the movement to reverse the childhood obesity epidemic. For example, when healthy foods like fruits and vegetables are more affordable, children are less likely to gain excess weight. And leading health authorities, including the Institute of Medicine, recommend new policies to reduce overconsumption of sugar-sweetened beverages, which are one of the top sources of calories in the American diet.
The resources below, from RWJF grantees and partners, explore the possible health and economic impacts of using pricing strategies to promote consumption of healthy foods and beverages and discourage consumption of unhealthy products.
| 107
|
The island nation of Rhodes was an immensely successful commercial power of the Aegean. They strategically sided with Athens but, when in jeopardy, negotiated favorable surrenders to Sparta, Alexander the Great, and Demetrios Poliorketes, allowing them to sustain at least some degree of independence based on neutrality. Rhodes was named for the rose, depicted on the reverse of this coin. A portrait of Helios, the sun god, wearing a crown of rays graces the obverse. This image was based upon the head of the great statue of Helios, better known as the Colossus of Rhodes, one of the Seven Wonders of the Ancient World. The Colossus memorialized the war with Demetrios Poliorketes. He unsuccessfully attacked the city for over a year, around 305-304 B.C. Eventually, a settlement was negotiated and Poliorketes abandoned his siege towers constructed from valuable timbers that the Rhodeans in turn sold off for shipbuilding. The proceeds from this sale were invested in the huge (105 ft.) bronze statue of Helios standing at the entrance of the harbor. While the famous statue served as a reminder of the city's fortitude during the siege and became a tourist attraction, it only stood until 227 B.C. when it was toppled by an earthquake.
How many hands have touched a coin in your pocket or your purse? What eras and lands have the coin traversed on its journey into our possession? As we reach into our pockets to pull out some change, we rarely hesitate to think of who touched the coin before us, or where the coin will venture to after us. More than money, coins are a symbol of the state that struck them, of a specific time and place, whether currency in the age we live or an artifact of a long forgotten empire. This ancient coin is more than an artifact; it is a memorial to the glories of Rhodes passed from the hands of civilization to civilization, from generation to generation. - (C.7323)
| 880
|
05-11-2009 / Fluvirin – Fighting Against Influenza
04-30-2009 / Fluzone For Children
04-14-2009 / Flu Vaccines And Protein Chemistry
04-10-2009 / X Ray, Viral Structure And Flu Vaccines.
04-07-2009 / The History Of Influenza
04-06-2009 / Flu Vaccine Contamination And Quality Control
03-22-2009 / Egg Allergies And Flu Vaccines
03-12-2009 / Australian Discovery May Improve Existing Vaccines
03-06-2009 / How Humidity Affects Flu Outbreaks
03-05-2009 / Preservative-Free Flu Shots
03-04-2009 / Flu Strains Developing Resistance to Tamiflu
03-03-2009 / The History Of Influenza
02-23-2009 / Preservative-Free Flu Vaccines
Egg Allergies And Flu Vaccines
The majority of flu vaccines are made using processes that employ fertilized hen eggs. This process of manufacturing causes the vaccines to pick up small amounts of egg proteins. Even the nasal mist flu vaccine has as much or more egg protein as the injectable variety.
For most people, this is no problem. The majority of us pick up hundreds of times more egg protein from our diet than in a flu shot. Even excluding the obvious source, eggs, we get exposed to egg proteins from baked goods or foods processed with eggs on a regular basis. But there is a small, yet real section of the population that has an allergy to eggs. For those individuals, the flu shot can be a serious problem.
This is not a matter of one manufacturer against another: the fertilized hen's egg method is used throughout the vaccine industry. There are some different vaccine development methods on the horizon, however, that will use bacteria that have been genetically engineered to make flu vaccines. But until that day, the egg method and the egg protein in flu vaccines are here to stay.
The amount of egg protein varies from one vaccine manufacturer to another. If you are concerned about allergic reactions, you should talk with your doctor and request a vaccine with the least egg protein available. A skin test to determine whether or not you will have an allergic reaction to the vaccine is a good idea before getting a flu shot. If it is determined that you can have the vaccine, the best course of action is not to leave the medical facility for at least 30 minutes afterwards, so that the staff can observe you and respond promptly to any distress that you may have.
If your allergies are too severe for vaccination, there’s still a way to minimize the effects of the vaccine. You can have the vaccine administered in small doses over a period of time, instead of getting the full dose in one sitting. This, of course, should also take place under close medical supervision. In the cases of the most severe egg allergies, it will probably be safer to not have the vaccine at all and take antiviral medications if you are exposed to or catch the flu.
Sign Up For Your
| 59
|
Three-dimensional printing is being used to make metal parts for aircraft and space vehicles, as well as industrial uses. Now NASA is building engine parts with this technique for its next-generation heavy-lift rocket.
The agency says that its Space Launch System (SLS) will deliver new abilities for science and human exploration outside Earth's orbit by carrying the Orion Multi-Purpose Crew vehicle, plus cargo, equipment, and instruments for science experiments. It will also supply backup transportation to the International Space Station, and it will even go to Mars.
NASA is using 3D printing to build engine parts for its next-generation Space Launch System. Shown here is the first test piece produced on the M2 Cusing Machine at the Marshall Space Flight Center.
(Source: NASA Marshall Space Flight Center/Andy Hardin)
NASA's Marshall Space Flight Center is using a selective laser melting (SLM) process to produce intricate metal parts for the SLS rocket engines with powdered metals and the M2 Cusing machine, built by Concept Laser of Germany. NASA expects to save millions in manufacturing costs and reduce manufacturing time. SLM, a version of selective laser sintering, is known for its ability to create metal parts with complex geometries and precise mechanical properties.
The SLS will weigh 5.5 million pounds, stand 321 feet tall, and provide 8.4 million pounds of thrust at liftoff. Its propulsion system will include liquid hydrogen and liquid oxygen. Its mission will launch Orion without a crew in 2017; the second will launch Orion with up to four astronauts in 2021. NASA's goal is to use SLM to manufacture parts that will be used on the first mission.
The rocket's development and operations costs will be reduced using tooling and manufacturing technology from programs such as the space shuttle. For example, the J-2X engine, an advanced version of J-2 Saturn engines, will be used as the SLS upper stage engine. Some SLM-produced engine parts will be structurally tested this year and used in J-2X hot-fire tests.
In a NASA video, Andy Hardin, engine integration hardware lead for the Marshall Space Flight Center SLS engines office, discusses the initial testing and building stages:
We do a lot of engineering builds first to make sure we have the process [worked] out. There's always weld problems that you have to deal with, and there's going to be problems with this that we will have to work out, too. But this has the potential to eliminate a lot of those problems, and it will have the potential to reduce the cost by as much as half in some cases on a lot of parts.
Since final parts won't be welded, they are structurally stronger and more reliable, which also makes for a safer vehicle.
Ken Cooper, advanced manufacturing team lead at the Marshall Space Flight Center, says in the video that the technique is especially useful for making very complex shapes that can't be built in other ways, or for simplifying the building of complex shapes. But geometry is not the deciding factor; whether the machine can do it or not is decided by the size of the part.
| 666
|
Do you cringe at the word “xeriscape”? Does that mean boring, thin-leaved, un-colorful plants to you? Well think again. Xeriscape gardening can look lush, colorful and be a snap to maintain. To help people learn how to have a beautiful garden while being water consious, Colorado Springs Utilities has created two xeriscaped demonstration gardens for denizens to learn about and apply water thriftiness in their own gardens. The display at these gardens provides a lesson for everyone, even if you don’t live in a drought prone area. Can you imagine not having to water much *at all* between rains?
(Editor's Note: This article was originally published on September 1, 2008. Your comments are welcome, but please be aware that authors of previously published articles may not be able to promptly respond to new questions or comments.)
Even if you don't live in Colorado Springs, the Rocky Mountains, or even in a dry climate, the suggestions from these demonstration gardens will help you save time watering and money on your next water bill. Colorado and areas like it have wonderfully dry climates due to low humidity and unfortunately sometimes high winds.For most gardeners, this means both dry skin and dry soil.Besides a good moisturizer, here's how to cope.Plant water-wise plants and irrigate intelligently!
Once the plants on this list get established (one-two weeks) they don't need much water, if any, at all.With the growing water shortage in Colorado and many other states, this is an important feature for sustainability in the coming years.The Mesa Road Xeriscape Garden (one of CSU's featured demonstration gardens) is actually a lush, colorful and peaceful place that requires very little water.Come along for a tour and a xeriscaping lesson, Colorado style.
Why should you xeriscape?
The biggest and most widely misunderstood lesson of xeriscaping is planting in water-usage zones, or hydrozones. That is, putting plants that need lots of water next to ones that need lots of water, and putting plants that can don't need any water with the like. Seems logical right? If you put an iris next to a nasturtium, you are bound to do one of three things: overwater the iris, underwater the nasturtium, or kill both. It just makes sense to put your irises with your Blanketflowers and your Nasturtium next to your cannas. That doesn't sound all that boring, right?
When you really get down to it though, xeriscaping does mean planting flora that does not require much additional water than the average rainfall of your area. Once established, they should practically grow themselves.
CSU (the utilities company, not the university) has created two demonstration gardens to showcase how good xeriscaping can look: the Mesa Xeriscape Demonstration Garden and the Cottonwood Creek Park Xeriscape Garden in Colorado Springs. Along with some planting tips and the right plants, xeriscaping never has to be humdrum.
Bee Balm - Monarda
Blanketflower - Gaillardia
Texas Red Yucca surrounded by other colorful xerics
Lavender and Creeping Thyme
Plants you'll find at the Mesa Xeriscape Demonstration Garden (and you should try!)
Silver Blade Evening Primrose
Dwarf Garden Phlox
Rocky Mountain Sumac
Mullein 'Southern Charm'
Mexican Feather Grass
Dwarf Goldenrod 'Goldenbaby'
Blue Mist Spirea
Pale Purple Coneflower
Gray Creeping Germander
California Fuchsia 'Orange Carpet'
Some more hardy xeric plants to check out:
Red Hot Poker
Autumn Joy Sedum
The Mesa Garden also has a demonstration rock garden, which usually are made up of small xeric plants. Some great xeric rock garden plants are sedums, Penstemon, Dwarf Barberry, Edelweiss, Lamb's Ear, Oregano, Hens and Chicks, Rupturewort, and Cattail Iris. The list of attractive rock garden plants is just about endless.
For more information on fantastic, hardy xeric plants, here is a great link:
Xeric gardening does not have to be colorless or boring as you can see from all of these beautiful flowers showcased at the Mesa Xeriscape Demonstration Garden. Even if you just put in a few xerics, you can benefit from the decreased water usage.
All photos taken at Mesa Xeriscape Demonstration Garden, Colorado Springs, Colorado. Copyrighted to Susanne and Kyle Talbert
About Susanne Talbert
I garden in beautiful Colorado Springs, half a mile from Garden of the Gods. Since we bought our first house two years ago, I have been busy revamping my 1/4 acre of ignored decomposed granite.
My garden passions include water gardening, vines, super-hardy perennials, and native xerics. By day, I am a high school ceramics teacher as well as a ceramicist and painter.
| 263
|
While Americans and others may celebrate Easter only on a Sunday, that occasion is marked in Spain by an entire week of festivities known as Semana Santa, or Holy Week. The Holy Week festivals that take place all over the country are especially marked in Andalucia. Here, the emphasis is less on doleful repentance and more on celebration of the central figures of Christianity.
The central feature of all those activities are the notable processions that take place in every town in the region. The times, participants and style varies considerably from town to town. But almost all have some common themes.
Typically, traffic is closed off to allow for the floats and the hundreds or sometimes even thousands of people that precede or follow them. The festival is marked by religious figures dressed in Church finery carrying candles and directing the parade. Encircled by them is the main focus of the event: the floats.
Every float is a unique and creative construction that contains a figure of Jesus, Mary or an important saint.
A small statue of Saint Rocco, for example, may be held aloft on one platform, displaying his bare leg. He is regarded as a patron of the sick and the hopeful will often toss money onto the float, seeking relief or improvement.
Other floats will depict scenes from the Bible, early Christian stories, or any of a thousand different images that evoke memories of tales passed down through hundreds of generations. There is the Gitano del Polvorín, the Virgen de la Victoria and the Señor de Sevilla, among many more.
The processions, like the stories, are an ongoing tradition that has its origins in the early Middle Ages. As far back as 1,500 years ago the faithful annually walked with the platforms to celebrate the Annunciation, the Sermon on the Mount, the Rising from the Dead and other well-known scenes from the Bible.
For centuries – during the period Spain was ruled by Islamic Moors, Berbers, Arabs and others from North Africa – the festivals were forbidden. Not surprisingly, therefore, after the Reconquest by Christian kings and the re-establishment of Christianity in the country, the festivals started anew. They have been a regular event in Andalucia, with few interruptions, ever since.
But the natives are not the only participants. Church officials from Rome and elsewhere, along with people from around the world merely wanting to take part, celebrate the Holy Week festivals, too.
The celebrations ramp up a notch the final few days before Easter, making that an especially good time to visit in order to see or participate in Semana Santa. Musicians will play and sing as others carry banners, followed by Nazarenos dressed in tunics and masks.
Whether taking place in Cadiz or Cordoba, or any of the other dozens of cities large and small in Andalucia, the scene is similar. At the end of the procession, which often takes place from dusk to the following dawn, the float enters its individual sponsoring church and a hush comes over the crowd, signifying the culmination of Semana Santa.
| 744
|
This week’s illustration post focuses on perhaps the most popular illustrated genre of the Renaissance: the emblem book. Emblem books were a genre developed in the early 16th century as digestible and curious works which combined an image, a motto and an explanatory text. This genre remained popular on the continent well into the 17th century, although, strangely, it never found any real footing in the U.K. The illustrations were most commonly woodcuts and the mottoes Latin, however Greek and vernacular mottoes were common as well.
The intention of most emblem books was to deliver a moral lesson through text and image, but often the connection between these two elements is obscure. There are many projects that have developed over the last decade to provide a digital repository for these books worth exploring (Emblematica Online [Univeristy of Illinois and Herzog August Bibliothek], University of Glasgow and the Emblem Project Utrecht are good examples).
Although St Andrews 17th century holdings remain largely unexplored, I did come across a wonderful example of a Spanish emblem book published at the height of the genre’s popularity. This week’s post features the emblems from Sebastián de Covarrubias y Orozco’s Emblemas morales (1610). This work was published under the direction of Don Francisco Gómez de Sandoval, 1st Duke of Lerma, shortly after Covarrubias had recovered from a serious illness. This is perhaps Covarrubias’s most popular publication on emblems, but he is most well-known for his Tesoro de la lengua castellana o española (1611).
From the beginning of this book it is quite obvious that the author’s roots, both as a canon in the Catholic Church and as a keen lexicographer, are quite influential. However, Covarrubias does stray off into the weirder and esoteric world of the emblem books of the day: dragons, serpents, snake-eating deer and oversized-tops (above, which look like the ship from Flight of the Navigator to this blogger). Covarrubias also draws on everyday scenes of fishermen and farmers toiling in the field to root this emblem book in the real-world.
I’ve sampled here some of my favourites from this book, however a full scan is available at Emblematica Online. This work was not Covarrubias’s first foray into the world of emblem books, in 1589 he published a work, also entitled Emblemas morales, which provides almost 100 pages worth of text about emblems and their origins and then provides 100 examples afterword (none of which are repeated in his 1610 work). This work was republished in 1604 with the same text but with new woodcuts.
Covarrubias’s 1610 Emblemas morales, however, is a completely new work, featuring new emblems and mottoes with shorter verse explanations.
| 100
|
Words Are Powerful
Speech -- the ability to convey ideas and feelings through words -- is unique to human beings. It can be a tremendous blessing, but it's also ripe for abuse.
The words our children choose to use in expressing themselves help create their personal window on the world.
Positive forms of expression can help our children grow into positive, optimistic people who view the world around them with generous and hopeful eyes. Negative forms of expression, such as defamatory, mean-spirited speech, will cultivate in them a negative, cynical view of the world.
Speech That Is Evil
It's easy to fall prey to a destructive pattern of speaking badly about others and gossiping -- to the point where it becomes a recreational activity! In order to enjoy the many people in our lives, we have to stop verbalizing the negativity and focus on their positive virtues. This takes a lot of effort but is essential in raising happy children.
If we're always finding fault, we will naturally be dissatisfied, disappointed and displeased, and so will be our children.
Raising happy children requires us to impart to our offspring the ability to look at everything positively --situations, places and material objects. Most important of all is how they view people.
Torah calls "evil language" anything negative, even if it's true.
The Hebrew term for speaking badly of others is called lashon hara, literally "evil language." Interestingly, the Torah calls "evil language" anything negative, even if it's true. (Slander -- malicious, false information is called motzi shem ra, literally "giving another a bad name.")
In sharp contrast to the Western adage about sticks and stones not hurting, Judaism looks very gravely upon misuse of speech. Our tradition teaches that lashon hara can destroy many lives, even unintentionally, in one fell swoop:
- the person speaking,
- the person spoken about,
- and the person spoken to.
Let's look at why.
- The person speaking: Although you briefly become the center of attention when you dish out a juicy piece of gossip, in the long run people start mistrusting you. "Gee, I wonder what she says about me when I'm not around." People don't trust gossips and will avoid confiding in you. In the end, you're killing your own reputation. Furthermore, because you are misusing the gift of speech that God gave you, you are also lessened in His eyes.
- The person spoken about: The person under discussion is, of course, being killed in everyone's eyes. Whether the information is true or false, it is hard to take back defamatory words already spoken and undo the character assassination already committed. That person's reputation is forever blemished.
- The person spoken to: Interestingly enough, this is the person who is the most culpable, even though s/he is seemingly the innocent one. All s/he did was listen! But the Talmud says that listening to lashon hara is even worse than speaking it; the person had the power to stop it and didn't. Now the transgression is complete.
Exceptions to the Rule
Of course, there are times we are all owed to speak share negative information about others; in fact, there are times it is an obligation to do so. For example - when a friend is about to be become financially involved with a person we know to be unethical, or seriously dating a person we know to be abusive or otherwise unsuitable. Or when a child has information that will prevent harm from occurring.
Beware of the excuses children and adults often use for speaking lashon hara:
- "But it's true!" Lashon hara specifically refers to sharing derogatory information when it is true. Spreading vicious lies is far worse!
- "If she were here I would say it to her face." Maybe you would, and maybe you wouldn't. In any case, it is still forbidden.
- "Everyone knows about it." Does this justify you adding fuel to the fire? Even if it is on the front page of the newspaper, you are still forbidden to speak about it.
Teaching our children to avoid speaking lashon hara takes a concerted effort. Experiment with the following tools:
- Teach by example. Showing children that it's a priority for you is perhaps the most important lesson. Don't let them hear your gossiping with your friends or relatives. Don't let them hear you laughing at other people's expense. Even better than "don't let them hear you" is not doing it -- whether they're in earshot or not.
- Discuss the importance of avoiding lashon hara. Help your children identify what is and isn't proper speech. Talk about how improper speech can hurt others and how it hurts the person speaking lashon hara. There are a number of excellent Jewish books that can help you.
- Discourage "tattling." When your kids come to "tell on" someone, tell them you aren't interested in reports of someone else's bad behavior, but that you're available if they need help or advice.
- Get in the habit of not using names. There's no need for you to know the names of problem students at school unless you'll have a direct role in addressing the issue. Focus the discussion on your child's feelings, worries and concerns. If he or she needs protection that requires your intervention, tell him or her that it is proper to tell you the name of the offending child.
- Don't fall into the trap of casual lashon hara. At dinner and at other family times, bring books to the table to discuss or talk about current events. When you discuss what happened in each person's day, focus on what they learned that day and how they felt. Show your children that there are more interesting things to talk about than other people's poor behavior.
- Give positive reinforcement. Be sure to commend your kids when they manage to tell you about school or neighborhood problems without mentioning who was involved. Let them know that you're proud of them - and that God is too.
- Reminders! Tape a reminder to the telephone: "No Lashon Hara!" Put up signs on the fridge and in other prominent locations around the house.
- Study it. Read a small section of the laws of lashon hara each day during dinner or at your Shabbat table. Encourage discussion and examples.
| 443
|
Since 1993, RAN’s Protect-an-Acre program (PAA) has distributed more than one million dollars in grants to more than 150 frontline communities, Indigenous-led organizations, and allies, helping their efforts to secure protection for millions of acres of traditional territory in forests around the world.
Rainforest Action Network believes that Indigenous peoples are the best stewards of the world’s rainforests and that frontline communities organizing against the extraction and burning of dirty fossil fuels deserve the strongest support we can offer. RAN established the Protect-an-Acre program to protect the world’s forests and the rights of their inhabitants by providing financial aid to traditionally under-funded organizations and communities in forest regions.
Indigenous and frontline communities suffer disproportionate impacts to their health, livelihood and culture from extractive industry mega-projects and the effects of global climate change. That’s why Protect-an-Acre provides small grants to community-based organizations, Indigenous federations and small NGOs that are fighting to protect millions of acres of forest and keep millions of tons of CO2 in the ground.
Our grants support organizations and communities that are working to regain control of and sustainably manage their traditional territories through land title initiatives, community education, development of sustainable economic alternatives, and grassroots resistance to destructive industrial activities.
PAA is an alternative to “buy-an-acre” programs that seek to provide rainforest protection by buying tracts of land, but which often fail to address the needs or rights of local Indigenous peoples. Uninhabited forest areas often go unprotected, even if purchased through a buy-an-acre program. It is not uncommon for loggers, oil and gas companies, cattle ranchers, and miners to illegally extract resources from so-called “protected” areas.
Traditional forest communities are often the best stewards of the land because their way of life depends upon the health of their environment. A number of recent studies add to the growing body of evidence that Indigenous peoples are better protectors of their forests than governments or industry.
Based on the success of Protect-an-Acre, RAN launched The Climate Action Fund (CAF) in 2009 as a way to direct further resources and support to frontline communities and Indigenous peoples challenging the fossil fuel industry.
Additionally, RAN has been a Global Advisor to Global Greengrants Fund (GGF) since 1995, identifying recipients for small grants to mobilize resources for global environmental sustainability and social justice using the same priority and criteria as we use for PAA and CAF.
Through these three programs each year we support grassroots projects that result in at least:
| 613
|
Sign up for our newsletter
Gardening in the Rainy Zone.
Lilium 'White Stargazer'
Sunset zones: All.
USDA zones: 3-10.
Heat zones: 8-1.
Height: 30-35 inches (76-88 cm).
Star-shaped, white petals recurved at the tips.
Humus rich, slightly acidic, moist, well-drained soil.
In spring, when lilies are at the spear stage of growth (like asparagus), fertilize with a complete organic fertilizer. Do not feed again for the rest of the year; excessive fertilizing can promote disease and soft growth.
As soon as the foliage dies back, dig bulbs and remove scales, bulbils and offsets from bulb; immediately replant.
Remove dried stems only after foliage dies back.
Rainy Side Notes
Lilium 'White Stargazer' is the most popular of the white upright lilies and used by many for forcing. In fact, Peter Schenk bred it with forcing and the cut flower trade in mind. However, L. 'Siberia' and L. 'Jan de Graaf', two other upright whites, are not far behind in popularity. Like most Oriental lilies, white stargazer lily is deliciously fragrant.
Cut flowers when the first flower is first opening. The flower lasts from 5-9 days in the vase. Pull the pollen sacks off when the flower opens to keep pollen from staining linens beneath the vase. If pollen should get on clothes or linens, let it dry before wiping it off. If wiped when wet, the pollen will stain.
To grow lilies well, mulch with composted manures or compost and shredded bark to keep the roots cool while adding humus to the soil.
Article: Oriental Lilies for the Summer Garden:
Photographed in author's garden.
| 304
|
Sign up for our newsletter
Gardening in the Rainy Zone.
Pronounced: AL-lee-um ka-ra-tah-vee-EN-see
Sunset zones 1-24.
USDA zones: 5-9.
Heat zones: 9-5.
Height: 4-10 inches (10-25 cm).
Two to three-inch diameter spherical umbel of 50 or more, flowers on a 6-inch stem.
Just above ground level the thick, leathery, pleated, 6-inch long leaves. 'Ivory White' has a pale, pewter tone.
Full sun to partial shade.
Dry, well-drained, ordinary garden soil.
Remove offsets in autumn and plant.
Pests and Diseases:
Bulb rot can occur during our damp conditions of fall through spring. Onion fly and thrips may be a problem.
Rainy Side Notes
Allium karataviense is a spunky plant with a large flower globe. It is a dwarf ornamental onion compared to the giant species in the Melanocrommyum group of alliums, in which it belongs. It grows exceptionally well in our Mediterranean climate, as the plants go dormant by the time our annual drought comes around. Some gardeners grow the species and its cultivars because it has the most attractive foliage in the entire genus. Its horizontal, long, leathery, pleated foliage is green with a striking purple cast. As the blossom first opens, it is nestled on top of the attractive foliage; the flowering stems continue to grow. It pushes the large, spherical umbels up and away from the leaves as they begin to look shabby. When you plant bulbs en masse in a garden bed or container, use companion perennials that are late in filling out, or annuals to fill in any bare soil the bulb's dormant state leaves behind.
Sensitive to excess moisture, these alliums are prone to rot, so grow them in well-drained soil in the ground or in deep containers with excellent drainage. Some A. karataviense cultivars available are 'Ivory Queen', 'Lucky' and 'Red Globe'.
Photographed in author's garden.
| 341
|
Joint Base Pearl Harbor-Hickam (JBPHH)
Joint Base Pearl Harbor-Hickam on the left, and HNL on the right share the Reef Runway.
Honolulu International Airport and Joint Base Pearl Harbor-Hickam (JBPHH) have been neighbors and partners for many years. In 1956 the Air Force and the Hawaii Aeronautics Commission (predecessor to the State of Hawaii Department of Transportation) agreed to a single airport complex that would be operated under a Joint Use Agreement.
Today, JBPHH and HNL share runways, safety networks, and a partnership that is more than 50 years old.
Hickam Field, as it was then known, was completed and officially activated on September 15, 1938. It was the principal army airfield in Hawaii and the only one large enough to accommodate the B-17 bomber.
During World War II the base played a major role in pilot training and aircraft assembly work, in addition to service as a supply center for both air and ground troops. Hickam served as the hub of the Pacific aerial network, supporting transient aircraft ferrying troops and supplies to, and evacuating wounded from, the forward areas, not only during World War II but also during the Korean conflict and the Vietnam War.
During the Korean Conflict, 19 percent of HNL operations were military. During the Vietnam War that increased to 20-50%.
Today, military operations are five to six percent of total operations.
| 872
|
John Langley Howard was a revolutionary regionalist painter known for depicting labor and industry in California as well as his reverence for the natural world. Howard took a strong stance on social and environmental issues and used his art to communicate his strong emotional response toward each of his subjects.
Table of Contents
John Langley Howard was born in 1902 into a respected family of artists and architects. His father, John Galen Howard relocated the family to California in 1904 to become campus architect of the University of California, Berkeley. It was only after attending the very same campus his father helped to create, that Howard suddenly decided he wanted to pursue a career as an artist and not an engineer as previously planned. Following this decision, Howard enrolled in the California Guild of Arts and Crafts in Oakland and then transferred to the Arts Students’ League in New York City.
At the school, he met Kenneth Hayes Miller who supported Howard’s attitude because the “taught the bare rudiments of painting and composition, and stressed the cultivation of the ultra-sensitive, intuitive approach” (Hailey 56). After saving his money, Howard travelled to Paris for six months to seek out his own artistic philosophy. However, it quickly became apparent to Howard that he placed more value on pure talent than professional training. In 1924, Howard left art school to pursue his career and marry his first wife, Adeline Day. He had his first one-person exhibition at the Modern Gallery in San Francisco in 1927. Shortly after, he attempted portraiture.
Following the start of the Depression, Howard found himself appalled by the social conditions and began to follow “his own brand of Marxism.” Howard and his wife began to attend meetings of the Monterey John Reed Club, discussing politics and social concerns. Soon, the artist became determined to communicate society’s needs for the betterment of the future. His landscapes began to include industry and its effects to the surrounding region. In 1934, Howard was hired through the New Deal Public Works Art Project to create a mural for the inside of Coit Tower on Telegraph Hill in San Francisco depicting California industry. The project called for twenty-seven artists to be hired to paint frescos inside the newly erected monument funded by philanthropist Lillie Hitchcock Coit. Each artist was to depict a scene central to California living, including industry, agriculture, law, and street scenes of San Francisco.
Howard’s completed fresco drew notorious attention for showing an unemployed worker reading Marxist materials, a gathered group of unemployed workers, and a man panning for gold while watching a wealthy couple outside of their limousine. In a nearby mural by Bernard Zakheim (1896-1985), Howard himself was used as a model. He is shown crumpling a newspaper and grabbing a Marxist book from a library shelf. This soon led to the artists being linked to a local group of striking dock workers. They were accused of attempting to lead a Communist revolution. Howard’s murals as well as the work of Clifford Wight (1900-1960) and Zakheim became highly scrutinized, and the uproar over the works led to a delay in opening Coit Tower. In order to protect their work from being defaced or completely destroyed, the muralists chose to sleep outside the tower. The SF Art Commission ultimately cancelled the opening of Coit Tower as a result of the controversy and did not open it until months later.
During this time, Howard relocated his family to Santa Fe, New Mexico citing his son’s health concerns for almost two years before returning to Monterey in 1940. Following the onset of World War II, he had a renewed interest in landscape and soon ceased to include social commentary within his work, thus removing the human figure from his paintings. The artist divorced his first wife in 1949. In 1951, Howard’s art took another turn when the artist painted The Rape of the Earth which rallied against the destruction of nature by technology, making Howard one of the first “eco-artists.” During the same year he also married sculptor Blanche Phillips (1908-1976). He began illustrating for Scientific American Magazine and used this medium to refine his technique.
Howard’s landscapes began turning to “magic realism” or “poetic realism” as Howard preferred to call it. This method is described as the use of naturalistic images and forms “to suggest relationships that cannot always be directly described in words” (Aldrich 184). His aim was to communicate a poetic and spiritual connection with the landscape depicted. Overall, Howard lived in more than 20 different locations during his career.
In 1997, Howard attended the dedication of Pioneer Park at Coit Tower and was the only surviving member of the twenty-seven muralists included in the original project. The murals were restored by the City of San Francisco in 1990 after water damage and age dictated the need for restoration. Howard died at the age of 97 in his sleep at his Potrero Hill home in 1999.
II. AN ANALYSIS OF THE ARTIST'S WORK
“I think of painting as poetry and I think of myself as a representational poet. I want to describe my subject minutely, but I also way to describe my emotional response to it…what I’m doing is making a self-portrait in a peculiar kind of way.” – John Langley Howard
John Langley Howard was widely considered a wanderer and a free spirit. While Howard did receive academic training from the California Guild of Arts and Crafts in Oakland and the Arts Students’ League in New York City, he chose to align himself with instructors whose opinions of art education matched his predetermined beliefs. These teachers included Kenneth Hayes Miller (1876-1952) who valued an analytical, bare bones approach to art instruction and supported greater personal development of intuitive talent. Howard expressed this viewpoint stating that:
“I want everything to be meaningful in a descriptive way. I want expression and at the same time I want to control it down to a gnat’s eyebrow. I identify with my subject. I empathize with my subject” (Moss 62).
In the 1920s, Howard became known as a Cezanne-influenced landscape artist and portraitist. Tempera, oil, and etching became his primary media while his subject matter turned to poetic and often spiritually infused imagery which would resurface later in his career. Earth tones and very small brushstrokes were utilized, allowing Howard to refine his images.
Howard exhibited frequently with his brothers Charles Howard (1899-1978) and Robert Howard (1896-1983). Critic Jehanne Bietry wrote of their joint Galerie Beaux Arts show that: “of (the Howard brothers), John Langley is the poet, the mystic and the most complex…there predominates in his work a certain quality, an element of sentiment that escapes definition but is the unmistakable trait by which one recognizes deeper art” (Hailey 60). It is significant that a critic would accurately take note of Howard artistic aims at such an early stage because what Bietry describes ultimately became the primary focus of Howard’s career.
Howard experienced a dramatic change in medium when he was commissioned to paint a mural for the Coit Tower WPA project in 1934. The project was Howard’s first and only mural and provided the artist with an outlet for his newly discovered Marxist social beliefs. While Howard supported a political agenda rather explicitly in his image, his focus on deeper subject matter permeates throughout the work. Most important to Howard is “the idea of human conflict that [he] pictorializes and deplores – man’s tragic flaw manifest again in this particular situation” (Nash 79). Howard’s work had progressed steadily into the realm of social realism until the backlash against the Coit Tower murals led him in a new direction.
Howard abandoned explicit statements of social commentary and returned to his roots as a landscape painter. However, this did not prevent the artist from illustrating important issues because he then became one of the first “eco-artists.” Through his painting, Howard investigated the role of technology on the environment and used the San Francisco Bay Area as well as Monterey to demonstrate his point of view. He continued following his original artistic tendencies by delving into “magic realism” or “poetic realism” which utilized the spiritual connection that Howard sought to find within his work. Art critic Henrietta Shore recognized the balance that Howard achieved within his work, stating that he “is modern in that he is progressive, yet his work proves that he does not discard the traditions from which all fine art has grown” (Hailey 65). Overall, Howard’s career presents a unique portrait of individual expression and spiritual exploration.
1902 Born in Montclair, New Jersey
1920 Enrolls as an Engineering major at UC Berkeley
1922 Realizes he wants to be an artist
1923-24 Attends Art Students’ League in New York
1924 Leaves art school
1924 Marries first wife, Adeline Day
1927 First one-person exhibition held at The Modern Gallery, San Francisco
1928 First child, Samuel born
1930 Daughter Anne born
1934 Commissioned to Paint Coit Tower mural, San Francisco
1940 Studies ship drafting and worked as a ship drafter during World War II
1942 Serves as air raid warden in Mill Valley, CA
1949 Divorces his first wife
1950 Teaches at California School of Fine Arts, San Francisco
1951 Marries second wife, sculptor Blanche Phillips
1951 Moves to Mexico
1951 Paints The Rape of the Earth communicating his eco-friendly stance
1953-1965 Illustrates for Scientific American magazine
1958 Teaches at Pratt Institute Art School, Brooklyn, NY
1965 Moves to Hydra, Greece
1967 Moves to London
1970 Returns to California
1979 Blanche Phillips dies
1980 Marries Mary McMahon Williams
1999 Died in his sleep at home San Francisco, California
California Palace of the Legion of Honor, CA
City of San Francisco, CA
IBM Building, New York, NY
The Oakland Museum, CA
The Phillips Collection, Washington D.C.
San Francisco Museum of Modern Art, CA
Security Pacific National Bank Headquarters, Los Angeles, CA
Springfield Museum of Fine Arts
University of Utah, UT
1927 Modern Gallery, San Francisco, CA
1928 Beaux Arts Gallery, San Francisco, CA
1928 East-West Gallery, San Francisco, CA
1928-51 San Franciso Art Association, CA
1935 Paul Elder Gallery, San Francisco, CA
1936 Cincinnati Art Museum, OH
1936 Museum of Modern Art, San Francisco, CA
1939 Golden Gate International Exposition, Department of Fine Arts, Treasure Island, CA
1939 Museum of Modern Art, San Francisco, CA
1941 Carnegie Institute, Pittsburgh, PA
1943 Corcoran Gallery, Washington D.C.
1943 M. H. de Young Memorial Museum, San Francisco, CA
1946-47 Whitney Museum, NY
1947 Rotunda Gallery, City of Paris, San Francisco, CA
1952 Carnegie Institute, Pittsburgh, PA
1956 Santa Barbara Museum of Art, CA
1973 Capricorn Asunder Gallery, San Francisco, CA
1974 Lawson Galleries, San Francisco, CA
1976 de Saisset Art Gallery and Museum, CA
1982 San Francisco Museum of Modern Art Rental Gallery, San Francisco, CA
1983 California Academy of Sciences, CA
1983 Monterey Museum of Art, CA
1986 Charles Campbell Gallery, San Francisco, CA
1987 Martina Hamilton Gallery, NY
1988 Oakland Museum, CA
1989 Tobey C. Moss Gallery, CA
1991 M. H. de Young Memorial Museum, San Francisco, CA
1992 Tobey C. Moss Gallery, CA
1993 Tobey C. Moss Gallery, CA
California Society of Mural Painters’ and Writers’ and Artists’ Union
Carmel Art Association
Club Beaux Arts
San Francisco Art Association
Society of Mural Painters
Marin Society of Artists
Monterey John Reed Club
Anne Bremer Memorial Award for Painting, San Francisco Art Association
First Prize, Pepsi-Cola Annual “Portrait of America”
First Prize, San Francisco Art Association
Award, City of San Francisco Art Festival
Citation for Merit, Society of Illustrators, New York
- 1. Aldrich, Linda. “John Langley Howard.” American Scene Painting: California, 1930s and 1940s. Irvine, Westphal Publishing: 1991.
- 2. Hailey, Gene. “John Langley Howard…Biography and Works.” California Art Research Monographs, v. 17, p.54-92. San Francisco: Works Progress Administration: 1936-1937.
- 3. Moss, Stacey. The Howards, First Family of Bay Area Modernism. Oakland Museum: 1988.
- 4. Nash, Steven A. Facing Eden: 100 Years of Landscape Art in Bay Area. University of California Press: 1995.
IX. WORKS FOR SALE BY THIS ARTIST
| 149
|
Protect your investment
Protected cropping can be as beneficial for small fruit production as it is for vegetables. Yields of some berries may be two to three times greater in protected cultivation than outside in the field. Build customer loyalty by having fresh fruit first in the spring, long into fall, and by growing the highest quality berries.
Strawberry plants, which are damaged by temperatures below 12F/-11C, require winter mulch. Although hay and straw are the traditional mulching materials, many strawberry growers use row covers because they require less labor to install and remove. Heavier row covers, weighing 1.25 oz./sq.yd. or more, are recommended.
Fall-bearing raspberries and blackberries that normally stop producing at the first frost will continue fruiting for months longer in an unheated hoophouse. They will fruit again earlier in spring than those in the field, commanding a much higher sales price.
Strawberry transplants from runners produced over the summer can be planted in an unheated hoophouse in September. They will produce fruit in the fall, continuing until December, and then fruit again in early spring.
Strawberry plasticulture is supplanting the traditional matted row system on many farms. Plant on black plastic mulch from mid-July to September and cover with row cover in fall. Fruits are harvested the following spring.
Protect fruits from marauding birds with Johnny's bird netting which won't damage fruit or bend branches.
Getting started with fruit
By Lynn Byczynski
For many market growers, fruits are the final frontier of horticultural expertise. Growing fruit is an interesting challenge for a vegetable grower because fruits require different systems for planting, cultivating, harvesting and post-harvest handling. But there are many reasons to take up the challenge.
- The primary reason is that people love fruit. Farmers market customers flock to vendors with berries and grapes for sale. CSA members develop stronger ties to farms that can supply a wide range of fresh produce. Chefs who tout their connections to local farmers are delighted to be able to list local fruit on their dessert menus. At home, even the pickiest eaters are usually happy to snack on berries and grapes.
- Consumption of berries and grapes is rapidly increasing worldwide, thanks to recent discoveries about the health benefits of these fruits. The pigments that give berries and red or purple grapes their deep colors contain phytochemicals that help prevent cancer, cardiovascular disease, and age-related mental decline. People feel good about eating grapes and berries!
- From the farmer's and gardener's perspective, berries and grapes are easier to grow than ever before. New varieties, production practices, and products are increasing the options for growers in every region. Berries are popular crops for the hoophouse, for example, because protection from wind and rain produces extraordinary yields of high-quality fruits. Plastic and paper mulches reduce the need for year-round weeding of these perennial plants. And because Johnny's offers plants in small quantities, growers can trial numerous commercial varieties without spending a lot of money.
- Although most berry and grape plants won't produce fruit for 1 to 3 years after planting, the wait is worthwhile. Commercial growers can charge a premium for fresh, ripe fruits. And home gardeners can save money by growing their own.
- What do you need to get started with fruits? First, if you aren't sure about the suitability of your climate for small fruits, contact your state Extension service for recommendations. Some regions of the country may not have enough cold (chilling hours) for certain varieties, while others may be too cold for the plants or too hot for the fruits. Good soil preparation is essential for successful fruit production. So is an irrigation system. Most small fruits don't compete well with weeds, so a mulch of hay, straw, or wood chips is beneficial. Grapes need a strong trellis, which should be erected when the vines are planted. A living mulch in the paths between rows will help reduce weed pressure and improve soil fertility. You’ll find products and information about living mulches in the cover crops section on the web and in the catalog.
By Lynn Byczynski
Growing grapes may appear complicated to the beginner, and with good reason. Although grapes will grow anywhere, there are many kinds of training and trellising systems, and choosing the right one requires some study before planting.
Training and trellising go hand-in-hand because the kind of structure you build to hold your grape vines will affect how you prune them. The structure, in turn, depends somewhat on the type of grapes you grow because some are more vigorous and need stronger supports.
In general, a grape trellis needs to be able to support the weight of the crop and withstand high winds. It also should be designed to last 20 years, as that's how long you can expect your vines to produce.
Home gardeners planting just a few vines can use a fence that fits into the landscape or, better still, an arbor that provides shade in summer as well as support for the grape vines. To get good fruit production from an arbor planting, pruning becomes the key. Texas Extension has a nicely illustrated manual on arbor training.
Commercial growers with larger aspirations need to set up a trellis in the field. The main ingredients for a vineyard trellis are strong end posts with braces, earth anchors, or deadmen; posts along the length of the trellis to support the wires; and high-tensile galvanized steel wire to support the vines.
The most common type of trellis is the single curtain trellis with either one or two wires and posts every 16 to 24 feet apart, depending on the training system. With this type of trellis, various training styles are possible. Another popular type of trellis, especially in northern areas, is the double curtain, which allows the vines to spread horizontally across two wires.
The recommended trellis and training system varies by climate. Northern growers with shorter growing seasons usually choose training systems that expose more leaf surface to the sun, but those can be inappropriate to warm climates. To learn more about the best training and trellising system for your location, check the list below of state viticulture guides and choose the state nearest your own. Or, contact your state Extension service for recommendations.
California: Viticulture and Enology Home Page
Colorado: Grape Growers Guide
Idaho, Oregon, Washington: Northwest Berry & Grape Information Network
Iowa: Viticulture Home Page
Kansas: Commercial Grape Production
Michigan: MSU Grape Information
Missouri: Home Fruit Production: Grape Training Systems
New York: Cornell Viticulture
Ohio: Midwest Grape Production Guide
Oklahoma: Viticulture and Enology
Pennsylvania: Wine Grape Network
South Dakota: Viticulture in South Dakota
Texas: Winegrape Network
Vermont: Cold Climate Grape Production
Wisconsin: Growing Grapes
By Lynn Byczynski
Strawberries are one of the most popular fruits in American gardens and market farms. They can be grown in many places, from hanging baskets to fields to hoophouses. The trick is to match the growing system to the type of strawberry you want to grow. Some varieties need plenty of space, whereas others can be grown in containers.
June-bearing varieties initiate fruit buds in fall and blossom the following spring. They are the earliest type to fruit. They produce one crop and then spend their energy sending out runners (also called daughter plants) that will fruit the following year. June-bearing strawberries are usually grown in a matted row system, in which the mother plants are planted in spring, spaced 18-24" apart in rows that are 3-4' apart. The first year, flowers are pinched off to stimulate the plants to send out runners that fill in the spaces within the row and between the rows. Plants produce fruit the second spring. A variation of this system is to prune runners to one or two per plant so that they stay in a line and don't spread out between the rows. This obviously requires a lot more labor, but may result in better yields because of reduced competition. Matted-row systems can be renovated to keep plants producing for many years. Another system is called the ribbon row system, in which strawberry crowns are planted in fall and allowed to bloom and fruit the following spring. As runners form, they are removed to increase fruit size. Once the crop is done, runners are allowed to develop and fill in the bed to a matted row system.
Day-neutral varieties produce fruit all summer. They can be grown as annuals: plant early in spring and pinch off flowers for two months to let the plants get established, and then let them fruit the rest of the summer. Day-neutral strawberries are good for container production on a deck or patio. Some varieties, including 'Seascape', will fruit on unrooted runners so they make attractive hanging baskets, with the runner plants cascading over the sides of the basket. Day-neutral strawberries can also be grown in a hill system, with 12 inches between plants.
Alpine strawberries produce small but intensely flavorful berries. They do not send out runners and are usually grown from seed. They are a good choice for strawberry pots and other containers, or as edging in the vegetable garden. They also can be grown with less than full sun, so they are a good choice for many home gardeners.
Region-specific growing information is available from most state Extension services. ATTRA has a publication on Organic Production of Strawberries.
By Lynn Byczynski
Strawberry quality, yield, and earliness is greatly improved in a hoophouse. Penn State researchers found that in their climate, hoophouse strawberries produced fruit 3 weeks earlier in spring than those grown outside, with about a 25% yield increase.
Most commercial hoophouse strawberries are grown using an annual plasticulture system that includes raised beds, drip irrigation, plastic mulch, and floating row cover. Plugs are planted in late summer on beds covered with plastic mulch, with drip tape beneath the mulch. As the weather gets cold, the young plants are covered with floating row cover to maintain the warmer soil temperatures needed for establishment. The plants grow slowly during winter in the protected environment of the hoophouse; then, as the weather warms, they flower and produce berries for several weeks. The crop is then finished for the year. Strawberry plants can either be removed to make way for other crops; or they can be left to produce a second year if berry prices or other factors justify tying up the space for a year.
Plugs are available from outside suppliers, or they can be produced on the farm in summer. To grow your own, detach unrooted daughter plants (runners) from the mother plant in July and stick them in potting mix in 72-cell flats under intermittent mist until roots protrude from the bottom of the cell. Then place on a greenhouse bench and grow until September, when they can be planted into the hoophouse. Plants that are rooted in July are likely to flower and fruit in fall in warmer climates, but that won't affect their yield the following spring.
For more information on hoophouse strawberries:
Growing Strawberries in High Tunnels in Missouri
Production of Vegetables, Strawberries, and Cut Flowers Using Plasticulture is a book about all aspects of horticultural plastics, and includes extensive information about hoophouse strawberries.
| 124
|
FDA OKs Nexium for Kids 1-11
FDA Approves Nexium for Short-Term Treatment of Gastroesophageal Reflux Disease (GERD) in Young Children
Feb. 28, 2008 -- The FDA today approved short-term use of the drug Nexium for children aged 1-11 with gastroesophageal
reflux disease (GERD).
The FDA OK'd Nexium in two forms -- a delayed-release capsule and a liquid
form -- in doses of 10 milligrams or 20 milligrams for kids aged 1-11.
Nexium was already approved for children aged 12-17 in 20 mg or 40 mg
"This approval provides important information for appropriate dosing for
children ages 1-11 years with GERD," Julie Beitz, MD, director of the FDA's
Office for Drug Evaluation III in the Center for Drug Evaluation and Research.
"Children prescribed this drug should be monitored by their physicians for
any adverse drug reactions."
Nexium is part of a class of drugs known as proton pump inhibitors (PPIs).
PPIs decrease the amount of acid produced in the stomach and help heal erosions
in the lining of the esophagus, a condition known as erosive
The FDA approved Nexium's use in children aged 1-11 for short-term treatment
of GERD based data extrapolated from previous studies done in adults, as well
as studies done in children.
In one study, 109 GERD patients aged 1-11 were treated with Nexium once
daily for up to eight weeks to evaluate Nexium's safety and tolerability. Most
of the patients showed healing of their esophageal erosions after eight weeks
The most common adverse reactions in children treated with Nexium were headache, diarrhea, abdominal pain, nausea, gas, constipation, dry mouth, and
Nexium isn't approved for children younger than 1. The drug's safety and
efficacy hasn't been established in children in that age range, the FDA
Nexium is made by the drug company AstraZeneca.
| 711
|
From Abracadabra to Zombies | View All
N'kisi & the N'kisi Project
N'kisi (pronounced ‘‘in-key-see’’) is a captive bred eight or nine-year-old hand raised African Grey Parrot whose owner, Aimée Morgana, thinks uses language. She doesn't think he just sounds out words. She thinks he communicates with her in language, which would in effect make N'kisi a rational parrot. For example, N'kisi utters "pretty smell medicine" when he wants to describe the aromatherapy oils that Aimée uses.* Furthermore, Aimée says her parrot has a fine sense of humor and knows how to laugh. Imagine having conversations with a humorous parrot. Think of all the things you could talk and joke about, besides aromatherapy. You could discuss the fame that would come to anyone who had a parrot that can think and converse in intelligent discourse, like pretty smell medicine and look at my pretty naked body.* And when some nasty skeptic makes fun of you, the two of you can joke about it.
I'm afraid that this story stretches the boundaries of reasonable credibility, though stories of rational parrots go back at least to the 17th century. John Locke, for example, relates a tale of a Portuguese-speaking parrot of some note in his Essay Concerning Human Understanding (II.xxvii.8). These cases are more likely cases of self-deception, delusion, and gullibility than of language-using parrots. Listen to this audio clip of N'kisi, Aimée, and a toy that "talks" when a button is pushed. First listen without reading the transcript. Some of it is intelligible, especially after the fourth or fifth repetition, but it is difficult to understand the "conversation," especially with the toy making its sounds as Aimée stimulates her parrot. Some of the tape sounds like gibberish until you are told what to listen for. When you listen while reading the transcript something amazing happens: you can hear just what you're reading. Why is that? The same thing happens when you listen to audio tapes played backward. When you just listen without anyone telling you what to listen for, you usually don't understand anything intelligible. But as soon as someone shows or tells you what to listen for, you can hear the message. Such is the power of suggestion and the way of audio perception. Hearing is a constructive process, like vision, in that bits of sensory data are "filled in" by the brain to produce a visual or auditory perception that is clear and distinct, and in accord with your expectations. Consider the following from an interview with Dr. Irene Pepperberg, Morgana's inspiration, who has been studying Alex, an African Grey Parrot, for many years:
We were doing demos at the Media Lab [at MIT] for our corporate sponsors; we had a very small amount of time scheduled and the visitors wanted to see Alex work. So we put a number of differently colored letters on the tray that we use, put the tray in front of Alex, and asked, "Alex, what sound is blue?" He answers, "Ssss." It was an "s", so we say "Good birdie" and he replies, "Want a nut."
Well, I don't want him sitting there using our limited amount of time to eat a nut, so I tell him to wait, and I ask, "What sound is green?" Alex answers, "Ssshh." He's right, it's "sh," and we go through the routine again: "Good parrot." "Want a nut." "Alex, wait. What sound is orange?" "ch." "Good bird!" "Want a nut." We're going on and on and Alex is clearly getting more and more frustrated. He finally gets very slitty-eyed and he looks at me and states, "Want a nut. Nnn, uh, tuh."
Not only could you imagine him thinking, "Hey, stupid, do I have to spell it for you?" but the point was that he had leaped over where we were and had begun sounding out the letters of the words for us. This was in a sense his way of saying to us, "I know where you're headed! Let's get on with it," which gave us the feeling that we were on the right track with what we were doing.*
Dr. Pepperberg thinks the bird is responding cognitively to her questions rather than simply responding to a stimulus. She thinks the bird is getting frustrated, but she has stipulated earlier in the interview:
I never claim that Alex has full-blown language; I never would. I'm not going to be able to put Alex on a "T" stand and have you interview him the way you interview me.
So, whereas you or I might say "give me the nut or this interview is over" were we parrots with intentionality and language, the parrot's movements and sounds have to be less direct and more complex, so that they have to be interpreted for us by Pepperberg. In her view, Alex is "clearly getting more frustrated" and his frustration culminates with a "very slitty-eyed" expression. But this is Pepperberg's interpretation, as is her hearing the bird sound out the letters of the word 'nut'. It could have been a stutter for all we know, but Pepperberg is facilitating Alex's communication by telling us what she hears. The final paragraph indicates that Pepperberg is having a hard time drawing the line between imagining what a parrot might be thinking and projecting those thoughts into the parrot's movements and sounds. She's also having a hard time getting grant money (NIH turned her down), so she started her own private foundation, the Alex Foundation.
When news of N'kisi broke on the pages of BBC online, there was no mention in the article by Alex Kirby of the parrot having conversations with people other than Aimée Morgana. (The story was originally told in USA Today in the February 12, 2001, edition.) Despite the headline "Parrot's oratory stuns scientists," there was no evidence given that the parrot had stunned anyone during a conversation. It seems that Aimée is to her parrot what the facilitator is to her client in facilitated communication, except that the parrot is actually providing data to interpret and is more like clever Hans, the horse that responded to unconscious movements of his master, than a disabled human who may not be providing any content or direction at all to the facilitator. It is Aimée who gives intentionality to the parrot's sounds. She is the one who attributes 'laughter' to his shrieks and conscious awareness to his responses, though those responses could be due to any one of many stimuli, consciously or unconsciously provided by Aimée or items in the immediate environment. Nevertheless, Dr. Jane Goodall, who studies chimpanzees, met N'kisi and said that he provides an "outstanding example of interspecies communication." There is some evidence, however, that much of the work with language-using primates also mistakes subjective validation by scientists for complex linguistic abilities of their animal subjects (Wallman 1992).
According to Mr. Kirby, N'kisi not only uses language but has been tested for telepathy and he passed the test with flying colors:
In an experiment, the bird and his owner were put in separate rooms and filmed as the artist opened random envelopes containing picture cards.
Analysis showed the parrot had used appropriate keywords three times more often than would be likely by chance.
Kirby doesn't provide any details about the experiment, so a reader might misinterpret this claim as implying that this parrot did about twice as well as people did in the ganzfeld telepathy experiments. In those experiments, subjects in separate rooms were monitored as one tried to telepathically send information from a picture or video to the other. Typically, there was a 20% chance of guessing what the item was but results as high as 38% were reported in some meta-analyses. If the parrot scored three times better than chance, then he would have gotten 60% correct. The odds of a parrot randomly blurting out words that match up 60% of the time with pictures being looked at simultaneously in another room are so high that there is virtually no way that this could happen by chance. However, as you might suspect, Kirby's claim is a bit misleading.
I assume that Kirby was writing about an experiment that was part of the N'kisi project, a joint effort by Morgana and Rupert Sheldrake to test not only the parrot's language-using abilities but his telepathic talents as well. Sheldrake has already validated the telepathic abilities of a dog and thinks the "findings [of this experiment] are consistent with the hypothesis that N'kisi was reacting telepathically to Aimée's mental activity."*
The full text of Sheldrake's study published in the peer reviewed Journal of Scientific Exploration is available online. The title of the paper would send most journal editors to their grave, killed by laughter: "Testing a Language-Using Parrot for Telepathy." Fortunately for Sheldrake and his associates there will always be a sympathetic editor for another story like that of J. B. Rhine and the telepathic horse, "Lady Wonder." At least Sheldrake's protocols show some measure of sophistication, unlike Rhine's. Even so, as the editor at the Journal of Scientific Exploration commented: "once again, we have suggestive results, a level of statistical significance that is less than compelling, and the devout wish that further work with refined protocols will ensue."* So, we'll just have to wait and see whether further study of N'kisi supports the telepathic hypothesis.
Anyway, here is how Sheldrake set up the experiment. He first compiled a list of 30 words from the bird's vocabulary that "could be represented by visual images." A package of 167 photos from a stock supplier was used for the test. Since only 20 of the photos corresponded to words on the list, the word list was reduced to 20. The word 'camera' was removed from the list because 'N’kisi "used it so frequently to comment on the cameras used in the tests themselves." Thus, they were left with 19 words.
During the tests, N’kisi remained in his cage in Aimée’s apartment in Manhattan, New York. There was no one in the room with him. Meanwhile, Aimée went to a separate enclosed room on a different floor. N’kisi could not see or hear her, and in any case, Aimée said nothing, as confirmed by the audio track recorded on the camera that filmed her continuously. The distance between Aimée and N’kisi was about 55 feet. Aimée could hear N’kisi through a wireless baby monitor, which she used to gain ‘‘feedback’’ to help her to adjust her mental state as image sender.
Both Aimée and N’kisi were filmed continuously throughout the test sessions by two synchronized cameras on time-coded videotape. The cameras were mounted on tripods and ran continuously without interruption throughout each session. N’kisi was also recorded continuously on a separate audio tape recorder. (Sheldrake and Morgana 2003)
According to Sheldrake:
We conducted a total of 147 two-minute trials. The recordings of N’kisi during these trials were transcribed blind by three independent transcribers....He scored 23 hits: the key words he said corresponded to the target pictures....If N’kisi said a key word that did not correspond to the photograph, that was counted as a miss, and if he said a key word corresponding to the photograph, that was a hit. (Sheldrake and Morgana 2003)
However, sixty of the trials were discarded because in those trials N'kisi either was silent or uttered things that were not key words, i.e., showed no signs of telepathy. A few other trials were discarded because the transcribers did not agree on what N'kisi said. In short, Sheldrake's statistical conclusions are based on the results of 71 of the trials. I'll let the reader decide whether it was proper to omit 40% of the data because the parrot didn't utter a word on the key word list during those trials. Some might argue that those sessions should be counted as misses and that by ignoring so much data where the parrot clearly did not indicate any sign of telepathy is strong evidence that Sheldrake was more interested in confirming his biases than in getting at the truth.
N'kisi's misses were listed at 94. Ten of the 23 hits were on the picture that corresponded to the word 'flower', which N'kisi uttered 23 times during the trials. The flower image, selected randomly, was used in 17 trials. The image corresponding to water was used in 10 of the trials. The bird said 'water' in twelve trials and got 2 hits. It seems oddly biased that almost one-third of the images and more than half the hits came from just 2 of the 19 pictures.
One of the peer reviewers thought that the fact that the flower word and picture played so heavy a role in the outcome that the paper's results were distorted and that the paper should not be published. The other reviewer accepted Sheldrake's observation that even if you throw out the flower data, you still get some sort of statistical significance. This may be true. However, since the bird allegedly had a vocabulary of some 950 words at the time of the test, omitting sessions where the bird said nothing or said something not on the key list, is unjustifiable. Furthermore, there is no evidence that it is reasonable to assume that when the parrot is by itself uttering words that it is trying to communicate telepathically with Morgana. Or are we to accept Sheldrake's assumption that the parrot turns his telepathic interest off and on, and it was on only when he uttered a word on the key list? That assumption is no more valid that Morgana's belief that the telepathy doesn't work as well when she makes an effort to send a telepathic message to her parrot. In any case, I wonder why Sheldrake didn't do a baseline study, where the parrot was videotaped for two-minutes at a time while Morgana was taking an aromatherapy bath or meditating or doing something unrelated to the key word pictures. Had he made several hundred such clips, he could then have randomly selected 71 and compared them to the 71 clips he used for his analysis. If there was no significant difference between the randomly selected clips and the ones that emerged during the experiment, then the telepathy hypothesis would not be supported. On the other hand, if he found a robust statistically significant difference, then the telepathy hypothesis would be supported. I suggest he do something along these lines when he attempts to replicate his parrot telepathy test.
In some trials, N’kisi repeated a given key word. For example, in one trial N’kisi said ‘‘phone’’ three times, and in another he said ‘‘flower’’ ten times, and in the tabulation of data the numbers of times he said these words are shown in parentheses as: phone (3); flower (10). For most of the statistical analyses, repetitions were ignored, but in one analysis the numbers of words that were said more than once in a given trial were compared statistically with those said only once for both hits and misses. For each trial, the key word or words represented in the photograph were tabulated. Some images had only one key word, but others had two or more. For example, a picture of a couple hugging in a pool of water involved two key words, ‘‘water’’ and ‘‘hug.’’ (Sheldrake and Morgana 2003)
He calculated 51 hits and 126 misses when repetitions were included. I'm not going to bother with any more detail because by now the overall picture should be clear. Once the statisticians went to work on the data, they were able to provide support for the claim that the data were consistent with the telepathic hypothesis. But nowhere in Sheldrake's paper can I find a claim that the parrot did three times better than expected by chance. In any case, I have to agree with the editor who published Sheldrake's parrot paper: the results have a statistical significance that is less than compelling. However, unlike that editor, my devout wish is that when such studies as these are published in the future, responsible journalists continue to ignore them and recognize them for the rubbish they are. On the other hand, if you happen to think your parrot is psychic, drop Dr. Sheldrake a line. He's set up a page just for you.
Sheldrake has responded to this article. His comments and my responses are posted here.
books and articles
new Grey parrots use reasoning where monkeys and dogs can’t "Christian Schloegl and his team at the University of Vienna, let six parrots choose between two containers, one containing a nut. Both containers were shaken, one eliciting a rattling sound and the other nothing. The parrots preferred the container that rattled, even if only the empty container was shaken....Thus, grey parrots seem to possess ape-like reasoning skills...." [/new]
Last updated 16-Aug-2012
| 50
|
Kidney Disease of Diabetes
On this page:
- The Burden of Kidney Failure
- The Course of Kidney Disease
- Diagnosis of CKD
- Effects of High Blood Pressure
- Preventing and Slowing Kidney Disease
- Dialysis and Transplantation
- Good Care Makes a Difference
- Points to Remember
- Hope through Research
- For More Information
The Burden of Kidney Failure
Each year in the United States, more than 100,000 people are diagnosed with kidney failure, a serious condition in which the kidneys fail to rid the body of wastes.1 Kidney failure is the final stage of chronic kidney disease (CKD).
Diabetes is the most common cause of kidney failure, accounting for nearly 44 percent of new cases.1 Even when diabetes is controlled, the disease can lead to CKD and kidney failure. Most people with diabetes do not develop CKD that is severe enough to progress to kidney failure. Nearly 24 million people in the United States have diabetes, 2 and nearly 180,000 people are living with kidney failure as a result of diabetes.1
People with kidney failure undergo either dialysis, an artificial blood-cleaning process, or transplantation to receive a healthy kidney from a donor. Most U.S. citizens who develop kidney failure are eligible for federally funded care. In 2005, care for patients with kidney failure cost the United States nearly $32 billion.1
African Americans, American Indians, and Hispanics/Latinos develop diabetes, CKD, and kidney failure at rates higher than Caucasians. Scientists have not been able to explain these higher rates. Nor can they explain fully the interplay of factors leading to kidney disease of diabetes—factors including heredity, diet, and other medical conditions, such as high blood pressure. They have found that high blood pressure and high levels of blood glucose increase the risk that a person with diabetes will progress to kidney failure.
1United States Renal Data System. USRDS 2007 Annual Data Report. Bethesda, MD: National Institute of Diabetes and Digestive and Kidney Diseases, National Institutes of Health, U.S. Department of Health and Human Services; 2007.
2National Institute of Diabetes and Digestive and Kidney Diseases. National Diabetes Statistics, 2007. Bethesda, MD: National Institutes of Health, U.S. Department of Health and Human Services, 2008.
The Course of Kidney Disease
Diabetic kidney disease takes many years to develop. In some people, the filtering function of the kidneys is actually higher than normal in the first few years of their diabetes.
Over several years, people who are developing kidney disease will have small amounts of the blood protein albumin begin to leak into their urine. This first stage of CKD is called microalbuminuria. The kidney's filtration function usually remains normal during this period.
As the disease progresses, more albumin leaks into the urine. This stage may be called macroalbuminuria or proteinuria. As the amount of albumin in the urine increases, the kidneys' filtering function usually begins to drop. The body retains various wastes as filtration falls. As kidney damage develops, blood pressure often rises as well.
Overall, kidney damage rarely occurs in the first 10 years of diabetes, and usually 15 to 25 years will pass before kidney failure occurs. For people who live with diabetes for more than 25 years without any signs of kidney failure, the risk of ever developing it decreases.
Diagnosis of CKD
People with diabetes should be screened regularly for kidney disease. The two key markers for kidney disease are eGFR and urine albumin.
eGFR. eGFR stands for estimated glomerular filtration rate. Each kidney contains about 1 million tiny filters made up of blood vessels. These filters are called glomeruli. Kidney function can be checked by estimating how much blood the glomeruli filter in a minute. The calculation of eGFR is based on the amount of creatinine, a waste product, found in a blood sample. As the level of creatinine goes up, the eGFR goes down.
Kidney disease is present when eGFR is less than 60 milliliters per minute.
The American Diabetes Association (ADA) and the National Institutes of Health (NIH) recommend that eGFR be calculated from serum creatinine at least once a year in all people with diabetes.
Urine albumin. Urine albumin is measured by comparing the amount of albumin to the amount of creatinine in a single urine sample. When the kidneys are healthy, the urine will contain large amounts of creatinine but almost no albumin. Even a small increase in the ratio of albumin to creatinine is a sign of kidney damage.
Kidney disease is present when urine contains more than 30 milligrams of albumin per gram of creatinine, with or without decreased eGFR.
The ADA and the NIH recommend annual assessment of urine albumin excretion to assess kidney damage in all people with type 2 diabetes and people who have had type 1 diabetes for 5 years or more.
If kidney disease is detected, it should be addressed as part of a comprehensive approach to the treatment of diabetes.
Effects of High Blood Pressure
High blood pressure, or hypertension, is a major factor in the development of kidney problems in people with diabetes. Both a family history of hypertension and the presence of hypertension appear to increase chances of developing kidney disease. Hypertension also accelerates the progress of kidney disease when it already exists.
Blood pressure is recorded using two numbers. The first number is called the systolic pressure, and it represents the pressure in the arteries as the heart beats. The second number is called the diastolic pressure, and it represents the pressure between heartbeats. In the past, hypertension was defined as blood pressure higher than 140/90, said as "140 over 90."
The ADA and the National Heart, Lung, and Blood Institute recommend that people with diabetes keep their blood pressure below 130/80.
Hypertension can be seen not only as a cause of kidney disease but also as a result of damage created by the disease. As kidney disease progresses, physical changes in the kidneys lead to increased blood pressure. Therefore, a dangerous spiral, involving rising blood pressure and factors that raise blood pressure, occurs. Early detection and treatment of even mild hypertension are essential for people with diabetes.
Preventing and Slowing Kidney Disease
Blood Pressure Medicines
Scientists have made great progress in developing methods that slow the onset and progression of kidney disease in people with diabetes. Drugs used to lower blood pressure can slow the progression of kidney disease significantly. Two types of drugs, angiotensin-converting enzyme (ACE) inhibitors and angiotensin receptor blockers (ARBs), have proven effective in slowing the progression of kidney disease. Many people require two or more drugs to control their blood pressure. In addition to an ACE inhibitor or an ARB, a diuretic can also be useful. Beta blockers, calcium channel blockers, and other blood pressure drugs may also be needed.
An example of an effective ACE inhibitor is lisinopril (Prinivil, Zestril), which doctors commonly prescribe for treating kidney disease of diabetes. The benefits of lisinopril extend beyond its ability to lower blood pressure: it may directly protect the kidneys' glomeruli. ACE inhibitors have lowered proteinuria and slowed deterioration even in people with diabetes who did not have high blood pressure.
An example of an effective ARB is losartan (Cozaar), which has also been shown to protect kidney function and lower the risk of cardiovascular events.
Any medicine that helps patients achieve a blood pressure target of 130/80 or lower provides benefits. Patients with even mild hypertension or persistent microalbuminuria should consult a health care provider about the use of antihypertensive medicines.
In people with diabetes, excessive consumption of protein may be harmful. Experts recommend that people with kidney disease of diabetes consume the recommended dietary allowance for protein, but avoid high-protein diets. For people with greatly reduced kidney function, a diet containing reduced amounts of protein may help delay the onset of kidney failure. Anyone following a reduced-protein diet should work with a dietitian to ensure adequate nutrition.
Intensive Management of Blood Glucose
Antihypertensive drugs and low-protein diets can slow CKD. A third treatment, known as intensive management of blood glucose or glycemic control, has shown great promise for people with diabetes, especially for those in the early stages of CKD.
The human body normally converts food to glucose, the simple sugar that is the main source of energy for the body's cells. To enter cells, glucose needs the help of insulin, a hormone produced by the pancreas. When a person does not make enough insulin, or the body does not respond to the insulin that is present, the body cannot process glucose, and it builds up in the bloodstream. High levels of glucose in the blood lead to a diagnosis of diabetes.
Intensive management of blood glucose is a treatment regimen that aims to keep blood glucose levels close to normal. The regimen includes testing blood glucose frequently, administering insulin throughout the day on the basis of food intake and physical activity, following a diet and activity plan, and consulting a health care team regularly. Some people use an insulin pump to supply insulin throughout the day.
A number of studies have pointed to the beneficial effects of intensive management of blood glucose. In the Diabetes Control and Complications Trial supported by the National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK), researchers found a 50 percent decrease in both development and progression of early diabetic kidney disease in participants who followed an intensive regimen for controlling blood glucose levels. The intensively managed patients had average blood glucose levels of 150 milligrams per deciliter-about 80 milligrams per deciliter lower than the levels observed in the conventionally managed patients. The United Kingdom Prospective Diabetes Study, conducted from 1976 to 1997, showed conclusively that, in people with improved blood glucose control, the risk of early kidney disease was reduced by a third. Additional studies conducted over the past decades have clearly established that any program resulting in sustained lowering of blood glucose levels will be beneficial to patients in the early stages of CKD.
Dialysis and Transplantation
When people with diabetes experience kidney failure, they must undergo either dialysis or a kidney transplant. As recently as the 1970s, medical experts commonly excluded people with diabetes from dialysis and transplantation, in part because the experts felt damage caused by diabetes would offset benefits of the treatments. Today, because of better control of diabetes and improved rates of survival following treatment, doctors do not hesitate to offer dialysis and kidney transplantation to people with diabetes.
Currently, the survival of kidneys transplanted into people with diabetes is about the same as the survival of transplants in people without diabetes. Dialysis for people with diabetes also works well in the short run. Even so, people with diabetes who receive transplants or dialysis experience higher morbidity and mortality because of coexisting complications of diabetes-such as damage to the heart, eyes, and nerves.
Good Care Makes a Difference
People with diabetes should
- have their health care provider measure their A1C level at least twice a year. The test provides a weighted average of their blood glucose level for the previous 3 months. They should aim to keep it at less than 7 percent.
- work with their health care provider regarding insulin injections, medicines, meal planning, physical activity, and blood glucose monitoring.
- have their blood pressure checked several times a year. If blood pressure is high, they should follow their health care provider's plan for keeping it near normal levels. They should aim to keep it at less than 130/80.
- ask their health care provider whether they might benefit from taking an ACE inhibitor or ARB.
- ask their health care provider to measure their eGFR at least once a year to learn how well their kidneys are working.
- ask their health care provider to measure the amount of protein in their urine at least once a year to check for kidney damage.
- ask their health care provider whether they should reduce the amount of protein in their diet and ask for a referral to see a registered dietitian to help with meal planning.
Points to Remember
- Diabetes is the leading cause of chronic kidney disease (CKD) and kidney failure in the United States.
- People with diabetes should be screened regularly for kidney disease. The two key markers for kidney disease are estimated glomerular filtration rate (eGFR) and urine albumin.
- Drugs used to lower blood pressure can slow the progression of kidney disease significantly. Two types of drugs, angiotensin-converting enzyme (ACE) inhibitors and angiotensin receptor blockers (ARBs), have proven effective in slowing the progression of kidney disease.
- In people with diabetes, excessive consumption of protein may be harmful.
- Intensive management of blood glucose has shown great promise for people with diabetes, especially for those in the early stages of CKD.
Hope through Research
The number of people with diabetes is growing. As a result, the number of people with kidney failure caused by diabetes is also growing. Some experts predict that diabetes soon might account for half the cases of kidney failure. In light of the increasing illness and death related to diabetes and kidney failure, patients, researchers, and health care professionals will continue to benefit by addressing the relationship between the two diseases. The NIDDK is a leader in supporting research in this area.
Several areas of research supported by the NIDDK hold great potential. Discovery of ways to predict who will develop kidney disease may lead to greater prevention, as people with diabetes who learn they are at risk institute strategies such as intensive management of blood glucose and blood pressure control.
Participants in clinical trials can play a more active role in their own health care, gain access to new research treatments before they are widely available, and help others by contributing to medical research. For information about current studies, visit www.ClinicalTrials.gov.
For More Information
National Diabetes Information Clearinghouse
1 Information Way
Bethesda, MD 20892-3560
National Kidney Foundation
30 East 33rd Street
New York, NY 10016
Phone: 1-800-622-9010 or 212-889-2210
National Kidney and Urologic Diseases Information Clearinghouse
The National Kidney and Urologic Diseases Information Clearinghouse (NKUDIC) is a service of the National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK). The NIDDK is part of the National Institutes of Health of the U.S. Department of Health and Human Services. Established in 1987, the Clearinghouse provides information about diseases of the kidneys and urologic system to people with kidney and urologic disorders and to their families, health care professionals, and the public. The NKUDIC answers inquiries, develops and distributes publications, and works closely with professional and patient organizations and Government agencies to coordinate resources about kidney and urologic diseases.
Publications produced by the Clearinghouse are carefully reviewed by both NIDDK scientists and outside experts.
This publication is not copyrighted. The Clearinghouse encourages users of this publication to duplicate and distribute as many copies as desired.
NIH Publication No. 08-3925
Page last updated: September 2, 2010
| 353
|
Intellectual knowledge appears to be innate and privy to the few, but in fact, access to information, development of intellectual work skills, time investment, and the maintenance of intellectual appearances are key to being perceived as an intellectual.
To make education more equitable, professors must go beyond knowledge transmission and instruct students in the concrete skills of knowledge acquisition and knowledge presentation. Instruction in intellectual skills-acquisition implies the breakdown of the traditional professor-student relationship and of the academic intellectual hierarchy and professors must learn to cope with the consequences of adopting new pedagogies. If we wish to share the secrets of our professions, how do we prepare our students for such a democratic approach and at the same time maintain our professional status?
The author, a professor of Spanish language and literature, presents strategies for democratizing education and demystifying intellectual work through the application of skills-based pedagogical methodologies to the teaching of literature. The implications that these strategies have for a new type of learning and the impact that they have on social stratification will also be discussed.
|Keywords:||Democratizing Education, Demystifying Intellectual Work, Knowledge Acquisition Skills, Interpretative Skills, Teaching Literature, Skills Based Teaching|
Assistant Professor of Spanish, Department of Literature and Languages, Roosevelt University, Chicago, Illinois, USA
There are currently no reviews of this product.Write a Review
| 835
|
European Painting before 1900, Johnson Collection
Triptych showing scenes from the Martyrdom of Saint Barbara and scenes from the Life of Christ
Master of the Laufen High Altarpiece, Austrian (active Salzburg), dated work 1467
Oil and gold on panels
EW1993-127-2a--cPurchased with the W. P. Wilstach Fund, the George W. Elkins Fund, and Museum funds, 1993
As described in The Golden Legend, written in about 1267 by Jacopo da Voragine, Saint Barbara converted to Christianity against the wishes of her pagan father. Locked in a tower by him, she ordered workmen to construct a third window inthe building to symbolize the Christian trinity. Enraged, her father turned her over to the Roman authorities for torture, and when her execution was ordered,he himself beheaded her. The left and right panels on the front of this triptychillustrate episodes from Barbara’s gruesome martyrdom.
These graphic depictions are intended to help the viewer equate Barbara’s physical torments with those of Christ, whose crucifixion is depicted in the top center composition. The figures at the top left and right, probably the apostles John and Luke, hold sayings taken from the last words of Christ on the cross, which here apply also to Barbara’s death. The bottom center panel shows Barbara enthroned in heaven, wearing a crown and holding a palm frond as symbols of victory over death. A tower and chalice, her identifying attributes, appear nearby.
Barbara’s martyrdom is paralleled on the reverse by the resurrected Christ standing in the tomb. The left and right panels of the back, which are movable and can be folded shut to appear on the front, show the angel Gabriel announcing the incarnation of Christ to the Virgin Mary.
Social Tags [?]austrian [x] gabriel [x] gold [x] martyr [x] three [x] trinity [x] triptych [x] [Add Your Own Tags]
* Works in the collection are moved off view for many different reasons. Although gallery locations on the website are updated regularly, there is no guarantee that this object will be on display on the day of your visit.
| 142
|
How many times did you hear that today? Probably a lot. But why are adults so hung up on hand washing? Why are they so in love with lather?
Washing your hands is the best way to stop germs from spreading. Think about all of the things that you touched today — from the telephone to the toilet. Maybe you blew your nose in a tissue and then went outside to dig around the dirt.
Whatever you did today, you came into contact with germs. It's easy for a germ on your hand to end up in your mouth. Think about how many foods you eat with your hands.
You can't wear rubber gloves all day long, but you can wash your hands so those germs don't get a chance to make you or someone else sick. When germs go down the drain, they can't make anyone sick.
So when are the best times to wash your hands?
when your hands are dirty
before eating or touching food (like if you're helping cook or bake, for example)
before and after visiting a sick relative or friend
Now you have the when and the why of hand washing. But did you know that a lot of people don't know how to get their paws perfectly clean? The next time you're told to step up to the sink and scrub up, remember these handy hints:
Use warm water (not cold or hot) when you wash your hands.
Use whatever soap you like. Some soaps come in cool shapes and colors or smell nice, but whatever kind gets you scrubbing is the kind you should use. Antibacterial soaps are OK to use, but regular soap works fine.
Work up some lather on both sides of your hands, your wrists, and between your fingers. Don't forget to wash around your nails. This is one place germs like to hide. Wash for about 10 to 15 seconds — about how long it takes to sing "Happy Birthday." (Sing it quickly two times or just once if you go nice and slow.)
| 470
|
Gonorrhoea - the drugs don't work
Published: 23rd Dec 2011 08:41:37
The prospect of untreatable gonorrhoea has provoked alarm around the world, and there are no new classes of antibiotics in development.
In this week's Scrubbing Up column, Peter Greenhouse of the British Association for Sexual Health & HIV (BASHH) argues financial incentives will be needed to seek a new cure.
We're all familiar with stories about hospital-acquired superbugs - MRSA and the like - becoming more difficult to treat, and are fearful whenever an elderly relative needs in-patient care.
But now, with a report from Japan of multidrug-resistant gonorrhoea, and the festive season in full swing, the spectre of an untreatable sexually transmitted infection looms over us - and our teenagers - for the first time in a generation.
Since penicillin was first used to treat gonorrhoea in 1943, the organism has gradually developed novel means of evading control by each new antibiotic.
For treatment to be effective and practical, it must be simple to administer by mouth as a single dose, achieving a high enough concentration of the drug in the body to treat over 95% of infections.
If the efficacy drops below this figure, the treatment has to change.
But over-the-counter medication, widely available in Africa and Asia means people self-medicate often taking the wrong dose at the wrong time, perhaps with alcohol which further reduces the concentration of the drug.
Strains of gonorrhoea which need a higher concentration of a drug to kill them become the dominant ones. This keeps happening until the drug no longer works.
If gonorrhoea becomes untreatable in these countries, the effect on increasing HIV rates could be disastrous - because any sexually transmitted infection which causes inflammation and discharge increases the transmission efficiency of HIV.
On average, transmission is five times more likely to occur if gonorrhoea or chlamydia are present
In the UK, the situation is monitored annually by the Health Protection Agency, providing an essential early warning of drugs which are about to fail, allowing a switch of treatment regimes before they become ineffective.
There's a desperate world-wide demand for new antibiotics, yet the drug companies aren't interested”
Ciprofloxacin - a drug introduced in the mid-1980s after the failure of penicillin - lasted in the UK until 2002: This may have survived longer because of the world-wide drop in gonorrhoea rates following the arrival of HIV, when fear of the new virus meant people practised safe-sex and changed partners less.
But it had already failed in the Far East, some four years previously.
Resistance develops faster in homosexual men, not just because of high rates of partner change.
Most people don't realise that oral sex is an important route of transmission for gonorrhoea, which doesn't usually cause a sore throat.
Gonorrhoea mixes with organisms which live naturally in the rectum and throat, picking up new types of antibiotic resistance from these bugs.
The next drug, cefixime, was introduced around 2003, but lasted only six years in the UK before resistance rose suddenly, hitting 25% among homosexual men.
Now, their only treatment option is an injection (Ceftriaxone) which has recently failed in Japan.
But why isn't there a new drug in development?
Since the mid-1980s and the arrival of HIV, almost all drug company research has focused on antiviral medicines, with no new classes of antibiotics being produced since the 1970s, and none on the horizon.
There's a desperate world-wide demand for new antibiotics, yet the drug companies aren't interested, so how could we motivate them?
Financial reality dictates research policy: Why bother to develop a drug which works in one day or one week, when you could make one - such as an antidepressant, statin or antiviral - which must be taken for months, for years, or for life?
So either the new drug(s) would have to be seriously expensive, precluding their use where they would be most needed, or there would have to be a substantial reward offered, perhaps of a magnitude only affordable by a fund such as the Gates Foundation.
Yet even if novel drugs could be produced, the biology and transmission dynamics of gonorrhoea mean that each new regime would probably fail within five-to-ten years of its introduction, unless we use multi-dose, multi-drug regimes, which will be less practical and more expensive to administer.
Faced with this, what can we do to stay sexually healthy? Stay at home, or take your partner to the New Year party: If that's not possible, use condoms - meticulously, and visit your local clinic - frequently.
Harvard CitationBBC News, 2011. Gonorrhoea - the drugs don't work. [Online] (Updated 23 Dec 2011)
Available at: http://www.ukwirednews.com/news.php/212819-Gonorrhoea-the-drugs-dont-work [Accessed 14th May 2013]
At 07:53:07 in BusinessA group of international investors is interested in buying UK water supplier Severn Trent, the company has said....
At 07:49:21 in WalesMajor road safety works have started on a one-mile stretch of coast road at Flintshire which has been the scene of serious accidents....
At 07:48:53 in EnglandStuart Hazell is due to be sentenced later after admitting mid-trial to killing his partner's granddaughter....
At 07:47:03 in WalesAn independent TV company has issued redundancy notices to 10 members of staff due to a cut in work from Welsh language broadcaster S4C....
At 07:43:33 in WalesFurther forensic evidence is expected to be heard later in the trial of the man accused of murdering April Jones....
At 07:41:47 in WalesA Cardiff theatre was asked to explain by the Arts Council of Wales why there was no plan to tackle an £800,000 overspend, it has emerged....
At 07:40:22 in EntertainmentHollywood actress Angelina Jolie has undergone a double mastectomy to reduce her chances of getting breast cancer....
At 07:32:39 in HeadlinesA boat carrying Rohingya Muslims has capsized off western Burma, aid agencies say. ...
At 07:31:47 in BusinessIndia's top drugmaker Ranbaxy Laboratories is to pay a record fine in the US for lying to officials and selling badly made generic drug...
At 07:31:30 in Northern IrelandPolice have issued a description of a man they want to question about the rape of a teenage girl in west Belfast....
News In Other Categories
A group of international investors is interested in buying UK water supplier Severn Trent, the company has said....
Free wi-fi access is to be introduced at 25 of Scotland's busiest railway stations before the end of the year. ...
Major road safety works have started on a one-mile stretch of coast road at Flintshire which has been the scene of serious accidents....
What makes a great foreign minister? Some of those who have held the great office of state, including Lord Carrington, David Owen and David ...
It was a melodious spectacle....
With the doors to its brand new £1million training centre officially open, one of the UK's leading apprentice training providers, Bristol ba...
| 399
|
Issued in Shanghai, February 28, 1972
President Richard Nixon of the United States of America visited the People's Republic of China at the invitation of Premier Chou En-lai of the People's Republic of China from February 21 to February 28, 1972. Accompanying the President were Mrs. Nixon, U.S. Secretary of State William Rogers, Assistant to the President Dr. Henry Kissinger, and other American officials.
President Nixon met with Chairman Mao Tsetung of the Communist Party of China on February 21. The two leaders had a serious and frank exchange of views on Sino-U.S. relations and world affairs.
During the visit, extensive, earnest and frank discussions were held between President Nixon and Premier Chou En-lai on the normalization of relations between the United States of America and the People's Republic of China, as well as on other matters of interest to both sides. In addition, Secretary of State William Rogers and Foreign Minister Chi Peng-fei held talks in the same spirit.
President Nixon and his party visited Peking and viewed cultural, industrial and agricultural sites, and they also toured Hangchow and Shanghai where, continuing discussions with Chinese leaders, they viewed similar places of interest.
The leaders of the People's Republic of China and the United States of America found it beneficial to have this opportunity, after so many years without contact, to present candidly to one another their views on a variety of issues. They reviewed the international situation in which important changes and great upheavals are taking place and expounded their respective positions and attitudes.
The Chinese side stated: Wherever there is oppression there is resistance. Countries want independence, nations want liberation and the people want revolution-this has become the irresistible trend of history. All nations, big or small, should be equal; big nations should not bully the small and strong nations should not bully the weak. China will never be a superpower and it opposes hegemony and power politics of any kind. The Chinese side stated that it firmly supports the struggles of all the oppressed people and nations for freedom and liberation and that the people of all countries have the right to choose their social systems according to their own wishes and the right to safeguard the independence, sovereignty and territorial integrity of their own countries and oppose foreign aggression, interference, control and subversion. All foreign troops should be withdrawn to their own countries. The Chinese side expressed its firm support to the peoples of Viet Nam, Laos and Cambodia in their efforts for the attainment of their goal and its firm support to the seven-point proposal of the Provisional Revolutionary Government of the Republic of South Viet Nam and the elaboration of February this year on the two key problems in the proposal, and to the Joint Declaration of the Summit Conference of the Indochinese Peoples. It firmly supports the eight-point program for the peaceful unification of Korea put forward by the Government of the Democratic People's Republic of Korea on April 12, 1971, and the stand for the abolition of the "U.N. Commission for the Unification and Rehabilitation of Korea". It firmly opposes the revival and outward expansion of Japanese militarism and firmly supports the Japanese people's desire to build an independent, democratic, peaceful and neutral Japan. It firmly maintains that India and Pakistan should, in accordance with the United Nations resolutions on the India-Pakistan question, immediately withdraw all their forces to their respective territories and to their own sides of the ceasefire line in Jammu and Kashmir and firmly supports the Pakistan Government and people in their struggle to preserve their independence and sovereignty and the people of Jammu and Kashmir in their struggle for the right of self-determination.
The U.S. side stated: Peace in Asia and peace in the world requires efforts both to reduce immediate tensions and to eliminate the basic causes of conflict. The United States will work for a just and secure peace; just, because it fulfills the aspirations of peoples and nations for freedom and progress; secure, because it removes the danger of foreign aggression. The United States supports individual freedom and social progress for all the peoples of the world, free of outside pressure or intervention. The United States believes that the effort to reduce tensions is served by improving communication between countries that have different ideologies so as to lessen the risks of confrontation through accident, miscalculation or misunderstanding. Countries should treat each other with mutual respect and be willing to compete peacefully, letting performance be the ultimate judge. No country should claim infallibility and each country should be prepared to reexamine its own attitudes for the common good. The United States stressed that the peoples of Indochina should be allowed to determine their destiny without outside intervention; its constant primary objective has been a negotiated solution; the eight-point proposal put forward by the Republic of Viet Nam and the United States on January 27, 1972 represents a basis for the attainment of that objective; in the absence of a negotiated settlement, the United States envisages the ultimate withdrawal of all U.S. forces from the region consistent with the aim of self-determination for each country of Indochina. The United States will maintain its close ties with and support for the Republic of Korea; the United States will support efforts of the Republic of Korea to seek a relaxation of tension and increased communication in the Korean peninsula. The United States places the highest value on its friendly relations with Japan; it will continue to develop the existing close bonds. Consistent with the United Nations Security Council Resolution of December 21, 1971, the United States favors the continuation of the ceasefire between India and Pakistan and the withdrawal of all military forces to within their own territories and to their own sides of the ceasefire line in Jammu and Kashmir; the United States supports the right of the peoples of South Asia to shape their own future in peace, free of military threat, and without having the area become the subject of great power rivalry.
There are essential differences between China and the United States in their social systems and foreign policies. However, the two sides agreed that countries, regardless of their social systems, should conduct their relations on the principles of respect for the sovereignty and territorial integrity of all states, non-aggression against other states, non-interference in the internal affairs of other states, equality and mutual benefit, and peaceful coexistence. International disputes should be settled on this basis, without resorting to the use or threat of force. The United States and the People's Republic of China are prepared to apply these principles to their mutual relations.
With these principles of international relations in mind the two sides stated that:
--progress toward the normalization of relations between China and the United States is in the interests of all countries;
--both wish to reduce the danger of international military conflict;
--neither should seek hegemony in the Asia-Pacific region and each is opposed to efforts by any other country or group of countries to establish such hegemony; and
--neither is prepared to negotiate on behalf of any third party or to enter into agreements or understandings with the other directed at other states.
Both sides are of the view that it would be against the interests of the peoples of the world for any major country to collude with another against other countries, or for major countries to divide up the world into spheres of interest.
The two sides reviewed the long-standing serious disputes between China and the United States. The Chinese side reaffirmed its position: The Taiwan question is the crucial question obstructing the normalization of relations between China and the United States; the Government of the People's Republic of China is the sole legal government of China; Taiwan is a province of China which has long been returned to the motherland; the liberation of Taiwan is China's internal affair in which no other country has the right to interfere; and all U.S. forces and military installations must be withdrawn from Taiwan. The Chinese Government firmly opposes any activities which aim at the creation of "one China, one Taiwan" "one China two governments", "two Chinas", an "independent Taiwan" or advocate that "the status of Taiwan remains to be determined".
The U.S. side declared: The United States acknowledges that all Chinese on either side of the Taiwan Strait maintain there is but one China and that Taiwan is a part of China. The United States Government does not challenge that position. It reaffirms its interest in a peaceful settlement of the Taiwan question by the Chinese themselves. With this prospect in mind, it affirms the ultimate objective of the withdrawal of all U.S. forces and military installations from Taiwan. In the meantime, it will progressively reduce its forces and military installations on Taiwan as the tension in the area diminishes.
The two sides agreed that it is desirable to broaden the understanding between the two peoples. To this end, they discussed specific areas in such fields as science, technology, culture, sports and journalism, in which people-to-people contacts and exchanges would be mutually beneficial. Each side undertakes to facilitate the further development of such contacts and exchanges.
Both sides view bilateral trade as another area from which mutual benefit can be derived, and agreed that economic relations based on equality and mutual benefit are in the interest of the peoples of the two countries. They agree to facilitate the progressive development of trade between their two countries.
The two sides agreed that they will stay in contact through various channels, including the sending of a senior U.S. representative to Peking from time to time for concrete consultations to further the normalization of relations between the two countries and continue to exchange views on issues of common interest.
The two sides expressed the hope that the gains achieved during this visit would open up new prospects for the relations between the two countries. They believe that the normalization of relations between the two countries is not only in the interest of the Chinese and American peoples but also contributes to the relaxation of tension in Asia and the world.
President Nixon, Mrs. Nixon and the American party expressed their appreciation for the gracious hospitality shown them by the Government and people of the People's Republic of China.
| 568
|
Autoimmune hemolytic anemia, or immune-mediated hemolytic anemia (IMHA), is a complex disease in which hemolysis occurs because
of antierythrocyte antibody production. This article explores the pathophysiology of primary and secondary IMHA and diagnostic
and treatment options, as well as prognosis in dogs and cats. Our review of the recent literature regarding IMHA in veterinary
patients reveals a focus on individual cases and a lack of controlled clinical studies, which makes a detailed review of
IMHA triggers and treatment options difficult.
HEMOLYTIC ANEMIA VS. IMHA
Many causes of anemia exist in dogs and cats, so a clear distinction should be drawn between hemolytic anemia and IMHA.
Hemolytic anemias are conditions in which red blood cells (RBCs) are destroyed at an accelerated rate and a normal regenerative
response is seen in the bone marrow. In these non-immune-mediated conditions, RBCs can be destroyed as a result of inherited
membrane and enzyme defects, increased fragility from oxidative damage, or metabolic causes such as hypophosphatemia or water
intoxication.1 Traditional immune-mediated mechanisms (immunoglobulins, complement) do not mediate lysis in hemolytic anemia. Instead,
destruction occurs because of factors such as increased osmotic fragility, decreased RBC function in an alkalemic environment,
or increased clearance from oxidative damage.1 Unlike the treatment of immune-mediated anemia, immunosuppression is generally not used to treat hemolytic anemias. Thus,
it is imperative to investigate whether an anemia has an underlying cause before assuming it is immune-mediated.
Below we discuss some of the common causes of hemolytic anemia. A more exhaustive list can be found in Table 1.
Table 1: Selected Causes of Canine and Feline Non-immune-mediated Hemolytic Anemia
Zinc and copper toxicosis. One of the most common causes of hemolytic anemia in dogs is zinc toxicosis from the ingestion of zinc-containing objects.
High zinc concentrations can be found in pennies minted since 1983, board game pieces, zippers, zinc oxide ointment, and various
other sources. Zinc toxicosis can cause a severe intravascular hemolysis that is associated with small amounts of Heinz body
and spherocyte formation. Hemolysis from zinc toxicosis can be easily mistaken for IMHA if a survey abdominal radiographic
examination is not performed. Treatment consists of removing the foreign object and providing supportive care. Copper toxicosis
can also result in a marked intravascular hemolysis and methemoglobinemia.1
Heinz body anemia. Heinz bodies are dark-staining refractile material that indicate irreversibly denatured and precipitated hemoglobin in RBCs
and can occur from oxidative damage in animals that have ingested onions or received drugs such as methylene blue, dl-methionine,
or vitamin K3.1 In some cases of hemolytic anemia, eccentrocytes—cells in which the damaged hemoglobin is clustered together and shifted
to one side of the RBC membrane, leaving a clear crescent-shaped region—are also present. Eccentrocytes and cells containing
Heinz bodies have less deformability and more rigidity, making them more likely to be lysed or removed from the circulation
by the spleen.
Feline RBCs are especially sensitive to oxidative damage because of a high number of sulfhydryl groups in their hemogloblin.
In addition, feline spleens, because of their nonsinusal conformation, are less effective at removing Heinz bodies than are
canine sinusal spleens. Thus, Heinz body anemia is more commonly seen in cats and can be present during toxicosis as well
as in a variety of other diseases such as hyperthyroidism, diabetes mellitus, and lymphoma.1
Hypophosphatemia. Hypophosphatemia can also cause hemolytic anemia in patients being treated for diabetes mellitus, hepatic lipidosis, starvation,
and other conditions.1 In these cases, it is thought that an abrupt drop in plasma phosphate concentrations can cause a concurrent depletion of
RBC adenosine triphosphate, diphosphoglycerate, and reduced glutathione. These depletions lead to less deformability, more
osmotic fragility, and more susceptibility to oxidative injury in erythrocytes. A rapid drop in packed cell volume and a mild
Heinz body anemia can be seen in animals with hypophosphatemia. Treatment consists of phosphate supplementation (intravenous
in cases of severe depletion, oral in mild cases).1
| 320
|
Freighter (or barge) was the term given to any spacecraft that was used to transport freight or cargo (such as parts and supplies). Both legitimate businesspersons and smugglers could be found captaining such transports. Freighters were needed from the time space travel began because of the need for supplies to all reaches of the galaxy. Freighters usually traveled with hyperdrives because people would often pay more for quick delivery.
For obvious reasons, freighters were used mainly for trade. Although very few fought in great battles, many freighters did see action. Smugglers and legitimate traders alike had some tangles now and then, but freighters were often armed and shielded so that they could resist attempts on their cargo.
Smuggling was a very prominent criminal fringe activity in the galaxy that involved the transportation of contraband between planets. To do this, a freighter generally of small size such as the Millennium Falcon, was required to transport the contraband past planetary security forces. Some examples of contraband include spice, blasters and medical supplies.
Smugglers often added upgrades to their ships so that they could beat competitors and outwit planetary security forces. Almost every smugglers vessel had improved light speed and sub-light speed drives for increased speed as well as boosted weapons systems to fight their way out of tough situations. Another of the most prevalent modifications to a smugglers freighter were numerous secret compartments to hide contraband from security checks while legitimate cargo occupied the cargo bays. On the Millennium Falcon, these consisted of removable floor plates.
Some freighters became so heavily modified that the mess of cross wired and non traditional parts prevented starship mechanics from working on them effectively. The smugglers themselves were therefore required to have a detailed knowledge of freighter mechanics and electronics to be able to keep their vessels in working order. Most smugglers preferred it this way as they didn't trust anyone tampering with their prized possessions.
Most smugglers freighters required a crew of more than one and as such most smugglers hired a copilot, such as Chewbacca on the Millennium Falcon or the droid LE-BO2D9 aboard the Outrider. To a smuggler, their freighter was everything; their job, their home, their lives. Many smugglers lived in their freighters as they had no terrestrial home.
- Star Wars: The Old Republic
- Tempest Feud
- The Clone Wars: Decide Your Destiny: Crisis on Coruscant
- "The Heart"
- "Maze Run"—Star Wars Insider 131
- Star Wars Episode IV: A New Hope (First appearance)
- Choices of One
- Star Wars: Empire at War
- Star Wars: Empire at War: Forces of Corruption
- "Sandbound on Tatooine" — Star Wars Galaxy 10
- "Slaying Dragons"—Star Wars Adventure Journal 9
- Slave Ship
- Hard Merchandise
- The New Rebellion
- Vision of the Future
- Emissary of the Void
- Force Heretic I: Remnant
- X-Wing Miniatures Mission 4: Den of Thieves
In other languages
- freighter on Wikipedia
| 19
|
Nicaraguan Pottery Designs by David Sequeira
- Dates: December 1, 1939 through February 4, 1940
- Collections: Arts of the Americas
December 1, 1939: Eighty large color drawings by David Sequeira from Chorotegan pottery excavated by him in Nicaragua will be shown at the Brooklyn Museum from December 1 through February 4 in an exhibition arranged by Fr. Herbert J. Spinden, Curator of American Indian Art and Primitive Cultures. Sequira is of Nicaraguan birth and primarily a musican whose interest in musical themes of the Indians induced him to study and reproduce the brilliant designs on the pre-Colombian pottery of Nicaragua identified by Dr. Spinden as superb examples of Chorotegan art.
The drawings are circular and rectangular and will be shown in panels. The designs are in strong, warm colors, brick red, reddish brown, black and white, The artist has, in effect, transferred the patterns from concave and convex pottery surfaces to flat surfaces The patterns incorporate motives of conventionalized jaguars, crab, serpents, crocodiles, monkeys and birds, the last including the famous Quetzal which Dr. Spinden states is unusual to find represented in a culture as far south as Nicaragua. Human figures and geometrical or abstract designs both showing a strong Maya influence are also used in the patterns.
The drawings are shown more for their interest as sound designs than for their scientific interest, which is nevertheless important, as part of the Museum’s policy of making available basic creative designs of old cultures that can he adapted to today’s needs or inspire new efforts.
The circumstances of the preparation of these drawings are important. In the first place, they were made only a short time after the pottery was excavated while the colors were still at their full strength. The tones tend to fade when they come in contact with the air after their long burial. Secondly, the artist, to carry out the preservation of the designs as faithfully as possible, mixed his own watercolors from the kinds of materials that were probably used in coloring the pottery, usually earth colors.
According to Dr. Spinden, in his paper presented at the 21st International Congress of Arnericanists in Goteborg in 1924, the Chorotegan Culture Area from which these designs come "is a strip of territory beginning on the humid north coast of Honduras near Ceiba, widening with the wet lands and extending southward across Nicaragua and Costa Rica to about the limits of Panama. The word Chorotegan…is of Mexican origin with some such meaning as the Driven-out People.
“The culture in this area was built historically on Mayan ideas of the First Empire ending A.D. 630, yet the Chorotegas reached the height of their productivity on the Toltec horizon (1150-1350) and were intermediaries in trade between Mexico and Colombia.” He further states that the present tribes in this area are not of the same stock as the Chorotegans who produced the high culture.
In the region of Northern Costa Rica and southern Nicaragua “……the culture was intensively developed, especially in pottery….. The designs are painted and modelled, with the crocodile, the monkey, the jaguar, the fish, etc. serving as motives. One peculiar type of pottery shows the use of a purplish black luster paint made from manganese ore.” It is probable that the pieces of pottery from which the exhibited designs were taken fall in this category.
| 732
|
The Brazilian Supreme Court's recognition of same-sex unions in early May marks the latest victory for gay rights in Latin America. The Court's ruling grants equal legal rights to same-sex civil unions as those enjoyed by married heterosexuals, including retirement benefits, joint tax declarations, inheritance rights, and child adoption.
An Unlikely Victory
As the world's largest Roman Catholic country, Brazil was an unlikely venue for such a promising gay rights victory. The Roman Catholic Church has actively fought proposals for same-sex unions in Brazil, arguing that the Brazilian Constitution defines a "family entity" as "a stable union between a man and a woman."2 The Catholic Church responded to the recent ruling with outrage. As Archbishop Anuar Battisti put it, the Supreme Court's decision marked a "frontal assault" on the sanctity of the family.3
The Catholic Church is losing its power in Brazil, which helped pave the way for the Supreme Court's recent decision in favor of homosexuals. Nevertheless, homophobia retains a tenacious grip on Brazilian society. Despite the fact that the nation boasts the world's largest gay pride parade, the LGBT movement has been unable to achieve fundamental progress and quell discrimination at a societal level. For instance, Marcelo Cerqueira, the head of the Gay Group of Bahia, claims the country is "number one when it comes to assassination, discrimination and violence against homosexuals."4 Additionally, in a disconcerting report, the Gay Group of Bahia found that 260 Brazilian gay people were murdered in 2010, exemplifying the level of hostility towards homosexuals.5 Because of this discriminating environment, gay rights activists traditionally have had little success in Brazil. Most notably, Congress disregarded proposals for gay rights legislation for nearly ten years.
The Supreme Court’s recent ruling was therefore a major turning point after a history of protracted, unsuccessful struggles. The judicial decision was made in response to two lawsuits, one of which was filed by Rio de Janeiro Governor Sérgio Cabral and the other by the Office of the Attorney General. While Congress repeatedly ignored requests for equal rights for gay Brazilian citizens, the Supreme Court argued that "Those who opt for a homosexual union cannot be treated less than equally as citizens."6 In this way, by appealing to the judicial system, the LGBT movement was able to achieve success despite deep-seated hostility throughout Brazilian society and in other branches of the government.
Latin America's Gay Rights Revolution
Professor Omar Encarnación of Bard College calls the recent string of gay rights legislation in Latin America a "gay rights revolution."7 Brazil's ruling came on the heels of several other noteworthy gay rights victories in Latin America, such as Uruguay’s legalization of same-sex civil unions in 2007. Shortly thereafter, in 2010, Argentina became the first Latin American nation and eighth nation worldwide to legalize gay marriage. Other landmark decisions in the past few years include Uruguay's decision to allow all men and women, regardless of sexual orientation, to serve in the military and Mexico City's legalization of same-sex civil unions.
The recent surge in gay rights victories throughout Latin America is altogether stunning, considering the region has generally been regarded as very homophobic. The Catholic Church has traditionally been a formidable enemy to gay rights movements in the region, but the secularization of much of Latin America has led to the impressive expansion of opportunities for gay rights movements.
Yet this success of gay rights movements throughout Latin America cannot be attributed solely to the declining importance of religion in the region. It is equally important, if not more so, to recognize the vital roles played by gay activist groups and the dynamic strategies these groups employ. For instance, gay rights groups in Brazil were able to reverse legislation banning gays from the workplace by forming partnerships with progressive businesses. In recent years, the use of social media has provided much of the gay movement's momentum by enhancing activist groups' ability to communicate and spread information. For instance, as Javier Corrales notes, by simply posting a video of a hate crime in San Juan or of a gay wedding in Argentina on YouTube, gay rights groups have been able to reach thousands of people and garner support.8 These innovative strategies have brought success despite a notably hostile environment towards homosexuals.
Through a comparison with the United States, we can see how remarkable the success of gay rights in Latin America has been. Latin America is marked by a much more homophobic environment than the US, according to a survey conducted by Mitchell Seligson and Daniel Moreno Morales.9 However, although the US has lower levels of societal discrimination towards gays, it is hard to imagine that the United States would completely legalize same-sex civil unions or gay marriage on a national scale. The fact that this legalization occurred in several Latin American nations, despite the formidable opposition there, makes these recent rulings even more significant.
Furthermore, the recent victories for gay rights exemplify the considerable progress toward the region's consolidation of democracy. The three Latin American countries that have now legalized same-sex unions—Brazil, Argentina, and Uruguay—were each ruled by repressive military regimes just over two decades ago. Even Colombia, which is one of the region's worst human rights violators, granted same-sex unions equal rights regarding social security benefits and inheritance rights in 2007. The fact that gay liberation movements have been successful in these unlikely places is a testament to how far these countries have progressed in recent years.
Marilia Brocchetto and Luciani Gomes. "Same-sex unions recognized by Brazil's high court." 5 May 2011.
Yana Marull. "Brazil top court recognizes same-sex civil unions." American Free Press. 5 May 2011.
Omar Encarnación. "A Gay Rights Revolution in Latin America." Americas Quarterly. 17 May 2011.
Javier Corrales. "Latin American Gays: The Post-Left Leftists." Americas Quarterly. 19 March 2010.
Mitchell A. Seligson and Daniel E. Moreno Morales, "Gay in the Americas," Americas Quarterly, Winter 2010.
| 715
|
King cobra is the longest venomous snake in the world. Although its venom is not much powerful like Black mamba or other snakes. But his large size make him more deadly. This huge snake can grow up to (19ft) and 20 kg in weight which is a lot for venomous snake. In ancient times peoples worshiped king cobra and some peoples are still doing so.
King Cobra mostly lives in forested areas. This huge snake can be found in Southeast Asia and some parts of India. It normally eats other small snake that’s why its genus name is Ophiophagus , which means “snake- eater”. It can also eat pythons which is size able snakes. Its color is pale olive, tan, yellow olive color or something like this.
This huge snake mostly hunt at day time and very little at night time. It hunts with their long fangs which may be growing up to 0.5 inches which make this snake more deadly. It swallows its prey and his toxins help him in digestion of prey. King Cobra venom is also very dangerous. The bite taken from king cobra can lead you to death. Its mortality rate of a bite is 75%.
Defense Against Enemy
The great predator of King cobra is mongoose, which is quite resistance to neurotoxins. King cobra will try to fight with its hood flatten this help him in closed mouth strikes. If all fails, he will generally try to flee. But it is also hard for mongoose to defeat such a long snake.
Due to king cobra long size and deadly venom make him perfect predator. Although his venom is not much powerful like black mamba or Russell viper. But it is five times faster reactive venom then black mamba. With every bite king cobra inject 6 mil of venom to prey with the help of its fangs. Which is enough for killing elephant in three hours. Venom is produced in salivary glands. When this huge snake injects his venom to a prey his fastest venom paralysis it in two minutes, thus the prey go in comma due to respiratory system failure and prey died.
The female King cobra lay about 20 to 50 eggs but before laying eggs she make a nest or mound with leaf litter to lay eggs in it. She guards her eggs and prevents any big animal to come near it. When eggs hatch a new king cobras arise from it and the life cycle of king cobra remains continuous.
- King cobra is the world longest venomous snake.
- King Cobra average life span is 20 years.
- King Cobra fastest venom allows him to kill elephant in three hours.
- King Cobra is used by snake charmers for dancing and for fun.
- King Cobra can smell blood from far away.
- King Cobra is one of the snakes who create nest or mound.
- The natural predator of king cobra is mongoose which is small mammal.
- Ancient peoples worshiped king cobras some are still doing so.
| 731
|
This coin represent King Eukratides I of Bactria, a land in central Asia (Afghanistan) that was conquered by Alexander. The adoption of the title 'Great King', implying imperial rather than merely royal status, followed some years after Eukratides' consolidation of control over Bactria. He established himself as king of "the thousand cities of Bactria" and set out to enlarge his kingdom in all directions. While returning from one of his campaigns, he was killed by one of his sons, who seized the throne. Bactrian coins were made to the Greek Standard, and this is one of the most beautiful coins of the late Hellenistic period. Known for their strongly realistic portraits, this tetradrachm of Eukratides is no exception. He is depicted as a formidable warrior, wearing a cuirass and a bronze helmet over his diadem. His features are individualized, his expression stern. On the reverse, his parents, Heliocles and Laodice, have been memorialized.
How many hands have touched a coin in your pocket or your purse? What eras and lands have the coin traversed on its journey into our possession? As we reach into our pockets to pull out some change, we rarely hesitate to think of who touched the coin before us, or where the coin will venture to after us. More than money, coins are a symbol of the state that struck them, of a specific time and place, whether contemporary currencies or artifacts of long forgotten empires. This stunning hand-struck coin reveals an expertise of craftsmanship and intricate sculptural detail that is often lacking in contemporary machine-made currencies. This coin is a memorial an ancient king and his kingdom passed from the hands of civilization to civilization, from generation to generation that still appears as vibrant today as the day it was struck. - (LC.025)
| 407
|
Close to standard
|Points of interest||
Ancient Rakatan ruins
Tandun III was originally surveyed approximately 12,293 BBY by Doctor Beramsh, who led an expedition from Ord Mantell. What Beramsh and his team discovered was a lush beautiful world with close to standard gravity and was suitable for both oxygen breathing humans and humanoids. Further exploration of the planet also revealed ruins of ancient population centers which Beramsh believed were constructed by the Rakata.
Despite the survey report of ideal atmospheric conditions, its distance from the Hydian Way prevented Tandun III from ever being colonized. Historical records would indicate that a possible second survey was conducted generations later during Chancellor Finis Valorum's second term.
At some point between 29 BBY and 19 BBY, the Republic Group took an Insignia of Unity that once stood on the podium of the Galactic Senate Rotunda. The group believed the insignia could be used as a symbol to restore Republic honor to the galaxy and brought it to Tandun III for safe-keeping. Enlisting the help of the Antarian Rangers, the Republic Group built a secret storehouse in one of the planet's ancient ruins and made the Stellar Envoy the key to accessing the facility. In 19 BBY the Republic Group and the Antarian Rangers were considered enemies of the Empire and thus the secret facility and the Insignia of Unity were left abandoned and forgotten for over six decades.
During the Yuuzhan Vong's invasion of the galaxy, the Yuuzhan Vong settled the world and began Vongforming it to meet their needs. The land was transformed into cliffs of yorik coral, tampasis of s'teeni, with populations of scherkil hla, sparkbees, and other Yuuzhan Vong biots.
Over the course of twenty years the Vongforming had taken its toll on the planet. Intact forests abundant with life still covered the southern hemisphere, while surface temperatures had rendered landmasses in the northern hemisphere unsuitable for all but the most extremophilic sentient species. Extensive regions of extreme volcanic and tectonic activity griped the planet in catastrophic forces that were likely to destroy it. The sky was filled with powerful winds and icy clouds that produced storms with sheets of rain, lightening, and hailstones. And while still breathable, dweebit beetles filled the atmosphere with high concentrations of carbon dioxide, methane, and sulfur.
In 43 ABY the Solos, accompanied by Tobb Jadak and Flitcher Poste, came to Tandun III as part of their quest to uncover the history of the Millennium Falcon. They landed the Falcon inside of the abandoned Republic Group warehouse and discovered the Insignia of Unity. It was at that time that Lestra Oxic appeared and revealed that he had followed the Solos to Tandun III in order to claim the emblem for himself. He told all present the history behind how the emblem ended up at Tandun III and inspected it only to realize that the emblem before him was a fake.
At the same time of the meeting, groundquakes increased in severity and volatility finally ending in the planet's strongest quake. Oxic and his associates, now accompanied by Jadak and Poste, left to continue the search for the true Insignia of Unity, as the Solos themselves fled the planet.
It was shortly after the Millennium Falcon's return to space that Tandun III flared and erupted in a shock wave that hurtled enormous chunks of itself into the vacuum of space.
Behind the scenesEdit
C-3PO stated the first survey of the world took place around 12,293 BBY and that the expedition was launched from Ord Mantell. This is odd due to the fact that Ord Mantell was not colonized until 12,000 BBY.
Additionally, C-3PO said Dr. Beramsh believed the ancient population centers were Rakatan in origin. Yet, the Rakata were unknown to the galaxy at large until the end of the Jedi Civil War in 3,956 BBY.
He continued to claim that the world was not colonized most likely due to its distance from the Hydian Way which was pioneered in 3,705 BBY. Thus the reason why the world was never colonized in the thousands of years between its exploration and the founding of the Hydian Way remains a mystery.
| 685
|
Hexavalent chromium is a compound used to create pigments and prevent corrosion in dyes, paints, primers, inks and plastics. The chemical is used in the production of stainless steel, leather tanning and wood preservation.
The Hexavalent Chromium Panel was formed in October 2009 to represent former and current users of hexavalent chromium. The panel’s primary activities include sponsoring research to fill the scientific database informing the risk levels for hexavalent chromium in drinking water and communicating the findings of this research.
In the past year, the panel has undertaken an effort to sponsor an independent study. Guided by highly qualified, leading experts, research is being conducted to determine the threshold at which the chemical is not toxic. This data will inform EPA and other policymakers as they develop a drinking water standard for hexavalent chromium.
| 38
|
Photo: James Duncan Davidson
Drawn into controversy
Wearing his wide-brimmed hat, climate scientist James Hansen starts his TEDTalk by asking, ”What do I know that would cause me, a reticent midwestern scientist, to get arrested in front of the White House, protesting?”
Hansen studied under professor James Van Allen, who told him about observations of Venus — there was intense microwave radiation — because it’s hot, and it was kept that way by a thick C02 atmosphere. He was fortunate enough to join NASA and send an instrument to Venus. But while it was in transit, he became involved in calculating what would be the effect of the greenhouse effect here on Earth.
It turns out the atmosphere was changing before our eyes and, “A planet changing before our eyes is more important, it affects and changes our lives.” The greenhouse effect has been understood for a century. Infrared radiation is absorbed by a layer of gas, working like a blanket to keep heat in.
He worked with other scientists and eventually published an article in Science in 1981. They made several predictions in that paper: There would be shifting climate zones, rising sea levels, an opening of the northwest passage, and other effects. All of these have happened or are underway.
That paper was reported on the front page of the NY Times, and led to him testifying to congress. He told them it would produce varied effects, heat waves and droughts, but also (because warmer atmosphere holds more water vapor) more extreme rainfall, stronger storms and greater flooding.
All the global warming ‘hoopla’ became too much, and was distracting him from doing science. In addition, he was upset that the White House had altered his testimony, so he decided to leave communication to others.
The future draws him back in
The problem with not speaking was that he had two grandchildren. He realized he did not want them to say, “Opah understood what was happening, but he didn’t make it clear.”
So he was drawn more and more into the urgency.
Adding carbon to the air is like throwing a blanket on the bed. “More energy is coming in than is going out, until Earth is warm enough to raiate to space as much energy as it recieves from the Sun.” The key quantity is the imbalance, so they did the measurements. It turns out that continents to depths of tens of meters were getting warmer, and the Earth is gaining energy from heat. That amount of energy is equivalent to dropping 400,000 Hrioshima bombs every day, over a year, and there is as much in the pipeline as has already occurred.
If we want to restore energy balance and prevent further warming, we need to reduce the carbon levels from 391 parts per million to 350.
The arguments against
Deniers contend that it’s the sun driving this change. But Hansen notes the biggest change occurred during the low point of the solar cycle — meaning that the effect from the sun is dwarfed by the warming effect.
There are remarkable records in the Earth of what has come before, and we’ve studied them extensively. There is a high correlation between the overall temperature, carbon levels, and sea level. The temperature slightly leads carbon changes by a couple centuries. Deniers like to use that to trick the public. But these are amplifying feedbacks, even through it’s instigated by small effect, a cycle is set up that feed in on itself: More sun in the summer means that ice sheets melt, which means a darker planet, which means more warming. These amplifying feedbacks account for almost entire paleoclimate changes.
The same amplifying feedbacks must occur today. Ice sheets will melt, carbon and methane will be releaseed. “We can’t say exactly how fast these effects will happen, but it is certain they will occur. Unless we stop the warming.”
The view of the future
Hansen presents data showing that Greenland and Antarctica are both losing mass, and that methane is bubbling from the permafrost. That does not bode well. Historically, even at today’s level of carbon, the sea level was 15 meters higher than it is now. We will get least one meter of that this century.
We will have started a process that is out of humanity’s control. There will be no stable shoreline, and the economic implications of that are devastating — not to mention the spectacular loss of speices. It’s possible that 20-50% of all species could be extinct by end of century if we stay on fossil fuels.
Changes have already started. The Texas, Moscow, Oklahoma and other heat waves in recent memory were all exceptional events. There is clear evidence that these were caused by global warming.
Hansen’s grandson Jake is super-enthusiastic, “He thinks he can protect his 2 and a half day old little sister. It would be immoral to leave these people with a climate system spiraling out of control.”
The tragedy is that we can solve this. It could be addressed by collecting a fee for carbon emissions, distributed to all residents. That would stimulate the economy and innovation, and would not enlarge the government. Instead of doing this, we are subsidizing fossil fuels by $400-500 billion per year worldwide.
This, says Hansen, is a planetary emergency, just as important as an asteroid on its way. “But we dither, taking no action to divert the asteroid, even though the longer we wait, the more difficult and expensive it becomes.”
“Now you know some of what I know that is sounding me to sound this alarm. Clearly I haven’t gotten this message across. I need your help. We owe it to our children and grandchildren.”
| 175
|
European Decorative Arts and Sculpture
Arched Entrance from the Chapter House of the Carmelite Convent in Limoges (Les Grands Carmes des Arènes)Made in Limoges, western France, France, Europe
Second half of 13th century
Artist/maker unknown, French
1928-76-1Purchased with funds contributed by Mr. and Mrs. Isaac Tatnall Starr, 1928
LabelThis arched entrance served as one of two passages connecting the contemplative space of the convent's cloister to the communal area of the chapter house (a meeting room for the governing body of the monastery). During the French Revolution (1789–99), the convent—continuously occupied and much expanded since its foundation in the thirteenth century—was closed and the nuns expelled. The convent was then sold and many of the buildings demolished. The chapter house escaped destruction, and in the early twentieth century this arch was rediscovered in its original location and eventually came to the Museum.
Social Tags [?]european art 1100-1500 [x] european decorative arts and sculpture [x] fransiscan monestary too [x] of the child of jesus (carmelite) [x] st. terasea of the little flower [x] the st. francis [x] [Add Your Own Tags]
* Works in the collection are moved off view for many different reasons. Although gallery locations on the website are updated regularly, there is no guarantee that this object will be on display on the day of your visit.
| 245
|
War: The Jacobite Rebellion of 1745
Date: 21st September 1745 (Old Style) (2nd October 1745 New Style). The dates in this page are given in the Old Style. To translate to the New Style (current dating system) add 11 days.
“The Highland Attack”
Place: South East of Edinburgh in Scotland
Combatants: The Highland Army of Prince Charles and the Royal Troops of King George II
Generals: Prince Charles, Lord George Murray and Sir John Cope
Size of the armies: Royal Army: 2,300 men and 6 guns. Highlanders: 2,500 men.
Winner: Prince Charles’ Army
“A contemporary print showing Sir John Cope arriving at Berwick to announce his defeat at Prestonpans”
British Regiments: This battle is not a battle honour for British Regiments. The regiments present at the battle were: Gardiner’s (13th) and Hamilton’s (14th) Dragoons, Guise’s (6th), Lee’s (44th), Murray’s (46th) and Lascelles (47th) Foot
On 25th July 1745 Prince Charles landed near Moidart in the Highlands of Scotland with seven companions. He raised his standard at Glenfinnan and assembled an army from the clans that supported his bid for the throne. This army marched into Edinburgh on 17th September 1745. The two royal dragoons regiments fled at the highland approach in the infamous “Colterbrigg canter”.
General Sir John Cope, the commander of the small royal force in Scotland, had marched to Inverness with his four regiments of foot. Cope brought his troops south to Dunbar by sea and met up with the dragoons. None of his troops, dragoons or foot, were experienced or even adequately trained. Cope’s artillery can only be described as a “scratch” force comprising invalids and seamen under headed by one aged gunner. Cope marched North along the coast road towards Edinburgh.
The cavalry found the rebel army to be inland and to the south, causing Cope to form his army against the sea behind a marsh. During the night of 20th September 1745 the rebels made use of a path through the marsh to come up on the left flank of the royal army.
Cope reformed his line to the left with the foot in the centre, the guns and mortars on their right and dragoon regiments on each end of the line. The highland army launched a charge at which the gunners fled leaving two officers to fire the six guns and six mortars.
The Battle of Prestonpans
On being threatened the dragoon regiments also fled and the foot began to give way. Finally under the impact of the highland attack the whole royal army, other than small groups of men under officers such as Lieutenant Colonel Peter Halkett, fled the field. Only the dragoons were able to get away in any numbers. All the foot bar some 170 were killed, wounded or captured. The injuries inflicted by the highlanders using broad swords and bill hooks are reported to have been horrific.
Casualties: The royal casualties are said to have been: around 300 killed, 400 to 500 wounded and 1,400 to 1,500 captured. Only 170 of the foot got away.
The highlanders probably lost less than 30 killed and 70 wounded.
Following the battle most of Scotland was in Prince Charles’ hands bar Edinburgh Castle held by General Guest and Stirling Castle held by the stalwart General Blakeney.
Regimental anecdotes and traditions:
| 750
|
Statement of Purpose
THE CIVIL WAR CENTER
The Civil War was arguably the most important event in the history of the United States. The conflict fundamentally altered the relationship of the states to the federal government, freed four million slaves, and changed the socio-economic development of the nation. Few states were affected more by this tumultuous period than Georgia. Georgia furnished substantial natural resources, manpower, and political and military leadership to the Southern Republic’s war effort. Important figures such as Vice President of the Confederacy Alexander Stephens, General William J. Hardee, and General James Longstreet led the way in the South’s fight for its independence. Critical battles at Dalton, Resaca, Dallas, New Hope Church, and Kennesaw Mountain were fought only a short distance from our own Kennesaw State University. Georgia’s society – men and women, white and black, slave and free – experienced change on a scale never witnessed before or since. However, despite Georgia’s key role in the war, no research center exists in the state where scholars, students, and interested citizens can go to find out about Georgia’s experience in the nation’s greatest calamity.
The Civil War historian in Kennesaw State University’s Department of History and Philosophy recognized this void and secured the support of the College of Humanities and Social Science to begin developing the Center. After nearly a year of planning and designing, the Civil War Center became a functioning unit within the college. Thus, the mission and purpose of the Center is clear: compile and organize key resources and help educate the public on all aspects of the South's role in the Civil War period.
No other event in American history affected the development of the United States more than the Civil War. The Civil War era is defined broadly to include primarily the nineteenth century, yet events before and after are related to the war as well. The nineteenth century witnessed dramatic social, economic, and political changes that reshaped American life and culture. While the Civil War is central to the experience, the decades before and after contain the seeds and bitter harvest of the national ordeal. The Center will, therefore, include resources on the causes, course, and repercussions of the war within the historic South. This region encompasses the states of Alabama, Arkansas, Delaware, Florida, Georgia, Kentucky, Louisiana, Maryland, Mississippi, Missouri, North Carolina, Oklahoma, South Carolina, Tennessee, Texas, Virginia, and West Virginia.
To summarize, then, the Civil War Center is committed to the following:
WHY KENNESAW STATE UNIVERSITY IS AN
The Atlanta area is an ideal location to house such a center because of its transportation networks, historic sites, and the abounding enthusiasm the Civil War generates among the citizens of the state and region. First and foremost an educational repository, the Civil War Center will offer the following advantages to instructors and students in the area:
KENNESAW STATE UNIVERSITY’S ROLE
Kennesaw State University is the third largest university out of thirty-five institutions in the University System of Georgia. It supports a growing student population of approximately 20,000 from 132 countries. KSU lies in the center of an area replete with key resources, events, and expertise with which to study Georgia's and the historic South's role in the Civil War. Critical battles at Dalton, Resaca, Dallas, New Hope Church, and Kennesaw Mountain were fought only a short distance from our own KSU, making the university a logical site to house the Civil War Center. Enthusiasm for the period abounds among the citizens of the area, and KSU's proximity to Atlanta, the Kennesaw Mountain National Battlefield Park, nearby historic Civil War sites, and interstate access will no doubt attract visitors from across the state, region, and country.
The Center will, in both the present and future, enhance KSU's reputation through the following:
KSU and the College of Humanities and Social Science are committed to supporting the Center through funding from the capital campaign, staffing as needed and budget allows, purchasing of equipment and other materials, and housing for the collection.
Created in 2006, the Center has already generated a flurry of interest and excitement in the area. Fundraising efforts are gaining momentum, and the archival collections are growing rapidly. In its brief tenure, the Center has made great strides in perpetuating its mission to educate the public on the importance of the Civil War to the nation's history.
| 652
|
America's oil and natural gas industry is committed to protecting the environment and to continuously improving its hurricane preparation and response plans. After any hurricane or tropical storm, the goal is to return to full operations as quickly and as safely as possible. For the 2012 hurricane season, the industry continues to build upon critical lessons learned from 2008's major hurricanes, Gustav and Ike, as well as other powerful storms, such as 2005's Katrina and Rita and 2004's Ivan.
API plays two primary roles for the industry in preparing for hurricanes. First, it helps the industry gain a better understanding of the environmental conditions in and around the Gulf of Mexico during hurricane or tropical storm activity and then assists industry in using that knowledge to make offshore and onshore facilities less vulnerable. Second, API collaborates with member companies, other industries and with federal, state and local governments to prepare for hurricanes and return operations as quickly and as safely as possible.
API member companies also independently work to improve preparedness for hurricanes and other natural or manmade disasters. They have, for example, reviewed and updated emergency response plans, established redundant communication paths and made pre-arrangements with suppliers to help ensure they have adequate resources during an emergency.
The API Subcommittee on Offshore Stuctures, the International Association of Drilling Contractors, and the Offshore Operators Committee, serve as a liaison to regulatory agencies, coordinate industry review of critical design standards and provide a forum for sharing lessons learned from previous hurricanes.
These combined efforts are critical since the Gulf of Mexico accounted for about 23 percent of the oil and 8 percent of total natural gas produced in the United States (approximately 82 percent of the oil supply comes from deepwater facilities), and the Gulf Coast region is home to almost half of the U.S. refining capacity.
Upstream (Exploration and Production)
During the major 2005 hurricanes, waves were higher and winds were stronger than anticipated in deeper parts of the Gulf so the industry moved away from viewing it as a uniform body of water. Evaluating the effects of those and other storms, helped scientists discover that the Central Gulf of Mexico was more prone to hurricanes because it acts as a gathering spot for warm currents that can strengthen a storm.
The revised wind, wave and water current measurements ("metocean" data) prompted API to reassess its recommended practices (RPs) for industry operations in the region.
- The upstream segment continues to integrate the updated environmental (metocean) data on how powerful storms affect conditions in the Gulf of Mexico into its offshore structure design standards. This effort led to the publication in 2008 of an update to RP 2SK, Design and Analysis of Stationkeeping Systems for Floating Structures, that provides guidance for design and operation of Mobile Offshore Drilling Unit (MODU) mooring systems in the Gulf of Mexico during the hurricane season. API RP 95J, Gulf of Mexico Jack-up Operations for Hurricane Season, which recommends locating jack-up rigs on more stable areas of the sea floor, and positioning platform decks higher above the sea surface, was also updated.
API publications are available at our (Search and Order
API in the past six years also has issued a number of bulletins to help better prepare for and bring production back online after Gulf hurricanes. These include:
Production and Hurricanes (steps industry takes to prepare for and return after a storm)
- Bulletin 2TD, Guidelines for Tie-downs on Offshore Production Facilities for Hurricane Season, which is aimed at better-securing separate platform equipment.
- Bulletin 2INT-MET, Interim Guidance on Hurricane Conditions in the Gulf of Mexico, which provides updated metocean data for four regions of the Gulf, including wind velocities, deepwater wave conditions, ocean current information, and surge and tidal data.
- Bulletin 2INT-DG, Interim Guidance for Design of Offshore Structures for Hurricane Conditions, which explains how to apply the updated metocean data during design.
- Bulletin 2INT-EX, Interim Guidance for Assessment of Existing Offshore Structures for Hurricane Conditions, which assists owners/operators and engineers with existing facilities.
- Bulletin 2HINS, Guidance on Post-hurricane Structural Inspection of Offshore Structures, which provides guidance on determining if a structure sustained hurricane-induced damage that affects the safety of personnel, the primary structural integrity, or its ability to perform the purpose for which it was intended.
Refineries and Pipelines
- Days in advance of a tropical storm or hurricane moving toward or near their drilling and production operations, companies will evacuate all non-essential personnel and begin the process of shutting down production.
- As the storm gets closer, all personnel will be evacuated from the drilling rigs and platforms, and production is shut down. Drillships may relocate to a safe location. Operations in areas not forecast to take a direct hit from the storm often will be shut down as well because storms can change direction with little notice.
- After a storm has passed and it is safe to fly, operators will initiate "flyovers" of onshore and offshore facilities to evaluate damage from the air. For onshore facilities, these "flyovers" can identify flooding, facility damage, road or other infrastructure problems, and spills. Offshore "flyovers" look for damaged drilling rigs, platform damage, spills, and possible pipeline damage.
- Many offshore drilling rigs are equipped with GPS locator systems, which allow federal officials and drilling contractors to remotely monitor the rigs' location before, during and after a hurricane. If a rig is pulled offsite by the storm, locator systems allow crews to find and recover the rig as quickly and as safely as possible.
- Once safety concerns are addressed, operators will send assessment crews to offshore facilities to physically assess the facilities for damage.
- If facilities are undamaged, and ancillary facilities, like pipelines that carry the oil and natural gas, are undamaged and ready to accept shipments, operators will begin restarting production. Drilling rigs will commence operations.
Despite sustaining unprecedented damage and supply outages during the 2005 and 2008 hurricanes, the industry quickly and safely brought refining and pipeline operations back online, delivering to consumers near-record levels of gasoline and record levels of distillate (diesel and heating oil) in 2008. The oil and oil-product pipelines operating on or near the Gulf of Mexico continue to review their assets and operations to minimize the potential impacts of storms and shorten the time it takes to recover. While there have been some shortages caused by hurricanes, supply disruptions have been temporary despite extensive damage to supporting infrastructure, such as electric power generation and distribution, production shut-ins and refinery shutdowns. Pipelines need a steady supply of crude oil or refined products to keep product flowing to its intended destinations.
To prepare for future severe storms, refiners and pipeline companies have
Refineries and hurricanes (steps industry takes to prepare for and return after a storm)
- Worked with utilities to clarify priorities for electric power restoration critical to restarting operations and to help minimize significant disruptions to fuel distribution and delivery.
- Secured backup power generation equipment and worked with federal, state and local governments to ensure that pipelines and refineries are considered "critical" infrastructure for back-up power purposes.
- Established redundant communications systems to support continuity of operations and locate employees.
- Worked with vendors to pre-position food, water and transportation, and updated emergency plans to secure other emergency supplies and services.
- Provided additional training for employees who have participated in various exercises and drills.
- Reexamined and improved emergency response and business continuity plans.
- Strengthened onshore buildings and elevated equipment where appropriate to minimize potential flood damage.
- Worked with the states and local emergency management officials to provide documentation and credentials for employees who need access to disaster sites where access is restricted during an emergency.
- Participated in industry conferences to share best practices and improvement opportunities.
Pipelines and hurricanes (steps industry takes to prepare for and return after a storm)
- Refiners, in the hours before a large storm makes landfall, will usually evacuate all non-essential personnel and begin shutting down or reducing operations.
- Operations in areas not forecast to take a direct hit from the storm often are shut down or curtailed as a precaution because storms can change direction with little notice.
- Once safe, teams come in to assess damage. If damage or flooding has occurred, it must be repaired and dealt with before the refinery can be brought back on-line.
- Other factors that can cause delays in restarting refineries include the availability of crude oil, electricity to run the plant and water used for cooling the process units.
- Refineries are complex. It takes more than a flip of a switch to get a refinery back up and running. Once a decision has been made that it is safe to restart, it can take several days before the facility is back to full operating levels. This is because the process units and associated equipment must be returned to operation in a staged manner to ensure a safe and successful startup.
- If facilities are undamaged or necessary repairs have been made, and ancillary facilities - like pipelines that carry the oil and natural gas - are undamaged and ready to accept shipments, operators will begin restarting production.
- Pipeline operations can be impacted by storms, primarily through power outages, but also by direct damage.
- Offshore pipelines damaged require the hiring of divers, repairs and safety inspections before supplies can flow. Damaged onshore pipelines must be assessed, repaired and inspected before resuming operations.
- Without power, crude oil and petroleum products cannot be moved through pipelines. Operators routinely hold or lease back-up generators but need time to get them onsite.
- If there is no product put into pipelines because Gulf Coast/Gulf of Mexico crude or natural gas production has been curtailed, or because of refinery shutdowns, the crude and products already in the pipelines cannot be pushed out the other end.
- Wind damage to above ground tanks at storage terminals can also impact supplies into the pipeline.
: The 2008 hurricane season was very active, with 16 named storms, of which eight became hurricanes and five of those were major hurricanes. For the U.S. oil and natural gas industry, the two most serious storms of 2008 were Hurricane Ike, which made landfall in mid-September near Baytown, Texas, and Hurricane Gustav, which made landfall on September 1 in Louisiana.
Hurricane Gustav, a strong Category 2 storm, kept off-line oil and natural gas delivery systems and production platforms that had not yet been fully restored from a smaller storm two weeks earlier, and brought significant flooding as far north as Baton Rouge. Hurricane Ike, another strong Category 2 hurricane, caused significant portions of the production, processing, and pipeline infrastructure along the Gulf Coast in East Texas and Louisiana to shut down. Ike caused significant destruction to electric transmission and distribution lines, and these damages delayed the restart of major processing plants, pipelines, and refineries. As many as 3.7 million customers were without electric power following the storm, with about 2.5 million in Texas alone.
At the peak of disruptions, more than 20 percent of total U.S. refinery capacity was idled. The Minerals Management Service - now called Bureau of Ocean Energy Management, Regulation and Enforcement (BOEMRE)
estimated that 2,127 of the 3,800 total oil and natural gas production platforms in the Gulf of Mexico were exposed to hurricane conditions, with winds greater than 74 miles per hour, from Hurricanes Gustav and Ike. A total of 60 platforms were destroyed as a result of Hurricanes Gustav and Ike. Some platforms which had been previously reported as having extensive damage were reassessed and determined to be destroyed. The destroyed platforms produced 13,657 barrels of oil and 96.5 million cubic feet of natural gas daily or 1.05 percent of the oil and 1.3 percent of the natural gas produced daily in the Gulf of Mexico.
: The 2005 hurricane season was the most active in recorded history, shattering previous records. According to the Department of Energy, refineries in the path of hurricanes Katrina and Rita accounting for about 29 percent of U.S. refining capacity were shut down at the peak of disruptions. Offshore, the Minerals Management Service (MMS) estimated 22,000 of the 33,000 miles of pipelines and 3,050 of the 4,000 platforms in the Gulf were in the direct paths of the two Category 5 storms. Together the storms destroyed 115 platforms and damaged 52 others.
Even so, there was no loss of life among industry workers and contractors. An MMS report found "no accounts of spills from facilities on the federal Outer Continental Shelf that reached the shoreline; oiled birds or mammals; or involved any discoveries of oil to be collected or cleaned up".
: Hurricane Ivan was the strongest hurricane of the 2004 season and among one of the most powerful Atlantic hurricanes on record. It moved across the Gulf of Mexico to make landfall in Alabama. Ivan then looped across Florida and back into the Gulf, regenerating into a new tropical system, which moved into Louisiana and Texas.
The MMS estimated approximately 150 offshore facilities and 10,000 miles of pipelines were in the direct path of Ivan. Seven platforms were destroyed and 24 others damaged. The oil and natural gas industry submitted numerous damage reports to MMS, including for mobile drilling rigs, offshore platforms, producing wells, topside systems including wellheads and production and processing equipment, risers, and pipeline systems that transport oil and gas ashore from offshore facilities.
| 682
|
Discover the cosmos! Each day a different image or photograph of our fascinating universe is featured, along with a brief explanation written by a professional astronomer.
2012 June 23
Explanation: As seen from Frösön island in northern Sweden the Sun did set a day after the summer solstice. From that location below the arctic circle it settled slowly behind the northern horizon. During the sunset's final minute, this remarkable sequence of 7 images follows the distorted edge of the solar disk as it just disappears against a distant tree line, capturing both a green and blue flash. Not a myth even in a land of runes, the colorful but elusive glints are caused by atmospheric refraction enhanced by long, low, sight lines and strong atmospheric temperature gradients.
Authors & editors:
Jerry Bonnell (UMCP)
NASA Official: Phillip Newman Specific rights apply.
A service of: ASD at NASA / GSFC
& Michigan Tech. U.
| 774
|
Moderating your food consumption is the most important component of a healthy, balanced diet. Incorporating concepts such as food combining and food rotation can also be beneficial when addressing poor digestion or food allergies. Following these simple guidelines below will help support your digestion and a strong immune system.
Eat Whole Natural Foods:
Whole natural foods that come direct from gardens, farms and orchards are the best choices for optimal nutrients, energy and vitality. Fresh fruits, vegetables, nuts, seeds, whole grains and legumes should ideally constitute 60-70% of your diet. Although processed food has an undeniable convenience factor and longer shelf life, they typically contain unnecessary additives and chemicals that can be detrimental to your health in the long run.
While not always the budget-friendly option, there are many benefits to eating organic whenever possible. Many believe that organically grown foods have a higher nutrient value than commercially grown produce, while others disagree. There are many studies to prove either side. What cannot be disputed is the undeniable benefit associated with the minimization of chemicals sprayed on organically grown produce, and therefore the reduced toxic load our bodies have to deal with.
Eat Naturally-Raised Meats (Fish, Seafood, Poultry, Beef, Lamb and Eggs):
Protein is an important component in diets, as it is the building block of all of our cells – and therefore our muscles, tissues and organs. And it is a tremendous source of vitamin B12 and iron. Due to its high fat content, it is important to consider only lean sources of protein. And to minimize our chemical and hormone exposure, look for naturally raised sources of protein.
Eat Raw or Lightly Steamed Vegetables:
Fruits and vegetables are best eaten as fresh as possible, though many can be stored well for several weeks. To maximize the nutrient value of fruits and vegetables, it’s best to eat them raw or lightly steamed. When cooked, many of the vitamins, some minerals and certainly the enzymes in fruits and vegetables are lost, thereby reducing their nutrient value.
Eat in Moderation:
As a general guideline, every meal and snack should have 40% lean protein, 20% fat and 40% complex carbohydrates for optimum nutrition and satisfaction. One of the best things you can do is to always eat a healthy balanced breakfast, setting up your metabolism properly for the day. If you have time constraints, consider options such as protein smoothies, healthy muffins, hardboiled eggs or granola with fruit, nuts and seeds.
Eliminate or Minimize Caffeine:
Caffeine is a stimulant that helps us feel temporarily energized, but it has many negative effects to the body. It has a laxative effect that may cover up sluggish bowels; it’s also a diuretic that causes frequent urination and dehydration, which can lead to many physical symptoms if taken in excess, including insomnia, high blood pressure and headaches.
Use Natural Sweeteners in Moderation:
Generally, it’s best to minimize sugar due to its well-documented impact on health, including hypoglycemia, weight gain, cravings, high cholesterol and dental cavities to name a few. When a sweetener is required, consider natural sources such as raw honey and maple syrup instead.
Eliminate or Limit Alcohol:
The negative effects of excessive alcohol use are well documented. Even in moderate amounts, alcohol is a diuretic and also uses up B vitamins as it is processed from the body, so it’s best to eliminate or strictly limit consumption.
Get Plenty of Sleep:
The importance of sleep is often overlooked. During sleep, our muscles and tissues are repaired and revitalized, and our minds process experiences and things we have learned during the day. Although we all have different sleep patterns and needs, the average person requires a minimum of seven to eight hours of uninterrupted sleep per night.
To achieve optimal sleep, avoid consumption of caffeine, heavy meals, exercise or alcohol within a few hours before bed. Before bed, consider reading, journaling or meditation to relax, as the day’s anxieties and an overactive mind can also impact your ability to sleep. And finally, ensure that your sleeping environment is set up for the best quality of sleep – minimize noise and keep the bedroom as dark as possible.
Exercise – Move your Body!:
Many feel that exercise only includes weekly classes, trips to the gym or running. The reality is that any form of exercise is beneficial, including morning or evening walks, or even taking the stairs whenever possible. Ideally, find an activity you enjoy doing and do it as frequently as possible. If finding the time becomes a challenge, use simple methods such as parking your car farther away from your destination and even taking business meetings outdoors and making them “walking meetings”. The key is to find every opportunity in a day to get your body moving.
| 154
|
Inflammation Key to the Asthma-Sinus Connection
People with asthma frequently experience problems with their sinuses. And more than half of those who have chronic sinusitis also suffer from asthma.
Is there a connection? Each condition is marked by inflammation, which is to blame for both the symptoms of asthma—cough, chest tightness, shortness of breath, and wheezing—and sinusitis. Researchers speculate inflammation in the lungs or sinuses affects both. As a result, people with lung symptoms likely will eventually get symptoms in the nose, and vice versa. Plus, sinusitis can trigger asthma attacks.
What is sinusitis?
Sinusitis occurs when air-filled spaces behind the nose, forehead, cheeks, and eyes become inflamed and blocked with mucus. Often, infection results.
Symptoms usually occur after a cold that fails to improve or gets worse after five to seven days. They include:
People with asthma who experience sinus problems should talk with their doctor about treatment. Studies show that resolving sinusitis often improves asthma and decreases the need for asthma medication.
How you can protect yourself
Those with asthma need to be vigilant about colds and the flu. Viral respiratory infections often worsen asthma. Some prevention tips:
Avoid smoke and pollution.
Drink plenty of fluids.
Take decongestants for upper respiratory infections.
Get help for allergies.
Use a humidifier to increase moisture in nose and sinuses.
| 287
|
Carbon Monoxide Detector Placement - Carbon Monoxide - Carbon Monoxide Poisoning
Proper placement of a carbon monoxide (CO) detector is important. If you are installing only one carbon monoxide detector, the Consumer Product Safety Commission (CPSC) recommends it be located near the sleeping area, where it can wake you if you are asleep. Additional detectors on every level and in every bedroom of a home provides extra protection against carbon monoxide poisoning.
Homeowners should remember not to install carbon monoxide detectors directly above or beside fuel-burning appliances, as appliances may emit a small amount of carbon monoxide upon start-up. A detector should not be placed within fifteen feet of heating or cooking appliances or in or near very humid areas such as bathrooms.
When considering where to place a carbon monoxide detector, keep in mind that although carbon monoxide is roughly the same weight as air (carbon monoxide's specific gravity is 0.9657, as stated by the EPA; the National Resource Council lists the specific gravity of air as one), it may be contained in warm air coming from combustion appliances such as home heating equipment. If this is the case, carbon monoxide will rise with the warmer air.
Installation locations vary by manufacturer. Manufacturers' recommendations differ to a certain degree based on research conducted with each one's specific detector. Therefore, make sure to read the provided installation manual for each detector before installing.
CO detectors do not serve as smoke detectors and vice versa. However, dual smoke/CO detectors are also sold. Smoke detectors detect the smoke generated by flaming or smoldering fires, whereas CO detectors can alarm people about faulty fuel burning devices to prevent carbon monoxide poisoning. Carbon monoxide is produced from incomplete combustion of fossil fuels. In the home CO can be formed, for example, by open flames, space heaters, water heaters, blocked chimneys or running a car inside a garage.
Since CO is colorless, tasteless and odorless (unlike smoke from a fire), detection and prevention of carbon monoxide poisoning in a home environment is impossible without such a warning device. In North America, some state, provincial and municipal governments require installation of CO detectors in new units - among them, the U.S. states of Illinois, Massachusetts, Minnesota, New Jersey, and Vermont, the Canadian province of Ontario, and New York City.
According to the 2005 edition of the carbon monoxide guidelines, NFPA 720, published by the National Fire Protection Association, sections 184.108.40.206 and 220.127.116.11, all CO detectors 'shall be centrally located outside of each separate sleeping area in the immediate vicinity of the bedrooms,' and each detector 'shall be located on the wall, ceiling or other location as specified in the installation instructions that accompany the unit.'
When carbon monoxide detectors were introduced into the market, they had a limited lifespan of 2 years. However technology developments have increased this and many now advertise 5 or even 6 years. Newer models are designed to signal a need to be replaced after that timespan although there are many instances of detectors operating far beyond this point.
Although all home detectors use an audible alarm signal as the primary indicator, some versions also offer a digital readout of the CO concentration [see side-bar], in parts per million. Typically, they can display both the current reading and a peak reading from memory of the highest level measured over a period of time.
The digital models offer the advantage of being able to observe levels that are below the alarm threshold, learn about levels that may have occurred during an absence, and assess the degree of hazard if the alarm sounds. They may also aid emergency responders in evaluating the level of past or ongoing exposure or danger of carbon monoxide poisoning.
Battery-only carbon monoxide detectors tend to go thru batteries more frequently than expected. Plug-in detectors with a battery backup (for use if the power is interrupted) provide less battery-changing maintenance.
Some CO detectors are available as system-connected, monitored devices. System-connected detectors, which can be wired to either a security or fire panel, are monitored by a central station. In case the residence is empty, the residents are sleeping or occupants are already suffering from the effects of CO, the central station can be alerted to the high concentrations of CO gas and can send the proper authorities to investigate possible carbon monoxide poisoning.
Page 1 of 3: Next Page »
Carbon Monoxide Detectors w/ PPM Displays and Battery Backup
Although all home carbon monoxide detectors use an audible alarm signal as the primary indicator, some versions also offer a digital readout of the CO concentration, in parts per million.
Typically, they can display both the current reading and a peak reading from memory of the highest level measured over a period of time.
The digital models offer the advantage of being able to observe levels that are below the alarm threshold, learn about levels that may have occurred during an absence, and assess the degree of hazard if the alarm sounds.
They may also aid emergency responders in evaluating the level of past or ongoing exposure or danger.
| 705
|
Find More Accuracy in the Treatment of Brain Tumor by Stereotactic Radiosurgery
Stereotactic radiotherapy (also called stereotaxy) is a type of minimally invasive surgical intervention. It uses radiation treatment to deliver a large and precise radiation dose to the tumor area in a single session. This type of surgery applies a three-dimensional coordinates system to locate small targets inside the body. With the help of gamma knife, radiation beams are focused at the tumor from various angles for a short period of time. Stereotactic radiosurgery (SRS) can be performed by several other machines such as X-Knife, CyberKnife and Clinac. It is used to treat brain tumors and other brain disorders that are difficult to be treated by regular surgery.
This type of surgery works under three main components. The first one is stereotactic planning system; it includes multimodality image matching tools. The stereotactic atlas is a series of cross sections of anatomical structure. In most atlases, the three dimensions are as latero-lateral (x), dorso-ventral (y) and rostro-caudal (z). Another one is a stereotactic device or apparatus which uses a set of three coordinates (x, y and z). And the last one is stereotactic localization and placement procedure. This therapy works in the same way as other forms of radiation treatment. Instead of removing the tumor, stereotactic radiotherapy damages the DNA of tumor cells. In this way malignant and metastatic tumors may shrink more rapidly.
Stereotactic radiosurgery can effectively treat different types of tumors including benign, malignant, primary, metastatic, single and multiple. A study in 2008 by The University of Texas M. D. Anderson Cancer Center revealed that SRS and Whole brain radiation therapy (WBRT) for the treatment of metastatic brain tumors have more that twice the risk of developing leaning and memory problems than treated with SRS alone.
| 17
|
A large-scale screen in Drosophila cells has shown how RNA interference can provide insights into the pathways controlling cell morphology.
When thinking of genetic screens in Drosophila, the Nobel-prize-winning screens for embryonic developmental phenotypes spring to mind. Loss-of-function mutational analysis has proved to be a powerful approach for dissecting complex processes such as morphogenesis and wing development. But historically the fruitfly community has focused much less on the study of cells in culture. All that may now be about to change. In this issue of Journal of Biology , Amy Kiger and colleagues describe the results of a powerful strategy for conducting high-throughput systematic loss-of-function screens in Drosophila tissue-culture cells (see 'The bottom line' box for a summary of the work).
Picking apart signaling pathways
The study was led by Norbert Perrimon, a Howard Hughes Medical Institute investigator who runs a Drosophila laboratory at Harvard Medical School in Boston, USA. Perrimon has focused his lab on developing large-scale genetic screens to dissect signal transduction pathways by analysis of mutant phenotypes. The Perrimon lab has a collection of hundreds of mutants with interesting developmental defects, each one waiting for a creative postdoc to track down the gene and work out what it does. "One can assume that genes with similar phenotypes are related and act in the same pathways," says Perrimon. This approach had allowed the lab to define groups of mutants, which have then been used to dissect conserved signal transduction pathways, such as those involving the Wnt family of intercellular signaling molecules and the Jak-Stat intracellular signal transducers.
Completion of the Drosophila melanogaster genome-sequencing project three years ago encouraged the fly community to explore new approaches. "Now we had 16,000 pieces of the puzzle, and working out what they each did was going to be an enormous challenge," says Perrimon. When RNA-interference (RNAi) technology appeared on the scene as a new way of inactivating genes whose sequence is known (see the 'Background' box), it seemed an appealing way to speed up the task of dissecting complex genetic pathways: genes could be systematically inactivated and loss-of-function phenotypes studied on a large scale. But applying the RNAi approach - which had worked so spectacularly in Caenorhabditis elegans [3,4] - to Drosophila proved not to be so straightforward, and early attempts at injecting interfering RNAs into embryos were plagued by technical problems and inconsistent results. It was not until Jack Dixon's group at the University of California in San Diego showed that gene silencing by RNAi works well in fly cells in culture that the stage was set for a screen on the scale that fly geneticists are used to.
"The cell biology community has a real 'nuts and bolts' view of the way that nature works," says Buzz Baum, who was a postdoctoral fellow in the Perrimon group. "Lots of cell biologists were beginning to use RNAi [when we began this work], but they still hadn't learnt to think as geneticists. What I wanted to do when I joined Norbert's lab was to take genetics to cell biology." So, he and Amy Kiger began to set up the methodologies for a large-scale screen. "We began exploring ways in which RNAi in cell culture could be used in a much more far-reaching way," recalls Amy Kiger, a postdoc in the Perrimon lab and lead author on the study.
Screening for cell morphology
"We needed to set up a pilot screen to work out the methodologies," says Perrimon. "We were interested in genes involved in morphogenesis, so we decided to focus on changes in cell morphology and genes that regulate the cytoskeleton." They hoped that some of the genes identified would be on the same pathways as those that they had already studied in flies. The Perrimon group was fortunate to be surrounded by a number of laboratories with experience of cell-based screening and robotic technologies, including automated microscopes.
"The thing that can be difficult about taking an assay to high throughput is that the way that the cells and microscopes behave can be quite different when you scale up," says Baum, who now runs his own laboratory at the Ludwig Institute for Cancer Research in London, UK. "The most important thing was finding cell lines where we could study the sort of things that we were interested in. We had to learn quite a bit about fly cell culture. We got cell lines from all over the world and tested them with thirty-odd RNAs. We got all these specific cellular phenotypes - it was really quite magical!" They selected two hemocyte cell lines for the screen; these were chosen because of their different shapes - Kc167 cells are small and round, whereas the S2R+ cells are large, flat and strongly adherent (Figure 1). It seemed likely that defining the sets of genes that generated the cell shape in these two cell lines would give clues about basic cell morphology and fly morphogenesis.
Perrimon and colleagues then scaled up operations by constructing a library of double-stranded RNAs (dsRNAs) representing around a thousand genes that could be used as a 'tool kit' for exploring cell morphology phenotypes; the kit included genes thought to be involved in regulating the cytoskeleton, such as putative GTPases and kinases. The library of dsRNAs was introduced into cells in 384-well optical-bottom plates, to allow image-based screening. The team recorded all visible changes three days after introduction of the RNA, using automated microscopy. Around 16% of the genes gave visible phenotypes in at least one of the two cell lines.
The group then had to work out a formalized process for dealing with all the information that came out of the screen. After familiarizing themselves with the range of phenotypes that appeared, they developed a system of 'phenotypic annotation' to record the effects of silencing each one of hundreds of genes, starting out by defining seven major phenotypic categories on the basis of defects in actin filaments, microtubules, DNA, cell shape, cell size, cell number and cell viability. Each class was then further subdivided into a number of categories that describe specific morphological attributes; for example, changes in cell shape were categorized using descriptions such as round or flat, retracted, bipolar, spiky or stretchy. Obviously these phenotypes overlap in many cases, so that the effects of each RNAi can be defined in terms of its phenotypic profile. Christophe Antoniewski, at the Institut Jacques Monod in Paris, notes that "the use of a 'phenotype matrix' provides an opportunity for statistical analysis of the results using clustering approaches on qualitative traits instead of quantitative traits." As with all genetic screens, genes with similar profiles might be expected to participate in common morphogenetic functions. For example, RNA specific for the Rho1 GTPase and the Pebble Rho-GTP exchange factor, as well as for an uncharacterized predicted kinase (CG10522), all generated enlarged, multinucleated cells, reflecting defects in cytokinesis; Kiger et al. suggest that CG10522 may be a novel effector, downstream of Rho1 and with a role in cytokinesis.
The most commonly seen phenotypes were changes in actin organization and cell shape. Several of the genes identified are thought to limit the rate of formation of filamentous (F) actin and encode proteins with F-actin capping functions. Some of these regulators seemed to play similar roles in both the adherent or non-adherent fly cell lines. But by comparing the adherent S2R+ cells with the round Kc167 cells, Kiger et al. were able to study genes involved in maintaining a particular (round or flat) cell shape. They expected to identify a set of genes that were differently expressed in the two cell lines and were responsible for their distinct morphologies. Indeed, they found that 78% of the morphological RNAi phenotypes were seen in only one of the cell types. For example, dsRNAs targeting genes involved in actin filament formation caused Kc167 cells to flatten, whereas dsRNAs for genes involved in cell-matrix adhesion functions caused S2R+ cells to round up and detach. Analysis of the levels of expression of the integrin adhesion receptors, which are known to mediate cell-matrix adhesion, showed that they are not expressed at lower levels in Kc167 cells, although the expression of the cytoskeletal component talin was significantly lower. The authors conclude that both integrin-mediated adhesion and reorganization of the cortical F-actin network are necessary to determine cell spreading.
The Perrimon group was very satisfied to see how successfully this pilot screen worked. Having established the 'proof of principle' they have now scaled up to perform whole-genome screens in Drosophila cells in culture (see the 'Behind the scenes' box for more on the development of the work). "It's really changing completely the way we do science," says Perrimon enthusiastically. "Everything now comes down to the assays that we design. You can be as imaginative as you want. We need to do more and more screens and to compile databases of annotated information so that we can build correlations between components of different pathways. The long-term experiment is to see how much of the complexity in the cell can actually be reduced by finding these kinds of correlations."
Others in the field agree. "It's exciting to see RNAi technology in cultured cells being adapted to a high-throughput format and being used to screen for genes involved in cell morphogenesis and cytoskeletal function," says Matthew Welch, a cell biologist working at the University of California, Berkeley. "Although the classical genetic approach has been used very successfully for many years to study cell morphogenesis in yeast and other organisms, the ability to systematically silence a large number of genes in cultured animal cells using RNAi has brought this approach to new systems and will help answer outstanding questions in new ways." And Antoniewski concurs. "This type of approach is fascinating because it actually combines both 'forward' and 'reverse' genetics in a single screen. Kiger et al. have coupled systematic knock-down technology to a phenotypic screen, so systematic 'reverse genetics' is guided by the functional assay. This is new and important." Baum predicts that deciphering the meaning of such large datasets will probably require new ways of thinking and the help of mathematicians and physicists. These approaches are being incorporated into the broad field of 'systems biology'.
To help the application of this technology throughout the fly and cell biology communities, Perrimon and collaborators have set up the Drosophila RNAi Screening Center (DRSC) , to make the relatively expensive and sophisticated technology accessible to all interested researchers; it also has the advantage of permitting valuable functional comparisons across many studies and the creation of an information database with a standardized format. The DRSC is currently performing screens for cell growth and viability and is accepting applications for potential collaborative screens. "We may need many different assays to dissect a single pathway," says Perrimon. He is interested in any proposal with a good cell-based assay. The Center is able to perform a whole-genome screen in just one week. But at a cost of US$10,000 per screen the DRSC will be carefully selecting screens with a high chance of success.
The cell biology community is keen to use the RNAi approach to probe basic questions about cell functions. "This type of approach, like classical genetics, is very flexible and can be applied to a wide range of cellular processes," says Welch. "What's also exciting is that open access to information from large-scale efforts like this one will be extremely useful for other researchers interested in a range of issues, from the function of their favorite gene or protein to more global issues of how gene function is integrated and coordinated during complex processes like cell morphogenesis." And as Antoniewski points out, "in a third of cases, an RNAi phenotype identified a previously uncharacterized gene that lacked a corresponding mutant allele in Drosophila. This is extremely important, because it shows that RNAi screens can identify new functions that could not have been identified by classical genetics using mutagenesis." An additional aspect of the potential for this type of approach is highlighted by Kiger. "Now that large-scale reverse-genetic or 'functional genomic' approaches are possible in yeast, worms and fly cells, it is interesting to consider how we might eventually be able to make functional comparisons across species that could shed light on common (or contrasting) cellular mechanisms throughout evolution."
Other groups are using RNAi technology in a more focused way to tackle biological questions. Ronald Vale's group at the University of California in San Francisco has used RNAi targeting specific gene families to investigate cytoskeletal function in Drosophila S2 cells. They targeted dynein and all 25 Drosophila kinesins to investigate the role of molecular motors during mitosis , and also identified sets of genes important for regulating actin dynamics during lamella formation . Rogers et al. suggest, "If proper cues are provided to these cells, cell migration and cell polarity may be amenable to investigation as well." A combination of genome-wide screens and studies focused on specific gene families will be needed to identify and then characterize components of the specific pathways involved in complex cellular processes. It will also be interesting to apply the tricks of the classic geneticists' trade, such as screens for modifiers and suppressors.
"I think the power of the RNAi screens will come when we combine RNAi screening data with data from other sources, like genomic sequence, microarray data, proteomic data, and so on," adds Baum. "This opens up the possibility of real systems biology: generating a more global understanding of the logical circuitry that underlies cell behavior. This approach is going to have a big impact - because I think that a lot of the apparent complexity in cell biology is a product of over-expression studies, and loss-of-function data can clarify some of these situations."
"And then we will have to go back to the fly," says Perrimon. "The next step will be to take what we have learned from the cell-based assays and validate them in vivo, either using defined existing mutations or by trying to generate them using classical methods or a promising gene-knockout methodology. Or we can also express [RNA] hairpins, which basically give an in vivo RNAi effect." Many of the lessons learnt by classical fly genetics have subsequently been confirmed and explored further in cellular systems. Now it looks as though researchers can begin by doing genetics in cells and then return to the whole fly to investigate further.
Gonczy P, Echeverri G, Oegema K, Coulson A, Jones SJ, Copley RR, Duperon J, Oegema J, Brehm M, Cassin E, et al.: Functional genomic analysis of cell division in C. elegans using RNAi of genes on chromosome III.
| 127
|
The election of Egypt’s president Muhammad Mursi momentarily threw a spotlight on the long-forgotten Palestinians exiled to Gaza after the Israelis’ infamous siege of the Church of the Nativity in Bethlehem in the West Bank 10 years ago.
Still kicking their heels in Gaza the exiles called on Mursi to continue efforts to end the squabble between political rivals Hamas in Gaza and Fatah in the West Bank. They were optimistic that the new Egyptian president would work towards easing Israel’s blockade on the Gaza Strip and press for Palestinians rights, including the right of exiles to return home.
It is expected Mursi will at least allow greater freedom to travel across Gaza’s Rafah crossing into Egypt, the besieged enclave’s only door to the outside world.
How did the exiles find themselves in the prison Gaza has become? In 2002 a young girl from a refugee camp triggered events that led to a 40-day siege of the Church of the Nativity in Bethlehem. This is probably the oldest Christian church in the world, built by Constantine the Great and dating from AD330. A member of the girl’s family had been killed by Israeli occupation troops. Grief-stricken, she took revenge by turning herself into a suicide bomber.
The Israelis responded by sending 250 tanks and armoured personnel carriers, F-16 fighter jets, Apache gunships and hundreds of soldiers into West Bank towns like Nablus, Jenin and Bethlehem late at night. In Bethlehem they cut the electricity supply and invaded the old township with helicopter gunships and occupied all key points around Manger Square. Many innocent Palestinians were killed by shelling and army snipers, and the market and some shops were set on fire as troops tried to hunt down suspected ‘fighters’. Civilians tried desperately to hide from the troops and a large number of people took refuge or arrived for other reasons at the Church and found themselves trapped, unable to leave.
A few years ago I interviewed one of the survivors, who recalled that “248 took refuge there. They included 1 Islamic Jihad, 28 Hamas, 50 to 60 Al-Aqsa Martyrs. The remainder were ordinary townsfolk and included 100 uniformed Palestinian Authority workers, also 26 children and 8 to10 women and girls. The Israeli soldiers would not allow them to leave, but they escaped in the first week by a back door.”
Priests and nuns – Armenian, Greek and Catholic – from the adjoining monasteries brought the number to over 300 at the beginning. “Some of them went back to the monasteries but some stayed with us every day for the 40 days.”
The Vatican was outraged. The Greek Orthodox Church of Jerusalem called on Christians worldwide to make the upcoming Sunday a “solidarity day” for the people in the Church and the Church itself, and urged immediate intervention to stop what it called the “inhuman measures against the people and the stone of the Church”.
The Israelis set up cranes on which were mounted robotic machine-guns under video control. According to eye-witnesses eight defenders, including the bell-ringer, were murdered, some by the armchair button-pushers playing with their video joysticks and some by regular snipers.
From the start, said my survivor, the Israeli troops used psychological warfare methods – for example, disorienting noise to deprive them of sleep, bright lights and concussion grenades. They paraded the families of the besieged in front of the Church to pressure them to surrender. They also used illegal dum-dum bullets which cause horrendous wounds and trauma. “Most of those who were killed… it was because of the dum-dums… so much bleeding, and it took so long to arrange to send them to a hospital.”
He said the soldiers fired tracer rounds into two of the monasteries and set the ancient fabric of the buildings alight.
15 days into the siege those inside managed to recharge their cellphones using the mains that supplied the Church towers and call for help. The Israelis had overlooked the fact that this was a separate supply coming from the Bethlehem municipality. Friends responded by sending food to the medical centre. From there it went by ambulance, along with authentic casualties, and was delivered to houses near the Church. At night young girls carried the food in plastic bags from house to house until supplies reached the dwellings nextdoor to the Church. The bags were then thrown from roof to roof. This went on for 6 days until one girl dropped a bag, which the soldiers found. The Israelis, now alerted, shot and paralysed another young man. It put an end to the food operation.
“Inside the Church we vowed not to harm the soldiers unless they actually broke in. When soldiers did gain access and killed one of the resisters, 4 of them were shot.”
Those trapped inside the Church were surprised to discover an old lady living within the complex. She had a small horde of olives and wheat, with which they made bread. So they managed to eke out the food for 28 days.
The Governor of Bethlehem and the Director of the Catholic Society were among those holding out in the Church. According to my survivor’s first-hand account, those inside only opened the door if someone died or was injured. He recalled watching through a peephole and seeing people approaching across the forecourt. “They were from the Peace Movement, 28 of them. By now the world media were watching. 17 were arrested but 11 took a big risk, managing to bluff their way in and bringing food in their rucksacks, which lasted another 4 days, and basic medicines.
The worst time, he said, was the final week – no food and only dirty water from the well. They resorted to boiling leaves and old chicken legs into a soup. He ate only lemons and salt for 5 or 6 days. “Many were so ill by this time that they were passing blood.”
Outside some 15 civilians had been indiscriminately shot in the street or in their homes. The Israelis refused to allow the dead in the Church to be removed for decent burial. “In the end, the Governor decided it was better to be in jail than die. So we opened the door and surrendered on the 40th day. 148 had survived. We were promptly arrested and interrogated.
“13 were exiled to the EU, 26 were exiled to Gaza, 26 were wounded, 26 had surrendered because they were under-age. 8 were killed inside the Church, and with Samir (the bellringer) makes 9. They shot Samir in front of the Church as he came out to surrender.”
The rest were allowed home, including my survivor. “The Israelis said to me, ‘Do you know why you are going home? Because America wants it’.” The adverse publicity had prodded the CIA and EU into taking a hand in deciding the fate of the survivors.
The whole disgraceful episode would no doubt have ended in more carnage if the world’s media hadn’t tuned in and ten international activists, including members of the International Solidarity Movement, hadn’t managed to enter the Church.
I hear that the exiles have not been allowed to work since or receive visits from their families. According to some reports they were not even allowed to say goodbye to their loved ones before being packed off.
What exactly were they guilty of? They may have been Palestinian gunmen but the last time I checked it was perfectly OK to put up armed resistance against an illegal military occupation. Israel’s gunmen happen to wear uniform and are equipped with the best weaponry American tax dollars can buy. They are fond of saying, “We have a right to defend ourselves.” So do the Palestinians. Obviously.
So why did America and the EU lend themselves to this shameful act of exiling… a helpful little boost to Israel’s ongoing programme of ethnic cleansing of the West Bank?
And having got their hands dirty isn’t it time, after 10 years, they cleaned up and insisted that these forgotten men be re-united with their families?
A few weeks ago the Israeli press was practising their usual distortions and telling readers that “the terrorists took shelter in the famous church, and used about 40 priests and nuns as a shield, knowing Israel would not take a chance on inadvertently hurting priests and nuns”.
But for Israel’s gunslingers it had been open season on bellringers and other innocents.
(Stuart Littlewood / www.eurasiareview.com / 30.06.2012)
Filed under: Terrorism | Tagged: Palestine, Revolution, Terrorism | Leave a Comment »
| 322
|
Molecular Imaging: Seeing Cancer in Action Inside the Body
Oct. 8, 2009 – Molecular imaging can reveal how cancer cells communicate, how they move, what they need to survive and where they are vulnerable to attack. In this edition of Breakthroughs in Cancer Research, David Piwnica-Worms, MD, PhD, professor of developmental biology and radiology at Washington University School of Medicine, explains how molecular imaging is different from more traditional radiology techniques and discusses new discoveries coming out of the medical school’s Molecular Imaging Center.
Other Siteman web resources.
TRANSCRIPT OF AUDIO FILE
On this edition of Breakthroughs in Cancer Research, we’ll talk about how scientists are making it possible to see cancer in action inside the body. That’s providing new information on how to destroy cancer cells.
Host: Thanks for downloading this podcast from the Siteman Cancer Center at Barnes-Jewish Hospital and Washington University School of Medicine in St Louis. I’m Gwen Ericson. With techniques collectively called molecular imaging, scientists are finding new ways to detect cancer early, monitor cancer treatment and develop better cancer drugs. Molecular imaging can reveal cancer’s secrets, showing how cancer cells communicate, how they move, what they need to survive and where they are vulnerable to attack. To tell us more is David Piwnica-Worms. He is director of the Molecular Imaging Center at the Mallinckrodt Institute of Radiology and the Siteman Cancer Center. Dr. Piwnica-Worms, thank you for joining us.
Piwnica-Worms: Well thank you, Gwen. I’m happy to talk with you today.
Host: Most of us have had experience with traditional kinds of medical imaging – things like X-rays, PET scans and ultrasound scans. But we’ve probably never encountered molecular imaging. Tell us about the distinction between traditional medical imaging and molecular imaging.
Piwnica-Worms: In most of the traditional medical imaging that we may have experienced as patients or with family members, the focus of the imaging is on the anatomy and, in the context of cancer, on the mass or the size of the mass. But in molecular imaging, we’re actually interested in the details of the molecules that are inside the cancer. That is the biology or the biochemistry or metabolism of the cancer. We’re actually trying to image changes in those molecules and biochemistry that may even precede changes in the anatomy.
Host: Give me an example of a molecular imaging technique.
Piwnica-Worms: One strategy is in the general area that we call genetically encoded reporters. And that means we may take some genetic material from one organism and transfer it into another that may have favorable imaging properties. One such example that we make common use of is firefly luciferase. That is the same firefly that we see flying around in our backyards on a July day in St Louis. And the gene that actually makes the fireflies glow is called luciferase. We can actually use standard molecular biology and cloning techniques to transfer that luciferase gene from fireflies into a cancer cell, and we can watch the cancer glow and move around the body. Or it may tell us how it glows in response to different genes being expressed, or maybe how a protein, one of the building blocks, is processed. And those signals can tell us something about the metabolism and the gene expression inside those cancer cells using noninvasive imaging cameras.
Host: Tell me more about what techniques like that reveal about cancer.
Piwnica-Worms: One way to think of this is that cancer is like a baseball game where we don’t understand the rules. And you can imagine if you were taking a snapshot of Pujols at bat in the first inning, and then sometime later in the third inning you had a snapshot of Holliday running around a base path. And then at some other time, you had a snapshot of Ankiel making a long throw to third base from the outfield. What are the underlying rules that allow that to happen? If you knew nothing about the game of baseball and you had to try to put these together, dynamic views, continuous videos and movie loops would allow you to begin understanding those rules of why sequential events occur in a more dynamic and easy way than a series of random snapshots. And we hope that analogy sort of lets you know how we can look dynamically, continuously, over time with these noninvasive molecular imaging strategies at the baseball game of cancer that’s going on inside the cells.
Host: In medical terms, what can molecular imaging do for us?
Piwnica-Worms: A lot of the activities do relate to preclinical and basic type of cancer biology work. These help us better understand the cancer pathways, the rules that guide the metabolism of the cancer cell. These represent the new targets of tomorrow, where new drugs and combinations of drugs can actually help patients. And in some special cases, maybe even in a human, we can tag the drug and understand, is it actually hitting the cancer? Is it binding to its molecular target? And we can image that noninvasively and repetitively over time. And in another area that relates more directly to patients, there are some techniques like fluorodeoxyglucose (FDG) PET. And this is a special subset of molecular imaging that is available today that does look at sugar metabolism in cancer. It can help us detect cancer at early stages or help guide therapeutic choices. That’s one area of molecular imaging that’s actually in the clinics today.
Host: What are some of the latest discoveries to come out of the Molecular Imaging Center?
Piwnica-Worms: We have a lot of activities going on both within our core groups and collaborators. Two areas of interest that have recently been published and explored involve one of these special pathways. It’s known scientifically as the beta-catenin pathway. It’s very, very important in colon cancer, and it also affects how cells grow and metastasize. And by using one of these luciferase strategies, we can now follow the time course of the regulation of the beta-catenin protein inside colon cancer cells.
Another area that’s been very interesting for us is we developed a new optical technique to look directly at inflammation in living animals. In inflammation that occurs around cancer, there is an enzyme known as myeloperoxidase. And it ends up that a small molecule that many people are familiar with known as luminal – which is the same small molecule that you see on CSI Miami that’s sprayed on blood – has a very highly sensitive response to the enzyme myeloperoxidase, which is one of the key enzymes in inflammation. We’ve developed some techniques now using that luminal where we can now look at this myeloperoxidase-mediated inflammation in mouse models, and maybe someday we might be able to look at that in humans and help us follow inflammatory processes related to cancer.
Host: Tell me about what the Molecular Imaging Center itself is like.
Piwnica-Worms: Well, the Molecular Imaging Center itself is really two main components – a physical center, if you will, and a virtual center. The physical center is the founding faculty investigators who are physically in our laboratory space, all interacting on a daily basis with our laboratories, ourselves and the imaging cores that are in my lab and the investigators’ labs around. And then in addition, we’re tied together by a National Institutes of Health Molecular Imaging Center grant that helps support the cores and the broad array of investigators. We have projects that are ongoing with collaborators in cancer biology and immunology and pharmacology and biochemistry – all interested in studying biology noninvasively through these molecular imaging strategies so that we can try to pry basic biology done with traditional reductionist approaches together with these new dynamic strategies in living cells and live animals and ultimately patients.
Host: Dr. Piwnica-Worms, thank you for joining us.
Piwnica-Worms: You’re welcome.
Host: If you are interested in learning more about the Molecular Imaging Center, please visit its Web site at mic.wustl.edu.
| 255
|
NINE BANDED ARMADILLO
Photo Credit: U. S. Fish and Wildlife Service
SCIENTIFIC NAME: Dasypus novemcinctus
OTHER NAMES: Armadillo, Common Long-Nosed Armadillo
DESCRIPTION: The nine-banded armadillo (Dasypus novemcinctus) cannot easily be confused with any other North American wild mammal. The armadillo’s body is covered with an armored carapace or shell. The carapace is a double layer of horn and bone, segmented into three main divisions: an anterior scapular shield covering the shoulder; a posterior pelvic shield covering the hip region; and a middle section comprised of a series of bands connected by soft, infolded skin between the bands. The head and legs are covered with thick scales, and the tail is encased in a series of bony rings. Coloration of nine-banded armadillos is generally grayish brown, with yellowish-white scales along the side of the carapace. The armadillo has a long, pointed snout, small eyes, and large, cylindrical ears. The armadillo’s pointed snout, short, stout legs, and heavy claws are well suited for digging and burrowing. Armadillos have a limited number of vocalizations: a low, wheezy grunt associated with digging and rooting; a wheezy grunt uttered by recently captured individuals; an audible buzzing noise given when highly alarmed or fleeing, a pig-like squeal given by frightened individuals; and a weak purring given by young attempting to nurse from an unrelated female. Total length ranges from 24 to 31 inches and weights vary from 8 to 15 pounds. There are six subspecies of Dasypus novemcinctus in Central and South America, but only one subspecies, D. n. mexicanus occurs in North America.
DISTRIBUTION: Dasypus novemcinctus mexicanus’ original distribution was from the lower Rio Grande Valley between Mexico and Texas, southward through Mexico and Central America to northwestern Peru on the west side of the Andes, and all of South America to northern Argentina east of the Andes, including the islands of Grenada, Trinidad, Tobago, and Margarita. The range of the nine-banded armadillo has undergone rapid expansion into the southern United States since the late 1800s. The recent rapid expansion of the armadillo’s range was facilitated by a number of factors: reduction in the number of large carnivores; climatic and biotic changes; and accidental and deliberate relocations of animals to unoccupied areas. Armadillos now occur throughout the southern and southeastern U.S., as far north as Missouri, Kansas, Colorado, and Nebraska. These animals are common throughout most of Alabama, but less common in several northeastern counties.
HABITAT: The armadillo is very adaptable and does well in most habitat types found in Alabama. They generally avoid or are scarce in very wet or very dry habitats. Habitat suitability likely depends more on the characteristics of the substrate or soils, rather than vegetation type due to the armadillo’s feeding and burrowing behavior.
FEEDING HABITS: A major portion of the armadillo’s time spent outside its burrow is devoted to feeding. They typically start foraging as they emerge from their burrow and move at a slow pace following an often erratic course. Prey is apparently detected by smell, although sound also may play a role. Typical foraging behavior involves quickly probing with the nose and occasionally pausing to dig for prey. Armadillos are opportunistic feeders and consume a wide variety of food items. Invertebrates, primarily insects, make up roughly 90 percent of their diet. Small vertebrates and plant material make up the remainder of their diet. Researchers also have seen evidence of armadillos feeding on small reptiles and amphibians, the eggs of ground-nesting birds, and carrion.
LIFE HISTORY AND ECOLOGY: Armadillos seem to exhibit a polygynous mating system, with most females paired with a single male and most males paired with more than one female. Den burrows have an enlarged nest chamber and are more complicated than a burrow dug for other purposes. The nest is a bulky mass of dried plant debris crammed into the nest chamber without any obvious structure. Armadillos in areas with poorly drained soils will construct above ground nests of dry plant material. Most breeding among armadillos occurs during the summer (June-August). The normal gestation period is 8 to 9 months, with most young born between February and May. The armadillo exhibits monozygotic polyembryony in which a single fertilized egg normally gives rise to four separate embryos at the blastula stage of development. This results in a litter of four genetically identical haploid clone offspring. Dasypus is the only genus of vertebrates in which this reproductive phenomenon occurs. The offspring are precocial and begin accompanying the female outside of the burrow at about 2 to 3 months of age. By 3 to 4 months, the young are self-sufficient. Most males reach sexual maturity between 6 to 12 months of age, but females do not become sexually mature until they are 1 to 2 years old.
REFERENCES: Author: Chris Cook, Wildlife Biologist, June 2005
Armstrong, J. Controlling Armadillo Ddamage in Alabama. ANR-773. Alabama Cooperative Extension System. 2pp.
Layne, J. N. 2003. Armadillo. Pages 75-97 in G. A. Feldhamer, B. C. Thompson, and J. A. Chapman, eds. Wild Mammals of North America: Biology, Management, and Conservation. Second edition. The Johns Hopkins University Press, Baltimore, MD and London, U.K.
Nowak, R. M. 1999. Walker’s Mammals of the World, sixth edition, volume one. The Johns Hopkins University Press, Baltimore, MD and London, U.K. 903 pp.
Outdoor Alabama Magazine Article, Nine-banded Armadillo
Watchable Wildlife Article
| 598
|
RIVERSIDE, Calif. – Biologists at the University of California, Riverside have found that voluntary activity, such as daily exercise, is a highly heritable trait that can be passed down genetically to successive generations.
Working on mice in the lab, they found that activity level can be enhanced with "selective breeding" – the process of breeding plants and animals for particular genetic traits. Their experiments showed that mice that were bred to be high runners produced high-running offspring, indicating that the offspring had inherited the trait for activity.
"Our findings have implications for human health," said Theodore Garland Jr., a professor of biology, whose laboratory conducted the multi-year research. "Down the road people could be treated pharmacologically for low activity levels through drugs that targeted specific genes that promote activity. Pharmacological interventions in the future could make it more pleasurable for people to engage in voluntary exercise. Such interventions could also make it less comfortable for people to sit still for long periods of time."
In humans, activity levels vary widely from couch-potato-style inactivity to highly active athletic endeavors.
"We have a huge epidemic of obesity in Western society, and yet we have little understanding of what determines variation among individuals for voluntary exercise levels," Garland said.
Study results appear online Sept. 1 in the Proceedings of the Royal Society B.
The researchers began their experiments in 1993 with 224 mice whose levels of genetic variation bore similarity to those seen in wild mouse populations. The researchers randomly divided the base population of mice into eight separate lines – four lines bred for high levels of daily running, with the remaining four used as controls – and measured how much distance the mice voluntarily ran per day on wheels attached to their cages.
With a thousand mice born every generation and four generations of mice each year, the researchers were able to breed highly active mice in the four high-runner lines by selecting the highest running males and females from every generation to be the parents of the next generation. In the control lines, breeders were chosen with no selection imposed, meaning that the mice either changed or did not change over time purely as a result of random genetic drift.
By studying the differences among the replicate lines, the researchers found that mice in the four high-runner lines ran 2.5-3-fold more revolutions per day as compared with mice in the four control lines. They also found that female and male mice evolved differently: females increased their daily running distance almost entirely by speed; males, on the other hand, increased speed but they also ran more minutes per day.
The study is an example of an "experimental evolution" approach applied rigorously to a problem of biomedical relevance. Although this approach is common with microbial systems and fruit flies, it has rarely been applied to vertebrates due to their longer generation times and greater costs of maintenance. The results of such studies can inform biologists about fundamental evolutionary processes as well as "how organisms work" in a way that may lead to new therapeutic strategies.
"This study of experimental evolution confirms some previous observations and raises new questions," said Douglas Futuyma, a distinguished professor of ecology and evolution at Stony Brook University, New York, who was not involved in the research. "It shows that 'there are many ways to skin a cat': different ways in which a species may evolve a similar adaptive characteristic – running activity, in this case. Garland and coauthors go further by beginning to explore the detailed ways in which an adaptive feature, such as muscle size or metabolic rate, may be realized and by showing sex differences in the response to selection. It would be fascinating to know, and challenging to find out, if any one of these different responses is adaptively better than others."
Garland was joined in the research by Scott Kelly, Jessica Malisch, Erik Kolb, Robert Hannon, Brooke Keeney, Shana Van Cleave and Kevin Middleton, all of whom work in his lab.
The study was supported primarily by a grant to Garland from the National Science Foundation.
Details of the experimental set-up
The mice run on wheels attached to their cages. Wheel running is a completely voluntary behavior for the mice. They can sit in their cages and not run at all. If they do run, they can get off the wheels at any time. For the experiments, each mouse was given access to the wheels for only six days of their lives. A computer recorded every minute how much distance (revolutions) the mice ran for the six days. The researchers selected breeders depending on how much distance the mice ran on days 5 and 6.
About Theodore Garland Jr.
Garland received his doctoral degree in biological sciences from UC Irvine. Before joining UCR in 2001, he was a faculty member at the University of Wisconsin-Madison. He is trained in comparative physiology and evolutionary biology, as well as quantitative genetics with emphasis on exercise physiology. He is co-editor of Experimental evolution: concepts, methods, and applications of selection experiments (University of California Press, 2009). On the editorial boards of several scientific journals, he is the author/coauthor of nearly 200 peer-reviewed publications.
The University of California, Riverside (www.ucr.edu) is a doctoral research university, a living laboratory for groundbreaking exploration of issues critical to Inland Southern California, the state and communities around the world. Reflecting California's diverse culture, UCR's enrollment of about 18,000 is expected to grow to 21,000 students by 2020. The campus is planning a medical school and has reached the heart of the Coachella Valley by way of the UCR Palm Desert Graduate Center. The campus has an annual statewide economic impact of more than $1 billion.
A broadcast studio with fiber cable to the AT&T Hollywood hub is available for live or taped interviews. To learn more, call (951) UCR-NEWS.
AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert! system.
| 870
|
To create a true net-zero building, one that literally generates as much or more energy than it consumes, is no easy task. Still, it’s a task that makes good business sense. After all, buildings consume a huge amount of energy, which cuts into profit margins. This simple equation finally hit home with Hines, the international real estate firm, and equity partner J.P. Morgan Asset Management. The two are partnering to build a new 13-story, 415,000-square-foot building at La Jolla Commons in San Diego that will become the nation’s largest carbon-neutral office building to date.
In order to achieve this rare feat, the building will utilize combination of high-performance building design, directed biogas and on-site fuel cells that annually will generate more electricity than tenants will use. The fuel cells, made by Bloom Energy, will generate approximately 5.0 million KWh of electricity annually, which is roughly equivalent to the electricity required to power 1,000 San Diego homes. Methane needed to power the fuel cells will be acquired from carbon-neutral sources, such as landfills and wastewater plants, and placed into the national natural gas pipeline system. The building’s exterior is predominately a glass curtainwall system incorporating highly efficient, insulated, double-paned glass with a clear, low-emissive coating.
Hines views its newest building, which will also contain a highly efficient under-floor air system, as a sort of ongoing R&D project. “Our net-zero project at La Jolla Commons gives us a great foundation for furthering the use of carbon-neutral technologies and fuels,” said Gary Holtzer, Hines’ global sustainability officer. ”Our next step is to adapt what we have learned and apply it to an existing urban property in a less temperate environment.”
Construction on the carbon neutral building began in April 2012 and completion is scheduled for mid-2014.
| 190
|
1938 Buick Y Job
Concept cars have been a mainstay of automotive design and marketing for more than a century. While Henry Leland's Osceola is considered the world's first concept car that actually made it into production, the 1938 Buick Y Job is another early example of a true concept car.
Built under the direction of General Motors' first design director, Harley Earl, the Y-Job was never intended for production but instead foreshadowed the styling and engineering cues Earl and his team hoped to use on future GM vehicles. In its day, the Y-Job earned praise for its modern style that included integrated fenders, hidden headlamps, and no running boards. The positive reaction helped several of its design cues to make it into production, including the stubby tail fins that would appear on the iconic 1948 Cadillacs and the grille design that continues to influence Buick design.
Although the Y-Job didn't make production, it remains an example of the good a concept car can do for a company and the industry. We can't help but wonder what would have happened, though, if Buick put the Y-Job into production. The company would have been in an even better position postwar.
| 115
|
What orbits the Earth at 17,500 miles per hour, is 360 feet wide, 260 feet long and has a crew of 3? It is the International Space Station, also known as the I.S.S. It is going to be finished in the year 2006. The countries that are putting it together are the United States, Russia, Japan, Canada, Brazil, Austria, Belgium, Denmark, Finland, France, Germany, Great Britain, Ireland, Italy, the Netherlands, Norway, Spain, Sweden and Switzerland. There will be 100 different pieces sent up to space to create the I.S.S.
What You Want to Know About The I.S.S.
When it is completed, the I.S.S. will be the size of two football fields. You will be able to see it from Earth without a telescope. The I.S.S. will be much bigger than previous space stations. Other space stations, like Mir, were cramped. You also had to eat dehydrated food (food without water). The I.S.S. will have a refrigerator with fresh foods.
The I.S.S. is like living on earth except one thing-you are weightless! When you are sleeping, you have to be strapped to the wall or else you will float out of your bed.
The first crew went up in orbit on October 31, 2000. The size of the crew was decided by how big the escape pod was. That was 3.
Did you know that every time we breathe or exercise we add moisture to the air? This extra moisture has to be removed from the air on the ISS so the moisture doesnít collect on the ISS equipment. This moisture will be recycled.
The crew will recycle all of the water on board from the moisture in the air to the crewís urine and wash water. I know that sounds really gross, but the water is going to be purified and going to be cleaner than the water we drink here on Earth. There are 3 steps to purifying the water:
The crew will be up in space for a long time, but when they come back, they wonít be able to walk right away. There is no gravity in space so you donít use your legs that much, and when you go to use them itís just like trying to run after you rollerblade.
The people that are training to get on the space station are going to train hard. They will have to train for 2 years. Some of that training will be in the Canadian forest in the middle of winter to prepare them for problems they might run into up in space.
What are Some Modules Up in Space?
In 1999,the first U.S. built station component, the Unity connecting module, was moved to the launch pad. It was loaded onto the Space Shuttle, Endeavour.
More than six major components are in the processing facility. At the end of the year 2000, more than 500,000 pounds of U.S. and international station equipment was completed. Thatís the weight of 250 pick-up trucks.
Unity is a six-sided connecting module to which all-future U.S. station modules will attach. Unity will serve as a passageway to various parts of the station. Attached to Unity are two adapters. One to serve as a permanent connection to the Russian station and another that will serve as a shuttle docking port.
What is Going to Put Some Of the International Space Station Together?
Some of the I.S.S. is going to be put together from the outside. One of the dangers is that you could be hit by a micrometeoroid. Some micrometeoroids are the size of a grain of sand, but others can be the size of basketballs and go right through you. The countries that are building the I.S.S. have built robots to put together some of the I.S.S. so that no one gets hurt. Remember, the I.S.S. is made up of 100 different modules so it will take a long time to put it together.
Space: Everything You Want to Know and Beyond. </J0112388> Last Visited: December, 2001.
NASA.<http://www.nasa.gov> Last Visited: January, 2002.
Space in the Spotlight Novi Meadows Elementary 2002
All pictures courtesy of NASA unless otherwise noted
| 559
|
||On this site, kids can enter
their spelling words and the site will create word search puzzles. They
can select varying levels of difficulty. The students can complete
these puzzles online or on paper.
||Once the customized spelling
list is entered, students can play numerous spelling games to reinforce
their vocabulary. Games are available for different ages and
||This website is a great way for
kids to practice their spelling words. They can be taught the
words, be tested on the words or play games to reinforce the word
|| Kids can play math games
to practice their skills. They also can review lessons on
specific mathematical concepts to supplement their learning.
||This site was created by the Oswego City School
District. It gives the students timed tests on their chosen math
facts. If desired, the kids can choose to be tested on a mix of
||This site provides lessons to help children better
their math concepts. It also offers math games and puzzles.
||An overview of different animals is provided.
can select specific animals to review general facts. Printable
worksheets also are available for additional learning.
||Run by the BBC, this site allows children to study
topics related to living things. They can play games, read and
then be quizzed on the material.
|Yahoo! Kids||This site contains facts on animals. Kids can
specific animals to see photos and statistics. They can play
games and get homework help in the study zone.
| 180
|
From cave paintings to the internet NOV 03 2008
A fantastically extensive timeline of recorded information "from cave paintings to the internet". It's an expanded version of the timeline that appears in the book, From Gutenberg to the Internet (more info), which seems really interesting.
From Gutenberg to the Internet presents 63 original readings from the history of computing, networking, and telecommunications arranged thematically by chapters. Most of the readings record basic discoveries from the 1830s through the 1960s that laid the foundation of the world of digital information in which we live. These readings, some of which are illustrated, trace historic steps from the early nineteenth century development of telegraph systems -- the first data networks -- through the development of the earliest general-purpose programmable computers and the earliest software, to the foundation in 1969 of ARPANET, the first national computer network that eventually became the Internet. The readings will allow you to review early developments and ideas in the history of information technology that eventually led to the convergence of computing, data networking, and telecommunications in the Internet.
(via design observer)
| 301
|
Name ID 1199 The first professional hunters came in 1913. They found the wildlife plentiful, especially the lions, but saw no elephants. Seven years later, an American arrived in a strange new contraption known as a Ford motor-car and news of the wonders of the Serengeti had reached the outside world. Because the Hunting of lions made them so scarse (they were considered 'vermin'), it was decided to make a partial Game Reserve in the area in 1921 and a full one in 1929. With the growing awareness of the need for conservation, it was expanded and upgraded to a National Park in 1951. Eight years later the Ngorongoro Conservation Area was established in the south-east as a separate unit. Arusha: A Brochure of the Northern Province and its Capital Town Page Number: 13-15-17 Extract Date: 1929 It is safe to say that Tanganyika holds a front place among our East African Colonies for the number and variety of its game animals. The belt from Tanga through to Lake Victoria is where game is most numerous. There is an abundance of the commoner antelope, and in certain parts the rarer species such as the Greater and Lesser Kudu, Gerenuk, etc., are still fairly plentiful. Big game like the Elephant, Rhinocerous, Lion and Buffalo, all of which hold for the hunter a new thrill and experience, are to be found in this area in such a variety of country and cover that the Hunting of no two animals is ever alike. Here the hunter passes through most interesting country; Kilimanjaro with its snow-capped dome, running streams and dense forests, across the plains to the Natron Lakes and the Great Rift Wall with its volcanic formation and on to the great Crater, Ngorongoro. In his travels he will come into contact with some of the most interesting and picturesque tribes that inhabit Africa such as the Masai, Wambulu, etc., each with their own quaint customs and histories. The Ngorongoro Crater, the greatest crater in the world, measuring approximately 12 miles in diameter, seen from the Mbulu side, is a delight to the eye with its teeming herds of game ; Wildebeest alone running into tens of thousands. This scene conveys to one the idea of a great National Park. Nature has provided the crater with a precipitous rock fence for tns most part and with lakes and streams to slake the thirst of the great game herds which inhabit it. The unalienated part of the crater is now a complete game reserve in which a great variety of game is to be found such as Rhinoceros, Hippopotamus, Lion, and all the smaller fry. The Elephant although not in the crater is to be found in the forests nearby. The Serengetti Plains lying away to the northwest of the crater holds its full share of animal life and here the sportsman has the widest possible choice of trophies. The Lion in this area holds full sway and is still to be seen in troops of from ten to twenty. Recently, Serengetti and Lion pictures have become synonymous. The commoner species of game are here in abundance and the plains are second only to the crater for game concentration. The country lying between the Grumeti River - Orangi River and the Mbalangeti from Lake Victoria to the Mou-Kilimafetha Road has recently been declared a game reserve. Game animals that inhabit the northern area are well protected an'd their existence is assured to posterity by the great game sanctuaries and regulations which govern the Hunting or photographing of game. In the Northern area there are six complete reserves and two closed areas. These are as follows: (2) Mount Meru. (3) Lake Natron (4) Northern Railway. The closed areas are : Pienaar's Heights, near Babati and Sangessa Steppe in the Kondoa district. The boundaries for these are laid down in the Game Preservation Ordinance No. 41 of 1921. There are, however, vast areas open to the hunter and the abovementioned sanctuaries do not in any way detract from the available sport which the Northern Tanganyika has to offer. The following game licences are now in force (Shillings) :Visitor's Full Licence - 1500 Visitor's Temporary Licence (14 days) - 200 Resident's Full Licence - 300 Resident's Temporary Licence (14 days) - 60 Resident's Minor Licence - 80 Giraffe Licence - 150 Elephant Licence 1st. - 400 2nd. - 600 To hunt the Black Rhinoceros in the Northern Province it is now necessary to hold a Governor's Licence, the fee for which is 150/-. This entitles the holder to hunt one male Rhinoceros. Elephant, Giraffe, and Rhinoceros Licences may only be issued to holders of full game licences. Now that the Railway is through to Arusha it is not too much to hope that with the assistance of a healthy public opinion the Sanya Plains may become restocked with game which would be a great source of interest and an attraction to the traveller visiting these parts. Extract Date: 1931 Herne, Brian White Hunters: The golden age of African Safaris Page Number: 375 Extract Date: 1965 Safari Hunting in East Africa was forever changed by the masterly blueprint of Brian Nicholson, a former white hunter turned game warden. The disciple and successor of C.I. P. Ionides, the "Father of the Selous game reserve," Nicholson conceived a plan for administering Tanzania's expansive wildlife regions. In 1965 he changed most of the vast former controlled Hunting areas, or CHAs, into Hunting concessions that could be leased by outfitters from the government for two or more years at a time. Nicholson also demarcated the Selous game reserve's 20,000 square miles of uninhabited country into 47 separate concessions. Concessions were given a limited quota of each game species, and outfitters were expected to utilize quotas as fully as possible, but not exceed them. Nicholson's plan gave outfitters exclusive rights over Hunting lands, providing powerful incentives for concession holders to police their areas, develop tracks, airfields, and camps, and, most importantly, preserve the wild game. When the system was put into effect, it was the larger outfitting organizations - safari outfitters who could muster the resources to bid and who had a clientele sufficient to fulfill the trophy quotas Nicholson had set (done in order to provide government revenue by way of fees for anti-poaching operations, development, and research) - that moved quickly to buy up the leases on the most desirable blocks of land. Smaller safari companies who could not compete on their own banded together and formed alliances so that they, too, could obtain Hunting territories. Herne, Brian White Hunters: The golden age of African Safaris Page Number: 389 Extract Date: 1973 Sep 7 By the end of 1973 Kenya was the sole remaining tourist destination in East Africa. While the neighboring country of Uganda was still in the throes of military anarchy, Tanzania surprised the world on September 7 by issuing an overnight ban on all Hunting and photographic safaris within its territory. Government authorities moved quickly to seize and impound foreign-registered Land Cruisers, supply trucks, minibuses, aircraft, and equipment. The stunned collection of safari clients as well as sundry mountain climbers, bird-watchers, and beachcombers who had been visiting the country at the time of the inexplicable edict were summarily escorted to Kilimanjaro airport outside of Arusha to await deportation. The residue of tourists stranded without flights were trucked to the northern town of Namanga where they were left on the dusty roadside to cross into Kenya on foot. All tourist businesses, including the government-owned Tanzania Wildlife Safaris, were closed down. No government refunds were ever made to tourists or to foreign or local safari outfitters Africa News Online Extract Date: 2000 June 5 Panafrican News Agency Frequent acrimony, currently depicting the relationship between game Hunting companies and rural communities in Tanzania, will be a thing of the past after the government adopts a new wildlife policy. Designated as wildlife management areas, the communities will benefit from the spoils of game Hunting, presently paid to local authorities by companies operating in those areas. The proposed policy seeks to amend Tanzania's obsolete Wildlife Act of 1975, and, according to the natural resources and tourism minister, Zakia Meghji, 'it is of utmost priority and should be tabled before parliament for debate soon'. She said the government would repossess all Hunting blocks allocated to professional hunters and hand them over to respective local authorities. In turn local governments, together with the communities, would be empowered to allocate the Hunting blocks to whichever company they prefer to do business with. 'Guidelines of the policy are ready and are just being fine-tuned,' she said. Communities set to benefit from this policy are chiefly those bordering rich game controlled areas and parks. They include the Maasai, Ndorobo, Hadzabe, Bahi, Sianzu and Kimbu in northeastern Tanzania. Members of these communities are often arrested by game wardens and fined for trespassing on game conservation areas. As a result, they have been extremely bitter about being denied access to wildlife resources, which they believe, naturally, belong to them. Under the new policy, Meghji said, the government will ensure that people undertake increased wildlife management responsibilities and get benefits to motivate them in the conservation of wildlife resources. The East African Extract Author: John Mbaria Extract Date: February 4, 2002 KENYA COULD end up losing 80 per cent of its wildlife species in protected areas bordering Tanzania to hunters licensed by the Tanzanian government. The hunters have been operating for about a decade in a section of the migratory route south from Kenya to Tanzania's Serengeti National Park. They shoot large numbers of animals as they move into the park during the big zebra and wildebeest migration between July and December. There are fears that the Maasai Mara National Park and most of Kenya's wildlife areas bordering Tanzania could lose much of their wildlife population, threatening the country's Ksh20 billion ($256 million) a year tourism industry. Kenya banned Hunting in 1977 but the sport is legal in Tanzania, where it is sold as "Safari Hunting." "The product sold is really the experience of tracking and killing animals, the services that go with this and the prestige of taking home the trophies," says a policy document from the Tanzania Wildlife Corporation (Tawico). Tanzania wildlife officials said wild animals that cross over from Kenya are hunted along their migratory routes in the Loliondo Game Controlled Area in Ngorongoro district of Arusha region, 400 km northwest of Arusha. The area was designated by the British colonial power as a sports Hunting region for European royalty. The officials said the area is now utilised by a top defence official from the United Arab Emirates (UAE), trading as Ortelo Business Company (OBC), through a licence issued in 1992 by former Tanzania President Ali Hassan Mwinyi. The permit allows the company to hunt wild game and trap and take some live animals back to the UAE. Safari Hunting earns big money for the Tanzanian government, which charges each hunter $1,600 a day to enter the controlled area. A hunter is also required to pay fees for each kill, with an elephant costing $4,000, a lion and a leopard $2,000 each and a buffalo $600. The document has no quotation for rhinos. The sport is organised in expeditions lasting between one and three weeks in the five Hunting blocks of Lake Natron Game Controlled Area, Rungwa Game Reserve, Selous Mai, Selous U3 and Selous LU4. For the period the hunters stay in each of the Hunting blocks, they pay between $7,270 and $13,170 each. Part of this money is shared out among the many Ujamaa villages, the local district councils and the central government. Although Tawico restricts the number of animals to be culled by species, poor monitoring of the activities has meant indiscriminate killing of game. "Some of the animals are snared and either exported alive or as meat and skins to the United Arab Emirates and other destinations," local community members told The EastAfrican during a recent trip to the area. They claimed the hunters were provided with "blank Hunting permits," giving them discretion over the number of animals to be hunted down. Kenya wildlife conservation bodies are concerned that big game Hunting in the Ngorongoro area is depleting the wildlife that crosses the border from Kenya. "Kenya is losing much of its wildlife to hunters licensed by the Tanzanian government," the chairman of the Maasai Environmental and Resource Coalition (MERC), Mr Andrew ole Nainguran, said. MERC was set up in 1999 to sensitise members of the Maasai community in Kenya and Tanzania to the benefits of wildlife conservation. Kenya and Tanzania wildlife authorities have regularly discussed the problem of security and poaching in Arusha. However, the KWS acting director, Mr Joe Kioko, said legalised Hunting has never been discussed in any of the meetings. The hunters are said to fly directly from the UAE to the area using huge cargo and passenger planes which land on an all-weather airstrip inside the OBC camp. The planes are loaded with sophisticated Hunting equipment, including four-wheel drive vehicles, weapons and communication gadgets. On their way back, the planes carry a variety of live animals, game trophies and meat. Employees at the camp said the hunters are sometimes accompanied by young Pakistani and Filipino women. The International Fund for Animal Welfare regional director, Mr Michael Wamithi, said Kenya and Tanzania should discuss the negative impact of the sport Hunting on Kenya's conservation efforts. "The two neighbours have a Cross-Border Law Enforcement Memorandum of Understanding where such issues could be dealt with." Kenya seems to be alone in adhering to strict protection of wildlife, a policy famously demonstrated by President Daniel arap Moi's torching of ivory worth $760,000 in 1989. Although the country has made significant progress in securing parks from poachers, it is yet to embrace a policy on "consumptive utilisation" of animals advocated by Kenyan game ranchers and Zimbabwe, which wants the international trade ban on ivory lifted. The animals in the Hunting block have been reduced to such an extent that the OBC camp management has been spreading salt and pumping water at strategic places to attract animals from Serengeti and the outlying areas. "We will not have any animals left in the vicinity unless the Hunting is checked," a local community leader, Mr Oloomo Samantai ole Nairoti, said, arguing that the area's tourism economy was being jeopardised. Mysterious fires in the area to the south of Serengeti have also forced animals to seek refuge in the Hunting blocks. Locals said the camp is exclusively patronised by Arab visitors. The camp is usually under tight security by Tanzanian police. The permit granted by Mr Mwinyi has raised controversy in Tanzania and was at one stage the subject of a parliamentary probe committee because members of UAE's royal family were not entitled to the Hunting rights in the country. "Only presidents or monarchs are entitled to hunt in the area," an official said, adding that the UAE royal family had abused their permit by killing animals outside their given quotas or specified species. The government revoked the licence in 1999 after realising that OBC was airlifting many wild animals to the Middle East, only to renew the permit in 2000. The current permit runs until 2005. The withdrawal of the permit followed the recommendations of a 1994 parliamentary probe commission set up to "investigate the Hunting behaviour" of the UAE company. Sources said permanent Hunting is prohibited in the Loliondo Game Controlled Area for fear of depleting animals from the four parks, which host the bulk of the region's tourist resorts. The area is in a natural corridor where wild animals cross while roaming between the Ngorongoro Conservation Area and Serengeti National Park in Tanzania and Maasai Mara Game Reserve and Amboseli National Park in Kenya. The late founding president of Tanzania, Mwalimu Julius Nyerere, took to himself the powers to issue Hunting permits for Loliondo when Tanzania became independent in 1961, but he never granted any. After obtaining the permit, the UAE hunters created Hunting blocks in the area covering over 4,000 sq km. No other Hunting companies have been granted permits, the source said. The UAE royal family has donated passenger aircraft to the Tanzania army and a number of vehicles to the Wildlife Division. The 1974 Wildlife Act set up five categories of wildlife conservation areas. These are national parks, game reserves, partial game reserve, open areas and Ngorongoro Conservation Area. Hunting is prohibited in the national parks and Ngorongoro Conservation Area, but allowed in other areas during the seasonal Hunting period from July to December. Additional reporting by Apolinari Tairo in Dar es Salaam Tomlinson, Chris Big game hunting threatening Africa Extract Author: Chris Tomlinson Extract Date: 2002 03 20 Loliondo GAME CONTROL AREA, Tanzania - At a dirt airstrip in rural Tanzania, a desert camouflaged cargo plane from the United Arab Emirates air force taxis up to pallets stacked with large coolers full of game meat, the harvest of a successful Hunting season. As Tanzanian immigration and customs officials fill out documents under a thatched shelter, brand-new, four-wheel-drive trucks and dune buggies drive to and from a nearby luxury campsite, the base for one of Tanzania's most expensive - and secretive - game Hunting operations, Otterlo Business Corp. Hundreds of members of Arab royalty and high-flying businessmen spend weeks in the Loliondo Game Control Area each year Hunting antelope, lion, leopard and other wild animals. The area is leased under the Otterlo name by a member of an emirate royal family who is a senior officer in the UAE defense ministry. While neighboring Kenya outlawed big game Hunting in 1978, the Tanzanian government says Hunting is the best use of the land and wildlife. But villagers and herders say big money has led government officials to break all the Hunting rules, resulting in the destruction of most of the area's non-migratory animals and putting East Africa's most famous national parks under threat. Loliondo is on the main migratory route for wildlife north of Ngorongoro Crater, east of Serengeti National Park and south of Kenya's Masai Mara National Reserve. The summer Hunting season coincides with the migration of wildebeest and zebra through the area, where they eventually cross into the Serengeti and the Masai Mara. Predatory animals follow the migration. During the colonial era, Loliondo was set aside for European royalty as a Hunting area. Since independence, Loliondo has remained a Hunting reserve, but it is supposed to be managed by area residents for their benefit. Local leaders, who refuse to speak publicly because they fear retribution, say they have not been consulted about the lease that was granted in 1995 by national officials in Tanzania's political capital, Dodoma. They say government officials have tried to silence criticism. "The lease was given by the government and the Maasai landowners were not involved," said one Maasai leader. "All the resident animals have been killed ... (now) they carry out Hunting raids in the Serengeti National Park, but the government closes its eyes." Maasai warriors told The Associated Press that hunters give cash to anyone who can lead them to big game, especially leopards. They also said that Otterlo officials have begun pumping water into some areas to attract more animals and that what the warriors call suspicious fires in the Serengeti have caused animals to move into Loliondo. An Otterlo manager, who gave his name only as Khamis, initially agreed to an interview with AP but later did not return repeated phone calls. In an interview with the newspaper, The East African, Otterlo managing director Juma Akida Zodikheri said his company adheres to Tanzanian law, and he denied hunters killed animals indiscriminately. He said the owner of the company is Maj. Gen. Mohammed Abdulrahim al Ali, deputy defense minister of the UAE. While Tanzania has strict rules on game Hunting, Maasai who have worked at the lodge say guests are never told of the limits and hunt as much as they want. Tanzanian officials deny that. Col. A.G.N. Msangi, district commissioner for Ngorongoro District, said all applicable rules are enforced. He accused the Maasai of rumor-mongering in an effort to discredit Otterlo. The company "is following the system the government wants," Msangi said. "OBC has invested more money here than any other company in the district." Msangi said Hunting companies request permission to kill a certain number of animals. Once the request is approved by wildlife experts at the Ministry of the Environment, the company pays a fee based on that number whether they actually kill the animals or not, he said. "We have police and ministry people making sure they don't exceed what they have paid for," Msangi said. The tourists are also required to employ professional hunters to ensure no female or young animals are killed, he added. Compared to the numbers in Serengeti National Park, very few large animals were seen during a three-hour drive through Loliondo. But without any independent survey of the animal population, it is impossible to know whether Msangi's conservation efforts are working. Msangi described his main duty as balancing the needs of people, animals and conservation. He said not only does Hunting revenue finance wildlife conservation, but Otterlo, like most tourism companies, also makes charitable donations to help pay for schools and development projects and it provides badly need jobs. Also appeared in http://www.washtimes.com/world/20020801-22110374.htm 1 Aug 2002 Internet Web Pages Extract Author: Lifer Extract Date: April 16 2002 Posted - April 16 2002 : 20:53:22 The East African Newspaper of 4-10 February 2002 carried an article titled "Game Carnage in Tanzania Alarms Kenya", written by John Mbaria with supplement information from Apolinari Tairo of Dar es Salaam. The article was on The Ortello Business Hunting Company, which started to hunt in the Loliondo Game Controlled Area in 1992. The following are issues raised in the article: a) Hunting activities carried out in Liliondo Game Controlled Area near the Tanzania / Kenyan border causes loses of 80% of the Kenyan wildlife. b) Hunting is conducted in the migratory route in the south between Kenya and Serengeti National Park. The animals are hunted during the migratory period as they move to Kenya and on their way back to Tanzania in July to December. c) Hunting is threatening the Kenyan tourism industry, which earns the country USD 256.0 annually. d) The Hunting kills animals haphazardly, without proper guidance and monitoring of actual number of animals killed and exported outside the country. e) Airplanes belonging to Ortello Business Corporation (OBC) carry unspecified type of live animals and birds from Loliondo on their way back to UAE. Further more, the air planes fly directly in and out of Loliondo without stopping at Kilimanjaro International Airport (KIA). The following are responses to the issues raised: 2.0 Conservation of wildlife in Tanzania Tanzania is among the top ten countries in the world rich in biodiversity. Tanzania is also leading in wildlife conservation in Africa. It has 12 National Parks, including the famous Serengeti National Park, 34 Game Reserves and 38 Game Controlled Areas. The wildlife –protected areas cover 28% of the land surface area of Tanzania. In recognition of the good conservation works, Tanzania was awarded a conservation medal in 1995 by the Safari Club International whose headquarters is in the United States of America. Tanzania has a number of important endangered animal species in the world. Such animal species are: Black Rhino, Wild Dog, Chimpanzee, Elephant and Crocodile (Slender Snorted Crocodile). In 1998, the Government of Tanzania adopted a Wildlife Policy, which gives direction on conservation and advocate sustainable use of wildlife resources for the benefit of the present and future generations. 3.0 Tourist Hunting Regulated tourist Hunting or any other type of Hunting that observes conservation ethics does not negatively affect wild animal populations. This is because Hunting ethics is based on selective Hunting and not random shooting of animals. Hunting was banned in Tanzania from 1972 to 1978. The resultant effect was increased poaching and reduced government revenue from wildlife conservation. Low revenue caused low budgetary allocations to wildlife conservation activities and the lack of working gear and equipment. When the tourist Hunting resumed Elephant population increased from 44,000 (in 1989) to 45,000 (in 1994). Elephant is a keystone species in the Hunting industry and is a good indicator in showing population status of other animal species in their habitat. In 1989 to 1993 the government revenue from the Hunting industry increased from USD 2,422,500.00 to USD 7,377,430.00. The government earned a total of USD 9.3
| 687
|
In a study led by UCSF team an existence of relationship between asthma and composition of bacterial species, which are present in bronchial airways was revealed.
This finding could suggest new treatment or even potential cures for the common inflammatory disease.
Using new detection methods, researchers learned that the diversity of microbes inside the respiratory tract is far vaster than previously suspected - creating a complex and inter-connected microbial neighborhood that appears to be associated with asthma, and akin to what has also been found in inflammatory bowel disease, vaginitis, periodontitis, and possibly even obesity.
"People thought that asthma was caused by inhalation of allergens but this study shows that it may be more complicated than that - asthma may involve colonization of the airways by multiple bacteria," said study co-author Homer Boushey, MD, a UCSF professor of medicine in the division of Pulmonary and Critical Care Medicine.
In their three-year pilot project, the scientists collected samples from the airway linings of 65 adults with mild to moderate asthma and 10 healthy subjects. Then, using a tool that can identify approximately 8,500 distinct groups of bacteria in a single assay, the scientists profiled the organisms present in each sample to look for relationships between bacterial community composition and clinical characteristics of the patients' asthma.The researchers found that bronchial airway samples from asthmatic patients contained far more bacteria than samples from healthy patients. The scientists also found greater bacterial diversity in the asthmatic patients who had the most hyper-responsive or sensitive airways (a feature of asthma).
The study has been published online in the Journal of Allergy and Clinical Immunology.
| 849
|
Story posted January 16, 2008
Brads, nails, and spikes take on the characteristics of line, mass and color in sculptures by artist and Lecturer in Art John Bisbee. Covering his more than 20-year career, Bright Common Spikes: The Sculpture of John Bisbee, features 20 large-scale abstract sculptures of nails that are welded and forged into various organic configurations. The exhibition will be on view at the Portland Museum of Art January 24, 2008, through March 23, 2008.
The January 20, 2008, edition of The Boston Globe features an in-depth profile of Bisbee and his art. Click here to read the article and watch a video slideshow.
Bisbee welds, cuts, hammers, forges, and bends his nails and spikes into a variety of forms for his sculptures. Since 1988, when he overturned a five gallon bucket of nails that had rusted and fused together, he has been creating small- to large-scale sculptures comprising mainly brads, nails and spikes.
Over the years, as Bisbee has gradually increased the size of his nails from brads to 10- and 12-inch spikes, the overall scale of his work has become more expansive. Bisbee began to explore the implications of monumentality in more serious ways with a series he calls "Tons." Each piece is made up of one ton of nails, and to date he has completed 11 "Tons." They vary in type of fabrication and assembly — groups of nails are welded together and piled in various configurations into corners, spread across the floor, or affixed to the wall.
Some "Tons" pieces vary with each new installation, and Bisbee does not determine the exact configuration until the moment a piece is installed. He arrives at the gallery with barrels filled with the component pieces and lets the spirit of the day and the idiosyncrasies of the site suggest the rest. Bisbee's intuitive working method comes closest to the surrealist process of automatic drawing, allowing the subconscious mind to direct the hand.
In addition to the "Tons" series, Bisbee has recently produced more intricately patterned pieces. These works share the decorative flair that characterizes a great deal of contemporary painting and sculpture. For his most recent sculptures, Bisbee uses a forge, press, and hand tools to flatten and bend his nails into even more elaborate shapes that resemble calligraphic lines. Some of these flattened nails are then welded together to form larger mandala-like shapes; others are piled freely on the floor.
Bisbee's solo museum exhibitions include the Albright-Knox Art Gallery in Buffalo, N.Y., and the Kemper Museum of Contemporary Art in Kansas City, Mo.
A recent recipient of a 2006 Joan Mitchell Foundation Grant, Bisbee has also received a Maine State Individual Artist Grant (2000) and The Rappaport Prize (2003), administered by the DeCordova Museum and Sculpture Park. He has also been awarded the Purchase Prize and the William E. Thon Jurors' Prize from the Portland Museum of Art's Biennial exhibition (1998, 2001). His work has been reviewed in Art in America, ARTnews, Sculpture Magazine, The New Yorker, The New York Times and The Boston Globe.
John Bisbee and his art were the focus of a Spring 2005 Bowdoin magazine feature. Read it here.
| 657
|
In a commentary in the February issue of the journal Nature, a team of scientists from University of California, San Francisco has suggested that sugar should be regulated like alcohol and cigarettes.
In an article entitled ‘The toxic truth about sugar’, the UCSF group suggests that like alcohol and tobacco, sugar is a toxic, addictive substance that should be highly regulated with taxes, laws on where and to whom it can be advertised, and even age-restricted sales.
In response, the American Beverage Association issued the following statement:
"The authors of this commentary attempt to address the critical global health issue of non-communicable diseases such as heart disease and diabetes. However, in doing so, their comparison of sugar to alcohol and tobacco is simply without scientific merit. Moreover, an isolated focus on a single ingredient such as sugar or fructose to address health issues noted by the World Health Organization to be caused by multiple factors, including tobacco use, harmful alcohol use, an unhealthy diet and lack of physical activity, is an oversimplification.
“There is no evidence that focusing solely on reducing sugar intake would have any meaningful public health impact. Importantly, we know that the body of scientific evidence does not support that sugar, in any of its various forms - including fructose, is a unique cause of chronic health conditions such as obesity, diabetes, hypertension, cardiovascular disease or metabolic syndrome."
Source: American Beverage Association
- If you enjoyed this article, you may also like this: ABA puts calorie information at your fingertips
- Caffé Culture 2013, in pictures
- Tetra Pak receives DuPont Continuing Innovation Award
- Pom-Bear Zoo snacks from Intersnack
| 622
|
10 Tips for Parents of Prospective College Students
1. Choosing a Career/Choosing a Major
Ultimately, your son or daughter should make the choice. Of course, you may want to mention factors to consider, such as job market demand, salary ranges, long-range opportunities, skills required, etc. Just because an occupation is "hot" now does not mean it will be equally in demand in 10 years or that your child has the aptitude or motivation for it.
2. Choosing to Double Major/Choosing a Major and Minor
Most employers do not place a premium on a double major. It usually requires an extra one or two semesters to obtain a second major and does not particularly enhance a student's marketability. Exceptions would be a second major or a major and minor chosen for a specific career, such as English and chemistry for technical writing, or a health policy major and business minor for hospital administration.
3. Grade Point Average (GPA)
Some students who get off to a rocky start eventually pull up their grades; however, this can be very difficult to do. Many employers stress the student's overall background: work experience, number of hours worked during the school year to finance college, leadership activities, etc. Encourage your son or daughter to make academics a high priority beginning with his or her freshman year. It is important to remember that it may take him or her a while to adjust to the rigorous academic demands of college.
4. Obtaining Marketable Skills
Most employers today put more emphasis on graduates' skills than on their academic majors. Encourage your son or daughter to develop strengths in at least two or three of the following areas:
- Computer skills (e.g., programming, word processing, spreadsheets, data base management, e-mail, Internet);
- Quantitative skills (e.g., accounting, statistics, economics);
- Communication skills (e.g., written and oral);
- Marketing/selling skills (e.g., sales, publicity, fundraising);
- Scientific skills (e.g., lab skills, scientific research);
- Foreign language skills (e.g., especially Spanish, Portuguese, Chinese, or Russian);
- Leadership skills (e.g., supervisory, extracurricular leadership roles, teamwork/team leader).
5. Leadership Activities
Many employers rate leadership activities even more highly than GPA. Students who were very active in high school activities may be less involved in college extracurricular activities. However, employers regard high school as "ancient history" for a college senior. It is more valuable for a student to be involved in a few meaningful leadership roles on campus than to be in a "laundry list" of many campus clubs.
You may want your son or daughter to work in his or her hometown every summer. However, the experience gained as a lifeguard or ice cream shop counter clerk does not compare to that which comes from an internship (paid or unpaid) in the career field that he or she aspires to enter. Future employers will seek graduates with relevant, real-world work experience. Some students have little to write about on a resume if their summers were spent in school, traveling, or working at low-level jobs. We strongly suggest that students seek career-related experience for their sophomore and junior summers even if they must live away from home or accept an unpaid internship. Students needing financial support can combine an unpaid internship with a paid job such as waiter/waitress, etc.
7. Graduating Early, Graduating Late
Some students graduate early through advanced placement credits, heavy course loads, and summer school courses. The advantages are lower educational expenses and the ability to start employment or graduate school earlier. The disadvantages may include the sacrifice of academic honors, work experience, and extracurricular and volunteer activities that may contribute to a student's maturity level and qualifications. Other students graduate late due to light course loads, academic difficulties, changing majors, poor academic advising, lack of direction, or reluctance to leave the cocoon of the college environment. Advantages to late graduation include the ability to improve grades with light class loads, extra time to change majors, the ability to take additional electives to improve marketability, and extra time to gain more career-related or leadership experience. Disadvantages to late graduation are increased college costs and possible disapproval of employers and graduate schools.
8. Planning for Graduate/Professional School
About 88 percent of the nation's college freshmen indicated in a recent survey that they plan to go to graduate or professional school, yet only about 24 percent do so within a year of completing their bachelor's degree. Students aspiring to graduate or professional school should: Be clear about the reasons they want to go on for further education; research the qualifications required for admission and be realistic about their chances of acceptance; and always have a "Plan B" or back-up plan in case they are not accepted. Students should discuss their interest in graduate or professional school well before their senior year with their academic adviser; the college's graduate or professional school adviser (e.g., the pre-law or pre-med adviser); and a college career adviser to obtain advice and guidance from three different perspectives.
9. Taking Time Off
Many students want to take time off after college graduation from college before attending graduate school or taking a career-related job. Future employers will want to know how the student has spent the intervening time. Do activities during this period demonstrate relevance to future career goals and/or a good work ethic? While short-term travel may be personally broadening, it does not increase a student's marketability to employers unless it is seen as career related. Therefore, the time off may result in a longer job search. For example, management trainee programs, which often begin shortly after graduation and hire large numbers of new graduates, may be filled by the time your child is ready to begin a job search.
10. Using the College Career Services Office
Students should begin using their campus career office in their freshman year. Virtually all career offices provide individual career counseling/advising, career planning workshops, internship assistance, and career fairs and programs. Your son or daughter should seek help early with choosing a career and preparing for it. Competition for good jobs, particularly in certain fields, is stiff. The career office can advise students about how to become a strong candidate for their field of interest.
Source: Article by Marcia B. Harris and Sharon L. Jones
| 602
|
American Heritage® Dictionary of the English Language, Fourth Edition
- n. A soft, silvery, easily oxidized metallic element that ignites spontaneously in air when finely divided. Strontium is used in pyrotechnic compounds and various alloys. Atomic number 38; atomic weight 87.62; melting point 769°C; boiling point 1,384°C; specific gravity 2.54; valence 2. See Table at element.
Century Dictionary and Cyclopedia
- n. Chemical symbol, Sr; atomic weight, 87.37; specific gravity, 2.54. A dark-yellow metal, less lustrous than barium, malleable, and fusible at a red heat. When heated in air, it burns with a bright flame to the oxid. It decomposes water at ordinary temperatures, evolving hydrogen, and uniting with the oxygen of the water to form the oxid strontia. It does not occur native. The chief strontium minerals are the carbonate (strontianite) and the sulphate (celestine). Strontium also occurs as a silicate in the mineral brewsterite. It has been detected in the waters of various mineral springs, as well as in sea-water, and in the ashes of some marine plants. Salts of strontium are chiefly used in pyrotechny, imparting an intense red color to flames.
- n. A metallic chemical element (symbol Sr) with an atomic number of 38.
GNU Webster's 1913
- n. (Chem.) A metallic element of the calcium group, always naturally occurring combined, as in the minerals strontianite, celestite, etc. It is isolated as a yellowish metal, somewhat malleable but harder than calcium. It is chiefly employed (as in the nitrate) to color pyrotechnic flames red. Symbol Sr. Atomic weight 87.3.
- n. a soft silver-white or yellowish metallic element of the alkali metal group; turns yellow in air; occurs in celestite and strontianite
- Named in a pseudo-Latin manner for the name of the Scottish town Strontian. (Wiktionary)
- From New Latin strontia, strontium oxide, from English strontian; see strontianite. (American Heritage® Dictionary of the English Language, Fourth Edition)
“But the water may contain other harmful substances, such as beta-ray emitting strontium, which is known for causing bone cancers.”
“The strontium is the secret," said Li - although exactly what it is doing is unclear.”
“One of the procedures is called strontium isotope analysis, a technique that measures the ratio of strontium isotopes in a person's tooth enamel to determine where they grew up.”
“Answer: A drug called strontium ranelate is commonly prescribed in several European countries.”
“Recent health department tests on National Gypsum board found it lacked a compound called strontium sulfide - something found in all of the Chinese drywall samples.”
“When a POS is attacked, when it hits the end of its shields, it goes into 'reinforced mode', and becomes invunerable for a period of time depending on how much 'strontium' it is fueled up with.”
“Other hazards such as strontium are insignificant.”
“A: The non-patentable forms of strontium supplements, such as strontium citrate, should dissolve and deliver strontium to the bones just as efficiently as strontium ranelate.”
These user-created lists contain the word ‘strontium’.
A list of words that are odd or words that I have looked up.
includes words of the "Prodcom list"
A list of chemical elements
I'm wading through Patrick O'Brian's Aubrey/Maturin novels one by one, and someday, I'll wade through them again and list all the words I learned while reading them.
Edit: I started ma...
Looking for tweets for strontium.
| 517
|
Problems of Philosophy
Chapter 5 - Knowledge by Acquaintance and Knowledge by Description
After distinguishing two types of knowledge, knowledge of things and knowledge of truths, Russell devotes this fifth chapter to an elucidation of knowledge of things. He further distinguishes two types of knowledge of things, knowledge by acquaintance and knowledge by description. We have knowledge by acquaintance when we are directly aware of a thing, without any inference. We are immediately conscious and acquainted with a color or hardness of a table before us, our sense-data. Since acquaintance with things is logically independent from any knowledge of truths, we can be acquainted with something immediately without knowing any truth about it. I can know the color of a table "perfectly and completely when I see it" and not know any truth about the color in itself. The other type of knowledge of things is called knowledge by description. When we say we have knowledge of the table itself, a physical object, we refer to a kind of knowledge other than immediate, direct knowledge. "The physical object which causes such-and-such sense-data" is a phrase that describes the table by way of sense-data. We only have a description of the table. Knowledge by description is predicated on something with which we are acquainted, sense-data, and some knowledge of truths, like knowing that "such- and-such sense-data are caused by the physical object." Thus, knowledge by description allows us to infer knowledge about the actual world via the things that can be known to us, things with which we have direct acquaintance (our subjective sense-data).
According to this outline, knowledge by acquaintance forms the bedrock for all of our other knowledge. Sense-data is not the only instance of things with which we can be immediately acquainted. For how would we recall the past, Russell argues, if we could only know what was immediately present to our senses. Beyond sense-data, we also have "acquaintance by memory." Remembering what we were immediately aware of makes it so that we are still immediately aware of that past, perceived thing. We may therefore access many past things with the same requisite immediacy. Beyond sense-data and memories, we possess "acquaintance by introspection." When we are aware of an awareness, like in the case of hunger, "my desiring food" becomes an object of acquaintance. Introspective acquaintance is a kind of acquaintance with our own minds that may be understood as self-consciousness. However, this self-consciousness is really more like a consciousness of a feeling or a particular thought; the awareness rarely includes the explicit use of "I," which would identify the Self as a subject. Russell abandons this strand of knowledge, knowledge of the Self, as a probable but unclear dimension of acquaintance.
Russell summarizes our acquaintance with things as follows: "We have acquaintance in sensation with the data of the outer senses, and in introspection with the data of what may be called the inner sense—thoughts, feelings, desires, etc.; we have acquaintance in memory with things which have been data either of the outer senses or of the inner sense. Further, it is probable, though not certain, that we have acquaintance with Self, as that which is aware of things or has desires towards things." All these objects of acquaintance are particulars, concrete, existing things. Russell cautions that we can also have acquaintance with abstract, general ideas called universals. He addresses universals more fully later in chapter 9.
Russell allocates the rest of the chapter to explaining how the complicated theory of knowledge by description actually works. The most conspicuous things that are known to us by description are physical objects and other people's minds. We approach a case of having knowledge by description when we know "that there is an object answering to a definite description, though we are not acquainted with any such object." Russell offers several illustrations in the service of understanding knowledge by description. He claims that it is important to understand this kind of knowledge because our language uses depends so heavily on it. When we say common words or proper names, we are really relying on the meanings implicit in descriptive knowledge. The thought connoted by the use of a proper name can only really be explicitly expressed through a description or proposition.
Bismarck, or "the first Chancellor of the German Empire," is Russell's most cogent example. Imagine that there is a proposition, or statement, made about Bismarck. If Bismarck is the speaker, admitting that he has a kind of direct acquaintance with his own self, Bismarck might have voiced his name in order to make a self-referential judgment, of which his name is a constituent. In this simplest case, the "proper name has the direct use which it always wishes to have, as simply standing for a certain object, and not for a description of the object." If one of Bismarck's friends who knew him directly was the speaker of the statement, then we would say that the speaker had knowledge by description. The speaker is acquainted with sense-data which he infers corresponds with Bismarck's body. The body or physical object representing the mind is "only known as the body and the mind connected with these sense-data," which is the vital description. Since the sense-data corresponding to Bismarck change from moment to moment and with perspective, the speaker knows which various descriptions are valid.
Still more removed from direct acquaintance, imagine that someone like you or I comes along and makes a statement about Bismarck that is a description based on a "more or less vague mass of historical knowledge." We say that Bismarck was the "first Chancellor of the German Empire." In order to make a valid description applicable to the physical object, Bismarck's body, we must find a relation between some particular with which we have acquaintance and the physical object, the particular with which we wish to have an indirect acquaintance. We must make such a reference in order to secure a meaningful description.
To usefully distinguish particulars from universals, Russell posits the example of "the most long-lived of men," a description which wholly consists of universals. We assume that the description must apply to some man, but we have no way of inferring any judgment about him. Russell remarks, "all knowledge of truths, as we shall show, demands acquaintance with things which are of an essentially different character from sense-data, the things which are sometimes called 'abstract ideas', but which we shall call 'universals'." The description composed only of universals gives no knowledge by acquaintance with which we might anchor an inference about the longest-lived man. A further statement about Bismarck, like "The first Chancellor of the German Empire was an astute diplomatist," is a statement that contains particulars and asserts a judgment that we can only make in virtue of some acquaintance (like something heard or read).
Statements about things known by description function in our language as statements about the "actual thing described;" that is, we intend to refer to that thing. We intend to say something with the direct authority that only Bismarck himself could have when he makes a statement about himself, something with which he has direct acquaintance. Yet, there is a spectrum of removal from acquaintance with the relevant particulars: from Bismarck himself, "there is Bismarck to people who knew him; Bismarck to those who only know of him through history" and at a far end of the spectrum "the longest lived of men." At the latter end, we can only make propositions that are logically deducible from universals, and at the former end, we come as close as possible to direct acquaintance and can make many propositions identifying the actual object. It is now clear how knowledge gained by description is reducible to knowledge by acquaintance. Russell calls this observation his fundamental principle in the study of "propositions containing descriptions": "Every proposition which we can understand must be composed wholly of constituents with which we are acquainted."
Indirect knowledge of some particulars seems necessary if we are to expressively attach meanings to the words we commonly use. When we say something referring to Julius Caesar, we clearly have no direct acquaintance with the man. Rather, we are thinking of such descriptions as "the man who was assassinated on the Ides of March" or "the founder of the Roman Empire." Since we have no way of being directly acquainted with Julius Caesar, our knowledge by description allows us to gain knowledge of "things which we have never experienced." It allows us to overstep the boundaries of our private, immediate experiences and engage a public knowledge and public language.
This knowledge by acquaintance and knowledge by description theory was a famous epistemological problem-solver for Russell. Its innovative character allowed him to shift to his moderate realism, a realism ruled by a more definite categorization of objects. It is a theory of knowledge that considers our practice of language to be meaningful and worthy of detailed analysis. Russell contemplates how we construct a sense of meaning about objects remote from our experience. The realm of acquaintance offers the most secure references for our understanding of the world. Knowledge by description allows us to draw inferences from our realm of acquaintance but leaves us in a more vulnerable position. Since knowledge by description also depends on truths, we are prone to error about our descriptive knowledge if we are somehow mistaken about a proposition that we have taken to be true.
Critics of this theory have held that Russell's hypothesis of knowledge by description is confusing. His comments when defining sense-data, that the physical world is unknowable to us, contradict his theory of knowledge by descriptions. He implies that "knowledge by description" is not really a form of knowledge since we can only know those things with which we are acquainted and we cannot be acquainted with physical objects. Russell's theory amounts to the proposition that our acquaintance with mental objects appears related in a distant way to physical objects and renders us obliquely acquainted with the physical world. Sense-data are our subjective representations of the external world, and they negotiate this indirect contact.
While innovative, Russell's theory of knowledge by description is not an attractive theory of knowledge. It is clearly unappealing because our impressions of the real world, on his view, are commensurate with muddy representations of reality. Though we have direct access to these representations, it seems impossible to have any kind of direct experience of reality. Reality, rather, consists in unconscious, inferential pieces of reasoning.
Readers' Notes allow users to add their own analysis and insights to our SparkNotes—and to discuss those ideas with one another. Have a novel take or think we left something out? Add a Readers' Note!
| 889
|
Choose Privacy Week
Privacy is a particularly slippery and amorphous issue, about which people hold a wide variety of opinions and beliefs, particularly in the post-9-11 world in which we live.
Libraries and library workers think about privacy issues a lot - and want our customers to think critically about privacy issues, too. Personal privacy issues touch everyone at virtually every stage of life, and even into death (e.g., access to the Social Security Death Index online), raising a universe of hard questions to be answered.
Sponsored by the ALA’s Office for Intellectual Freedom (OIF), Choose Privacy Week is an annual initiative inviting library users of all ages and backgrounds into a national conversation about privacy rights in a digital age. The theme for this year's Choose Privacy Week is "Freedom from Surveillance."
Libraries have been interested in maintaining the privacy of individuals. Beaufort County Library Board of Trustees adopted the American Library Association Core of Ethics years ago. It's part of our core values as library workers. Why? Because freedom of speech is meaningless without the freedom to read.
Click here [http://www.privacyrevolution.org/images/uploads/Trina_UserHandout.pdf] for handout explaining why librarians and libraries insist upon empowering our customers to explore, research, and make choices based upon their individual needs.
Q: Where are the lines drawn between "right of privacy" and "right to know" today ?
The major source used in the preparation of this entry was http://www.privacyrevolution.org/. Please explore the website - and think hard about where you stand on the issue of individual privacy rights.
| 45
|
History of Initiative & Referendum in Arizona
|Laws • History|
|List of measures|
The History of Initiative & Referendum in Arizona began when acquired statewide initiative, referendum, and recall rights at the time of statehood in 1912. The first initiative in the state was for women's suffrage. It was a landslide victory, passing by a margin of greater than two to one on Nov. 5, 1912.
Then, in 1914, Arizona saw of 15 qualified initiatives, which held the record until 2006 when 19 initiatives were passed. Four of the 1914 initiatives passed because of the efforts of organized labor. One prohibited blacklisting of union members; a second established an "old age and mothers' pension"; another established a state government contract system, and a fourth limited businesses employment of non-citizens. Lastly, the voters in 1914 passed an initiative that barred the governor and legislature from amending or repealing initiatives.
In response, the legislature tried to pass a constitutional amendment that would make it more difficult to pass initiatives. Because this amendment needed the approval of voters, the Arizona Federation of Labor waged a campaign against the measure. The amendment was narrowly defeated in 1916.
- This chart includes all ballot measures to appear on the Arizona ballot in the year indicated, not just initiated measures. See also Arizona ballot measures.
|Year||Propositions on ballot||How many were approved?||How many were defeated?|
Arizonans owe many of their reforms to John Kromko. Kromko, like most Arizonans, is not a native; he was born near Erie, Pennsylvania, in 1940 and moved to Tucson in the mid-1960s. He was active in protests against the Vietnam War, and in the 1970s and 1980s he was elected to the lower house of the state legislature several times. By night, he was a computer-programming instructor; by day, he was Arizona’s "Mr. Initiative."
Kromko’s first petition was a referendum drive to stop a Tucson city council ordinance banning topless dancing, arguing for free speech. In 1976 Kromko was among the handful of Arizonans who, in cooperation with the People’s Lobby Western Bloc campaign, succeeded in putting on the state ballot an initiative to phase out nuclear power. The initiative lost at the polls, but Kromko’s leadership on the issue got him elected to his first term in the legislature.
Repealing the sales tax on food
Once elected, Kromko set his sights on abolishing the sales tax on food, a "regressive" tax that hits the poor hardest. Unsuccessful in the legislature, Kromko launched a statewide initiative petition and got enough signatures to put food tax repeal on the ballot. The legislature, faced with the initiative, acted to repeal the tax.
After the food tax victory, Kromko turned to voter registration reform. Again the legislature was unresponsive, so he launched an initiative petition. He narrowly missed getting enough signatures in 1980, and he failed to win re-election that year.
Undaunted, he revived the voter registration campaign and turned to yet another cause: Medicaid funding. Arizona in 1981 was the only state without Medicaid, since the legislature had refused to appropriate money for the state's share of this federal program.
In 1982, with an initiative petition drive under way and headed for success, the legislature got the message and established a Medicaid program. Kromko and his allies on this issue, the state’s churches, were satisfied and dropped their petition drive.
Motor Voter initiative
The voter registration initiative, now under the leadership of Les Miller, a Phoenix attorney, and the state Democratic Party, gained ballot placement and voter approval. In the ensuing four years, this "Motor Voter" initiative increased by over 10 percent the proportion of Arizona’s eligible population who were registered to vote.
Late legislative career
Kromko, re-elected to the legislature in 1982, took up his petitions again in 1983 to prevent construction of a freeway in Tucson that would have smashed through several residential neighborhoods. The initiative was merely to make freeway plans subject to voter approval, but Tucson officials, seeing the campaign as the death knell for their freeway plans, blocked its placement on the ballot through various legal technicalities. Kromko and neighborhood activists fighting to save their homes refused to admit defeat. They began a new petition drive in 1984, qualified their measure for the ballot, and won voter approval for it in November 1985.
Arizona’s moneyed interests poured funds into a campaign to unseat Kromko in 1986. Kromko not only survived but also fought back by supporting a statewide initiative to limit campaign contributions, sponsored by his colleague in the legislature, Democratic State Representative Reid Ewing of Tucson. Voters passed the measure by a two to one margin.
Kromko’s initiative exploits have made him the most effective Democratic political figure, besides former governor Bruce Babbitt, in this perennially Republican-dominated state. And Babbitt owes partial credit for one of his biggest successes - enactment of restrictions on the toxic chemical pollution of drinking water - to Kromko. Early in 1986 Kromko helped organize an environmentalist petition drive for an anti-toxic initiative, while Babbitt negotiated with the legislature for passage of a similar bill. When initiative backers had enough signatures to put their measure on the ballot, the legislature bowed to the pressure and passed Babbitt's bill. Even today, Kromko is still active in politics, writing letters to the editor about immigration policies.
Petition drive problems in 2008
2008 was a tough year for ballot initiatives in Arizona. Nine citizen initiatives filed signatures to qualify for the November 2008 Arizona ballot by the state's July 3 petition drive deadline. In the end, only six of the initiatives were certified, with three initiatives disqualified as a result of an historically high number of problems with flawed petition signatures. When the November vote was held, of the six that qualified for the ballot, only one was approved.,
Criticisms of process
After 19 were proposed in 2006, legislators were worried about "ballot fatigue" or overuse of the initiative system. This led to legislators considering steps to limit or otherwise exert more control over the initiative process. Ironically, any attempt to alter the initiative and referendum process would require an amendment to the state constitution, and thus in itself be put forth as a referendum.
This article is significantly based on an article published by the Initiative & Referendum Institute, and is used with their permission. Their article, in turn, relies on research in David Schmidt's book, Citizen Lawmakers: The Ballot Initiative Revolution.
Also portions of this article were taken from Wikipedia, the free encyclopedia under the GNU license.
- ↑ Arizona Daily Star, "'Clown' takes some serious initiative", July 20, 2007
- ↑ Arizona Republic, "'Flawed' election petitions face review", September 13, 2008
- ↑ Phoenix New Times, "Citizen initiatives have been kicked off the ballot this year in record numbers, and the problems could go much deeper than invalid signatures", August 21, 2008
- ↑ Legislators seeking more control over initiatives, Arizona Republic, Feb. 13, 2007
- ↑ History of Arizona's initiative
- ↑ Citizen Lawmakers: The Ballot Initiative Revolution Temple University Press, 352 pp., ISBN-10: 0877229031, October 1991
History of I&R
Alaska · Arizona · Arkansas · California · Colorado · Florida · Idaho · Illinois · Kentucky · Maine · Maryland · Massachusetts · Michigan · Mississippi · Missouri · Montana · Nebraska · Nevada · New Mexico · North Dakota · Ohio · Oklahoma · Oregon · South Dakota · Utah · Washington · Wyoming
Direct Legislation by the Citizenship Through the Initiative and Referendum · Citizen Lawmakers: The Ballot Initiative Revolution · Direct Legislation: Voting on Ballot Propositions in the United States
| 819
|
Hold the salt: UCLA engineers develop revolutionary new desalination membrane
Process uses atmospheric pressure plasma to create filtering 'brush layer'
Desalination can become more economical and used as a viable alternate water resource.
By Wileen Wong Kromhout
Originally published in UCLA Newsroom
Researchers from the UCLA Henry Samueli School of Engineering and Applied Science have unveiled a new class of reverse-osmosis membranes for desalination that resist the clogging which typically occurs when seawater, brackish water and waste water are purified.
The highly permeable, surface-structured membrane can easily be incorporated into today's commercial production system, the researchers say, and could help to significantly reduce desalination operating costs. Their findings appear in the current issue of the Journal of Materials Chemistry.
Reverse-osmosis (RO) desalination uses high pressure to force polluted water through the pores of a membrane. While water molecules pass through the pores, mineral salt ions, bacteria and other impurities cannot. Over time, these particles build up on the membrane's surface, leading to clogging and membrane damage. This scaling and fouling places higher energy demands on the pumping system and necessitates costly cleanup and membrane replacement.
The new UCLA membrane's novel surface topography and chemistry allow it to avoid such drawbacks.
"Besides possessing high water permeability, the new membrane also shows high rejection characteristics and long-term stability," said Nancy H. Lin, a UCLA Engineering senior researcher and the study's lead author. "Structuring the membrane surface does not require a long reaction time, high reaction temperature or the use of a vacuum chamber. The anti-scaling property, which can increase membrane life and decrease operational costs, is superior to existing commercial membranes."
The new membrane was synthesized through a three-step process. First, researchers synthesized a polyamide thin-film composite membrane using conventional interfacial polymerization. Next, they activated the polyamide surface with atmospheric pressure plasma to create active sites on the surface. Finally, these active sites were used to initiate a graft polymerization reaction with a monomer solution to create a polymer "brush layer" on the polyamide surface. This graft polymerization is carried out for a specific period of time at a specific temperature in order to control the brush layer thickness and topography.
"In the early years, surface plasma treatment could only be accomplished in a vacuum chamber," said Yoram Cohen, UCLA professor of chemical and biomolecular engineering and a corresponding author of the study. "It wasn't practical for large-scale commercialization because thousands of meters of membranes could not be synthesized in a vacuum chamber. It's too costly. But now, with the advent of atmospheric pressure plasma, we don't even need to initiate the reaction chemically. It's as simple as brushing the surface with plasma, and it can be done for almost any surface."
In this new membrane, the polymer chains of the tethered brush layer are in constant motion. The chains are chemically anchored to the surface and are thus more thermally stable, relative to physically coated polymer films. Water flow also adds to the brush layer's movement, making it extremely difficult for bacteria and other colloidal matter to anchor to the surface of the membrane.
"If you've ever snorkeled, you'll know that sea kelp move back and forth with the current or water flow," Cohen said. "So imagine that you have this varied structure with continuous movement. Protein or bacteria need to be able to anchor to multiple spots on the membrane to attach themselves to the surface — a task which is extremely difficult to attain due to the constant motion of the brush layer. The polymer chains protect and screen the membrane surface underneath."
Another factor in preventing adhesion is the surface charge of the membrane. Cohen's team is able to choose the chemistry of the brush layer to impart the desired surface charge, enabling the membrane to repel molecules of an opposite charge.
The team's next step is to expand the membrane synthesis into a much larger, continuous process and to optimize the new membrane's performance for different water sources.
"We want to be able to narrow down and create a membrane selection system for different water sources that have different fouling tendencies," Lin said. "With such knowledge, one can optimize the membrane surface properties with different polymer brush layers to delay or prevent the onset of membrane fouling and scaling.
"The cost of desalination will therefore decrease when we reduce the cost of chemicals [used for membrane cleaning], as well as process operation [for membrane replacement]. Desalination can become more economical and used as a viable alternate water resource."
Cohen's team, in collaboration with the UCLA Water Technology Research (WaTeR) Center, is currently carrying out specific studies to test the performance of the new membrane's fouling properties under field conditions.
"We work directly with industry and water agencies on everything that we're doing here in water technology," Cohen said. "The reason for this is simple: If we are to accelerate the transfer of knowledge technology from the university to the real world, where those solutions are needed, we have to make sure we address the real issues. This also provides our students with a tremendous opportunity to work with industry, government and local agencies."
A paper providing a preliminary introduction to the new membrane also appeared in the Journal of Membrane Science last month.
Published: Thursday, April 08, 2010
| 157
|
Includes bibliographical references (p. -491) and indexes.
PART ONE: FRAMEWORK Guide to This Book Qualitative Research: Why and How To Do It Qualitative and Quantitative Research Ethics of Qualitative Research PART TWO: FROM THEORY TO TEXT Making Use of the Literature in Qualitative Research Theoretical Positions Underlying Qualitative Research Epistemological Background: Construction and Understanding of Texts PART THREE: RESEARCH DESIGN The Qualitative Research Process Research Questions Entering the Field Sampling How to Design Qualitative Research: An Overview PART FOUR: VERBAL DATA Interviews Narratives Focus Groups Verbal Data: An Overview PART FIVE: OBSERVATION AND MEDIATED DATA Observation and Ethnography Visual Data: Photography, Film, and Video Using Documents as Data Qualitative Online Research: Using the Internet Observation and Mediated Data: An Overview PART SIX: FROM TEXT TO THEORY Documentation of Data Coding and Categorizing Analyzing Conversation, Discourse, and Genres Narrative and Hermeneutic Analysis Using Computers in Qualitative Analysis Text Interpretation: An Overview PART SEVEN: GROUNDING AND WRITING QUALITATIVE RESEARCH Quality Criteria in Qualitative Research The Quality of Qualitative Research: Beyond Criteria Writing Qualitative Research PART EIGHT: QUALITATIVE RESEARCH: INTEGRATION AND OUTLOOK Qualitative Research At Work I: Grounded Theory Qualitative Research At Work II: Triangulation Qualitative Research: State of the Art and the Future Glossary References Author index Subject index About the Author.
(source: Nielsen Book Data)
'The fourth edition of Uwe Flick's "Introduction to Qualitative Research" remains the most comprehensive and thorough text in qualitative research. It is student - and user-friendly, thoroughly up-to-date in terms of the latest developments in the field, imminently practical. It is the single most important introductory book on qualitative inquiry in the social sciences today' - Norman K. Denzin, University of Illinois. The new edition of Uwe Flick's bestselling textbook has been fully revised, expanded and updated. "An Introduction to Qualitative Research" guides the student step-by-step through the research process of qualitative research. This classic text covers all of the main theoretical approaches to qualitative research, and provides unmatched coverage of the full range of different qualitative methods and approaches now available to researchers. A range of new features have been added to the new edition including: new structure to better meet the needs of teaching qualitative research; a new chapter on Grounded Theory plus updated coverage on the full range of other qualitative methods; a summary section discussing the state-of-the-art in qualitative research; a glossary; and, updated cases studies, exercises and guided questions. This new edition will continue to ensure that. "An Introduction to Qualitative Research" remains an essential introductory text for all students of qualitative research. (source: Nielsen Book Data)
| 673
|
Newegg.com - A great place to buy computers, computer parts, electronics, software, accessories, and DVDs online. With great prices, fast shipping, and top-rated customer service - once you know, you Newegg.
If you are reading this message, Please click this link to reload this page.(Do not use your browser's "Refresh" button). Please email us if you're running the latest version of your browser and you still see this message.
Table of contents
Bluetooth is an industrial specification for wireless data transfer. Bluetooth connectivity is often found in high-end keyboards and mice. Bluetooth generally provide an operating range of up to 30 feet and is less prone to interference in comparison to RF technology.
DPI and FPS
DPI (dots per inch) and FPS (frames per second) are the number of counts in an inch of movement and the number of times the sensor reads the surface in a second respectively. These figures are measures of the amount of information recorded by the mouse's sensor. The greater the amount of information that is gathered, the more accurately and precisely the surface can be tracked. To detect movement, optical and laser mice use sensors to read beams of light as they are reflected from the tracking surface.
Currently 400 and 800 DPI optical mice as well as 800 DPI laser mice are very popular, but some high-end models are capable of 1000, 1600 or even 2000 DPI tracking speeds.
The Personal System/2 or PS/2 was the designation for IBM's second generation of personal computers. The PS/2 keyboard and mouse ports were introduced with it. PS/2 ports connect the keyboard and mouse to a computer and are usually color-coded on today's systems - purple for keyboards and green for mice. Most desktop motherboards still provide PS/2 ports, but an increasing number of keyboards and mice are using USB ports.
Radio Frequency (RF) is a wireless communication technology. Using RF technology allows keyboards and mice to computers without annoying cables.
The USB (Universal Serial Bus) port is a popular I/O interface used for connecting computers and peripherals or other devices. It is capable of supporting up to 127 daisy-chained peripheral devices simultaneously. The latest USB 2.0 specification can deliver 480Mbps data transfer bandwidth. In addition, USB provides plug-and-play capabilities to allow device changes while the computer is powered on. Today, many keyboard and mice use the USB interface.
| 171
|
No flood, no money. What a trade!
When asked about his future, Nguyen Van Nghe, a fisherman in Dong Thap Province answered: “No flood, no money. What a trade!” He was referring to the fact that over the last 10 years the construction of high dykes in Dong Thap and the Long Xuyen Quadrangle has blocked the annual replenishment of freshwater, nutrients, and sediment on which fish depend. In turn, this has destroyed the wild fisheries on which the landless, including Nghe, who make up 20% of the Mekong Delta’s population, depend. Wild fish production has declined by 40% over the last 10 years and species that were once used as fertilizer now sell for VND180,000/kg. The predatory snakehead, which once occupied the top of the aquatic food chain, has disappeared. The high dykes have also greatly reduced the annual flushing. This has resulted in the accumulation of pathogens and toxins in the surface water and growing public health problems.
High dykes were built to allow a third, or autumn-winter, rice crop to be grown because of the high prices this off-season crop can fetch on the international market. In 2011, 560,000 hectares of autumn-winter rice were planted, up from 520,000 hectares in 2010. But because of the loss of sediment, rice productivity can only be maintained through the heavy use of fertilizer. Nguyen Huu Thien, a wetland specialist based in Can Tho, questions whether the third rice crop is profitable once you take into account the increased use of fertilizer and pesticide, the cost of dyke maintenance, the loss of wild fish, and, inevitably, the cost of dyke failure: when dykes failed in 2011, 50 people were killed and tens of millions of dollars of houses, roads, and other infrastructure was destroyed. The intensification of rice production has also resulted in the virtual extinction of the traditional long-stem floating rice varieties, which in Brazil sell for $3,500/ton, almost ten times the price of autumn-winter rice.
Dr. Ngo Van Be of the Dong Thap Muoi Institute of Research and Development says that the floods that used to be “mild” are now “fierce” and unpredictable. In hydrological terms, what the high dykes have done is to separate the Mekong River from its 1.5 million-hectare floodplain. According to Dr. Le Anh Tuan of Can Tho University, these dykes have narrowed the floodplain during the peak October-November flood from 150 kilometers to a few tens of kilometers. This has accelerated the water flow and displaced flooding to residential areas downstream. Reduction of the flooded area has also reduced groundwater recharge, reduced river base flows, and increased dry season saline intrusion, which increases the cost of drinking water supplies. The violent floods of 2011 call into question the value of the third rice crop and instead argue for a more natural hydrology that provides multiple benefits, including greater resilience to climate change, which is likely to result in more intense rainfalls and flash flooding.
To learn more about these issues, watch this 30-minute file produced by the Center for Water Resources Conservation and Development (WARECOD) and VTC16.
| 16
|
U.S. Navy in Florida
The presence of the U.S. Navy in Florida officially began with the transfer of the peninsula by Spain to the U.S. in 1821, and the subsequent establishment of Florida as a U.S. Territory. Two of the earliest American military vessels which sank in Florida waters were actually American privateers captured by Britain's Royal Navy during the American Revolution, and which later sank while in Royal Navy service. The captured American privateer Who's Afraid was placed into the Royal Navy and renamed HMS Mentor. The captured American privateer Independence was placed into the Royal Navy and renamed HMS Racer. HMS Mentor was burned by the British at Pensacola in 1781 to prevent her capture by invading Spanish forces. HMS Racer foundered in a gale in the Gulf of Mexico in 1814. The first commissioned Navy ship lost in Florida waters was USS Alligator, which wrecked in the Florida Keys in 1822, while assigned to combat piracy. Many naval, and other military craft have been sunk in Florida waters since 1821, through military action, foundering, grounding, fire and explosion, deliberate sinking to create artificial reefs, and other causes. Such shipwrecks are the major focus of the Florida Navy Legacy Shipwreck Project.
During its long history in Florida, the U.S. Navy has established a number of naval bases and stations. The most strategically significant of these have been at Pensacola, Key West, and Jacksonville. There has been a U.S. Navy presence at Pensacola since 1825, at Key West since 1822, and more recently at Jacksonville since the Second World War. There have been several military conflicts in Florida since 1821 in which the U.S. Navy has played an important role. These were the First Seminole War (1816-1818), the Second Seminole War (1835-1842), the Third Seminole War (1855-1858), the Civil War (1861-1865), the Spanish American War (1898), the First World War (1917-1918), the Second World War (1941-45), and events such as the Cuban Missile Crisis (1963).
During the Seminole Wars, the Spanish American War, and First and Second World Wars, there were a number of smaller Naval stations established along Florida's vast coastline. During the Seminole Wars, these stations were often operated as supply depots, intended for the joint Army-Navy operations being conducted. One example is the Naval Depot established on Key Biscayne in 1836, which lasted until 1926. During the First and Second World Wars, with new developments in aviation technology, these coastal naval stations were outfitted to serve as stations for lighter than air craft (dirigibles), conventional aircraft, and smaller surface craft, used to patrol Florida's coastline and offshore waters. Between the spring of 1942, and the fall of 1943, German submarines attacked and sank large numbers of ships on Florida's Atlantic and Gulf coasts. The Navy's system of coastal defense was ultimately one of the major factors which turned the tide against the U-Boats and caused Germany's Naval Command to redirect its effort to other locations where targets were easier to attack.
Since the Second World War, the U.S. Navy has continued to maintain a significant presence in Florida. The Key West Naval Station became a major staging area for the Navy during the Cuban Missile Crisis, for example. Recognizing the on-going strategic significance of the Key West Naval Station, the Navy invested large amounts of money in the early 1980s to upgrade its facilities. Since the stationing of the Navy's first aircraft carrier at Pensacola in 1925, Pensacola has become the focus for naval aviation activities and training. A Naval Training Center was in operation at Orlando from 1968 to 1994. Today in Florida, there are naval facilities at Pensacola, Panama City, Key West, Homestead, Mayport and Jacksonville, as well as some other smaller stations.
The U.S. Navy's history in Florida spans 175 years, from the earliest days of U.S. territorial expansion to the Space and Nuclear Age. Although much valuable work has already been done, much of this interesting chapter of U.S. Naval history remains to be studied and written by future historians.
| 511
|
You are here: PureEnergySystems.com > News > October E-Cat Test Validates Cold Fusion Despite Challenges The test of the E-Cat (Energy Catalyzer) that took place on October 6, 2011 in Italy has validated Andrea Rossi's claim that the device produces excess energy via a novel Cold Fusion nuclear reaction. Despite its success, the test was flawed, and could have been done in a way that produced more spectacular results -- as if confirmation of cold fusion is not already stunning enough. Andrea Rossi stands in front of his E-Cat apparatus, October 6, 2011 Photo by Maurizio Melis of Radio24 by Hank Mills Pure Energy Systems News Andrea Rossi has made big claims for the past year, about his cold fusion "E-Cat" (Energy Catalyzer) technology. He has claimed that it produces vast amounts of energy via a safe and clean low energy nuclear reaction that consumes only tiny amounts of nickel and hydrogen. A series of tests had been performed earlier this year that seemed to confirm excess energy is produced by the systems tested. Some of the tests were particularly impressive, such as one that lasted eighteen hours, and was performed by Dr. Levi of the University of Bologna. Unfortunately, the tests were not planned out as well as they could have been and had flaws. The most recent test that took place on October 6, 2011 in Bologna, Italy, was supposed to address many of the concerns about the previous tests, and be performed in a way that would put to rest many issues that had been discussed continually on the internet. Despite showing clear evidence of excess energy -- which is absolutely fantastic -- this most recent test failed to live up to its full potential. It was a big success in that it validated the claim the E-Cat produces excess energy via cold fusion, but it was not nearly as successful as it could have been. Or as successful as we, the outsiders looking in, would like for it to have been. The Inventor's Mindset One thing that should be stated is that inventors do not always think like the people who follow their inventions. They have their own mindset and way of looking at things. This should be obvious, because they are seeing *everything* from a different perspective. For example, when we think seeing the inside of an important component would be exciting and informative, they consider it a threat to their intellectual property. Or, for example, when we would like to see a test run for days, they are thinking that a few hours is long enough. In their mind, they know their technology works, and running it for hours, days, or weeks would be more of a chore to them than an exciting event. In Rossi's case, he has worked with these reactors for many years. He has tested them time and time again. In fact, he has built hundreds of units (of different models), and has tested every one of them. He is aware of how the units operate and how they perform. Actually, for a period of many months to a year or more, he had an early model of E-Cat heating one of his offices in Italy. Satisfying the curiosity of internet "chatters" by operating a unit for an extended period of time -- beyond what he thinks is needed to prove the effect -- is just a waste of his time, according to his thinking. He could spend the time getting the one megawatt plant ready to launch. Don't forget, Rossi is a busy person. In addition to finishing the one megawatt plant, he has a new partner company to find, a wife at home, and a life to live! We need to consider that he works sixteen to eighteen hours a day building units, testing them, addressing other issues about the E-Cat. Although he is a very helpful person in many ways (willing to communicate with people and answer questions), he simply does not have the time to grant all of the many requests made of him. If he did, he could not get any work done at all, and the E-Cat would never be launched, or ever make it to the market place! The Outsider's Mindset I consider myself an outsider. I have never built a cold fusion device, have never spent years working to develop a technology, and have never gone through the grueling process of trying to bring a product into the market place. Although I spend a lot of time researching various technologies on the internet, I don't work sixteen to eighteen hours a day. In addition, I have no vested interest in the success of any technology, other than simply wanting at least one to hit the market place, ASAP. As an outsider, I do not think like Rossi thinks. I don't think the majority of people think like Rossi thinks, because they are not in his shoes. They are not working to the point of exhaustion, and do not have years of their life invested in an exotic technology. Due to the fact we do not think like Rossi, his actions or sometimes lack thereof can seem strange, bizarre, or odd. Sometimes, they can make us want to smack ourselves, to make sure we are not in some sort of strange dream. The recent test on October 6, 2011 is an example of a situation in which outsiders would have liked to have seen a very different test. Here are examples of how an outsider would have liked to see the test performed, compared to Rossi's possible mindset. (Please note that I am making speculations about what Rossi is thinking, and his mindset. I do not know for sure if my guesses are accurate. If they are not, then I would like to apologize to Rossi, and give him the chance to respond in any way he sees fit.) In the recent test, the output producing capability of the reactors was throttled down for safety reasons. This may have been done by keeping the hydrogen pressure low, or adding less of the catalyst to the nickel powder. Also, only one out of the three reactors that were inside of the module, were used in the test. For an experimental test to prove the effect beyond a shadow of a doubt, I, as an outsider, would have loved to have seen the device fully throttled up, despite the safety risks. Even if it meant everyone that attended would have had to sign long legal disclaimers, it would have been worth it. I think it would have been great if all three reactors were utilized, and they all were adjusted to produce their maximum level of output. This would have increased the amount of output produced dramatically, and would have reduced the amount of input needed. The more heat produced by the system, the less heat would have needed to be input via the electric resistors. Rossi, on the other hand, probably thought throttling up the device to a high level was not worth the risk, and was not needed to prove that excess energy was being produced. It is true that an explosion causing injuries -- while probably VERY unlikely -- could result in a setback of his project, and possible legal ramifications. Also, in reality, the test proved excess energy was being produced even with only one, throttled down reactor being used. So even though a test of the device adjusted to operate at full power would have been useful and exciting, it was not absolutely needed for what Rossi wanted to accomplish. I would like to ask Rossi to consider performing a demonstration with a module both adjusted to operate at full power, and utilizing all three reactor cores. Even if he has to limit the number of people involved, perform the test remotely with cameras monitoring the module, utilize a blast shield, or only allow certain individuals (who have signed disclaimers) to go into the room in which the module is running. A Longer Self Sustain As an outsider, I have not had the chance to look at test data from these devices self-sustaining for long periods of time -- 12 hours, 24 hours, days, weeks, etc. I would really like to see one of these units self sustain for a *very* long period of time. This is not because I think the output of the E-Cat during the recent test was due to stored energy being released (the 'thermal inertia' theory being floated around the internet). In fact, I think that the flat line in NyTeknik's graph -- showing self sustain mode for three and a half hours without any drop in output temperature -- provides clear evidence against the thermal inertia theory. The reason I would like to see a longer period of self sustain, is that it would not only document a huge gain of energy, but one that no individual could rationally deny! Rossi has claimed that these devices represent an alternative energy solution that could change the world. I think this is true. However, to show just how much potential this technology has, an even more extended test of the E-Cat in self sustain mode (at full power or at least with all three reactors inside the module being used), would have been much more impressive. I am not saying the Oct. 6 test was not impressive -- it was very significant because it demonstrated excess energy and proof of cold fusion -- but that a longer test would have been better. It would have done more to shut up the cynics (a few of which will never change their minds), and help the technology get into the mainstream (dumbstream) media. I really don't think Rossi cares too much about showing off the technology's full potential, at this point. He also does not seem to appear to want the attention of the mainstream media, or at least any more than he thinks he needs. If he did, the test would have been far different, and would have produced such a gigantic amount of excess energy everyone's jaws would have dropped. My jaw dropped when I saw the flat line during self sustain mode (because it proved beyond a doubt the system was producing excess energy), but my jaw did not drop as far as it could have, if the period of operation had lasted longer. Interestingly, I have known inventors, of unrelated energy technologies, that purposely held back from showing the *best* version of their technology. They did not want to show off too much, because they did not want to deal with the fallout of attracting too much attention. Instead of performing an amazing demonstration, they performed one that proved the point -- at least to their satisfaction -- but would not attraction too much attention. I think Rossi may feel the same way. If he had his way, he would have never done a single test before the launch of the one megawatt plant. It was Focardi that convinced him to do a public test, because he feared that (due to health problems) he may not live long enough to see the technology be revealed to the world. A longer test (at least 12 hours) in self sustain mode would have been great, exciting, and would have produced even more excess energy. However, in Rossi's mind, it was not needed, for potentially valid reasons (at least from the perspective of someone on the inside). I would simply like to humbly plead with Rossi, to try and step in the shoes of the outsider, and at the next test allow the module to run for a longer period in self sustain mode. Modern Testing Methods and Tools I have looked at the data acquired during the test, but have not had a chance to study it as in depth as I would like to. The data shows a clear gain of energy in my opinion, and confirms that the E-Cat is producing excess energy. As I said before, the test was a success. However, it could have been performed in a more modern way. For example, all of the temperature measurements, power input measurements, and water flow measurements should have been fed into the same computer, to be recorded in a real time manner. This way all the data would have been automatically recorded into one data set, including the hour and second of every measurement. It seems data collection was not done this way at the test, and some of the data was actually taken by hand! Because the data was not all automatically recorded into one computer during the test, NyTeknik (who had the exclusive right to be the first to post a report on the test) has not yet posted a graph that charts all the measurements of all the factors of the test. What I would like to see is a single high resolution graph, that shows all of the measurements that were taken of every parameter of the test. If one graph showing everything would be too complex for a non-expert to easily interpret, then a series of graphs would be ideal. This would allow everyone to more simply determine the total energy in, and the total energy out. The data collected and the manner in which it was collected is good enough to show there was a significant amount of excess energy produced, especially during self sustain mode. It may also be good enough to show even more details about the excess energy produced. Sadly, I'm not an expert in scientific data interpretation, so it takes me more time to interpret data than an expert who does so full time (like Rossi). I hope that when I have had the time to examine the data in more depth, I will see that Rossi's claims about the results of the test (not just excess energy but a six fold gain of energy, in a worst case scenario) are accurate. At this point, I am not going to doubt him. He is the expert, and there are many people going over the data, and hopefully more data from the test will be coming in the near future. What I would like to do, is request that he upgrade his data acquisition methods for any upcoming public tests. However, from Rossi's perspective, the way the data was acquired was good enough, and proved the point he wanted to make. I respect his view, but I do hope that he will change his mind in the future. For the record, I am not stating that I think better data acquisition techniques are needed to verify his technology produces excess energy, and even significant amounts of it. I simply think it would make analysis of test data much simpler, quicker, and precise. One of the most useful tools in the scientific method is a control. A control is an object or thing that you do not try to change during the experiment. For example, if you were giving an experimental drug to a hundred people, you might want to have a number of additional people who do not receive the drug. You would compare how the drug effects the people who consumed it, to those who did not receive the drug at all. By comparing the two sets of people, those who consumed the drug and those who did not, you could more easily see the effectiveness of the drug -- or if it was doing harm. In Rossi's test, a control system would have been an E-Cat module that was setup in the exact same way, except it would have not been filled with hydrogen gas. It would have had the same flow of water going through it, the same electrical input, and it would have operated for the same length of time as the E-Cat unit with hydrogen. By comparing the two, you could easily see the difference between the "control" E-Cat (that was not having nuclear reactions take place), and the "real" E-Cat (that was producing excess heat). If a control had been used in the experiment, the excess heat would be even more obvious. It would have been so obvious, that it could have made the test go from a major success (with some flaws), to the most spectacular scientific test in the last hundred years. Yes, a control would have made that much of a difference! I understand that Rossi may not see the need for a control, when the test that was performed clearly showed excess energy without it. A control might have made the experiment so mind blowingly amazing, it could have attracted too much media attention, too many scientists that would want to get involved, and too many individuals wanting additional information. The result could have been that Rossi would not even have the time to finish his one megawatt plant. However, from the view point of an outsider, I think a control would have greatly benefited the experiment. If it created too much media attention, perhaps someone could volunteer to work for a month as an unpaid intern, filtering through all of the requests from media representatives, and taking care of many non-technical tasks, so Rossi could focus on getting the one megawatt plant ready! I sincerely hope that during the test of the one megawatt plant, and any tests before then, a control run will be performed, in which no hydrogen is placed in the reactors. Rossi's Statement about the Test Results Andrea Rossi responded to an email we sent him that had questions about the test. Here is the email, and his responses. THANK YOU FOR YOUR CONTINUOUS ATTENTION. PLEASE FIND THE ANSWERS IN BLOCK LETTERS ALONG YOUR TEXT: Dear Andrea Rossi, In regards to the latest test of the Energy Catalyzer, I have a number of questions I hope you can answer. 1) My understanding is that if a reactor core is not adjusted to be under-powered (below its maximum potential) in self-sustain mode, it can have a tendency to become unstable and climb in output. If the reactor is left in an unstable self-sustaining mode for too long, the output can climb to potentially dangerous levels. Can you provide some information about how the reactor core in the test was adjusted to self-sustain in a safe manner? NO, VERY SORRY a) For example, there was only one active reactor core in the module tested. How was the single reactor core adjusted to be under-powered? b) Is adjusting the reactor core as simple as lowering the hydrogen pressure? 2) What is the power consumption of the device that "produces frequencies" that was mentioned in the NyTeknik article? Although the power consumption of this device is probably insignificant, providing a figure could help put to rest the idea (that some are suggesting) that a large amount of power was being consumed by the frequency-generating device, and transmitted into the reactor. THE ENERGY CONSUMED FROM THE FREQUENCY GENERATOR IS 50 WH/H AND IT HAS BEEN CALCULATED, BECAUSE THIS APPARATUS WAS PLUGGED IN THE SAME LINE WHERE THE ENERGY-CONSUME MEASUREMENT HAS BEEN DONE a) Can you tell us anything more about this frequency generating device and its function? NO, SORRY, THIS IS A CONFIDENTIAL ISSUE b) Is the frequency-generating device turned on at all times when a module is in operation, or only when a module is in self-sustain mode? c) Some are suggesting that this device is "the" catalyst that drives the reactions in the reactor core. However, you have stated in the past that the catalyst is actually one or more physical elements (in addition to nickel and hydrogen) that are placed in the reactor core. Can you confirm that physical catalysts are used in the reactor? YES, I CONFIRM THIS 3) Does the reaction have to be quenched with additional water flow though the reactor, or is reducing the hydrogen pressure enough to end the reactions on its own? NEEDS ADDITIONAL QUENCHING a) If reducing the hydrogen pressure (or venting it completely) is not enough to turn off the module, could it be due to the fact some hydrogen atoms are still bonded to nickel atoms, and undergoing nuclear reactions? b) If there is some other reason why reducing hydrogen pressure is not enough to quickly turn off the module, could you please specify? Thank you for taking the time to answer these questions, and for allowing a test to be performed that clearly shows anomalous and excess energy being produced. Hopefully, the world will notice the significance of this test. THANK YOU VERY MUCH, AND, SINCE I HAVE ABSOLUTELY NOT TIME TO ANSWER (I MADE AN EXCEPTION FOR YOU) PLEASE EXPLAIN THAT BEFORE THE SELF SUSTAINING MODE THE REACTOR WAS ALREADY PRODUCING ENERGY MORE THAN IT CONSUMED, SO THAT THE ENERGY CONSUMED IS NOT LOST, BUT TURNED INTO ENERGY ITSELF, THEREFORE IS NOT PASSIVE. ANOTHER IMPORTANT INFORMATION: IF YOU LOOK CAREFULLY AT THE REPORT, YOU WILL SEE THAT THE SPOTS OF DRIVE WITH THE RESISTANCE HAVE A DURATION OF ABOUT 10 MINUTES, WHILE THE DURATION OF THE SELF SUSTAINING MODES IS PROGRESSIVELY LONGER, UNTIL IT ARRIVES TO BE UP TO HOURS. BESIDES, WE PRODUCED AT LEAST 4.3 kWh/h FOR ABOUT 6 HOURS AND CONSUMED AN AVERAGE OF 1.3 kWh/h FOR ABOUT 3 HOURS, SO THAT WE MADE IN TOTAL DURING THE TEST 25.8 kWh AND CONSUMED IN TOTAL DURING THE TEST 3.9 kWh. IN THE WORST POSSIBLE SCENARIO, WHICH MEANS NOT CONSIDERING THAT THE CONSUME IS MAINLY MADE DURING THE HEATING OF THE REACTOR DURING THE FIRST 2 HOURS, WE CAN CONSIDER THAT THE WORST POSSIBLE RATIO IS 25.8 : 3.9 AND THIS IS THE COP 6 WHICH WE ALWAYS SAID. OF COURSE, THE COP IS BETTER, BECAUSE, OBVIOUSLY, THE REACTOR, ONCE IN TEMPERATURE, NEEDS NOT TO BE HEATED AGAIN FROM ROOM TEMPERATURE TO OPERATIONAL TEMPERATURE. WARMEST REGARDS TO ALL, ANDREA ROSSI He claims that the test produced 25.8 kilowatt hours of power, and consumed only 3.9 kilowatt hours, not considering the losses from using two circuits of water and a heat exchanger. This would be very impressive for a system that is only using one reactor core (out of three), that has been adjusted to only produce a fraction of its maximum potential power. However, from my analysis of the data so far (still trying to wrap my head around it), I have not been able to confirm his claim of a COP of 6. I am not saying it is not the case, or not in the data. I simply have yet to fully examine the data, and I am waiting for more data to be released. Actually, I hope that someone will release all the data in one file and/or graph that will be easier to interpret. Perhaps NyTeknik, if they have not done so already, could contact Rossi or someone else who attended and recorded the data, and ask for any test data they are missing. Bottom Line - Cold Fusion Is Here The fact of the matter was that the October 6th test was a success in many ways. - It documented a gain of energy. - It documented a gain of energy in self-sustain mode. - It documented massive "heat after death." Most importantly, it proved beyond a doubt, that cold fusion is a reality. Italian scientific journalist Maurizio Melis of Il Sole 24 Ore, who witnessed the test in Bologna, "In the coming weeks Rossi aims to activate a 1MW plant, which is now almost ready, and we had the opportunity to inspect it during the demonstration of yesterday. If the plant starts up then it will be very difficult to affirm that it is a hoax. Instead, we will be projected suddenly into a new energetic era." The test could have been made better in many ways. It had flaws. However, it was the most significant test of the E-Cat so far, for one reason in particular.... This graph shows that the E-Cat is a device producing excess energy, because the red line does not go down until after the hydrogen is vented. - Some may legitimately argue about how much energy was produced, because we don't yet have all the test data in one easy to interpret graph or file. - Some may point out the flaws in the test, such as the lack of a control, the lack of another several hours of operation in self sustain mode. - Some may point out ways the test could
| 48
|
Cancer of the cervix (sir-vix) is one of the most common cancers in women. There seems to be a connection between cervical cancer and sexual activity at an early age, especially when multiple partners are involved. Cervical cancer grows without symptoms, that's why a yearly pap smear is so important. A pap smear can detect the presence of cancer cells at an early stage. When precancerous cells are found, usually called dysplasia (dis-PLAY-zha), they can be removed in the doctor's office using various procedures that burn or freeze the cells off the cervix. If the cancer has advanced, the recommended treatment usually includes a combination of chemotherapy, radiation and surgery, which will prevent a woman from bearing children. For more information about cervical cancer, contact your healthcare provider.
| 876
|
Journal Issue: Juvenile Justice Volume 18 Number 2 Fall 2008
The Prevalence of Mental Disorders among Adolescent Offenders
Two kinds of studies address questions about the social consequences of the links between mental disorders and delinquency. One type examines the degree of "overlap" between a community's population of youth with mental disorders and its population of youthful offenders. Knowing this overlap gives some notion of the risk of official delinquency for youth with mental disorders and the degree to which mental disorders of youth contribute to a community's overall delinquency. The second type of study examines the proportion of youth with mental disorders within juvenile justice facilities or programs. These studies provide information with which to formulate policy about treating and managing youth with mental disorders in juvenile justice custody.
It is important to recognize that these two types of research begin with very different populations, even though they both address the relation between mental disorder and delinquency. The first typically focuses on all delinquent youth in the community, while the second examines only delinquent youth placed in juvenile pretrial detention centers when they are arrested or in juvenile correctional facilities when they are adjudicated. This distinction is further complicated, as discussed later, by the fact that not all youth in juvenile justice facilities are necessarily delinquent.
Epidemiologic Studies of Mental Disorder and Delinquency
Some studies have identified a significant overlap between the populations of youth served by community mental health agencies and youth in contact with the community's juvenile court.23 These studies are few in number, but they have found that the risk of juvenile court involvement among a community's young mental health clients is substantial. For example, a study in one city found that adolescents in contact with the community's mental health system during a nine-month period were two to three times more likely to have a referral to the juvenile justice system during that period than were youth in the city's general population.24 Youth in contact with a mental health system's services, however, are not the sum of a community's youth with mental health needs because many receive no services. The results of the study above probably represent the proportion of more seriously disturbed youth who have juvenile justice contact. Even so, merely knowing that youth "have contact" with the juvenile justice system tells us little about their offenses or even whether they offended at all.
Very few studies have used samples that make it possible to identify both the proportion of delinquent youth in a community who have mental disorders and the proportion of youth with mental disorders who have been delinquent. The few that have, however, are large studies with careful designs.
One examined a community population (drawn from several cities) that identified youth with persistent serious delinquency (repeat offending) and youth with persistent mental health problems (manifested multiple times).25 About 30 percent of youth with persistent mental health problems were persistently delinquent. But among all persistently delinquent youth, only about 15 percent had persistent mental health problems.
Another recent study examined the relation between mental disorders during adolescence and criminal behavior when those youth became adults.26 Delinquencies and adult criminal arrests were recorded for a sample of youth in a large geographic region aged nine through twenty-one. The youth were also assessed for mental disorders three times between the ages of nine and sixteen. A diagnosis at any one of these three points identified the youth as having a mental disorder "sometime during childhood or adolescence."
In this study, youth who were arrested between the ages of sixteen and twenty-one included a considerably greater share of youth who had had mental disorders in adolescence than those who were not arrested—for males, 51 percent as against 33 percent. This finding does not mean that 51 percent of the arrested group had mental disorders at the time of their arrest, but that they had had a mental disorder sometime in adolescence. It also does not mean that the majority of youth who had mental disorders in adolescence were arrested in adulthood. A different statistical procedure in this study, called "population attributable risk," addressed that question. It showed that the risk of adult arrest among individuals who had mental disorders at some time during adolescence was about 21 percent for women and 15 percent for men.
These few studies suggest the following conclusions, all of which need further confirmation. First, consistent with the clinical research reviewed earlier, youth who have mental disorders are at greater risk of engaging in offenses than youth without mental disorders. It is possible that treating their disorders would reduce that risk. But most youth with mental disorders do not engage in offenses that involve them in juvenile or criminal justice systems. Second, youth with mental disorders represent only a minority of all youth who engage in delinquent behavior, although the share is somewhat disproportionately greater than their prevalence in the general community. If those youth received treatment that reduced their delinquency, it is possible that overall rates of delinquency in the community would fall somewhat, but the majority of delinquencies are not related to mental disorders.
Third, rates of delinquency are higher among youth with certain types of emotional disorders— for example, depression or anxiety co-morbid with substance use disorders— and among youth with chronic and multiple disorders (seriously emotionally disturbed youth). Finally, a few studies have suggested that youth with mental disorders make up a somewhat greater proportion (although still a minority) of youth who were arrested for more serious and violent delinquencies or crimes.27
Mental Disorder in Juvenile Justice Settings
Research on the subset of delinquent youth who enter juvenile pretrial detention centers and correctional programs cannot tell us the relation between mental disorder and delinquency, because most youth who engage in delinquencies are not placed in secure juvenile justice programs. Such studies, however, are extremely important for public policy, because they identify the scope and nature of mental disorder among youth for whom the juvenile justice system has custodial responsibility.
Until recently the precise prevalence of mental disorders among youth in juvenile justice custody was unknown. Estimates varied widely from study to study, largely because of inadequate research methods or differences from one study site to another.28 In the past decade, however, well-designed studies executed in a variety of sites have provided a reliable and consistent picture. Those studies have found that among youth in various types of juvenile justice settings—for example, pretrial detention centers where youth are taken soon after arrest—about one-half to two-thirds meet criteria for one or more mental disorders.29 The prevalence of mental disorders is much higher in juvenile justice settings than it is among youth in the U.S. general population, which is about 15 to 25 percent.30
Across these studies, the rate is higher for girls than for boys.31 The overall prevalence rate does not vary greatly between younger and older adolescents or for youth with various ethnic and racial characteristics, although age and race differences are sometimes found for specific types of disorders and symptoms.32 As described in the earlier clinical review, about two-thirds of youth in juvenile justice custody who meet criteria for a mental disorder (that is, about one-third to one-half of youth in custody) meet criteria for more than one disorder.33
I will focus later on the implications of these statistics for the juvenile justice system's best response to mental disorders among youth in its custody. The high prevalence of mental disorder in juvenile justice facilities does not necessarily define the need for treatment. Some youth who meet criteria for mental disorders are experiencing their disorders temporarily and need only emergency services, while a smaller share—about one in ten—represents a core group of youth with chronic mental illness who can be expected to continue to need clinical services into adulthood.34 Some are functioning fairly well despite their symptoms, while others are barely able to function at all. And some have mental health needs, such as learning disabilities, that were not even included in the recent studies of prevalence among youth in juvenile justice settings.
Reasons for the High Prevalence of Mental Disorders in Juvenile Justice Programs
Why are mental disorders so prevalent among adolescent offenders in juvenile justice settings? Three perspectives—clinical, socio-legal, and inter-systemic—help to explain. They are not competing explanations. All probably play a role, and no evidence suggests that one is more important than the others.
From a clinical perspective, it is likely that the same symptoms of mental disorder that increase the risk of aggression also increase the likelihood that youth will be placed in secure juvenile justice facilities for any significant period of time. When police officers arrest youth, usually those youth are not placed in pretrial detention. Nor is detention reserved for the most serious offenders—in fact, youth arrested for very violent offenses typically do not make up the majority of youth in detention. Those youth who are detained more than a few hours are those who have been more unruly or unmanageable at the time of their arrest, which satisfies detention criteria regarding a risk that they will be endangered, or might endanger others, if not detained.
Youth with mental disorders frequently have symptoms involving impulsiveness, anger, and cognitive confusion that can make them less manageable and a greater risk to themselves or others, especially under the stress associated with their offense and arrest. Thus, among youth who are detained, a significant share is likely to have mental disorders that create unmanageable behavior—more so than for youth without mental disorders and more so than their peers with less severe mental disorders. This likelihood makes it no surprise that youth with mental disorders contribute disproportionately to detention populations.
From a socio-legal perspective, recent changes in laws applied to youths' delinquencies may have increased the likelihood that youth with mental disorders will enter the juvenile justice system. Before the 1990s, law enforcement officers, juvenile probation departments, prosecutors, and judges typically had some discretion regarding whether they would arrest or prosecute youth with mental disorders when they engaged in illegal behaviors, especially if those behaviors involved minor offenses committed by younger adolescents without offense histories. But a wave of serious juvenile violence during the late 1980s caused virtually all states to revise their juvenile justice statutes during the 1990s to rein in this discretion.35 Under the new laws, certain charges or offenses required legal responses based on the nature of the offense alone, not the characteristics or needs of the individual youth. Penalties more often involved custody in secure juvenile facilities, thus reducing the likelihood that youth could receive mental health services in the community after their adjudication. An unintended consequence of these changes in law, therefore, was an increase in the share of youth with mental disorders coming into the system rather than being diverted on the basis of the juvenile court's discretion.
A final, inter-systemic, explanation involves the dynamic relation between systems that serve youth. During the 1990s, most states saw a reduction in the availability of public mental health services for children, especially inpatient services.36 It is possible that less adequate treatment contributed to increased delinquencies among youth with mental disorders. But it is certain that many communities began using the juvenile justice system to try to fill the gap caused by decreased availability of mental health services.
This phenomenon was documented in media articles, the observations of juvenile justice personnel, and government reports beginning in the mid-1990s and continuing into the early 2000s.37 Some parents of children with serious mental disorders began urging police to arrest their children, knowing that courts could "order" mental health services that were becoming nearly impossible for parents to get on their own. Soon the local juvenile pretrial detention center was becoming the community's de facto mental health center that provided emergency mental health services or simply acted as a holding place for seriously disturbed youth who had nowhere to go.
In summary, these three factors—clinical, socio-legal, and inter-systemic—may together produce a prevalence of mental disorder in juvenile justice settings that does not represent the actual relation between adolescent mental disorder and delinquency. That high prevalence does, however, represent a demand on the juvenile justice system to respond to youth in custody who have mental disorders, and the demand is almost overwhelming. Some of those youth are in secure custody because they have committed serious crimes, others because the legal system has widened the door to juvenile justice processing, and many because their symptoms make them difficult to handle and they have no place else to go.
The problem requires a solution, and the multiple causes of the problem as well as the various types of youth involved suggest that the solution will be complex. What have clinicians and researchers learned that can help us determine the appropriate response?38
| 759
|
Will our beaches become an example of the Tragedy of the Commons?
Click on image for full size
NCAR Digital Library
The Tragedy of the Commons
The term "Tragedy of the Commons" was coined by Garrett Hardin in a 1968 article in Science magazine. The concept, however, dates back to the days of Aristotle. Briefly, it states that a shared resource is inevitably ruined by uncontrolled use.
The metaphor that Hardin uses to explain the concept is that of a community common or park on which the town’s people bring their cows to be fed. In the back of everyone’s mind is the fact that the common is going to be ruined because the grass is going to be eaten to depletion. Still, everyone wants to get grass for their cows. No one thinks or cares about the consequences of so many cows eating the grass, and the Tragedy of the Commons occurs.
Human actions that many categorize as examples of the phenomenon include human-created air pollution; the hunting of the American buffalo to near extinction in the 1800s; the widespread abuse and destruction of rainforests and our oceans’ coral reefs; and human-induced climate change due largely to the burning of fossil fuels for energy use.
Some people believe that the Tragedy of the Commons can only be averted by making most commodities private property. But how does someone own the air or the ocean? And can either the air or ocean stay unpolluted with populations of 10 million or more in the world’s megacities? Others believe the "Tragedy of the Commons" can be avoided through laws and taxing devices which make it more costly to serve one’s self interest over the common good.
What almost everyone can agree on for now is that such vital resources need some form of control so that the world’s natural resources can be sustained and the Tragedy of the Commons can be avoided.
Last modified February 19, 2006 by Teri Eastburn.
Shop Windows to the Universe Science Store!Cool It!
is the new card game from the Union of Concerned Scientists that teaches kids about the choices we have when it comes to climate change—and how policy and technology decisions made today will matter. Cool It! is available in our online store
You might also be interested in:
What do smog, acid rain, carbon monoxide, fossil fuel exhausts, and tropospheric ozone have in common? They are all examples of air pollution. Air pollution is not new. As far back as the 13 th century,...more
Rainbows appear in the sky when there is bright sunlight and rain. Sunlight is known as visible or white light and is actually a mixture of colors. Rainbows result from the refraction and reflection of...more
The Earth travels around the sun one full time per year. During this year, the seasons change depending on the amount of sunlight reaching the surface and the Earth's tilt as it revolves around the sun....more
Scientists sometimes travel in specially outfitted airplanes in order to gather data about atmospheric conditions. These research aircraft have special inlet ports that bring air from the outside into...more
An anemometer is a weather instrument used to measure the wind (it can also be called a wind gauge). Anemometers can measure wind speed, wind direction, and other information like the largest gust of wind...more
Thermometers measure temperature. "Thermo" means heat and "meter" means to measure. You can use a thermometer to measure the temperature of many things, including the temperature of...more
Weather balloons are used to carry weather instruments that measure temperature, pressure, humidity, and winds in the lowest few miles of the atmosphere. The balloons are made of rubber and weigh up to...more
| 515
|
This is an old revision of ArchaeologicalFindsandtheClassroom from 2008-03-13 15:48:05.
West Yorkshire Archaeology Advisory Service
Archaeological Finds and the Classroom
As its name suggests this web-based presentation is intended primarily for use with schools. However, the ideas it contains will also be useful for those of you working with young people in a less formal setting, perhaps through a youth group or a Young Archaeologists Club. Archaeological Finds and the Classroom outlines activities which you can use to get children thinking and talking about archaeological artefacts. It also offers guidance on carrying out fieldwork (fieldwalking, garden pottery surveys and molehill surveys) and offers advice on how to make the best use of the results.
To find out more, visit the website (www.arch.wyjs.org.uk/AdvSrv/Classroom/introduction.htm∞).
| CBA, St Mary's House, 66 Bootham, York YO30 7BZ.
tel: +(44) (0)1904 671417 | fax: +(44) (0)1904 671384 | email:
| 628
|
The dangers of feeding lorikeets
Lorikeet feeding habits
Lorikeets are brightly coloured parrots of the family Psittacidae and sub-family Loriinae (lories and lorikeets). Six species are found in Australia, although further species occur in New Guinea, Indonesia and many South Pacific islands.
The tip of a lorikeet's tongue holds a brush-like appendage, enabling the bird to feed on pollen, nectar and fruits. Because lorikeets are mainly seen feeding on blossoms, it was long thought that nectar was their staple diet. However, studies have now shown that pollen is probably their major food. They also eat insects and insect larvae, together with other invertebrates found in foliage and blossom.
Lorikeets travel in small, swift flocks which attract attention with their piercing calls. The birds locate blossoms by sight and by the calls of other birds already feeding. When food is abundant, large numbers gather in the blossoms and foliage, clamouring noisily.
Lorikeets have a very simple digestive system. Proteins and fats are obtained from pollen and insects, and carbohydrates come from the sugars found in nectar. They spend a large amount of time foraging for food, which indicates that carbohydrates are extremely important to them. Their diet is high in energy, and low in protein.
The dangers of feeding lorikeets
Many people have taken to encouraging lorikeets into their gardens by giving them a variety of alluring foods. These foods are usually based on sugar, honey or jam, which will ensure regular visits to the garden by these wonderfully happy, colourful birds. However, it is this very act of 'kindness' that can cause the death of lorikeets at an unnaturally early age.
OEH officers became concerned at outbreaks of lorikeets dying in large numbers for no apparent reason, and decided to investigate. Autopsies were carried out by NSW Agriculture. With other investigations, these autopsies established that the lorikeets were dying from a disease called necrotizing enteritis. A bacteria called clostridia was isolated as the organism responsible.
Necrotizing enteritis is associated with poor hygiene and an inadequate diet, and it was found that the disease is not normally associated with lorikeets living a natural way of life. The scientists could directly relate the lorikeets' deaths to inadequate artificial diets, and to the unsanitary conditions of trays and other containers from which they were being fed.
Another disease is also having a serious impact on lorikeets. Called psittacine beak and feather disease (PBFD) or psittacine circovirus disease, it is considered the most important viral disease affecting cockatoos, parrots, lories, lorikeets and macaws around the world. Affecting the immune system of the birds, it often results in an AIDS-like condition. The keratin in the birds' bodies breaks down, causing deformed beaks and feathers, and feather loss. The disease is usually fatal, but infected birds may not display symptoms for many months or even years.
Infected birds can drop faeces, which contain the potentially infectious circovirus particles, into the environment to infect other birds. Unfortunately, there is a great potential for the virus to spread at places like feed stations. Clustering in large numbers, many birds may come into contact with contaminated perches and food dishes.
Preventing lorikeet deaths
Obviously the best solution to these problems is to stop feeding the birds altogether. The only feeding method recommended by OEH is to plant a variety of flowering native shrubs, such as grevilleas, callistemon (bottlebrushes) and banksias, around your garden. These plants not only look good; they also provide safe, healthy food for blossom-feeding birds such as lorikeets and honeyeaters.
Page last updated: 15 April 2011
| 480
|
There are two well-known approved apparitions of Saint Joseph - but of which occurred in the context of a Marian apparition.
Our Lady of Knock (1879)
On the evening of Thursday, 21 August 1879 (Eve of the Octave of the Assumption) in Knock, Ireland, three mysterious persons appeared. The one in the middle was immediately identifiable as the Blessed Virgin Mary, veiled, crowned and praying. Immediately on her right was Saint Joseph with a gray beard, bowing toward Mary. To Mary's left was one appearing as a bishop who wore a short mitre. This last saint was the virginal Apostle Saint John the Evangelist.
Our Lady of Fatima (1917)
As we discussed a few days about on the significance of the three Marian visions of the Miracle of the Sun (See: Miracle of the Sun Signifies Rosary Mysteries), the first vision included Jesus, Mary, and Joseph, as Lucia described:
When Our Lady disappeared in the immense distance of the sky, next to the sun we saw Saint Joseph holding the Child Jesus and Our Lady dressed in white with a blue mantle. Saint Joseph and the Child seemed to be blessing the world making the sign of the cross.I find it startling that Saint Joseph, although not a priest, would give the sacerdotal blessing of the sign of the cross in union with his foster-Son, Christ.
What can we observe from these two apparitions of Saint Joseph?
- In both visions, Joseph does not speak.
- Joseph appears along with Mary.
- In the Knock vision, John appears, and John is the "son" of Mary is a special way. Is he also a "son" of Joseph in a special way?
| 283
|
Conservation work completed on one of Glasgow’s most historic buildings
27 February 2013
Glasgow Cathedral has recently received £30,000 worth of conservation work by Historic Scotland to repair the damage inflicted by last year’s gales.
Despite being the only medieval cathedral on the Scottish mainland to have survived the 1560 Reformation virtually complete, the weathercock and spike on top of the landmark did not escape last year’s merciless strong winds unscathed.
The spike of the church was no longer vertical due to corrosion and a closer examination of the weathercock showed damage to some of the rivets with the tail being completely blown off. A further structural inspection also found an area of the steeple was in need of re-pointing.
Overseeing the conservation work was Historic Scotland's District Architect, Ian Lambie. While the conservation itself was a fairly simple task, it was made more complex by the height of the building and the narrow conditions.
Ian Lambie said: “The only way of accessing the spire head was by climbing up inside the steeple using steep narrow ladders and squeezing and crawling through small windows at the apex. Making repairs at this sort of height has its challenges but it's always worth it for the view!”
Now that the necessary repairs are complete, the weathercock should be able to withstand the worst of the wind and the re-pointing should hold for at least another ten years.
Notes for editors:
Historic Scotland around the web:
- Historic Scotland is an executive agency of the Scottish Government charged with safeguarding the nation’s historic environment. The agency is fully accountable to Scottish Ministers and through them to the Scottish Parliament.
| 595
|
The Future of Salmon and Trout Stocking in Lake Michigan
Five species of salmon and trout support a world-class recreational fishery in Lake Michigan. Stocking has played an important role in maintaining the balance between predators and forage fish, such as the non-native alewife, since the late 1960s. If too many salmon and trout are in the lake, forage fish decline and salmon starve or fall prey to disease. If too few salmon and trout are in the lake, the non-native alewife could foul beaches and affect native species populations.
Ongoing research is being used to investigate the possibility that changes to stocking policy could improve fisheries and limit the risk of predator-prey imbalance. Fisheries managers in Michigan, Wisconsin, Illinois and Indiana will set a stocking policy for Lake Michigan salmon and trout by fall of 2012.
LAKE MICHIGAN SALMON STOCKING SURVEY RESULTS
Survey results are now available and are being considered along with biological data and other factors as managers move toward a decision on future stocking policy.
See: Survey Results
Lake Michigan Salmon Stocking Workshop
Lake Michigan College, Benton Harbor, Michigan
Saturday, April 14, 2012
This half-day workshop was free and open to the public. Participants learned more about specific options for stocking policy and had the opportunity to speak with fisheries managers about the future of Lake Michigan fisheries. Check out the recorded webinar for more details.
- Lake Michigan Salmon Stocking Strategies Briefing (PDF)
Includes overview of collaborative decision-making process, status of fishery and forage base, and stocking options under consideration.
| 225
|
Individual differences |
Methods | Statistics | Clinical | Educational | Industrial | Professional items | World psychology |
Mindfulness-Based Cognitive Therapy (MBCT) is a method of therapy which blends features of two disciplines:
- Cognitive therapy aims to identify and alter cognitive distortions (warped or inaccurate thoughts);
- Mindfulness is a meditative practice from Buddhism, which aims to help people identify their thoughts, moment by moment, but without passing judgement on the thoughts.
In MBCT, the patient is urged to recognize and accept undesired feelings as they come and go instead of trying to push them away. Traditional cognitive therapy, or cognitive behavioral therapy (CBT), focuses on changing negative content of thoughts while MBCT emphasizes the process of paying attention to thoughts and feelings moment by moment and without judgment. Changing the patient's relationship to the suffering caused by negative thoughts is the key because there is no possible way to alleviate all suffering. No therapy or meditation will prevent unpleasant things from happening in our daily lives but the two practices combined may provide more objectivity from which to view these unpleasant things.
MBCT's main technique is based on the Mindfulness-Based Stress Reduction (MBSR) eight week program, developed by Jon Kabat-Zinn in 1979 at the University of Massachusetts Medical Center. Research shows that MBSR is enormously empowering for patients with chronic pain, hypertension, heart disease, cancer, and gastrointestinal disorders, as well as for psychological problems such as anxiety and panic. People often misunderstand the goal of therapy and especially mindfulness. Relaxation and happiness are not the aim, but rather a "freedom from the tendency to get drawn into automatic reactions to thoughts, feelings, and events" . Patients change the relationship to chronic pain so the pain becomes more manageable.
Mindfulness-Based Cognitive Therapy grew largely from Jon Kabat-Zinn's work. Zindel V. Segal, J. Mark G. Williams and John D. Teasdale helped adapt the MBSR program so it could be used with people who had suffered repeated bouts of depression in their lives. Currently, MBCT programs usually consist of eight-weekly two hour classes with weekly assignments to be done outside of session. The aim of this program is to enhance awareness so we are able to respond to things instead of react to them. "We can respond to situations with choice rather than reacting automatically. We do that by practicing to become more aware of where our attention is, and deliberately changing the focus of attention, over and over again". The structure of MBCT requires strong commitment and work on the clients' part but the rewards can be lasting.
Effectiveness of MBCT Edit
Research is now showing the effectiveness of mindfulness in the prevention of relapse. The UK National Institute of Clinical Excellence (NICE) has recently endorsed MBCT as an effective treatment for prevention of relapse. Research has shown that people who have been clinically depressed three or more times (sometimes for twenty years or more) find that taking the program and learning these skills helps to reduce considerably their chances that depression will return. In a study conducted with 145 participants, all the patients had previously recovered from depression and then relapsed. These sufferers were split randomly into groups providing different methods of treatment. Within a year, patients who were undergoing MBCT "reduced relapse from 66% (control group) to 37% (treatment group)". "Whereas most people might be able to ignore sad mood, in previously depressed persons a slight lowering of mood might bring about a potentially devastating change in thought patterns". The core skill of MBCT is to teach the ideas of recognizing these thought patterns in order to break away from the false constructs of our mind. Relapse is avoided because the onset of depression is recognized before it has fully developed. The vicious cycle is stopped before it even gets started.
Benefits of MBCT and mindfulness practice Edit
Mindfulness meditation is a useful tool in dealing with many different scenarios. Practicing mindfulness aids patients, laypersons, and therapists. This approach to meditation focuses our attention back to the present, to what is happening right now in this exact moment. When one is mindful, the attention is focused on the present so judgment cannot be placed. Often, our pain and mental discomfort are caused by the judgment placed on the present moment and not by what is actually happening. This judgment and negative thinking is what can possibly lead to depression. MBCT prioritizes learning how to pay attention or concentrate with purpose, in each moment and most importantly, without judgment.
Segal and his partners found that "thoughts and feelings could interact with each other in a damaging, vicious spiral". Through the practice of mindfulness, we recognize that holding onto some of these feelings is ineffective and mentally destructive. Viewing things mindfully requires not turning away from any feeling but instead being open to the experience while trying not to engage defense mechanisms. All thoughts are welcomed into the mind equally so that one does not judge the thought or herself for thinking the thought. Gaining perspective on one's own thoughts allows us to escape the mental grooves and ruminative thinking that plagues us. Through mindfulness practice the spiral of negative thought is stopped before one finds herself at the bottom looking up.
Not only is this practice helpful to laypersons but to the actual therapist doing this type of MBCT. As a therapist, mindfulness can be implemented into therapy sessions, and used as a means of self-care in the therapist's personal life. "Meditating therapists often report feeling more 'present', relaxed, and receptive with their patients if they meditate earlier in the day". Mindfulness incorporates not judging thought. By having that non-judgment, the therapist allows the patient to fully express true feelings by having that openness. "As the therapist learns to disentangle from her own conditioned patterns of thoughts that arise in the therapy relationship, the patient may discover the same emotional freedom". The concentration development from mindfulness also helps the therapist be able to stay fully engaged with the patient. The mind naturally wanders to other things but mindfulness is the answer to being unfocused. There is a degree of perspective that also comes with mindfulness meditation. This new perspective allows a therapist to see other solutions or options to a patient's problem he or she may not have been originally aware of. "Having this [perspective] enables the therapist to have some flexibility in finding a formulation that accords with the patient's understanding". As therapists help their patients come to these solutions and become more fully functioning, it may be easy to think they are powerful and all knowing. Maintaining perspective prevents therapists from 'buying their own press'.
As means of self-care, P. Fulton and his fellow authors would say "offering love and care to ourselves replenishes the physical and emotional reservoirs that are necessary to care for others" (p. 87). When looking at burn-out rates in the social service fields, one can see that self-care is absolutely necessary whether one thinks they need it or not. Meditating saturates these reservoirs so compassionate, sincere work can continue. Also by dealing with personal suffering through this practice, therapists develop greater empathy and become more openhearted to the needs of their clients.
Depression as the inspiration of MBCT Edit
Depression is a more serious problem than how it is presently seen. The World Health Organization (WHO) conducted a study and came up with the following projection for the year 2020: "of all diseases, depression will impose the second largest burden of ill health worldwide". Research shows that at any given time, ten percent of the United States has experienced this type of clinical depression in the last year alone. Women being affected at a significantly higher rate (20-25%) than men (7-12%). The people who are affected with this common mental disorder are also the least likely to get help or treatment.
Depression is a severe and prolonged state of mind in which normal sadness grows into a painful state of hopelessness, listlessness, lack of motivation, and fatigue. Depression can vary from mild to severe. When depression is mild, one may find himself brooding on negative aspects of himself or others. He may feel resentful, irritable or angry much of the time, feeling sorry for himself, and needing reassurance from someone. Various physical ailments could also occur that have no correlation to physical illness.
Depression is classified as clinical when the episode inhibits a person's ability to accomplish routine daily tasks for at least two weeks. If suddenly 'normal' activities become difficult to do or the interest to do them is lost completely for a sustained amount of time, clinical depression could be a possibility. A change in basic bodily functions may also be experienced. The usual daily rhythms seem to go 'out of kilter'. One can't sleep, or one sleeps too much. One can't eat, or one eats too much. Others may notice that the sufferer may become agitated or slowed down. One may find that required energy for activities that used to be enjoyed is now gone. He or she may even feel that life is not worth living, and begin to develop thoughts that he or she would be better off dead.
Currently the most commonly used treatment for major depression is antidepressant medication. These medications are relatively cheap, and easy for family practitioners (who treat the majority of depressed people) to prescribe. However, once the episode has past, and the client has stopped taking the antidepressants, depression tends to return, and at least 50% of those experiencing their first episode of depression find that depression comes back, despite appearing to have made a full recovery. After a second or third episode, the risk of recurrence rises to between 80% and 90%. Also, those who first became depressed before 20 years of age are particularly likely to suffer a higher risk of relapse and recurrence.
The main method for preventing this recurrence is the continuation of the medication, but many people do not want to stay on medication for indefinite periods, and when the medication stops, the risk of becoming depressed again returns. People are turning to new ways of helping them stay well after depression. To see what it is most helpful to do, we need to understand why it is that we may remain at high risk, even when we've recovered. Mindfulness-Based Cognitive Therapy seems to be a complementary method to treating acute and chronic depression.
Why do we remain vulnerable to depression? Edit
New research shows that during any episode of depression, negative mood occurs alongside negative thinking (such as 'I am a failure', 'I am inadequate, 'I am worthless') and bodily sensations of sluggishness and fatigue. When the episode is past, and the mood has returned to normal, the negative thinking and fatigue tend to disappear as well. However, during the episode a mental association has formed between the mood that was present at that time, and the negative thinking patterns.
This means that when negative mood happens again (for any reason) a relatively small amount of such mood can trigger or reactivate the old thinking pattern. Once again, people start to think they have failed, or are inadequate - even if it is not relevant to the current situation. People who believed they had recovered may find themselves feeling 'back to square one'. They end up inside a rumination loop that constantly asks 'what has gone wrong?', 'why is this happening to me?', 'where will it all end?' Such rumination feels as if it ought to help find an answer, but it only succeeds in prolonging and deepening the mood spiral. When this happens, the old habits of negative thinking will start up again, negative thinking gets into the same rut, and a full-blown episode of depression may be the result.
The discovery that, even when people feel well, the link between negative moods and negative thoughts remains ready to be re-activated, is of enormous importance. It means that sustaining recovery from such depression depends on learning how to keep mild states of depression from spiralling out of control.
Future of MBCT Edit
Further research is being conducted to identify all the different uses of MBCT. Significant decreases in anxiety, depression, with a resulting increased sense well being, have been found so far. Research being conducted will evaluate MBCT as a useful technique with patients who are diagnosed with cancer or haematological illness. Mindfulness practice is being done over various cultures and demographics. Ellen Langer has been focusing on the future of mindfulness.
- Jon Kabat-Zinn
- Dialectical behavior therapy
- Acceptance and Commitment Therapy
References & BibliographyEdit
- REDIRECT Template:Reflist
- Segal, Z.V., Williams, J.M.G. & Teasdale, J.D, 2002, Mindfulness–based Cognitive Therapy for Depression. A New Approach to Preventing Relapse. Guilford Press.
- Teasdale, JD, Segal, ZV, and Williams, JMC. How does cognitive therapy prevent depressive relapse and why should attentional control (mindfulness) help? Behav Res Ther (1995) 33:25-29.
- Teasdale, JD, Segal, ZV, Williams MG, Ridgeway, VA, Soulsby, JM, Lau, MA. Prevention of Relapse/Recurrence in Major Depression by Mindfulness-Based Cognitive Therapy. J. of Consulting and Clinical Psychology (2000) 68:615-623.
- Williams JMG, Teasdale JD, Segal ZV and Soulsby J. Mindfulness-based cognitive therapy reduces overgeneral autobiographical memory in formerly depressed patients. J Abnorm Psychol (2001).
- [http://www.mbct.co.uk/ Mindfulness-Based Cognitive Therapy (UK). Mindfulness-Based Cognitive Therapy Homepage. Very complete website with information on books, founders, and workshops.
- What is Mindfulness-Based Stress Reduction?
- The Centre for Mindfulness Research and Practice
- Oxford Cognitive Therapy Centre
- MBSR & MBCT
- How to do Mindfulness Meditation[[Category:Mindfulness
|This page uses Creative Commons Licensed content from Wikipedia (view authors).|
<ref>tags exist, but no
<references/>tag was found
| 26
|
|ARM Technical Support Knowledge Articles|
Applies to: ARM7TDMI
MAS[1:0] is used to indicate whether a word/half-word or byte access is to be performed, and is described in the ARM7TDMI Technical Reference Manual (1.8MB PDF).
The signal has the following states:
|Bit 1||Bit 0||Data Size|
The memory system must be able to handle word, half word and byte writes.
Memory systems supporting only word writes will have severe difficulties supporting C code because the compilers assume that the underlying access types of the ARM architecture are always available. Furthermore, it will not be possible to set software breakpoints in Thumb code using the EmbeddedICE Interface.
When in Thumb state, A is not driven, and will be held at whatever level it was last driven to, by a 'sticky latch'. Usually, this will be set following a BX instruction (with bit 0 of the register set), or a data transfer to/from an odd address. It would normally be cleared again following a data transfer to/from an even address.
The memory controller should ignore A for Thumb instruction fetches (nOPC=0 and MAS[1:0]=01), and A[1:0] for ARM instruction fetches (nOPC=0 and MAS[1:0]=10).
Article last edited on: 2008-09-09 15:47:35
Did you find this article helpful? Yes No
How can we improve this article?
| 707
|
Study Questions Antibiotics' Safety in Preterm Labor
There's important new information out today for pregnant women and their doctors: Giving antibiotics to some pregnant women who have gone into labor too early could be dangerous for the long-term development of their children, according to a major new study.
Now, it's important to make it clear right away that if a pregnant woman definitely has an infection she should take an antibiotic--an infection can be very dangerous to the mom and baby. But the new study says doctors should be very careful about giving a pregnant woman antibiotics if there is no clear evidence that she has an infection and her water hasn't broken.
This comes from a big international study involving more than 11,000 women known as the ORACLE study. British researchers started it to find out if antibiotics might help women who go into labor too soon. Infections can trigger premature labor without causing any other symptoms. So the idea was that antibiotics could help.
In 2001, the researchers reported that giving erythromycin to pregnant women whose water had broken prematurely reduced the risk their babies would develop infections, breathing problems and other short-term complications. But erythromycin didn't seem to help women in premature labor whose water had not broken.
The researchers then followed more than 6,000 of the children in the study who were born in England to see how they fared. In a pair of papers released yesterday by the journal The Lancet, the researchers reported that those whose mothers took erythromycin before their water broke were 18 percent more likely to have mild developmental problems. And, perhaps most disturbing, they were about three times as likely to have cerebral palsy.
The researchers do not know why antibiotics would cause those problems, but they speculate that perhaps keeping a fetus in the womb of a mother who has an infection may not be good for the child.
Now, it's important to note that current guidelines only recommend antibiotics for pregnant women whose water has broken or have clear signs of an infection, and the study shows that is the right thing to do.
But experts say it's important to get the word out in case some doctors are giving pregnant women who go into premature labor antibiotics even though their water has not broken and they have no obvious signs of infection.
The research also is a good reminder that just because something seems like a good idea you never know for sure until you do the research to find out. You could end up doing more harm than good.
September 18, 2008; 7:00 AM ET
Categories: Cancer , Motherhood
Save & Share: Previous: LIfe's Big Questions: How Can I Face This Financial Stress?
Next: Colon Screening: Ready to Go Virtual?
The comments to this entry are closed.
| 437
|
Moore's Law Is Alive And Well, And Intel Will Prove It Today
Intel, the world’s biggest manufacturer of computer chips, and by far the one with the most advanced manufacturing capabilities, is holding a big event in San Francisco which it described in an invitation to reporters as its “most significant technology announcement of the year.” It provided no further details.
This appears to be the announcement that CEO Paul Otellini alluded to during Intel’s quarterly earnings conference call last month. Intel has kept a pretty tight lid on the details, but I’ve talked to enough people who say this is one of those times when Intel will “open the kimono” on what will be going on inside its chip factories–or fabs–later this year. The big news will revolve around Intel’s disclosure of its 22-nanometer manufacturing process. It’s the sort of thing that gets people who know chips kind of excited and leaves others kind of cold. But in fact, everyone should be kind of excited about this.
Perhaps you’ve heard of Moore’s Law. This was the observation in 1965 by the Intel co-founder Gordon Moore (pictured at the Intel Museum in 2005) that the number of transistors that could be crammed onto a chip doubles–and the size of those chips tended to shrink–as manufacturing technology improved on a fairly regular basis: About every 18 to 24 months. That shrinking meant two things. Chipmakers could make a chip with the same computing power as the previous generation more cheaply, or they could make a more powerful one with more transistors for about the same cost.
It all comes down to how many transistors you can cram onto a chip, and how many useful chips you can get from a single silicon wafer. In both cases, more is better. Moore’s observation–which was first published 46 years ago this month–has held up remarkably well and has proven one of the most important engines of growth in the technology industry. All the computing oomph you take for granted in your notebook, your smart phone, in the cloud, and all around you happens in part because the chips inside the hardware have gotten smaller and yet ever more powerful every two years or so.
So back to today’s announcement. As I mentioned, it’s going to revolve around its 22-nanometer manufacturing process. A nanometer is a billionth of a meter, and its current factory processors turn out chips with transistors that are somewhat bigger–32 nanometers. Intel executives often refer to a process they call “tick-tock.” Today constitutes a tick, when in odd-numbered years, a new manufacturing process comes online and the previous generation chips are shifted to being built with the smaller transistors. A “tock” occurs in even-numbered years when Intel engineers come up with new chip designs that really show what the new factory processes are capable of. The implication is that it’s so regular you can almost set your watch by it. Intel’s long-term strategy can be summed up like so: Tick, tock, repeat.
On top of that there are likely to be disclosures about some of the advances in physics that Intel has had to make in order to get chips with transistors so small to work properly. When you’re dealing with things that small, the individual electrons flowing on the chip sometimes don’t behave as they should. For example, in 2007 Intel had to add the element Hafnium to its chip-making process in order to stop individual transistors from wasting electricity. (It was more complicated than that, but that in a nutshell was the problem.) Billions upon billions of transistors in billions of computers around the world wasting electricity is a bad thing, both financially–power is expensive–and environmentally.
What’s funny is that for years people have been saying that Intel–and indeed the entire chip industry–can’t continue on the Moore’s Law trajectory. At some point things get so small that you’re dealing with individual atoms and you can’t get any smaller than that. However, every time people have predicted its end, something happens to keep it going. A lot of companies have come up with some important advances that have kept it going. In the 1990s and early 2000s, IBM came up with some important advances that kept Moore’s Law on track. But more often that not it has been Intel that has kicked down the door when the experts said it was locked. Today it will probably kick down another.
This older video was created around the time that Intel unveiled its 45-nanometer process with Hafnium–kicking down one of those earlier doors. Perhaps there will be another today. Check in later as I cover the announcement.
| 857
|
Vol. 17 Issue 6
One-Legged (Single Limb) Stance Test
The One-Legged Stance Test (OLST)1,2 is a simple, easy and effective method to screen for balance impairments in the older adult population.
You may be asking yourself, "how can standing on one leg provide you with any information about balance, after all, we do not go around for extended periods of time standing on one leg?"
True, as a rule we are a dynamic people, always moving, our world always in motion, but there are instances were we do need to maintain single limb support. The most obvious times are when we are performing our everyday functional activities.
Stepping into a bath tub or up onto a curb would be difficult, if not impossible to do without the ability to maintain single limb support for a given amount of time. The ability to switch from two- to one-leg standing is required to perform turns, climb stairs and dress.
As we know, the gait cycle requires a certain amount of single limb support in order to be able to progress ourselves along in a normal pattern. When the dynamics of the cycle are disrupted, loss of balance leading to falls may occur.
This is especially true in older individuals whose gait cycle is altered due to normal and potentially abnormal changes that occur as a result of aging.
The One-Legged Stance Test measures postural stability (i.e., balance) and is more difficult to perform due to the narrow base of support required to do the test. Along with five other tests of balance and mobility, reliability of the One-Legged Stance Test was examined for 45 healthy females 55 to 71 years old and found to have "good" intraclass correlations coefficients (ICC range = .95 to .099). Within raters ICC ranged from 0.73 to 0.93.3
To perform the test, the patient is instructed to stand on one leg without support of the upper extremities or bracing of the unweighted leg against the stance leg. The patient begins the test with the eyes open, practicing once or twice on each side with his gaze fixed straight ahead.
The patient is then instructed to close his eyes and maintain balance for up to 30 seconds.1
The number of seconds that the patient/client is able to maintain this position is recorded. Termination or a fail test is recorded if 1) the foot touches the support leg; 2) hopping occurs; 3) the foot touches the floor, or 4) the arms touch something for support.
Normal ranges with eyes open are: 60-69 yrs/22.5 ± 8.6s, 70-79 yrs/14.2 ± 9.3s. Normal ranges for eyes closed are: 60-69 yrs/10.2 ± 8.6s, 70-79 yrs/4.3 ± 3.0s.4 Briggs and colleagues reported balance times on the One-Legged Stance Test in females age 60 to 86 years for dominant and nondominant legs.
Given the results of this data, there appears to be some difference in whether individuals use their dominant versus their nondominant leg in the youngest and oldest age groups.
When using this test, having patients choose what leg they would like to stand on would be appropriate as you want to record their "best" performance.
It has been reported in the literature that individuals increase their chances of sustaining an injury due to a fall by two times if they are unable to perform a One-Legged Stance Test for five seconds.5 Other studies utilizing the One-Legged Stance Test have been conducted in older adults to assess static balance after strength training,6 performance of activities of daily living and platform sway tests.7
Interestingly, subscales of other balance measures such as the Tinetti Performance Oriented Mobility Assessment8 and Berg Balance Scale9 utilize unsupported single limb stance times of 10 seconds and 5 seconds respectively, for older individuals to be considered to have "normal" balance.
Thirty percent to 60 percent of community-dwelling elderly individuals fall each year, with many experiencing multiple falls.10 Because falls are the leading cause of injury-related deaths in older adults and a significant cause of disability in this population, prevention of falls and subsequent injuries is a worthwhile endeavor.11
The One-Legged Stance Test can be used as a quick, reliable and easy way for clinicians to screen their patients/clients for fall risks and is easily incorporated into a comprehensive functional evaluation for older adults.
1. Briggs, R., Gossman, M., Birch, R., Drews, J., & Shaddeau, S. (1989). Balance performance among noninstitutionalized elderly women. Physical Therapy, 69(9), 748-756.
2. Anemaet, W., & Moffa-Trotter, M. (1999). Functional tools for assessing balance and gait impairments. Topics in Geriatric Rehab, 15(1), 66-83.
3. Franchignoni, F., Tesio, L., Martino, M., & Ricupero, C. (1998). Reliability of four simple, quantitative tests of balance and mobility in healthy elderly females. Aging (Milan), 10(1), 26-31.
4. Bohannon, R., Larkin, P., Cook, A., & Singer, J. (1984). Decrease in timed balance test scores with aging. Physical Therapy, 64, 1067-1070.
5. Vellas, B., Wayne, S., Romero, L., Baumgartner, R., et al. (1997). One-leg balance is an important predictor of injurious falls in older persons. Journal of the American Geriatric Society, 45, 735-738.
6. Schlicht, J., Camaione, D., & Owen, S. (2001). Effect of intense strength training on standing balance, walking speed, and sit-to-stand performance in older adults. Journal of Gerontological Medicine and Science, 56A(5), M281-M286.
7. Frandin, K., Sonn, U., Svantesson, U., & Grimby, G. (1996). Functional balance tests in 76-year-olds in relation to performance, activities of daily living and platform tests. Scandinavian Journal of Rehabilitative Medicine, 27(4), 231-241.
8. Tinetti, M., Williams, T., & Mayewski, R. (1986). Fall risk index for elderly patients based on number of chronic disabilities. American Journal of Medicine, 80, 429-434.
9. Berg, K., et al. (1989). Measuring balance in the elderly: Preliminary development of an instrument. Physio Therapy Canada, 41(6), 304-311.
10. Rubenstein, L., & Josephson, K. (2002). The epidemiology of falls and syncope. Clinical Geriatric Medicine, 18, 141-158.
11. National Safety Council. (2004). Injury Facts. Itasca, IL: Author.
Dr. Lewis is a physical therapist in private practice and president of Premier Physical Therapy of Washington, DC. She lectures exclusively for GREAT Seminars and Books, Inc. Dr. Lewis is also the author of numerous textbooks. Her Website address is www.greatseminarsandbooks.com. Dr. Shaw is an assistant professor in the physical therapy program at the University of South Florida dedicated to the area of geriatric rehabilitation. She lectures exclusively for GREAT Seminars and Books in the area of geriatric function.
APTA Encouraged by Cap Exceptions
New process grants automatic exceptions to beneficiaries needing care the most
Calling it "a good first step toward ensuring that Medicare beneficiaries continue to have coverage for the physical therapy they need," Ben F Massey, Jr, PT, MA, president of the American Physical Therapy Association (APTA), expressed optimism that the new exceptions process will allow a significant number of Medicare patients to receive services exceeding the $1,740 annual financial cap on Medicare therapy coverage. The new procedure, authorized by Congress in the recently enacted Deficit Reduction Act (PL 109-171), will be available to Medicare beneficiaries on March 13 under rules released this week by the Centers for Medicare and Medicaid Services (CMS).
"APTA is encouraged by the new therapy cap exceptions process," Massey said. "CMS has made a good effort to ensure that Medicare beneficiaries who need the most care are not harmed by an arbitrary cap."
As APTA recommended, the process includes automatic exceptions and also grants exceptions to beneficiaries who are receiving both physical therapy and speech language pathology (the services are currently combined under one $1,740 cap).
"We have yet to see how well Medicare contractors will be able to implement and apply this process. Even if it works well, Congress only authorized this new process through 2006. Congress must address this issue again this year, and we are confident that this experience will demonstrate to legislators that they must completely repeal the caps and provide a more permanent solution for Medicare beneficiaries needing physical therapy," Massey continued.
The therapy caps went into effect on Jan. 1, 2006, limiting Medicare coverage on outpatient rehabilitation services to $1,740 for physical therapy and speech therapy combined and $1,740 for occupational therapy.
The American Physical Therapy Association is a national professional organization representing more than 65,000 members. Its goal is to foster advancements in physical therapy practice, research and education.
New Mouthwash Helps With Pain
Doctors in Italy are studying whether a new type of mouthwash will help alleviate pain for patients suffering from head and neck cancer who were treated with radiation therapy, according to a new study (International Journal of Radiation Oncology*Biology*Physics, Feb. 1, 2006).
Fifty patients, suffering from various forms of head and neck cancer and who received radiation therapy, were observed during the course of their radiation treatment. Mucositis, or inflammation of the mucous membrane in the mouth, is the most common side effect yet no additional therapy has been identified that successfully reduces the pain.
This study sought to discover if a mouthwash made from the local anesthetic tetracaine was able to alleviate the discomfort associated with head and neck cancer and if there would be any negative side effects of the mouthwash. The doctors chose to concoct a tetracaine-based mouthwash instead of a lidocaine-based version because it was found to be four times more effective, worked faster and produced a prolonged relief.
The tetracaine was administered by a mouthwash approximately 30 minutes before and after meals, or roughly six times a day. Relief of oral pain was reported in 48 of the 50 patients. Sixteen patients reported that the mouthwash had an unpleasant taste or altered the taste of their food.
| 364
|
by Gerry Everding
St. Louis MO (SPX) Feb 12, 2013
Nominated early this year for recognition on the UNESCO World Heritage List, which includes such famous cultural sites as the Taj Mahal, Machu Picchu and Stonehenge, the earthen works at Poverty Point, La., have been described as one of the world's greatest feats of construction by an archaic civilization of hunters and gatherers.
Now, new research in the current issue of the journal Geoarchaeology, offers compelling evidence that one of the massive earthen mounds at Poverty Point was constructed in less than 90 days, and perhaps as quickly as 30 days - an incredible accomplishment for what was thought to be a loosely organized society consisting of small, widely scattered bands of foragers.
"What's extraordinary about these findings is that it provides some of the first evidence that early American hunter-gatherers were not as simplistic as we've tended to imagine," says study co-author T.R. Kidder, PhD, professor and chair of anthropology in Arts and Sciences at Washington University in St. Louis.
"Our findings go against what has long been considered the academic consensus on hunter-gather societies - that they lack the political organization necessary to bring together so many people to complete a labor-intensive project in such a short period."
Co-authored by Anthony Ortmann, PhD, assistant professor of geosciences at Murray State University in Kentucky, the study offers a detailed analysis of how the massive mound was constructed some 3,200 years ago along a Mississippi River bayou in northeastern Louisiana.
Based on more than a decade of excavations, core samplings and sophisticated sedimentary analysis, the study's key assertion is that Mound A at Poverty Point had to have been built in a very short period because an exhaustive examination reveals no signs of rainfall or erosion during its construction.
"We're talking about an area of northern Louisiana that now tends to receive a great deal of rainfall," Kidder says. "Even in a very dry year, it would seem very unlikely that this location could go more than 90 days without experiencing some significant level of rainfall. Yet, the soil in these mounds shows no sign of erosion taking place during the construction period. There is no evidence from the region of an epic drought at this time, either."
Part of a much larger complex of earthen works at Poverty Point, Mound A is believed to be the final and crowning addition to the sprawling 700-acre site, which includes five smaller mounds and a series of six concentric C-shaped embankments that rise in parallel formation surrounding a small flat plaza along the river. At the time of construction, Poverty Point was the largest earthworks in North America.
Built on the western edge of the complex, Mound A covers about 538,000 square feet [roughly 50,000 square meters] at its base and rises 72 feet above the river. Its construction required an estimated 238,500 cubic meters - about eight million bushel baskets - of soil to be brought in from various locations near the site. Kidder figures it would take a modern, 10-wheel dump truck about 31,217 loads to move that much dirt today.
"The Poverty Point mounds were built by people who had no access to domesticated draft animals, no wheelbarrows, no sophisticated tools for moving earth," Kidder explains. "It's likely that these mounds were built using a simple 'bucket brigade' system, with thousands of people passing soil along from one to another using some form of crude container, such as a woven basket, a hide sack or a wooden platter."
To complete such a task within 90 days, the study estimates it would require the full attention of some 3,000 laborers. Assuming that each worker may have been accompanied by at least two other family members, say a wife and a child, the community gathered for the build must have included as many as 9,000 people, the study suggests.
"Given that a band of 25-30 people is considered quite large for most hunter-gatherer communities, it's truly amazing that this ancient society could bring together a group of nearly 10,000 people, find some way to feed them and get this mound built in a matter of months," Kidder says.
Soil testing indicates that the mound is located on top of land that was once low-lying swamp or marsh land - evidence of ancient tree roots and swamp life still exists in undisturbed soils at the base of the mound. Tests confirm that the site was first cleared for construction by burning and quickly covered with a layer of fine silt soil. A mix of other heavier soils then were brought in and dumped in small adjacent piles, gradually building the mound layer upon layer.
As Kidder notes, previous theories about the construction of most of the world's ancient earthen mounds have suggested that they were laid down slowly over a period of hundreds of years involving small contributions of material from many different people spanning generations of a society. While this may be the case for other earthen structures at Poverty Point, the evidence from Mound A offers a sharp departure from this accretional theory. Kidder's home base in St.
Louis is just across the Mississippi River from one of America's best known ancient earthen structures, the Monk Mound at Cahokia, Ill. He notes that the Monk Mound was built many centuries later than the mounds at Poverty Point by a civilization that was much more reliant on agriculture, a far cry from the hunter-gatherer group that built Poverty Point. Even so, Mound A at Poverty Point is much larger than almost any other mound found in North America; only Monk's Mound at Cahokia is larger.
"We've come to realize that the social fabric of these socieites must have been much stronger and more complex that we might previously have given them credit. These results contradict the popular notion that pre-agricultural people were socially, politically, and economically simple and unable to organize themselves into large groups that could build elaborate architecture or engage in so-called complex social behavior," Kidder says.
"The prevailing model of hunter-gatherers living a life 'nasty, brutish and short' is contradicted and our work indicates these people were practicing a sophisticated ritual/religious life that involved building these monumental mounds."
Washington University in St. Louis
All About Human Beings and How We Got To Be Here
|The content herein, unless otherwise known to be public domain, are Copyright 1995-2012 - Space Media Network. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA Portal Reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. Advertising does not imply endorsement,agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement|
| 229
|
A grooved pulley wheel like that used for ropes is called a sheave. A single sheave mounted in a block and fixed in place simply changes the direction of force exerted on the rope passing over it. If the end of the rope that ordinarily would attach to the load is passed around a second, unfixed pulley and back to the fixed pulley, a load attached to the free pulley can be raised with half the effort, or with a mechanical advantage of 2. Thus arranged, the device is called a block and tackle. The number of pulley wheels mounted in the fixed and free blocks can be increased indefinitely to get a higher and higher mechanical advantage, the mechanical advantage equaling the number of strands running to the free pulley. Therefore if the rope is run over the first fixed pulley wheel, around the free pulley, over a second pulley wheel in the fixed block, and back to the free block, the mechanical advantage is 3. A 300-lb load can be raised by a pull of 100 lb on the free end of the rope. To raise the load 10 ft, however, the free end of the rope must be pulled 30 ft.
Disregarding friction, work output will always equal work input. If the action is reversed by attaching the load to the free end of the rope and pulling on the free block, the mechanical advantage becomes a mechanical disadvantage, but a speed advantage. A rope block and tackle is usually for hand operation. To lift larger loads by hand, a chain is substituted for the rope and the pulleys have grooves for gripping the links. A differential pulley consists of two pulleys of different radii connected and rotating as one on a common axle. The pulleys have their circumferences grooved and spiked so that a chain will run in them without slipping. Over the pulleys an endless chain is run, forming two hanging loops. In one loop is placed a movable block, whose pulley is shaped to take the chain. The load is attached to the movable block and is raised by pulling on the other loop of the chain.
Power-operated machinery usually has cables, as in vertical-lift drawbridges, power shovels, and cranes. Before the extensive use of electric motors, steam engines or water turbines often supplied the power for factory machinery. One engine or turbine might run a whole factory through a complicated system of shafts, pulleys, and belts. Pulleys for flat belts are crowned to keep the belt centered. Raised flanges will serve the same purpose, but they wear the edges of the belt. Drive pulleys for conveyor belts often have a covering, called lagging, to provide better grip. Individual electric motors usually provide drive by means of V belts, the pulleys having raised flanges to form slots that match the trapezoidal cross sections of the belts. Cone pulleys consist of a number of pulleys of varying diameters massed in the shape of a cone. They are used with belt drives for machines (e.g., lathes) requiring a variety of speeds.
In mechanics, a wheel that carries a flexible rope, cord, cable, chain, or belt on its rim. Pulleys are used singly or in combination to transmit energy and motion. In belt drives, pulleys are attached to shafts at their axes, and power is transmitted between the shafts by means of endless belts running over the pulleys. One or more independently rotating pulleys can be used to gain mechanical advantage, especially for lifting weights. The shafts around which the pulleys turn may attach them to frames or blocks, and a combination of pulleys, blocks, and rope is called a block and tackle. The pulley is considered one of the five simple machines.
Learn more about pulley with a free trial on Britannica.com.
A pulley (also called a sheave or block) is a wheel with a groove between two flanges around its circumference. A rope, cable or belt usually runs inside the groove. Pulleys are used to change the direction of an applied force, transmit rotational motion, or realize a mechanical advantage in either a linear or rotational system of motion.
A belt and pulley system is characterized by two or more pulleys in common to a belt. This allows for mechanical power, torque, and speed to be transmitted across axes and, if the pulleys are of differing diameters, a mechanical advantage to be realized.
A belt drive is analogous to that of a chain drive, however a belt sheave may be smooth (devoid of discrete interlocking members as would be found on a chain sprocket, spur gear, or timing belt) so that the mechanical advantage is given by the ratio of the pitch diameter of the sheaves only (one is not able to count 'teeth' to determine gear ratio).
Belt and pulley systems can be very efficient, with stated efficiencies up to 98%.
In a system of a single rope and pulleys, when friction is neglected, the mechanical advantage gained can be calculated by counting the number of rope lengths exerting force on the load. Since the tension in each rope length is equal to the force exerted on the free end of the rope, the mechanical advantage is simply equal to the number of ropes pulling on the load. For example, in Diagram 3 below, there is one rope attached to the load, and 2 rope lengths extending from the pulley attached to the load, for a total of 3 ropes supporting it. If the force applied to the free end of the rope is 10 lb, each of these rope lengths will exert a force of 10 lb. on the load, for a total of 30 lb. So the mechanical advantage is 3.
The force on the load is increased by the mechanical advantage; however the distance the load moves, compared to the length the free end of the rope moves, is decreased in the same proportion. Since a slender cable is more easily managed than a fat one (albeit shorter and stronger), pulley systems are often the preferred method of applying mechanical advantage to the pulling force of a winch (as can be found in a lift crane).
Pulley systems are the only simple machines in which the possible values of mechanical advantage are limited to whole numbers.
In practice, the more pulleys there are, the less efficient a system is. This is due to sliding friction in the system where cable meets pulley and in the rotational mechanism of each pulley.
It is not recorded when or by whom the pulley was first developed. It is believed however that Archimedes developed the first documented block and tackle pulley system, as recorded by Plutarch. Plutarch reported that Archimedes moved an entire warship, laden with men, using compound pulleys and his own strength.
These are different types of pulley systems:
The simplest theory of operation for a pulley system assumes that the pulleys and lines are weightless, and that there is no energy loss due to friction. It is also assumed that the lines do not stretch.
In equilibrium, the total force on the pulley must be zero. This means that the force on the axle of the pulley is shared equally by the two lines looping through the pulley. The situation is schematically illustrated in diagram 1. For the case where the lines are not parallel, the tensions in each line are still equal, but now the vector sum of all forces is zero.
A second basic equation for the pulley follows from the conservation of energy: The product of the weight lifted times the distance it is moved is equal to the product of the lifting force (the tension in the lifting line) times the distance the lifting line is moved. The weight lifted divided by the lifting force is defined as the advantage of the pulley system.
It is important to notice that a system of pulleys does not change the amount of work done. The work is given by the force times the distance moved. The pulley simply allows trading force for distance: you pull with less force, but over a longer distance.
In diagram 2, a single movable pulley allows weight W to be lifted with only half the force needed to lift the weight without assistance. The total force needed is divided between the lifting force (red arrow) and the "ceiling" which is some immovable object (such as the earth). In this simple system, the lifting force is directed in the same direction as the movement of the weight. The advantage of this system is 2. Although the force needed to lift the weight is only W/2, we will need to draw a length of rope that is twice the distance that the weight is lifted, so that the total amount of work done (Force x distance) remains the same.
A second pulley may be added as in diagram 2a, which simply serves to redirect the lifting force downward, it does not change the advantage of the system.
The addition of a fixed pulley to the single pulley system can yield an increase of advantage. In diagram 3, the addition of a fixed pulley yields a lifting advantage of 3. The tension in each line is W/3, and the force on the axles of each pulley is 2W/3. As in the case of diagram 2a, another pulley may be added to reverse the direction of the lifting force, but with no increase in advantage. This situation is shown in diagram 3a.
This process can be continued indefinitely for ideal pulleys with each additional pulley yielding a unit increase in advantage. For real pulleys friction among rope and pulleys will increase as more pulleys are added to the point that no advantage is possible. It puts a limit for the number of pulleys usable in practice. The above pulley systems are known collectively as block and tackle pulley systems. In diagram 4a, a block and tackle system with advantage 4 is shown. A practical implementation in which the connection to the ceiling is combined and the fixed and movable pulleys are encased in single housings is shown in figure 4b.
Other pulley systems are possible, and some can deliver an increased advantage with fewer pulleys than the block and tackle system. The advantage of the block and tackle system is that each pulley and line is subjected to equal tensions and forces. Efficient design dictates that each line and pulley be capable of handling its load, and no more. Other pulley designs will require different strengths of line and pulleys depending on their position in the system, but a block and tackle system can use the same line size throughout, and can mount the fixed and movable pulleys on a common axle.
Shape up your pulleys: minimize costly downtime and extend conveyor belt life by keeping wing pulleys free of entrapped material.(SPECIAL SECTION: CONVEYING & MATERIAL HANDLING)
Aug 01, 2009; Consider the frequent task of dislodging material from the standard wing pulley. It's necessary yet annoying and time...
| 37
|
What are managed lanes?
Highway facilities or a set of lanes where operational strategies are proactively implemented and managed in response to changing conditions.
Transportation agencies are faced with growing challenges of congestion and a limited ability to expand freeway capacity due to construction costs, right-of-way constraints, and environmental and societal impacts. Transportation officials are taking advantage of opportunities to address mobility needs and provide travel options through a combination of limited capacity expansion coupled with operational strategies that seek to manage travel demand and improve transit and other forms of ridesharing. The managed lanes concept is gaining interest around the country as an approach that combines these elements to make the most effective and efficient use of a freeway facility.
The distinction between managed lanes and other traditional forms of freeway lane management is the operating philosophy of "active management." Under this philosophy, the operating agency proactively manages demand and available capacity on the facility by applying new strategies or modifying existing strategies. The agency defines from the outset the operating objectives for the managed lanes and the kinds of actions that will be taken once pre-defined performance thresholds are met.
You will need the Adobe Acrobat Reader to view the PDFs on this page.
United States Department of Transportation - Federal Highway Administration
| 887
|
EPO is a nootropicWednesday, 10 September, 2008
Erythropoietin (EPO), a hormone that increases red blood cells and is used as a performance enhancer for athletic performance, has now been shown to enhance memory in normal, healthy mice. Mice that received EPO injections had enhanced memory for 3-4 weeks afterwards, which is longer than the elevation in red blood cell count lasts.
This effect isn’t actually novel, as other researchers had noticed that EPO improved brain function over 18 years ago (Grimm et al, 1990), and research into mental illness has also suggested that EPO has an effect on brain function (Ehrenreich et al, 2004). But it was always thought to be dependent on the change in red blood cells, but more recent evidence has suggested it works independently of effects on blood cells (Miskowiak et al, 2007). This mouse model confirms this.
Of course, the researchers have been focusing on this as a treatment, but anyone can see that this is a promising enhancement too. This mouse research showed that EPO enhanced memory and athletic function in healthy mice. It enhances both athletic and mental performance – how good is that?
Then again, if EPO becomes a common cognitive enhancer, it will mean that few of us normal people would ever be able to compete in the Olympics. It was only in 2004 that caffeine was allowed in professional competition, but pretty soon college students will be doping themselves with EPO as a biochemical study aid. It will be interesting when almost all normal people would not be able to pass an Olympic-level drug test.
The possibility exists, however, that we may want the cognitive boost without increasing our red blood cells too much. And now that we know the cognitive effects of EPO are independent of red blood cell production, this may be possible too. Make a drug that stimulates the brain like EPO does, but doesn’t effect an increase in red blood cells. And this study has gone a long way to unraveling the relevant effects of EPO on neuronal plasticity that underly the enhancement to memory circuitry in the brain, which means that we may be able to find drugs that do so more effectively than EPO or act on other brain functions.
| 374
|
EVEN a material 10 billion times as strong as steel has a breaking point. It seems neutron stars may shatter under extreme forces, explaining puzzling X-ray flares.
Neutron stars are dense remnants of stars gone supernova, packing the mass of the sun into a sphere the size of a city. Their cores may be fluid, but their outer surfaces are solid and extremely tough - making graphene, the strongest material on Earth, look like tissue paper by comparison.
These shells may shatter, though, in the final few seconds before a pair of neutron stars merges to form a black hole - a union thought to generate explosions known as short gamma-ray bursts.
David Tsang of the California Institute of Technology in Pasadena and colleagues have calculated how the mutual gravitational pull of such stars will distort their shape, creating moving tidal bulges. As the stars spiral towards each other, orbiting ever faster, they squeeze and stretch each other ever faster too.
A few seconds before the stars merge, the frequency of this squeezing and stretching matches the frequency at which one of the stars vibrates most easily. This creates a resonance that boosts the vibrations dramatically, causing the star's crust to crack in many places - just as a wine glass may shatter when a certain note is sung, the team says (Physical Review Letters, DOI: 10.1103/physrevlett.108.011102).
The star's gravity is too powerful to let the pieces fly away, but the sudden movement can disturb its magnetic field, accelerating electrons and leading to a powerful X-ray flare. That could explain observations by NASA's Swift satellite in which a blast of X-rays preceded some short gamma-ray bursts by a few seconds.
Combining observations of X-ray flares with those of gravitational waves emitted by the stars as they spiral together could fix the exact frequency at which the shattering occurs, which would reveal more about the stars' mysterious interiors, says Tsang.
If you would like to reuse any content from New Scientist, either in print or online, please contact the syndication department first for permission. New Scientist does not own rights to photos, but there are a variety of licensing options available for use of articles and graphics we own the copyright to.
Have your say
Only subscribers may leave comments on this article. Please log in.
Only personal subscribers may leave comments on this article
| 469
|
A notch or groove cut into a piece of material to allow two sections to be combined with a flush joint.
A woodcutting tool used to cut an L-shaped groove into a piece of material. see also Rabbet
A strong current in a stream or river.
1. An enclosed metal channel, usually fire-resistant, installed in a building to hold electrical wiring.
2. A chute that directs the flow of a material to a specific location in a device.
A channel holding electrical wiring that is designed to look like a piece of decorative trim or molding.
A channel holding electrical wiring designed to be installed on a floor. The unit has a low profile and sloping edges to facilitate walking over it.
The illegal practice of directing certain races away from some neighborhoods and into others.
1. A storage unit designed to hold various objects.
2. To cause a structure to shift so that it is out of plumb.
1. A force that causes a structure to shift so that it is out of plumb.
2. Installation of bricks or other masonry units so that each course is stepped back from the previous one.
Straight-line outward movement from a circle's center.
A power saw with a circular blade that is mounted on a moveable arm. The arm is lowered or raised to move the cutting blade to or away from the material to be cut.
A drill press with a moveable arm that can be swung to various positions on the work table.
An HVAC system with ductwork running outward from a central heating and/or cooling unit.
The surface of a log cut down the center.
Heating system where electrical or hot-water heating elements are installed in a concrete slab floor.
see Radiant heating
Use of radiation to generate heat such as with baseboard heating where the circulating hot water is radiated through conduction by thin metal fins at the bottom of the wall. The room is warmed by air circulating around the heating unit using convention.
Heating unit that is exposed and which transfers heat generated by hot eater or steam through conduction. When the air circulates around the radiator using convention, the room is heated.
The distance from the center of a circle to the circumference. One-half of the diameter of a circle.
A tool used for checking the radii of convex and concave surfaces.
Radioactive gas that seeps into some homes, from the ground, through sump pumps, cracks in the foundation, etc., it is considered a health hazard.
Any of the beams that slope from the ridge of a roof to the eaves to serve as support for the roof.
A metal fastener attached to the top plate of a wall to hold a rafter.
A rafter parallel to the gable end that projects out to form an overhang.
The end of a rafter extending beyond the line of a building's walls.
A guide used when cutting rafters.
The top plate of a building's walls. The rafters rest on the rafter plate.
The vertical cut made into a rafter so it will rest on the wall plate. see also Rafter Seat Cut
The horizontal cut made into a rafter so it will rest on the wall plate. see also Rafter Plumb Cut
Cutting a section off of the end of a rafter equal to one-half of the thickness of the ridge board (the rafter on the other side of the ridge board receives a similar cut).
Tables, often printed on a framing square, containing the data required to calculate angles and lengths of rafters for various roof types.
see Rafter Overhang
A horizontal structure used as a handhold or to block off a drop or other unsafe area.
1. Continuous metal bars on which wheeled vehicles travel (i.e. railroads).
2. The horizontal sections of a panel door.
3. The top and bottom sections of a window sash.
Waterproof cap, also called weatherheads, mast heads or entrance caps, which is placed at the upper part of an electrical mast at the point where the wires are run to the inside electrical meter. Wires hang from the pole to the entrance cap so that the entrance cap is not the low point in the downhill run from the pole because water will run to the low point before dripping to the ground. Wires enter the entrance cap at an upward angle through a tight insulator. Water is further stopped from getting through the entrance cap because of this entrance angle.
Wood where the fibers have swelled, usually because of becoming wet. Wood is often sanded with the grain raised to achieve an extremely smooth finished surface.
1. A fork-like tool used for gathering materials (i.e. leaves) or smoothing an area of soil.
2. A roof overhang on a building's gable end.
3. An angle between objects.
A masonry joint where a portion of mortar has been removed, creating a groove between masonry units. A raked joint if often used in brickwork.
Mortgage, most commonly used by the elderly who have substantial equity in their homes. A periodic payment is made to the borrower from the lender thus, increasing the loan balance, causing negative amortization.
A hydraulically powered piston used for driving a weight.
A sloping surface used to move from one elevation to another.
| 873
|
During this week's close encounter with Earth, astronomers tracked the aircraft-carrier-sized asteroid 2005 YU55 with radar instruments, infrared cameras, visible-light telescopes — and an ultraviolet-sensitive space telescope as well.
This video shows the view from NASA's Swift satellite, which trained its Ultraviolet/Optical Telescope on the space rock as it zoomed away from Earth at 24,300 mph on Wednesday. Swift is best-known for its observations of high-energy outbursts and cosmic explosions, but it turns out that the spacecraft has been involved in 10 asteroid-observing sessions as well.
Swift's scientists had the satellite watch a couple of patches of sky that YU55 was predicted to pass through, and during the second observing opportunity, the telescope got a good fix on the asteroid.
"We observed the asteroid with Swift's Ultraviolet/Optical and X-ray telescopes but, as expected, we saw it only in the UV," Dennis Bodewits, a Swift team member at the University of Maryland in College Park, said in today's video advisory from NASA.
During the 27-minute long exposure, Swift detected short-term variations in brightness caused by the asteroid's rotation.
"The result is a movie of 2005 YU55 at ultraviolet wavelengths unobtainable from ground-based telescopes," NASA reported today. "For planetary scientists, this movie is a treasure trove of data that will help them better understand how this asteroid is put together, information that may help make predictions of its motion more secure for centuries to come."
More about the asteroid encounter:
- Parting shots from the asteroid
- Passing asteroid puts on a show
- Your guide to the asteroid encounter
- How to save our planet from a killer asteroid
- Could the asteroid destroy the moon? (No)
- Why radar's the best for tracking near-Earth objects
- Interactive: Close encounters of the asteroid kind
Connect with the Cosmic Log community by "liking" the log's Facebook page, following @b0yle on Twitter or following the Cosmic Log Google+ page. You can also check out "The Case for Pluto," my book about the controversial dwarf planet and the search for new worlds.
| 768
|
In De vita et moribus Iulii Agricolae, Tacitus describes and praises the life of his father-in-law Gnaeus Julius Agricola, an eminent Roman general. It covers briefly the people and geography of Britain, where Agricola was stationed.
- Idque apud imperitos humanitas vocabatur, cum pars servitutis esset.
- Translation: Because they didn't know better, they called it 'civilization,' when it was part of their slavery.
- Book 1, paragraph 21.
- Longer variant: Step by step they were led to things which dispose to vice, the lounge, the bath, the elegant banquet. All this in their ignorance they called civilisation, when it was but a part of their servitude.
- Auferre, trucidare, rapere, falsis nominibus imperium; atque, ubi solitudinem faciunt, pacem appellant.
- Translation: To ravage, to slaughter, to usurp under false titles, they call empire; and where they make a desert, they call it peace. Oxford Revised Translation (at Project Gutenberg)
- Translation: They plunder, they slaughter, and they steal: this they falsely name Empire, and where they make a wasteland, they call it peace. — translation Loeb Classical Library edition
- Translation: To plunder, butcher, steal, these things they misname empire: they make a desolation and they call it peace. — translation by William Peterson
- More colloquially: They rob, kill and plunder all under the deceiving name of Roman Rule. They make a desert and call it peace.
- At the end of chapter 30.
- This is a speech by the Caledonian chieftain Calgacus addressing assembled warriors about Rome's insatiable appetite for conquest and plunder. The chieftain's sentiment can be contrasted to "peace given to the world" which was frequently inscribed on Roman medals. The last part solitudinem faciunt, pacem appellant (they make a desert, and call it peace) is often quoted alone. Lord Byron for instance uses the phrase (in English) as follows,
- Mark where his carnage and his conquests cease!
He makes a solitude, and calls it — peace.
- Lord Byron, Bride of Abydos (1813), Canto 2, stanza 20.
- Mark where his carnage and his conquests cease!
- Et maiores vestros et posteros cogitate.
- Translation: Think of your forefathers and posterity.
- Chapter 32.
- It belongs to human nature to hate those you have injured.
- Chapter 42; reported in Bartlett's Familiar Quotations, 10th ed. (1919).
- Tu vero felix, Agricola, non vitae tantum claritate, sed etiam opportunitate mortis.
- Translation: Thou wast indeed fortunate, Agricola, not only in the splendour of thy life, but in the opportune moment of thy death.
- Chapter 45.
- The Germans themselves I should regard as aboriginal, and not mixed at all with other races through immigration or intercourse. For in former times, it was not by land but on shipboard that those who sought to emigrate would arrive; and the boundless and, so to speak, hostile ocean beyond us,is seldom entered by a sail from our world.
- Chapter 2
- They even say that an altar dedicated to Ulysses, with the addition of the name of his father, Laertes, was formerly discovered on the same spot, and that certain monuments and tombs with Greek inscriptions, still exist on the borders of Germany and Rhaetia.
- Chapter 3.
- On the whole,one would say that their strength is in their infantry, which fights along with the cavalry; admirably adapted to the action of the latter is the swiftness of certain foot soldiers, who are picked from the entire youth of their country, and stationed in front of the line.
- Chapter 6.
- To abandon your shield is the basest of crimes; nor may a man thus disgraced be present at the sacred rites, or enter their council; many, indeed, after escaping from battle, have ended their infamy with the halter.
- Chapter 6.
- Mercury is the deity whom they chiefly worship, and on certain days they deem it right to sacrifice to him even with human victims.
- Chapter 9.
- Quanquam severa illic matrimonia
- Translation: However the marriage is there severe.
- Start of chapter 18.
- This is in the sense that the matrimonial bond was strictly observed by the Germanic peoples, this being compared favorably against licentiousness in Rome. Tacitus appears to hold the fairly strict monogamy (with some exceptions among nobles who marry again) between Germanic husbands and wives, and the chastity among the unmarried to be worthy of the highest praise. (Ch. 18).
- …ibi boni mores valent quam alibi bonae leges.
- Translation: …good habits are here more effectual than good laws elsewhere.
- End of chapter 19.
- No one in Germany laughs at vice, nor do they call it the fashion to corrupt and to be corrupted.
- Chapter 19.
- Indeed, the crowning proof of their valour and their strength is that they keep up their superiority without harm to others.
- Chapter 35.
- Their shields are black, their bodies dyed. They choose dark nights for battle, and, by the dread and gloomy aspect of their death-like host, strike terror into the foe, who can never confront their strange and almost infernal appearance.
- Chapter 43.
- All this is unauthenticated, and I shall leave it open.
- Chapter 46 (last text line).
- It is the rare fortune of these days that one may think what one likes and say what one thinks.
- Book I, 1.
- He possessed a peculiar talent of producing effect in whatever he said or did.
- Book II, 80.
- Once killing starts, it is difficult to draw the line.
- Book I, 39.
- Some might consider him as too fond of fame; for the desire for glory clings even to the best men longer than any other passion.
- Book IV, 6.
- Deos fortioribus adesse.
- Translation: The gods are on the side of the stronger.
- Book IV, 17.
- Corruptissima re publica plurimae leges.
- Translation: The more numerous the laws, the more corrupt the government.
- Variant: The more corrupt the state, the more laws.
- Original Quote: And now bills were passed, not only for national objects but for individual cases, and laws were most numerous when the commonwealth was most corrupt.
- Book III, 27.
- Conspicuous by his absence.
- Book III, 76; Lord John Russell, alluding to an expression used by him ("Conspicuous by his absence") in his address to the electors of the city of London, said, "It is not an original expression of mine, but is taken from one of the greatest historians of antiquity".
- The images of twenty of the most illustrious families—the Manlii, the Quinctii, and other names of equal splendour—were carried before it [the bier of Junia]. Those of Brutus and Cassius were not displayed; but for that very reason they shone with pre-eminent lustre.
- Book III, 76.
- He had talents equal to business, and aspired no higher.
- Book VI, 39.
- He upbraided Macro, in no obscure and indirect terms, "with forsaking the setting sun and turning to the rising".
- Book VI, 52, referring to Tiberius.
- What is today supported by precedents will hereafter become a precedent.
- Book XI, 24.
- So true is it that all transactions of preeminent importance are wrapt in doubt and obscurity; while some hold for certain facts the most precarious hearsays, others turn facts into falsehood; and both are exaggerated by posterity.
- Variant: So obscure are the greatest events, as some take for granted any hearsay, whatever its source, others turn truth into falsehood, and both errors find encouragement with posterity.
- Book III, 19.
- The desire for safety stands against every great and noble enterprise.
- A popular rendering of: “nisi impunitatis cupido retinuisset, magnis semper conatibus adversa”
- Variant: "but desire of escape, foe to all great enterprises, held him back." This of Subrius Flavus’ passing thought of assassinating Nero while the emperor sang on stage.
- Book XV, 50.
Quotes about Tacitus
- Abuse, if you slight it, will gradually die away; but if you show yourself irritated you will be thought to have deserved it.
- Liberty is given by nature even to mute animals.
- Great empires are not maintained by timidity.
| 659
|
a combination of one or more elementary reaction steps which start with the appropriate reactants and end with the appropriate product(s)
a description of the path, or sequence of steps, by which a reaction occurs
a description of the path that a reaction takes
a detailed description of how a chemical reaction occurs
a detailed description of the way a reaction occurs and is based on the known experimental data about the reaction
a detailed (theoretical) description of how we think the chemical reaction proceeds
a series of elementary reactions or elementary steps that lead from reactants to products
a set of steps at the molecular level
a step by step description of the separate steps that occur during a chemical reaction
a stepwise description of the reaction path
mechanism. A list of all elementary reactions that occur in the course of an overall chemical reaction.
In chemistry, a reaction mechanism is the step by step sequence of elementary reactions by which overall chemical change occurs.
| 194
|
A long time ago I wrote the article The Dull Case of Emissivity and Average Temperatures and expected that would be the end of the interest in emissivity. But it is a gift that keeps on giving, with various people concerned that no one has really been interested in measuring surface emissivity properly.
All solid and liquid surfaces emit thermal radiation according to the Stefan-Boltzmann formula:
E = εσT4
where ε=emissivity, a material property; σ = 5.67×10-8 ; T = temperature in Kelvin (absolute temperature)
and E is the flux in W/m²
More about this formula and background on the material properties in Planck, Stefan-Boltzmann, Kirchhoff and LTE.
The parameter called emissivity is the focus of this article. It is of special interest because to calculate the radiation from the earth’s surface we need to know only temperature and emissivity.
Emissivity is a value between 0 and 1. And is also depends on the wavelength of radiation (and in some surfaces like metals, also the direction). Because the wavelengths of radiation depend on temperature, emissivity also depends on temperature.
When emissivity = 1, the body is called a “blackbody”. It’s just the theoretical maximum that can be radiated. Some surfaces are very close to a blackbody and others are a long way off.
Note: I have seen many articles by keen budding science writers who have some strange ideas about “blackbodies”. The only difference between a blackbody and a non-blackbody is that the emissivity of a blackbody = 1, and the emissivity of a non-blackbody is less than 1. That’s it. Nothing else.
The wavelength dependence of emissivity is very important. If we take snow for example, it is highly reflective to solar (shortwave) radiation with as much as 80% of solar radiation being reflected. Solar radiation is centered around a wavelength of 0.5μm.
Yet snow is highly absorbing to terrestrial (longwave) radiation, which is centered around a wavelength of 10μm. The absorptivity and emissivity around freezing point is 0.99 – meaning that only 1% of incident longwave radiation would be reflected.
Let’s take a look at the Planck curve – the blackbody radiation curve – for surfaces at a few slightly different temperatures:
The emissivity (as a function of wavelength) simply modifies these curves.
Suppose, for example, that the emissivity of a surface was 0.99 across this entire wavelength range. In that case, a surface at 30°C would radiate like the light blue curve but at 99% of the values shown. If the emissivity varies across the wavelength range then you simply multiply the emissivity by the intensity at each wavelength to get the expected radiation.
Sometimes emissivity is quoted as an average for a given temperature – this takes into account the shape of the Planck curve shown in the graphs above.
Often, when emissivity is quoted as an overall value, the total flux has been measured for a given temperature and the emissivity is simply:
ε = actual radiation measured / blackbody theoretical radiation at that temperature
[Fixed, thanks to DeWitt Payne for pointing out the mistake]
In practice the calculation is slightly more involved, see note 1.
It turns out that the emissivity of water and of the ocean surface is an involved subject.
And because of the importance of calculating the sea surface temperature from satellite measurements, the emissivity of the ocean in the “atmospheric window” (8-14 μm) has been the subject of many 100′s of papers (perhaps 1000′s). These somewhat overwhelm the papers on the less important subject of “general ocean emissivity”.
Aside from climate, water itself is an obvious subject of study for spectroscopy.
For example, 29 years ago Miriam Sidran writing Broadband reflectance and emissivity of specular and rough water surfaces, begins:
The optical constants of water have been extensively studied because of their importance in science and technology. Applications include a) remote sensing of natural water surfaces, b) radiant energy transfer by atmospheric water droplets, and c) optical properties of diverse materials containing water, such as soils, leaves and aqueous solutions.
In this study, values of the complex index of refraction from six recent articles were averaged by visual inspection of the graphs, and the most representative values in the wavelength range of 0.200 μm to 5 cm were determined. These were used to find the directional polarized reflectance and emissivity of a specular surface and the Brewster or pseudo-Brewster angle as functions of wavelength.
The directional polarized reflectance and emissivity of wind-generated water waves were studied using the facet slope distribution function for a rough sea due to Cox and Munk .
Applications to remote sensing of sea surface temperature and wave state are discussed, including effects of salinity.
Emphasis added. She also comments in her paper:
For any wavelength, the total emissivity, ε, is constant for all θ [angles] < 45° [from vertical]; this follows from Fig. 8 and Eq. (6a). It is important in remote sensing of thermal radiation from space, as discussed later..
The polarized emissivities are independent of surface roughness for θ < 25°, while for θ > 25°, the thermal radiation is partly depolarized by the roughness.
This means that when you look at the emission radiation from directly above (and close to directly above) the sea surface roughness doesn’t have an effect.
I thought some other comments might also be interesting:
The 8-14-μm spectral band is chosen for discussion here because (a) it is used in remote sensing and (b) the atmospheric transmittance, τ, in this band is a fairly well-known function of atmospheric moisture content. Water vapor is the chief radiation absorber in this band.
In Eqs. (2)-(4), n and k (and therefore A and B) are functions of salinity. However, the emissivity value, ε, computed for pure water differs from that of seawater by <0.5%.
When used in Eqs. (10), it causes an error of <0.20°C in retrieved Ts [surface temperature]. Since ε in this band lies between 0.96 and 0.995, approximation ε= 1 is routinely used in sea surface temperature retrieval. However, this has been shown to cause an error of -0.5 to -1.0°C for very dry atmospheres. For very moist atmospheres, the error is only ≈0.2°C.
One of the important graphs from her paper:
Click to view a larger image
Emissivity = 1 – Reflectance. The graph shows Reflectance vs Wavelength vs Angle of measurement.
I took the graph (coarse as it is) and extracted the emissivity vs wavelength function (using numerical techniques). I then calculated the blackbody radiation for a 15°C surface and the radiation from a water surface using the emissivity from the graph above for the same 15°C surface. Both were calculated from 1 μm to 100 μm:
The “unofficial” result, calculating the average emissivity from the ratio: ε = 0.96.
This result is valid for 0-30°C. But I suspect the actual value will be modified slightly by the solid angle calculations. That is, the total flux from the surface (the Stefan-Boltzmann equation) is the spectral intensity integrated over all wavelengths, and integrated over all solid angles. So the reduced emissivity closer to the horizon will affect this measurement.
Niclòs et al – 2005
One of the most interesting recent papers is In situ angular measurements of thermal infrared sea surface emissivity—validation of models, Niclòs et al (2005). Here is the abstract:
In this paper, sea surface emissivity (SSE) measurements obtained from thermal infrared radiance data are presented. These measurements were carried out from a fixed oilrig under open sea conditions in the Mediterranean Sea during the WInd and Salinity Experiment 2000 (WISE 2000).
The SSE retrieval methodology uses quasi-simultaneous measurements of the radiance coming from the sea surface and the downwelling sky radiance, in addition to the sea surface temperature (SST). The radiometric data were acquired by a CIMEL ELECTRONIQUE CE 312 radiometer, with four channels placed in the 8–14 μm region. The sea temperature was measured with high-precision thermal probes located on oceanographic buoys, which is not exactly equal to the required SST. A study of the skin effect during the radiometric measurements used in this work showed that a constant bulk–skin temperature difference of 0.05±0.06 K was present for wind speeds larger than 5 m/s. Our study is limited to these conditions.
Thus, SST used as a reference for SSE retrieval was obtained as the temperature measured by the contact thermometers placed on the buoys at 20-cm depth minus this bulk–skin temperature difference.
SSE was obtained under several observation angles and surface wind speed conditions, allowing us to study both the angular and the sea surface roughness dependence. Our results were compared with SSE models..
The introduction explains why specifically they are studying the dependence of emissivity on the angle of measurement – for reasons of accurate calculation of sea surface temperature:
The requirement of a maximum uncertainty of ±0.3 K in sea surface temperature (SST) as input to climate models and the use of high observation angles in the current space missions, such as the 55° for the forward view of the Advanced Along Track Scanning Radiometer (AATSR) (Llewellyn-Jones et al., 2001) on board ENVISAT, need a precise and reliable determination of sea surface emissivity (SSE) in the thermal infrared region (TIR), as well as analyses of its angular and spectral dependences.
The emission of a rough sea surface has been studied over the last years due to the importance of the SSE for accurate SST retrieval. A reference work for many subsequent studies has been the paper written by Cox and Munk (1954)..
The experimental setup:
From Niclos (2004)
The results (compared with one important model from Masuda et al 1988):
From Niclos (2004)
Click on the image for a larger graphic
This paper also goes on to compare the results with the model of Wu & Smith (1997) and indicates the Wu & Smith’s model is a little better.
The tabulated results, note that you can avoid the “eye chart effect” by clicking on the table:
Click on the image for a larger view
Note that the emissivities are in the 8-14μm range.
You can see that the emissivity when measured from close to vertical is 0.98 – 0.99 at two different wind speeds.
Konda et al – 1994
A slightly older paper which is not concerned with angular dependence of sea surface emissivity is by Konda, Imasato, Nishi and Toda (1994).
They comment on a few older papers:
Buettner and Kern (1965) estimated the sea surface emissivity to be 0.993 from an experiment using an emissivity box, but they disregarded the temperature difference across the cool skin.
Saunders (1967b, 1968) observed the plane sea surface irradiance from an airplane and determined the reflectance. By determining the reflectance as the ratio of the differences in energy between the clear and the cloudy sky at different places, he calculated the emissivity to be 0.986. The process of separating the reflection from the surface irradiance, however, is not precise.
Mikhaylov and Zolotarev (1970) calculated the emissivity from the optical constant of the water and found the average in the infrared region was 0.9875.
The observation of Davies et al. (1971) was performed on Lake Ontario with a wave height less than 25 cm. They measured the surface emission isolated from sky radiation by an aluminum cone, and estimated the emissivity to be 0.972. The aluminum was assumed to act as a mirror in infrared region. In fact,aluminum does not work as a perfect mirror.
Masuda et al. (1988) computed the surface emissivity as a function of the zenith angle of observed radiation and wind speed. They computed the emissivity from the reflectance of a model sea surface consisting of many facets, and changed their slopes according to Gaussian distribution with respect to surface wind. The computed emissivity in 11 μm was 0.992 under no wind.
Each of these studies in trying to determine the value of emissivity, failed to distinguish surface emission from reflection and to evaluate the temperature difference across the cool skin. The summary of these studies are tabulated in Table 1.
The table summarizing some earlier work:
Konda and his co-workers took measurements over a one year period from a tower in Tanabe Bay, Japan.
They calculated from their results that the ocean emissivity was 0.984±0.004.
One of the challenges for Konda’s research and for Niclòs is the issue of sea surface temperature measurement itself. Here is a temperature profile which was shown in the comments of Does Back Radiation “Heat” the Ocean? – Part Three:
Kawai & Wada (2007)
The point is the actual surface from which the radiation is emitted will usually be at a slightly different temperature from the bulk temperature (note the logarithmic scale of depth). This is the “cool skin” effect. This surface temperature effect is also moderated by winds and is very difficult to measure accurately in field conditions.
Smith et al – 1996
Another excellent paper which measured the emissivity of the ocean is by Smith et al (1996):
An important objective in satellite remote sensing is the global determination of sea surface temperature (SST). For such measurements to be useful to global climate research, an accuracy of ±0.3K or better over a length of 100km and a timescale of days to weeks must be attained. This criterion is determined by the size of the SST anomalies (≈1K) that can cause significant disturbance to the global atmospheric circulation patterns and the anticipated size of SST perturbations resulting from global climate change. This level of uncertainty is close to the theoretical limits of the atmospheric corrections..
It is also a challenge to demonstrate that such accuracies are being achieved, and conventional approaches, which compare the SST derived from drifting or moored buoys, generally produce results with a scatter of ±0.5 to 0.7K. This scatter cannot be explained solely by uncertainties in the buoy thermometers or the noise equivalent temperature difference of the AVHRR, as these are both on the order of 0.2K or less but are likely to be surface emissivity/reflectivity uncertainties, residual atmospheric effects, or result from the methods of comparison
Note that the primary focus of this research was to have accurate SST measurements from satellites.
From Smith et al (1996)
The experimental work on the research vessel Pelican included a high spectral resolution Atmospheric Emitted Radiance Interferometer (AERI) which was configured to make spectral observations of the sea surface radiance at several view angles. Any measurement from the surface of course, is the sum of the emitted radiance from the surface as well as the reflected sky radiance.
- ocean salinity
- intake water temperature
- surface air temperature
- wind velocity
- SST within the top 15cm of depth
There was also independent measurement of the radiative temperature of the sea surface at 10μm with a Heimann broadband radiation thermometer “window” radiometer. And radiosondes were launched from the ship roughly every 3 hours.
Additionally, various other instruments took measurements from a flight altitude of 20km. Satellite readings were also compared.
The AERI measured the spectral distribution of radiance from 3.3μm to 20μm at 4 angles. Upwards at 11.5° from zenith, and downwards at 36.5°, 56.5° and 73.5°.
There’s a lot of interesting discussion of the calculations in their paper. Remember that the primary aim is to enable satellite measurements to have the most accurate measurements of SST and satellites can only really “see” the surface through the “atmospheric window” from 8-12μm.
Here are the wavelength dependent emissivity results shown for the 3 viewing angles. You can see that at the lowest viewing angle of 36.5° the emissivity is 0.98 – 0.99 in the 8-12μm range.
From Smith et al (1996)
Note that the wind speed doesn’t have any effect on emissivity at the more direct angle, but as the viewing angle moves to 73.5° the emissivity has dropped and high wind speeds change the emissivity considerably.
Henderson et al – 2003
Henderson et al (2003) is one of the many papers which consider the theoretical basis of how viewing angles change the emissivity and derive a model.
Just as an introduction, here is the theoretical variation in emissivity with measurement angle, versus “refractive index” as computed by the Fresnel equations:
The legend is refractive index from 1.20 to 1.35. Water, at visible wavelengths, has a refractive index of 1.33. This shows how the emissivity reduces once the viewing angle increases above 50° from the vertical.
The essence of the problem of sea surface roughness for large viewing angles is shown in the diagram below, where multiple reflections take place:
Henderson and his co-workers compare their results with the measured results of Smith et al (1996) and also comment that at zenith viewing angles the emissivity does not depend on the wind speed, but at larger angles from vertical it does.
A quick summary of their model:
We have developed a Monte Carlo ray-tracing model to compute the emissivity of computer-rendered, wind-roughened sea surfaces. The use of a ray-tracing method allows us to include both the reflected emission and shadowing and, furthermore, permits us to examine more closely how these processes control the radiative properties of the surface. The intensity of the radiation along a given ray path is quantified using Stokes vectors, and thus, polarization is explicitly included in the calculations as well.
Their model results compare well with the experimental results. Note that the approach of generating a mathematical model to calculate how emissivity changes with wind speed and, therefore, wave shape is not at all new.
Water retains its inherent properties of emissivity regardless of how it is moving or what shape it is. The theoretical challenge is handling the multiple reflections, absorptions, re-emissions that take place when the radiance from the water is measured at some angle from the vertical.
The best up to date measurements of ocean emissivity in the 8-14 μm range are 0.98 – 0.99. The 8-14 μm range is well-known because of the intense focus on sea surface temperature measurements from satellite.
From quite ancient data, the average emissivity of water across a very wide broadband range (1-100 μm) is 0.96 for water temperatures from 0-30°C.
The values from the ocean when measured close to the vertical are independent of wind speed and sea surface roughness. As the angle of measurement moves from the vertical around to the horizon the measured emissivity drops and the wind speed affects the measurement significantly.
These values have been extensively researched because the calculation of sea surface temperature from satellite measurements in the 8-14μm “atmospheric window” relies on the accurate knowledge of emissivity and any factors which affect it.
For climate models – I haven’t checked what values they use. I assume they use the best experimental values from the field. That’s an assumption. I’ve already read enough on ocean emissivity.
For energy balance models, like the Trenberth and Kiehl update, an emissivity of 1 doesn’t really affect their calculations. The reason, stated simply, is that the upwards surface radiation and the downward atmospheric radiation are quite close in magnitude. For example, the globally annually averaged values of both are 396 W/m² (upward surface) vs 340 W/m² (downward atmospheric).
Suppose the emissivity drops from 0.98 to 0.97 – what is the effect on upwards radiation through the atmosphere?
The upwards radiation has dropped by 4W/m², but the reflected atmospheric radiation has increased by 3.4W/m². The net upwards radiation through the atmosphere has reduced by only 0.6 W/m².
One of our commenters asked what value the IPCC uses. The answer is they don’t use a value at all because they summarize research from papers in the field.
Whether they do it well or badly is a subject of much controversy, but what is most important to understand is that the IPCC does not write papers, or perform GCM model runs, or do experiments – and that is why you see almost no equations in their many 1000′s of pages of discussion on climate science.
For those who don’t believe the “greenhouse” effect exists, take a look at Understanding Atmospheric Radiation and the “Greenhouse” Effect – Part One in the light of all the measured results for ocean emissivity.
On Another Note
It’s common to find claims on various blogs and in comments on blogs that climate science doesn’t do much actual research.
I haven’t found that to be true. I have found the opposite.
Whenever I have gone digging for a particular subject, whether it is the diurnal temperature variation in the sea surface, diapycnal & isopycnal eddy diffusivity, ocean emissivity, or the possible direction and magnitude of water vapor feedback, I have found a huge swathe of original research, of research building on other research, of research challenging other research, and detailed accounts of experimental methods, results and comparison with theory and models.
Just as an example, in the case of emissivity of sea surface, at the end of the article you can see the first 30 or so results pulled up from one journal – Remote Sensing of the Environment for the search phrase “emissivity sea surface”. The journal search engine found 348 articles (of course, not every one of them is actually about ocean emissivity measurements).
Perhaps it might turn out to be the best journal for this subject, but it’s still just one journal.
Broadband reflectance and emissivity of specular and rough water surfaces, Sidran, Applied Optics (1981)
In situ angular measurements of thermal infrared sea surface emissivity—validation of models, Niclòs, Valor, Caselles, Coll & Sànchez, Remote Sensing of Environment (2005)
Measurement of the Sea Surface Emissivity, Konda, Imasato, Nishi and Toda, Journal of Oceanography (1994)
Observations of the Infrared Radiative Properties of the Ocean—Implications for the Measurement of Sea Surface Temperature via Satellite Remote Sensing, Smith, Knuteson, Revercomb, Feltz, Nalli, Howell, Menzel, Brown, Brown, Minnett & McKeown, Bulletin of the American Meteorological Society (1996)
The polarized emissivity of a wind-roughened sea surface: A Monte Carlo model, Henderson, Theiler & Villeneuve, Remote Sensing of Environment (2003)
Note 1: The upward radiation from the surface is the sum of three contributions: (i) direct emission of the sea surface, which is attenuated by the absorption of the atmospheric layer between the sea surface and the instrument; (ii) reflection of the downwelling sky radiance on the sea, attenuated by the atmosphere; and (iii) the upwelling atmospheric radiance emitted in the observing direction.
So the measured radiance can be expressed as:
where the three terms on the right are each of the three contributions noted in the same order.
Note 2: 1/10th of the search results returned from one journal for the search term “emissivity sea surface”:
Remote Sensing of Environment - search results
Read Full Post »
| 770
|
Forgetfulness: Knowing when to Ask for Help
Maria has been a teacher for 35 years. Teaching fills her life and gives her a sense of accomplishment, but recently she has begun to forget details and has become more and more disorganized. At first, she laughed it off, but her memory problems have worsened. Her family and friends have been sympathetic but are not sure what to do. Parents and school administrators are worried about Maria's performance in the classroom. The principal has suggested she see a doctor. Maria is angry with herself and frustrated, and she wonders whether these problems are signs of Alzheimer's disease or just forgetfulness that comes with getting older.
Many people worry about becoming forgetful. They think forgetfulness is the first sign of Alzheimer's disease. Over the past few years, scientists have learned a lot about memory and why some kinds of memory problems are serious but others are not.
Age-related changes in memory
Forgetfulness can be a normal part of aging. As people get older, changes occur in all parts of the body, including the brain. As a result, some people may notice that it takes longer to learn new things, they don't remember information as well as they did, or they lose things like their glasses. These usually are signs of mild forgetfulness, not serious memory problems.
Some older adults also find that they don't do as well as younger people on complex memory or learning tests. Scientists have found, though, that given enough time, healthy older people can do as well as younger people do on these tests. In fact, as they age, healthy adults usually improve in areas of mental ability such as vocabulary.
Keeping your memory sharp
People with some forgetfulness can use a variety of techniques that may help them stay healthy and maintain their memory and mental skills. Here are some tips that can help:
Other causes of memory loss
Some memory problems are related to health issues that may be treatable. For example, medication side effects, vitamin B12 deficiency, chronic alcoholism, tumors or infections in the brain, or blood clots in the brain can cause memory loss or possibly dementia (see more on dementia, below). Some thyroid, kidney, or liver disorders also can lead to memory loss. A doctor should treat serious medical conditions like these as soon as possible.
Emotional problems, such as stress, anxiety, or depression, can make a person more forgetful and can be mistaken for dementia. For instance, someone who has recently retired or who is coping with the death of a spouse, relative, or friend may feel sad, lonely, worried, or bored. Trying to deal with these life changes leaves some people confused or forgetful.
The confusion and forgetfulness caused by emotions usually are temporary and go away when the feelings fade. The emotional problems can be eased by supportive friends and family, but if these feelings last for a long time, it is important to get help from a doctor or counselor. Treatment may include counseling, medication, or both.
More serious memory problems
For some older people, memory problems are a sign of a serious problem, such as mild cognitive impairment or dementia. People who are worried about memory problems should see a doctor. The doctor might conduct or order a thorough physical and mental health evaluation to reach a diagnosis. Often, these evaluations are conducted by a neurologist, a physician who specializes in problems related to the brain and central nervous system.
A complete medical exam for memory loss should review the person's medical history, including the use of prescription and over-the-counter medicines, diet, past medical problems, and general health. A correct diagnosis depends on accurate details, so in addition to talking with the patient, the doctor might ask a family member, caregiver, or close friend for information.
Blood and urine tests can help the doctor find the cause of the memory problems or dementia. The doctor also might do tests for memory loss and test the person's problem-solving and language abilities. A computed tomography (CT) or magnetic resonance imaging (MRI) brain scan may help rule out some causes of the memory problems.
Amnestic mild cognitive impairment (MCI). Some people with memory problems have a condition called amnestic mild cognitive impairment, or amnestic MCI. People with this condition have more memory problems than normal for people their age, but their symptoms are not as severe as those of Alzheimer's disease, and they are able to carry out their normal daily activities.
Signs of MCI include misplacing things often, forgetting to go to important events and appointments, and having trouble coming up with desired words. Family and friends may notice memory lapses, and the person with MCI may worry about losing his or her memory. These worries may prompt the person to see a doctor for diagnosis.
Researchers have found that more people with MCI than those without it go on to develop Alzheimer's within a certain timeframe. However, not everyone who has MCI develops AD. Studies are underway to learn why some people with MCI progress to AD and others do not.
There currently is no standard treatment for MCI. Typically, the doctor will regularly monitor and test a person diagnosed with MCI to detect any changes in memory and thinking skills over time. There are no medications approved for use for MCI.
Dementia. Dementia is the loss of thinking, memory, and reasoning skills to such an extent that it seriously affects a person's ability to carry out daily activities. Dementia is not a disease itself but a group of symptoms caused by certain diseases or conditions such as AD. People with dementia lose their mental abilities at different rates.
Symptoms may include:
Two of the most common forms of dementia in older people are Alzheimer's disease and vascular dementia. These types of dementia cannot be cured at present.
In Alzheimer's disease, changes to nerve cells in certain parts of the brain result in the death of a large number of cells. Symptoms of Alzheimer's disease begin slowly and worsen steadily as damage to nerve cells spreads throughout the brain. As time goes by, forgetfulness gives way to serious problems with thinking, judgment, recognizing family and friends, and the ability to perform daily activities like driving a car or handling money. Eventually, the person needs total care.
In vascular dementia, a series of strokes or changes in the brain's blood supply leads to the death of brain tissue. Symptoms of vascular dementia can vary but usually begin suddenly, depending on where in the brain the strokes occurred and how severe they were. The person's memory, language, reasoning, and coordination may be affected. Mood and personality changes are common as well.
It's not possible to reverse damage already caused by a stroke, so it's very important to get medical care right away if someone has signs of a stroke. It's also important to take steps to prevent further strokes, which worsen vascular dementia symptoms. Some people have both Alzheimer's and vascular dementia.
Treatment for dementia
A person with dementia should be under a doctor's care. The doctor might be a neurologist, family doctor, internist, geriatrician, or psychiatrist. He or she can treat the patient's physical and behavioral problems (such as aggression, agitation, or wandering) and answer the many questions that the person or family may have.
People with dementia caused by Alzheimer's disease may be treated with medications. Four medications are approved by the U.S. Food and Drug Administration to treat Alzheimer's disease. Donepezil (Aricept®), rivastigmine (Exelon®), and galantamine (Razadyne®) are used to treat mild to moderate Alzheimer's disease (donepezil has been approved to treat severe Alzheimer's disease as well). Memantine (Namenda®) is used to treat moderate to severe Alzheimer's disease. These drugs may help maintain thinking, memory, and speaking skills, and may lessen certain behavioral problems for a few months to a few years in some people. However, they dont stop Alzheimer's disease from progressing. Studies are underway to investigate medications to slow cognitive decline and to prevent the development of Alzheimer's disease.
People with vascular dementia should take steps to prevent further strokes. These steps include controlling high blood pressure, monitoring and treating high blood cholesterol and diabetes, and not smoking. Studies are underway to develop medicines to reduce the severity of memory and thinking problems that come with vascular dementia. Other studies are looking at the effects of drugs to relieve certain symptoms of this type of dementia.
Family members and friends can help people in the early stages of dementia to continue their daily routines, physical activities, and social contacts. People with dementia should be kept up-to-date about the details of their lives, such as the time of day, where they live, and what is happening at home or in the world. Memory aids may help. Some families find that a big calendar, a list of daily plans, notes about simple safety measures, and written directions describing how to use common household items are useful aids.
What you can do
If you're concerned that you or someone you know has a serious memory problem, talk with your doctor. He or she may be able to diagnose the problem or refer you to a specialist in neurology or geriatric psychiatry. Health care professionals who specialize in Alzheimer's can recommend ways to manage the problem or suggest treatment or services that might help. More information is available from the organizations listed below.
People with Alzheimer's disease, MCI, or a family history of Alzheimer's disease, and healthy people with no memory problems and no family history of Alzheimer's disease may be able to take part in clinical trials. Participating in clinical trials is an effective way to help in the fight against Alzheimers. To find out more about clinical trials, call the Alzheimer's Disease Education and Referral (ADEAR) Center toll-free at 1-800-438-4380 or visit the ADEAR Center website at www.nia.nih.gov/Alzheimers. More information is available at www.ClinicalTrials.gov.
For more information
Here are some helpful resources:
Alzheimer's Disease Education and Referral (ADEAR) Center
P.O. Box 8250
Silver Spring, MD 20907-8250
The National Institute on Aging's ADEAR Center offers information and publications in English and Spanish for families, caregivers and professionals on diagnosis, treatment, patient care, caregiver needs, long-term care, education and training, and research related to Alzheimer's disease.
225 North Michigan Avenue, Floor 17
Chicago, IL 60601-7633
National Library of Medicine
For more information on health and aging, contact:
Visit NIHSeniorHealth (www.nihseniorhealth.gov), a senior-friendly website from the National Institute on Aging and the National Library of Medicine. This website has health information for older adults. Special features make it simple to use. For example, you can click on a button to have the text read out loud or to make the type larger.
National Institute on Aging
National Institutes of Health
U.S. Department of Health and Human Services
Page last updated March 29, 2010
| 780
|
The riveting stories of two former slaves
Recently surfaced manuscripts recount the rigors of both slavery and freedom.
"Slave narratives, some of the most powerful records of our past, are extremely rare," David Blight tells us on the first page of A Slave No More. Only 55 post-Civil War narratives exist and of these, Blight notes, "a mere handful" are first-person accounts by slaves who ran away and freed themselves.Skip to next paragraph
Subscribe Today to the Monitor
So it is remarkable indeed that two such accounts should now surface – and almost at the same moment. Both had been handed down over the years by family and friends and both finally landed in the lap of Blight, who is director of Yale University's Gilder Lehrman Center for the Study of Slavery, Resistance, and Abolition.
Even for Blight, with his expertise in the field, these manuscripts were a revelation. Reading them, he says, "I have come to understand emancipation as never before."
"A Slave No More" contains both the unedited manuscripts and analysis by Blight, which gives the manuscripts context and offers additional information about the lives of the narrators, John Washington and Wallace Turnage.
As far as is known, Washington and Turnage never met. They have much in common – yet there is also much that divides them. Washington was an urban slave, living in Fredericksburg, Va. He was 24 on the day in April 1862, when he slipped across Union Army lines and won his freedom. Turnage was a field hand on an Alabama cotton plantation. Four times he attempted escape and finally, on his fifth try in 1864, he succeeded in reaching the Union Army in Mobile Bay. Both men had white fathers, although Turnage knew his while Washington did not.
Both men went on to live for many years as free men. Washington settled in Washington, D.C., and became a house painter, enjoying a reasonably comfortable life. Turnage had a more hardscrabble existence in New York City, where he worked as a watchman, waiter, and other occupations.
Their narratives are powerful and poignant and help to fill in the cracks of history in voices too rarely heard.
Washington describes the day the Union Army marched toward the Rappahannock River. Every white person in Fredericksburg, he says, was in the streets, attempting to flee, while all the black people were "out on the house top looking over the River at the Yankees for their glistening bayonets could Easily be seen."
Inside, Washington writes, "I could not begin to Express my new born hopes, for I felt already like I was certain of My freedom Now." When he crossed the river and met the Union soldiers, one asked him where his master was. In the Rebel Navy, he replied.
"Well you don't belong to anybody then," the soldiers told him, explaining that two days earlier the District of Columbia had freed its slaves. "I did not know what to say for I was dumb with joy and could only thank God and laugh."
Both men's narratives depict the ugliness of slavery, but Turnage's experience was far harsher than Washington's. Washington recalls, as a small boy, waiting all day for the commands of an elderly white woman, even as he could see other children playing outdoors.
But Turnage was whipped and beaten in the fields. He was also, for a time, forced to work in an auction house, where he saw his fellow slaves flogged, chained, cataloged, and sold. Yet he begins his account: "I do not mean to speak disparagingly of those who sold me, nor of those who bought me. Though I seen a hard time, it had an attendency to make a man of me."
While nothing can match the power of the men's own words, Blight's commentary does much to round out the portrait of the slave and former-slave experience. He tells of the "contraband colonies" that for many former slaves were the first stop after escape. He notes the eagerness of even the best-treated slaves to find their freedom, quoting a Georgia slave owner who confessed, "Those we loved best, and who loved us best – as we thought, were the first to leave us."
He also describes the netherworld of urban blight through which many former slaves – Turnage included – struggled to make their way after emancipation.
Washington offers a heartbreaking account of being separated from his mother as a boy. The night before, she came to his bed. "Her tears mingled with mine amid kisses and heart felt sorrow ... I would rather die" than leave her.
Both Turnage and Washington did eventually reunite with their mothers. Of Turnage's mother, Blight writes, "her story, like her son's so ordinary and yet so extraordinary, makes us wish we could know her even better."
Readers will agree – and yet will also be powerfully grateful for the fascinating bits and pieces that they are given here.
| 503
|
Wednesday, March 11, 2009
rothbard on the utility of quantitative easing
Economist Murray Rothbard ("America‟s Great Depression", Ludwig von Mises Institute 2000) made extensive analysis of why the inflationary policies of the US Federal Reserve failed during the Great Depression. Those policies proved to be counterproductive:“American citizens lost confidence in the banks and demanded cash.. for their deposits.. while foreigners lost confidence in the dollar and demanded gold.. The more that.. the Fed tried to inflate, the more worried the.. public became about the dollar, the more gold flowed out of the banks, and the more deposits were redeemed for cash.. The Fed purchase of government securities was a purely artificial attempt to dope the inflation horse..”
this has to be a core concern of every briton at the moment, with the bank of england having entered the gilt market to monetize government debt. the grand experiment intends to goose economic activity at the expense of the real value of the already-slashed pound, with some sincerely icelandic tail risk.
in normal conditions, a devalued currency could be expected to assist exporters as well. as mentioned by the financial times, there is considerable doubt as to the efficacy of that pathway amid a global demand collapse. so the primary intention is to increase the currency in circulation in the hopes that more on hand will mean more spending and (critically) more borrowing.
this clearly did not work well in the 1930s to the extent it was undertaken as banks chose to hoard cash (much as today). of course, the united states also experienced a series of devastating bank runs which reinforced that tendency. today, with the FDIC, bank runs are a much smaller problem. aren't they?
the hard answer is that "it depends". much of the catastrophe of 2008 can be seen as a run on the shadow banking system -- that is, the uninsured and off-balance-sheet extensions of the financial system. the contraction of that massive credit generator is analogous in its effect to a systemic bank run that has resulted in failed "banks", boggling capital losses and widespread (and justifiable) paranoia.
that run has at times extended to more visible elements of the banking system, forcing insurance to be put into place on an ad hoc basis -- such as the commercial paper and money market fund backstops put in place following the collapse of lehman brothers, which were hoped to be able to forstall a systemic run on those markets which was actually underway.
beyond that there is the question of the fate of crossborder deposits -- international depositors in american banks have heretofore been well treated by the FDIC, but their accounts are not legally protected. these accounts are a massive part of the deposits in money-center banks such as C or JPM. crossborder runs have already become a reality in some parts of the world during this crisis.
and there is finally the question of the FDIC itself. michael panzner forecast in his sometimes-eerily-prescient "financial armageddon" (and has repeated elsewhere) that eventually deposit insurance (along with many overpromised government-backed obligations) will likely have to be diluted or perhaps even eliminated in the crisis now underway as the fiscal position of government may not allow it to backstop all failed banks as the debt unwind progresses and the losses being assumed by the government through the FDIC and other pathways become overwhelming. the mere anticipation of that development could reintroduce the bank run to american society after a long absence, which is probably why sheila bair recently approached congress for a $500bn line to the treasury, a move many see as preparatory of the resolution of a money-center bank. this comes on top of a draconian assessment levied across all banks, leaving many smaller community bankers incensed at what some term "confiscating the capital of the industry" on the part of the government in service of the majors.
and this is not to mention hoarding in anticipation of runs on other elements of the financial system -- most notably of late the insurance sector, which rolfe winkler delves into today.
beyond simply engendering some healthy skepticism of the value of government insurance in a real crisis, the two-pronged point here is
- that quantitative easing may not work to promote economic activity through increased lending in part because the value of assurances against losses are, even and perpahs particularly where comprehensive, questionable;
- that quantitative easing, or rather monetization, may be the only resort to which government could really appeal in the case of multiple large-scale bank resolutions.
it is impossible to know how this would work out in detail -- but should we witness a real spate of dollar weakness in forex markets as things go along, government may be faced with some very stark choices about what it can really afford to backstop.
| 875
|
Seven miles to the south west of Sheffield there's a piece of wild moorland surrounded by roads. It's called Big Moor. I estimate that if you circumnavigated it on foot by the boundary roads it would take you over seven hours. In spite of its wild appearance, Big Moor was once home to generations of neolithic people. They built at least three significant stone circles, several cairns, dwelling houses and field systems. This was in days when there were still wolves in the British Isles, long before the Romans introduced rabbits and horse chestnut trees and in a time when there were no horses or wheat or potatoes. Life must have been very hard.
By the middle ages, the stone circle builders had gone but the country had very few roads and most travel was still on foot. Trading tracks began to criss-cross the land and on Big Moor several stone posts or stoops were erected to guide travellers and traders across the difficult terrain with its bogs and dips, rivulets and hardy moorland vegetation. Mick Jagger's surname harks back to the jaggers of yesteryear - hardy men who carried heavy packs - they were pedlars and hawkers. The verb "to jag" meant to pack or to carry. When good roads were built and roadworthy coaches developed, the jaggers disappeared but their paths and their guideposts were left behind.
In the nineteenth century, as the importance of clean, reliable water supplies became more obvious, many reservoirs were constructed across England and on Big Moor you can see the location of Barbrook Reservoir which is now disused.
|Barbrook II stone circle|
|Highland cattle avoiding the hot British sunshine|
|One of the remote guideposts on Big Moor|
|Another one - taller, more weathered|
|Perhaps an old cross in the middle of the moor - not marked on maps|
| 751
|
Engineer your first step to a four-year degree at Edmonds Community College
Engineers Make a Difference
Engineering is the art of applying scientific and mathematical principles, experience, judgment, and common sense to make things that benefit people. Engineers solve problems. They improve and develop products to meet consumer and societal needs. They find ways for existing products to work better, last longer, operate more safely and cost less.
They also look for innovative solutions to global problems. Engineers design bridges and important medical equipment as well as processes for cleaning up toxic spills and systems for mass transit.
It's a Fact
Engineering professions requiring a four-year degree are among the highest paying jobs in Washington State according to the office of Labor Market and Economic Analysis. Source: www.workforceexplorer.com
| 579
|
Tamoxifen use 'should double' to stop breast cancer returning
A decade's treatment with tamoxifen could stop the return of breast cancer for many more women than the current recommendation for five years, say researchers.
University of Oxford scientists say the doubling cuts both the risk of the cancer returning and deaths from the disease.
Writing in The Lancet, they say this outweighs the risk of side effects.
Cancer Research UK said the study added important clarity.
Tamoxifen tablets have long been part of the daily routine for women in the years after remission from 'ER-positive' breast cancer.
ER-positive tumour growth is accelerated by the female sex hormone oestrogen, and tamoxifen blocks the hormone.
An international group of scientists, led by Dr Christina Davies from the University of Oxford, looked at the effect of sustained tamoxifen on the likelihood that cancer would return.
They looked at 6,847 women with the ER-positive version of the disease, half of whom were given tamoxifen for the standard five year period, and half who continued with the drug until the 10 year mark.
The results found that cancer returned in a quarter of the women given five years of tamoxifen. That fell to 21.4% in those given ten years of the drug.
Deaths from breast cancer was also significantly reduced, from 15% of those on the five-year programme, to just over 12% of those given 10 years.
End Quote Dr Caitlin Palframan Breakthrough Breast Cancer
This trial is great news for women with this type of breast cancer as it suggests that taking the commonly used drug, tamoxifen, for longer could save more lives with minimal additional side effects”
While the changes are only a few percent, with more than 40,000 women diagnosed with breast cancer in the UK alone every year, a large proportion of whom have 'ER-positive' cancer, it could mean a significant decrease in the number whose cancer returns.
The report authors said: "Good evidence now exists that 10 years of tamoxifen in ER-positive breast cancer produces substantial reductions in rates of recurrence and in breast cancer mortality not only during the first decade, while treatment continues, but also during the second decade, long after it has ended."
The drug does cause side-effects in some patients, including, rarely, cancer of the lining of the womb, but the scientists said that the increased chance of preventing breast cancer returning far outweighed the risk.
Dr Caitlin Palframan, head of policy at Breakthrough Breast Cancer, said: "ER-positive breast cancer is the most common type of breast cancer and so improving treatment and getting the most out of the effective treatments we already have is key.
"This trial is great news for women with this type of breast cancer as it suggests that taking the commonly used drug, tamoxifen, for longer could save more lives with minimal additional side effects."
Other scientists say that 10 years could now become the standard for treatment.
Prof Trevor Powles, from Cancer Centre London, said that this, and other studies due to report next year, "should herald a change in practice".
The National Institute for Health and Clinical Excellence said: "We do not generally make recommendations relating to how long a patient should receive tamoxifen; that decision should be made by a doctor according to their professional judgement and in discussion with their patient, not least because there are continuing trials in this area.
"We review our guidance regularly to ensure it remains based on the best, most up-to-date evidence available and, when we come to update our guidelines on the diagnosis and management of breast cancer, we will take all new evidence into consideration."
Martin Ledwick, from Cancer Research UK which funded the study, said: "This important study adds further clarity to the question about the length of time women should take tamoxifen.
"Although treatment for hormone receptor positive breast cancer has become more complex in recent years with some women receiving aromatese inhibitors, these results will help in deciding the length of treatment for women who are prescribed tamoxifen alone."
| 874
|
Charles Heywood kept his cannon firing even as his ship sank.
Born in Waterville on Oct. 3, 1839, Heywood was commissioned a second lieutenant in the United States Marine Corps at New York City on April 5, 1858. Thirty-five months later he reported aboard the USS Cumberland, a 24-gun sloop of war then assigned to the Gosport Navy Yard in Portsmouth, Va.
Fearing the shipyard’s capture by Confederate militia, senior Navy officers ordered Gosport abandoned on April 20, 1861. After mining key installations with barrels of gunpowder, Heywood and his Marines lit the fuses and hastened aboard the USS Cumberland, already slipping seaward on the ebbing Elizabeth River.
The resulting explosions and fires caused significant damage, but Confederate engineers quickly repaired crucial machinery. They also raised the burned and scuttled USS Merrimac. Upon its hull arose an ironclad, the CSS Virginia. Measuring 263 feet in length and weighing 3,200 tons, the steam-engine ironclad mounted 10 massive cannons. The skipper was Capt. Franklin Buchanan.
He intended to sail into Hampton Roads and sink the Yankee warships stationed there. By the time Buchanan ordered his ship to sail on March 8, 1862, Heywood was a captain commanding the USS Cumberland’s Marine contingent.
Brilliant late-winter sunshine and warm air engulfed Hampton Roads as the CSS Virginia stood downriver that Saturday. Buchanan had already selected his first target. “I am going to ram the Cumberland,” which was equipped with “the new rifled guns, the only ones in their whole fleet we have cause to fear,” he told Chief Engineer H. Ashton Ramsay. “The moment we are out in the Roads, I’m going to make right for her and ram her.”
That morning, the USS Cumberland lay moored some 300 yards off Newport News on the north shore of Hampton Roads. Aboard the Cumberland, Cmdr. William Radford had gone ashore, leaving Lt. George Morris, the executive officer, in command.
At noon, sailors on watch aboard the Cumberland “discovered three vessels under steam, standing down the Elizabeth River toward Sewell’s Point,” Morris wrote on March 9. One ship belched black smoke; Morris, Heywood, and other USS Cumberland officers could see the smoke without using their spyglasses.
Buchanan’s underpowered ironclad steamed ponderously into Hampton Roads. After the CSS Virginia cleared Sewell’s Point about 1:30 p.m., Buchanan steered west to attack the USS Cumberland.
Aboard that ship, crewmen had already “double breeched the guns on the main deck, and cleared [the] ship for action,” Morris reported. He ordered the sloop pivoted on her anchor so that her starboard broadside would face the ironclad.
Assigned to his ship’s after gun division, Heywood stood with his gun crews and watched as the Virginia approached. At 2 p.m. aboard the ironclad, Lt. Charles Simms fired the bow pivot gun, a 7-inch Brooke rifle.
According to William C. Davis writing in “Duel Between the First Ironclads,” Simms’ cannon ball “screamed across the Roads and hit” the USS Cumberland “squarely, passing through the starboard-quarter rail” and hurling wood splinters into nearby Marines. After a Cumberland gun crew fired their ship’s forward 10-inch pivot gun and missed, Simms’ second shot exploded amidst that gun crew and killed all but two men.
“Our firing became at once very rapid from the few guns we could bring to bear” as the ironclad “approached slowly,” recalled Master Moses Stuyvesant, who commanded the sloop’s after gun division. Simms’ pivot gun pounded the Cumberland; as shell fragments and wood splinters struck down sailors and Marines alike, Heywood’s crews fired their cannons through the dense smoke.
At approximately 2:30 p.m., Cumberland pilot A.B. Smith thought the approaching ironclad resembled “a huge half-submerged crocodile” with its “iron ram projecting, straight forward.”
Steaming at 6 knots, the Virginia “stood on and struck us under the starboard fore channels” near the Cumberland’s bow, Morris reported. “She delivered her fire at the same time; the destruction was great. We returned the fire with solid shot with alacrity.”
The ironclad’s ram opened a hole about 7 feet across. Flooding with seawater, the USS Cumberland listed to starboard — and the ship’s weight bore the Virginia downward.
The ironclad was already reversing its engines as the tide pivoted the trapped Virginia alongside the sinking Cumberland. Then the ram broke off, freeing the ironclad from a potential watery grave. Buchanan backed his ship until it lay parallel to the Cumberland and only 20 feet away.
Cannons flashed and boomed for some 30 minutes as Heywood and his comrades heroically fought the Virginia amidst “a scene of carnage and destruction never to be recalled without horror,” Stuyvesant remembered. Blood and gore splattered across the main deck as Heywood shouted at sailors and Marines to drag wounded comrades to the ship’s port side.
Now the sea lapped at the Cumberland’s main deck as the bowsprit disappeared into Hampton Roads. Gun crews kept firing until their cannons submerged. Stationed near the stern, Heywood worked his guns even as another cannon broke loose and lurched across the deck to crush a sailor.
“At 3:35 [p.m.] the water had risen to the main hatchway, and the ship canted to port, and we delivered a parting fire, each man trying to save himself by jumping overboard,” Morris recalled. Acknowledging that severely wounded men taken below decks could only be left to drown, he credited specific officers for their coolness under fire.
“I can only say in conclusion that all did their duty and we sunk with the American flag at the [mast] peak,” Morris closed his report.
Even as the Cumberland’s deck sharply canted, Heywood remained with the aft guns. “The water began to swash over the upper deck, and still every unencumbered gun was hurling defiance at the foe,” wrote John S.C. Abbott in “The History of the Civil War in America, Vol. 1,” published in 1863.
“The ship careened upon one side. The last gunner [Heywood], knee-deep in water, pulled the trigger of the last gun,” Abbott wrote. Then “the majestic frigate, with all her dead and all her wounded, sank like lead.”
And Charles Heywood went overboard into Marine Corps lore.
His heroism went noticed. “I omitted to mention to you the gallant conduct of Lieutenant Charles Heywood … whose bravery upon the occasion of the fight with the Merrimack won my highest applause,” Morris wrote Navy Secretary Gideon Welles on April 12.
Heywood received a brevet promotion to major after the battle. Of the 46 Marines aboard the USS Cumberland, 14 died. Another 107 sailors were killed aboard the sloop on March 8.
The next day, the Monitor fought the Virginia in Hampton Roads in history’s first ironclad-to-ironclad sea battle. Most history books still call the Virginia the “Merrimac.”
Heywood would fight during another famous Civil War battle; he was commanding two gun crews aboard the USS Hartford at Mobile Bay, Ala. as Adm. David Farragut roared, “Damn the torpedoes! Full speed ahead!” on Aug. 5, 1864. Within an hour, Heywood and Farragut traded shot and shell with another Confederate ironclad, the CSS Tennessee — commanded by Adm. Franklin Buchanan.
This time Buchanan lost and Heywood won.
Named the ninth Marine Corps commandant on Jan. 30, 1891, Charles Heywood instituted important changes to the corps’ mission and strength. Promoted to major general a year before his 1903 retirement, he died at Washington, D.C., on Feb. 26, 1915. His wife, Caroline Bacon, outlived him by 12 years.
Heywood lies buried in Section 2, Lot 1115 at Arlington National Cemetery.
Brian Swartz is the BDN special sections editor. An avid Civil War buff, he has extensively explored and photographed Civil War battlefields throughout the South. Swartz may be reached at email@example.com or visit his blog at http://maineatwar.bangordailynews.com.
| 162
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.