text
stringlengths
301
138k
- The Endless Decline of the English Language - Etymological Fallacy - "The False Refinements in Our Style," by Jonathan Swift - Language Change - Language Maven - Linguistic Complaint - "Literature and the Schoolma'm," by H.L. Mencken - Purism and Prescriptive Grammar: DON'T! - Skunked Term - Triumph of the Barbarisms - Verbal Hygiene - What is a SNOOT? Examples and Observations: - "Like other tabooing practices, language purism seeks to constrain the linguistic behaviour of individuals by identifying certain elements in a language as 'bad.' Typically, these are words and word usage that are believed to threaten the identity of the culture in question--what 18th-century grammarians referred to as the 'genius' of the language. Authenticity has two faces: one is the struggle to arrest linguistic change and to protect it from foreign influences. But, as Deborah Cameron claims, the prescriptive endeavours of speakers are more complex and diverse than this. She prefers the expression verbal hygiene over 'prescription' or 'purism' for exactly this reason. According to Cameron, a sense of linguistic values makes verbal hygiene part of every speaker's linguistic competence, as basic to language as vowels and consonants." (Keith Allan and Kate Burridge, Forbidden Words: Taboo and the Censoring of Language. Cambridge Univ. Press, 2006) - "I am of this opinion that our own tung shold be written cleane and pure, unmixt and unmangeled with borowing of other tunges, wherein if we take not heed by tiim, ever borowing and never paying, she shall be fain to keep her house as bankrupt." (John Cheke, Regius Professor of Greek at Cambridge University, in a letter to Thomas Hoby, 1561) - "A certain Captain Hamilton in 1833 demonstrates the invective the British directed at the language used in America. He claims that his denunciation is 'the natural feeling of an Englishman at finding the language of Shakespeare and Milton thus gratuitously degraded. Unless the present progress of change be arrested by an increase of taste and judgment in the more educated classes, there can be no doubt that, in another century, the dialect of the Americans will become utterly unintelligible to an English man . . ..' Hamilton's vituperation exemplifies a purist view of language, which allows only one fixed, immutable, correct version [and] which sees difference and change as degradation." (Heidi Preschler, "Language and Dialect," in Encyclopedia of American Literature, ed. by Steven Serafin. Continuum, 1999) - "Despite the exacerbated protests of the upholders of authority and tradition, a living language makes new words as these may be needed; it bestows novel meanings upon old words; it borrows words from foreign tongues; it modifies its usages to gain directness and to achieve speed. Often these novelties are abhorrent; yet they may win acceptance if they approve themselves to the majority. . . . "To 'fix' a living language finally is an idle dream, and if it could be brought about it would be a dire calamity." (Brander Matthews, "What Is Pure English?" 1921)
Normal Cell Division A cell grows a bit larger and then divides into two cells To understand what breast cancer is, it is important to first understand how the body's cells normally work. The body is made up of tiny cells, such as skin cells, muscle cells, heart cells, nerve cells, and bone cells. When a baby grows, the number of cells increases very quickly. A cell becomes a bit larger, then divides into 2 "daughter" cells (see Figure 1). Each of these cells divides again, and so on. Once a child grows to adulthood, the cell division process still takes place, but more slowly - for example, dead skin flakes off as the skin constantly renews itself. Normally, this renewal process happens in a very ordered, systematic way. Each cell carries genetic "instructions" that control how fast the cell should divide and grow and when the cell should die. A balance between cells growing and dying keeps our bodies functioning normally. Sometimes a cell starts to grow without regard for the normal balance between cell growth and death, and a small, harmless lump of cells will form. These harmless growths are referred to as benign. A benign growth can occur in any part of the body, including the breast, skin, or intestine. In other cases a cell may grow and divide with complete disregard for the needs and limitations of the body. These cells have the potential to grow into large masses or to spread to other areas of the body. Cells that have this aggressive behaviour are called malignant. More commonly, a mass of such cells is called a cancer. A cancer that continues to grow can eventually overwhelm and destroy the part of the body or particular organ where it is located - in this case, the breast tissue. When clumps of these cells spread to other parts of the body they are termed metastases. Cancer cells also have the ability to stimulate the development of blood vessels to increase their own blood supply and enhance their growth. However, sometimes a cancer's wild growth rate backfires. The cancer may grow so rapidly that it can't get enough oxygen and nutrients from the blood vessels. When this happens a part of the cancer may suddenly die. This death of a group of cells within a cancer is known as necrosis. How Breast cancer may spread A cancer grows in the breast (A) and spreads by invitation to surrounding tissues of the breast (B) or through the lymph system to lymph nodes (C ) or through the blood stream to distant organs such as the bones, lungs, liver or brain (D) Most normal cells stay where they belong and don't spread to other parts of the body. Breast cancer cells, like other cancer cells, disregard this principle and may spread through the body (metastasize) in several ways (see Figure 2). They can directly invade and destroy the organ of origin, or spread through the lymph system or bloodstream to distant organs such as the lungs, liver, and bones. When a cancer spreads, it retains the properties of the original cancer. In other words, a breast cancer that spreads to the bones is still a breast cancer. Under the microscope it looks different from a cancer that started in the bones, and it responds to treatment like a breast cancer, not a bone cancer. The original cancer in the breast is called the primary cancer. A cancer that has spread to another site is called a secondary or metastatic cancer. The immune system consists of a group of cells called white blood cells, specialized to recognize and destroy "foreign" material in the body such as bacteria, viruses, and unfamiliar or abnormal cells. Cancer cells somehow manage to slip through this detection system without triggering the immune system to start fighting, whether at the primary cancer site, in the blood vessels, or at the site of distant spread. Breast cancer does not develop overnight Gradually the cells become more abnormal-looking or atypical. Eventually, the cells are recognized as being sufficiently abnormal to be called cancer cells. At first they are within the milk ducts (in situ cancer), and later become invasive breast cancer cells It can take years of cells dividing before a normal cell becomes a cancerous cell. The cell first undergoes very small changes in which it becomes slightly abnormal or atypical as seen under a microscope. It may also begin to divide, grow more quickly, and accumulate in excessive numbers (hyperplasia). Then, over the years the cells continue to change, become more abnormal-looking, and finally cancerous (see Figure 3). Initially the cancer cells are confined within the milk ducts (in situ cancers), but with time the cells develop the ability to invade surrounding tissues and travel into the blood and lymph system. This is called an invasive cancer. Unfortunately, it is not possible to detect one or a few abnormal or cancer cells. Current technology is only capable of detecting a small lump or mass of cancer cells that has already started to grow. By the time a cancer can be detected as a lump, it may have been growing slowly within the breast for anywhere from 2 to 8 years. Ivo Olivotto, MD, Karen Gelmon, MD, Urve Kuusk, MD Excerpt from Intelligent Patient Guide to Breast Cancer in association with the MediResource Clinical Team Did you find what you were looking for on our website? Please let us know.
This is an adaptation from an invited talk given at the tenth Annual International Workshop on "Micro Electro Mechanical Systems" in Nagoya, Japan, Proceedings IEEE Catalog Number 97CH36021, pages 9 to 13, The field of micromechanics (also called MEMS or Micro System Technology or Micro Machines) describes a whole new realm of human endeavor. This paper discusses: Why has human interest in the small taken so long to develop? What are the prospects? What can we do to ensure the health of this emerging field? Do small things have a Perhaps things normally start small, and grow. Man's habitats have grown from houses, to buildings, to skyscrapers. Our ability to travel has increased from a few miles on foot, to horses, to trains, and now we can encircle the world in a few days. Individually we work to make large accomplishments in hopes of enormous success. We are enthralled with the big and significant and substantial. The insignificant, insubstantial, and minuscule is usually beneath our concern. A dozen years ago, I was trying to persuade a machinist to build a very small structure. He listened patiently for awhile, and then said 'Why do you want something small, a toy? I can make you something that is big and good.' In his mind, most people's mind, small things were cheap, and no more than a toy. When H. A. Rowland (1848 to 1901, professor of physics at the Johns Hopkins University, Baltimore) went to make very small and accurate grooves for diffraction gratings, he used large machines, and buried them in even larger vaults for thermal stability. Ten years ago, an eminent colleague at Bell Laboratories looked me in the eye, and said 'Your micro things will never amount to anything. Large objects will always do a better job at a lower cost.' This was very strongly the feeling at this time. Even Feynman responded with good natured jesting to critics of small machines. In his famous talk, There's Plenty of Room at the Bottom, given at the American Physical Society meeting in 1959 he says "What would be the utility of such machines? Who knows? Of course, a small automobile would only be useful for the mites to drive around in, and I suppose our Christian interests don't go that far." And in his 1983 talk, Infinitesimal Machinery, at the Jet Propulsion Laboratory he says "I also talked in the 1960 lecture about small machinery, and was able to suggest no particular use for the small machines. You will see there has been no progress in that respect." And yet. There is something special about this field of small mechanical systems. Most advances represent a specific technology. The Scanning Tunneling Microscope for example, gives us the ability to detect and perhaps manipulate atoms. High temperature superconductors hold the promise of efficient power transmission, and novel electronic circuits. The diesel engine gives us a source of mechanical power. Each of these is an important advance of a single thing. The field we are contemplating here today is vast beyond our normal concerns. It is the science and engineering and development and commercialization of a whole new realm of human enterprise. I defy you to think of a large scale, macro, discipline in science or engineering that does not have a small scale, micro, equivalent. Your challenge, should you decide to accept it, is the imaging of the macro into the micro. You are the pioneers. How you behave, how you interact with your colleagues, what areas you pursue, your understanding of how to develop a new science, the clarity of your vision, will define how this field explodes forth. One thing giving me confidence in our future is the broad range of tools and techniques available. We are not dependent upon a single tool solving a problem, there are many ways to make micro devices. The designer has a host of fabrication techniques, actuators, and sensors at his command; and he should be free to choose the best for his purposes. To define this field as a single technology is limiting it. It is the richness of this field, and hopefully the collaboration between its practitioners, that will bear the sweetest fruit. The original ingenious and intelligent systems were mechanical, things such as clocks that chimed and displayed dancing figures on the hour. Electronics is now doing a superb job of providing this intelligence. Complex calculations and decisions have now become inexpensive. It is now the mechanical devices needed to interface electronics to the world that are expensive. Fortunately micromechanical devices integrate well with electronics: one providing the intelligence and one providing the hands. Electronics has lead much of the development of micromechanical devices by providing many of the tools and techniques, making the rapid advances possible. This partnership is to great advantage. Yet how did things insignificant in size gain a purpose? Perhaps Johann Gutenberg gave an indication of the usefulness of small mechanical devices. Gutenberg means good mountain, and indeed, in 1456 he set in motion a mountain of small mechanical devices (individual movable type) for the good of mankind. For years the watch makers art has represented the limits of our micro excursion. And the practitioners of the watch making art have succeeded admirably. For example, the motor on a wrist watch has high efficiency, runs for years (even after being dropped), and cost less than a cup of coffee. Yet, when I was talking with a gentleman who had designed many of the watches we wear, he said 'I have spent my life trying to make smaller mechanisms, and when you show me something really smaller, I do not know what to do with it.' This is a common response to motors the diameter of a human hair. Much that has happened in micromechanics was presaged by Feynman's remarkable talk, There's Plenty of Room at the Bottom. Later Feynman gave a version of this talk to my college, and he is responsible for igniting my interest in the possibilities for the small. If you have not, you should read the transcript of his presentation. Yet making things too small to manipulate with human hands was one serious barrier to the exploration of the micro. A number of techniques now make micro devices inexpensive to manufacture. In 1967 the paper, The Resonant Gate Transistor, described a structural-sacrificial fabrication technique. In these initial experiments gold was used as the structural material and photoresist was the sacrificial material. This is an extremely powerful technique that allows complete micro structures to be built without having to assemble components, a great advantage when dealing with components too small to see with the eye, or manipulate with the hand. The 1983 paper, Polycrystalline Silicon Micromechanical Beams, extended this work to the polysilicon-silicon dioxide, structural-sacrificial, material we now normally describe as surface micromachining. A second powerful fabrication technique for making small things was discussed in the 1978 paper Anisotropic Etching of Silicon. A year later the paper Fabrication of Hemispherical Structures Using Semiconductor Technology for Use in Thermonuclear Fusion Research describes how to make micro spheres filled with deuterium-tritium using isotropic etching. These anisotropic and isotropic etching techniques in silicon and related materials are now commonly referred to as bulk micromachining. The 1982 paper Production of Separation Nozzle Systems for Uranium Enrichment by a Combination of X-Ray Lithography and Galvano-plastics also presents an early application for microfabrication techniques in the nuclear power generation industry. This technique is now called LIGA (in German Lithographie, Galvanoformung, Abformung), and makes plastic and metal parts with spectacular accuracy. The above techniques demonstrate novel approaches to the manufacturing of micro parts: v Surface micromachining makes parts without assembly. It is as if an automobile was made by putting down alternate layers of steel and flour, and the last fabrication step was washing away the flour, leaving a completely assembled automobile, engine and all. v In most manufacturing, the metric (measurement system) is defined by the tools used. In contrast, bulk micromachining relies upon the metric in the material's crystalline structure to define the part. The part helps define itself. v LIGA, bulk micromachining, and surface micromachining use photons to shape the micro devices. Other micromachining techniques are clever extensions of macro manufacturing techniques. A nice description of these techniques is given in the paper Micro Machining by Machine Tools. Extensions of single point diamond machining can make micro parts with spectacular accuracy of less than 0.01 microns (10E-8 meters or 100 Angstroms ! ) Diamond machining works well on most materials, except a few materials like steel. Extensions of electro discharge machining, EDM, allows minute turbines, power generators, and orifices to be manufactured. Sacrificial techniques, similar to surface micromachining, enable complex interior and exterior shapes. Many other methods of making micro things are possible. One example is Stereo Lithography. A scanning laser beam writes the micro parts directly onto ultra violet curable epoxy, as discussed in the papers Real Three Dimensional Micro Fabrication using Stereo Lithography and Metal Molding and Photoforming Applied to Fine Machining. The race to more clever ways to machine micro parts has just begun. The earlier disdain for the small and insignificant is gone. Now I sense a growing excitement about the micro. My fears are gone that the micro field would grow on 'isn't that neat' and then die when no purpose was found. Enough people now recognize the importance of micro science and engineering and product development to ensure the field. Things insignificant in size do have a grand purpose. Yet, how to proceed? First, we should work as a community, and profit from other's insights. This workshop (MEMS '97) is an excellent place to bring together disparate disciplines and ideas. The hosting of this workshop by many different countries in the different regions of the world is strongly encouraged. Also having only one session at a time helps bring people with different ideas into the same room. And the support of students and their presentations help bring new ideas and people into the field. Publishing excellent articles in archival journals is critical to the orderly development of the field. Well written articles filled with new material help the authors and the field. Poorly written material slapped together with one or two incrementally new results has the unfortunate effect of risking the reputations of the authors, and clogging the literature. Many excellent research programs and developments have resulted from the collaborations between people with diverse backgrounds. I especially encourage people new to this field to form collaborations with established groups. The behavior of micro systems is substantially different from our normal macro experiences. Second, a science grows by the unfettered competition of ideas, not people. People should be free to suggest and work on the new. Unfortunately it is easy to disdain the unfamiliar. Scholarship requires an unbiased and careful evaluation of one's own and others work, an understanding of the previous literature, and honesty. Journal articles must be fairly refereed, with a special tolerance given to new approaches. Science is driven by people's excitement about learning. Third, we should build infrastructures that facilitate the growth of the field. Much work is needed developing the basic science and metrology of this new field. As the field prospers and grows, more resources should be devoted to a fundamental understanding. Multi-user runs, where many projects share a fabrication run, help give access and capabilities to groups far removed from the fabricators. Efficient use of journals, books, and the world wide web for the free exchange of information, ideas, experimental results, and computer modeling tools speeds the development of the field. Fourth, we should not forget our purpose. As a child, I read about the great scientists. Men and women who have structured our understanding of the universe by their discoveries. I wished I could have been with these great scientists, and shared in their adventure. We are fortunate. Though our work from day to day may seem insignificant in size, together our work is grand in purpose. I wish you well. R. Feynman, "There's Plenty of Room at the Bottom," Caltech's Engineering & Science magazine, February, 1960. (Reprinted in Micromechanics and Mems: Classic and Seminal Papers to 1990, Edited by W. Trimmer, the IEEE Press PC4390-QCL, ISBN 0-7803-1085-3, January 1997, page 3.) R. Feynman, "Infinitesimal Machinery," Journal of Microelectromechanical Systems, Volume 2, Number 1, pages 4 to 14, March, 1993. (Reprinted in Micromechanics and MEMS, page 10.) H. C. Nathanson, W. E. Newell, R. A. Wickstrom, and J. R. Davis, Jr., "The Resonant Gate Transistor," IEEE Transactions on Electron Devices, March, 1967. (Reprinted in Micromechanics and MEMS, page 21.) R. T. Howe and R. S. Muller, "Polycrystalline Silicon Micromechanical Beams," Journal of the Electrochemical Society: Solid State Science and Technology, June, 1983. (Reprinted in Micromechanics and MEMS, page 505.) K. E. Beam, "Anisotropic Etching of Silicon," IEEE Transactions on Electron Devices, October, 1978. (Reprinted in Micromechanics and MEMS, page 50.) K. D. Wise, T. N. Jackson, N. A. Masnari, M. B. Robinson, D. E. Solomon, G. H. Wuttke, and W. B. Rensel, "Fabrication of Hemispherical Structures Using Semiconductor Technology for Use in Thermonuclear Fusion Research," Journal of Vacuum Science Technology, May/June, 1979. (Reprinted in Micromechanics and MEMS, page 551.) E. W. Becker, H. Betz, W. Ehrfeld, W. Glashauser, A. Heuberger, H. J. Michel, D. Munchmeyer, S. Pongratz, and R.v. Siemens, "Production of Separation Nozzle Systems for Uranium Enrichment by a Combination of X-Ray Lithography and Galvano-plastics," Naturwissenschaften, 69, (1982), pages 520 to 523. T. Higuchi and Y. Yamagata, "Micro Machining by Machine Tools," MEMS 1993 Workshop on Micro Electro Mechanical Systems, Fort lauderdale, FL, USA, February, 1993, pages 1 to 6. K. Ikuta and K. Hirowatari, "Real Three Dimensional Micro Fabrication using Stereo Lithography and Metal Molding," MEMS 1993 Workshop on Micro Electro Mechanical Systems, Fort lauderdale, FL, USA, February, 1993, page 42. T. Takagi and N. Nakajima, "Photoforming Applied to Fine Machining," MEMS 1993 Workshop on Micro Electro Mechanical Systems, Fort lauderdale, FL, USA, February, 1993, page 173. Keywords: micromechanics field, history of micromachines, prospects for MEMS, opportunities for microelectromechanical systems, development of microsystem technology
By Serena Makofsky People who work in food preparation know that culinary skills must be practiced and refined, making it more of an art form than a mere hobby or a task. A master chef endeavors to not only stimulate the palate, but also the senses of smell and sight. There are many compelling facts and bits of trivia regarding the study and practice of the culinary arts. Learning the culinary arts was once the work of apprenticeship, with daughters learning recipes from mothers, and traditional recipes passed on via oral history. It was not until the year 1800 that the first academic program in the culinary arts opened for business. The Boston Cooking School was popular, drawing an international student body. One of its most famous students was Fannie Merrit Farmer, who published the world's first cookbook in 1896. "The Boston Cooking School Cookbook" is still a reference work for students of culinary arts today. The first televised cooking show was presided over by James Beard, airing in 1946. His illustrious work earned him the moniker "dean of American cookery," bestowed upon him by "The New York Times." People who work in the culinary arts often aspire to positions such as executive chef or sous chef. But there are plenty of other positions in the field, such as the saucier, who oversees sauces and gravies with a chemist-like precision; the garde manager, who may be responsible for all cold foods and often can integrate leftover ingredients into new dishes; pastry chefs, a major responsibility that often entails overseeing a whole pastry staff; and a sommelier, who is paid to professionally wine taste and share expertise with patrons. Less chef-oriented positions include dining room service, food and beverage inventory specialists, restaurant consultants, salespeople, food writers, food critics, food stylists, food photographers and research and development. Others may opt to train other chefs and enter into teaching. Finally, there are those who are entrepreneurs and decide to open their own restaurant or catering service. The culinary arts are an object of public fascination. Celebrity chefs and foodies such as Rachel Ray and Rick Bayless tour the world's cuisines, while the iron chefs demonstrate their artistry in a live, timed competition. Movies and novels such as "Like Water for Chocolate," "Heartburn," "The Wedding Feast," "Cocolat" and "Babette's Feast" reflect the seductive power of beautifully prepared food. The culinary arts even has its own cable television channel, The Food Network. A student of culinary arts does not just enter the kitchen chopping. Typical core coursework includes practicing the fundamentals of cooking (yes, chopping), researching theory, learning sanitary practices, preparing simple cold dishes and composing soups and sauces before ever learning how to make an entree. Advanced skills are required for classes such as baking, pastry creation, nutrition, menu planning, international cuisine, catering and the technology of preserving products. This last course inspires some amount of controversy, because both consumers and creators of food have divergent opinions on the incorporation of pre-prepared foods. Two of the more entertaining courses are the art of working with chocolate and sugar and wine appreciation. An examination of culinary arts would not be complete without mentioning its patron saint, Julia Child, who first made cooking history with her 1961 book "Mastering the Art of French Cooking." Her radio show and, later, her television program, "The French Chef" were enormously popular and introduced the public to revolutionary approaches in cookery. She had strong opinions about her craft, commenting "Non-cooks think it's silly to invest two hours' work in two minutes' enjoyment; but if cooking is evanescent, so is the ballet."
Chicago’s Museum of Science and Industry houses 5.6 hectares — that’s more than seven soccer fields — of hands-on exhibition space. One of the museum’s permanent exhibitions is Science Storms, which teaches visitors basic physics and chemistry by recreating weather phenomena, like the avalanche pictured above, on a smaller scale. To illustrate the larger scale, the museum established Earth Revealed. Using computers and video projectors along with National Oceanic and Atmospheric Administration and NASA data sets, the exhibition generates a revolving globe that displays planet Earth as a living system in real time. Beth Boston, a spokeswoman for the museum, said that by providing exhibitions such as Science Storms and Earth Revealed, the museum “hopes to inspire children to achieve their full potential in the sciences.” Smart Home: Green + Wired is the museum’s newest exhibition. Through guided tours of the “green” home, it provides environmentally friendly ideas for living and decorating. The Museum of Science and Industry offers podcasts from the experts behind these exhibitions, in addition to more traditional educational experiences such as field trips, lectures and science demonstrations.
Swimmers Itch: 8 Questions & Controlling It In Lakes & Ponds What is Swimmers Itch? Swimmers Itch (also known as Cercarial dermatitis, Lake Itch and Duck Itch) is an itchy rash that is caused by a certain type of parasite (shistosomes) that live in freshwater snails and waterfowl that gather among lake weeds. As part of their life-cycle, these parasites are released on warm sunny days and migrate through the water. They then reattach to swimmers by burrowing into the skin. Humans are not good hosts so the parasites soon die while still under the skin, which may cause an allergic reaction. Who get Swimmers Itch? All age groups and both sexes can get swimmers itch. Children are most often infected due to their habit of swimming or wading in shallow water and playing on the beach, allowing lake water to evaporate. How does Swimmers Itch spread? An individual may get the infection by swimming or wading in infested water and then allowing water to evaporate off the skin rather than drying skin with a towel. Person-to-person spread does not occur. What are the symptoms of Swimmers Itch? Repeated exposure increases a person’s allergic sensitivity to the parasite and increases the likelihood of rash development. Symptoms may appear within 1-2 hours of exposure. An initial tingling sensation may be felt associated with penetration of the parasite into the skin. The irritated raised papule, which reaches its maximum size after about 24 hours, may continue itching for several days. Symptoms should disappear within a week. What is the treatment of Swimmers Itch? There is no necessary treatment. Swimmers Itch is uncomfortable but usually short-lived and will clear within a few days. Over-the-counter medication will normally help control itching. When can you get Swimmers Itch? First outbreaks usually occur in late May or early June and last up to one month. What can be done to reduce the chances of getting Swimmers Itch? Toweling off immediately after swimming or wading in infested water. Swim in water away from the shore. Avoid swimming in areas where the snails have accumulated. Don’t encourage birds to stay near swimming areas by feeding them. How can I prevent or control Swimmers Itch? Copper Sulfate is toxic to snail hosts. Call Aquacide Company for assistance 800-328-9350. Physically remove lake weeds & muck where snail gather. Administer praziquantel to hatchling waterfowl. Topical application of a water-resistant insect repellent with DEET. Water resistant sun cream with .1-1% niclosamide has been shown to be a highly reliable protectant Swimmers Itch, Wikipedia, the free encyclopedia, Swimmers Itch, Wisconsin Department of Health Services, www.dhs.wisconsin.gov, Swimmers Itch, www.MayoClinic.com/health/Swimmers-itch
Add a new threat to the list already facing China’s cute, but highly endangered giant pandas: climate change. Pandas, who have already seen much of their forest habitats lost to human development, could see a rapid decline in bamboo, their primary food, due to global warming, Sky News Australia reports. A group of researchers used data from the Intergovernmental Panel on Climate Change (IPCC) to estimate the impact of climate changes on panda habitats. In the worst-case-scenario, the group calculated that the bamboo in existing panda territories could completely disappear by the end of the century. The researchers called for environmental planning to devise ways to secure a supply of bamboo to sustain panda populations in the event their predictions come to pass. Less than 2,000 pandas are known to exist in the wild. Eating up to 38 kilograms of bamboo each day, pandass would be hard pressed to survive if bamboo withers in a warmer climate.
In this class you will learn to construct electronics projects that can control installation artworks. You will master the techniques of wiring and soldering. You will gain understanding of the basic concepts of electronics such as the circuit loop. You will discover many different IC (Integrated Circuit) chips, including the 555 timing chip and the 74xx decision making series. You will learn binary logic, binary counting, and the binary representation of numbers. You will become fluent in applying various electrical devices, such as resistors, capacitors, diodes, and transistors. You will build your own regulated 5V power supply. You will have the opportunity to experiment with various input devices; such as switches & buttons, photoresistors, potentiometers (volume knobs), optical encoders, and contact mics, as well as output devices; such as LEDs, incandescent lights, relays, solenoids, and motors. Students will design and develop independent projects that utilize these new skills. This class is a primer for “Electronics in Installations II" (Microcontrollers), which will introduce you to miniature stand-alone computers that are programmable in BASIC or C. These microcontrollers could become the most powerful art-making tool in your arsenal! We will do some reading about contemporary culture and the new breed of artists that exploit the power of electronics in their work. We may go on a field trip or two. Although the lab has been outfitted with a set of tools and many electronics components, you will be required to purchase your own set of tools, as well as any excessive electronic components that you use. These can be found at Radio Shack, or can be ordered from one of the various catalogs that are in the lab. Students are encouraged to collaborate their orders to save money on postage and handling. You will be graded primarily on the quality of your completed projects. I will give you a letter grade on each of 5 projects that will be determined by these factors: how well it works, how well it was designed, how well it was constructed, how well you are mastering the techniques, and promptness. On the final project, concept as an artwork will also factor in to your grade. Projects that are completed late will lose one fraction of a grade per week (for example, a B+ project turned in one week late will be given a B). Other factors in your final grade include attendance, participation, and homework. There may be a pop quiz (be warned!). *** You should pick a LOCKER and call 592-2214 to tell Jess what locker number you have selected. · Getting Started in Electronics; By Forrest Mims; Paperback; You must purchase this book from a Radio Shack by the second week of classes. OTHER GOOD REFERENCES: · ** Basic Electronics Theory; By Delton T. Horn; Paperback; · Beginning Electronics Through Projects; By Andrew Singmin; Paperback; · TTL Cookbook,; By Donald E. Lancaster, Don Lancaster; Paperback; · Electronic Circuit Guidebook : Sensors; By Joseph J. Carr, Joe Carr; Paperback; · The Art of Electronics; By Paul Horowitz, Winfield Hill (Contributor); Hardcover; Good magazines to read: ARTBYTE, Leonardo, Art in America, ArtForum A good place to see Electronic Art: Postmasters Gallery, 19th Street, just East of 10th Avenue, Chelsea Fine web resources (I strongly urge you to scope out both of these sites): Technical -- http://www.bc1.com/users/sgl/html/jo4.htm Artistic -- http://userwww.sfsu.edu/~infoarts/links/wilson.artlinks2.html |Week 1:||Lecture: Introduction to Basic Electronics: The power loop, AC/DC, voltage, amperage, Ohm's| |law, resistors, capicitors, diodes, L.E.D.s, 5V Power supply circuit| |Homework: Purchase tool set, Mims book, and copper printed circuit board at Radio Shack.| |Read introductory chapters in Mims| |Week 2:||Lecture: Physical techniques - soldering, de-soldering, wiring| |Workshop: Construct 5V power Supply circuit| |Homework: Finish 5V power Supply| |Week 3:||Workshop: 5V power supply DUE, grading| |Lecture: The 555 Chip. A-Stable 555 timer design, breadboarding.| |Workshop: Breadboard A-Stable 555 timer| |Homework: Design transfer of breadboard to circuit board, begin construction| |Week 4:||Lecture: tips on IC circuit construction| |Workshop: Transfer 555 timer to circuit board| |Homework: Finish 555 timer on circuit board, Purchase AC socket and bulb| |Week 5:||Workshop: 555 timer DUE, grading| |Lecture: Transistors, Relays, Potentiometers, AC loops. AC relay circuit w/ Lightbulb| |Workshop: Use clip leads to try out AC relay circuit, modify 555 to include Pots| |Week 6:||Workshop: Complete AC relay circuit, testing| |Homework: Finish AC relay circuit| |Week 7:||Workshop: AC relay circuit DUE, grading| |Lecture: 74xx IC chips, 8-segment Counter circuit design| |Homework: Purchase any needed components for Counter circuit| |Week 8:||Lecture: Introduction to decision making 74xx chips, written logic assignment| |Workshop: Breadboard Counter circuit (w/partner), begin construction| |Homework: Written assignment, and finish breadboard if not yet completed| |Week 9:||Workshop: Turn in written homework, Construct and Debug Counter circuit on circuit board| |Homework: Finish Counter circuit| |Week 10:||Workshop: Counter circuit DUE, grading| |Homework: Brainstorm ideas for final project, write proposal| |Week 11:||Workshop: Discuss final project ideas. Possible field trip to Chelsea galleries| |Homework: Purchase supplies for final project| |Week 12:||Final project lab| |Week 13:||Final project lab| |Week 14:||Final project should be COMPLETED here. Debugging.| |Workshop: Create schedule for Final project critique| |Homework: Prepare display of Final project| |Week 15:||Final project class critique and grading| |As you can see, there is NO room for LATE final projects. Incomplete projects will suffer poor grades!
The coffee tree comes from a large family of evergreen shrubs; to call it a tree may be somewhat of an exaggeration. For much of coffee’s recent history, the species and variety of the tree was chiefly a concern of coffee farmers’ efforts to combat disease and pests along with increasing crop yields. Recently, coffee roasters have taken an interest in coffee varieties for a very different reason: marketing. While coffee originated in Ethiopia, it was in neighboring Yemen that Turkish farmers first cultivated it on a commercial scale. Undoubtedly, the Turks could have selected any number of species and varieties to plant, but the one they chose was most likely for its fine flavor since there were other hardier, and greater producing species to be found. The Turks closely guarded their treasure, and how seeds were finally secreted away is a source of legend. This species of coffee took its name from whence it came: Arabia; taking on the nomenclature Coffea Arabica, which today accounts for 80% of the world’s production. The spread of Arabica coffee around the world was based on a very limited number of trees. Seven berries were taken by Baba Budan to India; a small shipment was taken to the French colony of Reunion; and the tree taken from Java to Amsterdam in 1706, together with its offspring in Paris, which provided all the planting material for South and Central America. Consequently, the whole genetic base of the Arabica coffee industry is very narrow. As demand for coffee grew, large areas - sometimes entire countries – became dedicated to coffee production, revealing weaknesses inherent in propagating a single species. Since the coffee tree is an evergreen shrub having broad leaves, it is a tempting target for pests and disease. This became glaringly clear in Sri Lanka - once one of the world’s largest producers – that ultimately lost its entire crop to disease in the late 1800’s, never to recover. Experience has led farmers to seek alternatives through inbreeding to increase resistance, and to boost crop yield. Since the coffee tree is an inbreeder, natural mutations occasionally occurred. These mutations often displayed characteristics that allowed them to adapt to specific growing conditions. Inbreeding has led to a number of hybrids of the Arabica species that are referred to as varieties or cultivars. These cultivars all derive from the original typica variety from Yemen and the bourbon variety from the Reunion. The two varieties are considered identical from a herbarium perspective. Hybrid varieties have become so popular that in many countries they greatly outnumber the original typica and bourbons and, in some cases, have completely replaced them. For the most part, coffee connoisseurs were quite unaware of this change until recently. Hybridization began in earnest after the Second World War as part of the “Green Revolution” that occurred throughout the world’s agricultural industry. This movement was characterized with a change in growing practices to increase production and simplify labor practices through mono-cropping and heavy use of fertilizer and pesticides. Since coffee contributes significantly to the GNP of many equatorial nations, government sponsored research institutes and boards were created to assist farmers in the latest methods in what came to be known as technified farming. Technified farming encouraged the removal of traditional multi-story shade canopies in favor of specialized shade or full sun growing techniques.A new form of hybrid propagation was encouraged to combat disease and pest outbreaks due to intense mono-cropping. This new hybrid is referred to as interspecific hybrids. Interspecific hybrids are cross bred from outside the Arabica species, most notably the Robusta species. These hybrids display negative taste characteristics that are quite recognizable to the trained coffee professional. The success of interspecific hybrids may have increased yields and forged a (temporary) bulwark from disease and pests, but many professionals within the coffee industry believe that this success has come at a cost to flavor quality. The confusion between inbred cultivars and interspecific hybrids has led to a resurgence of interest in specific varieties, and what some may describe as the heirloom varieties of bourbon and typica, and an “Indiana Jones” style quest to find the ultimate cultivar. The recent interest in the Geisha variety best illustrates this phenomena. Some roasters, as a fashionable method to display their commitment to quality, now actively promote particular varieties. What is important to remember however is that the extreme narrow genetic base for the Arabica species makes taste variation quite negligible. It is difficult enough, if not impossible, to distinguish a flavor variance between the typica and bourbon variants, much less variants within cultivars. Any perceived taste variation is more likely resulting from external variables. Ultimately, flavor quality depends on microclimate variables, ecological stewardship and processing methodology. Flavor profiles are the result of specific locations, the best cultivar is the one that thrives best in its location. Coffees coming from technified plantations are thin, acidy, and flavorless due to the low nutrient quality of their environment. Pumping nitrogen fertilizer into poor soil conditions and highly variable temperature conditions does not make for good flavor quality. Coffees coming from rich, biodiverse environments, that are managed in a systematically caring fashion, whether through science or tradition, display fine flavor characteristics that are a snapshot of their environment. Rather than wasting time promoting varietals, roasters would do better promoting micro-climate. Not only would this better promote flavor characteristics, it would reward farms that are healthy stewards of biodiversity.
South Asians have a complex historical relationship with African Americans. Over time, Desis (South Asians) and Blacks have had multiple crossovers in philosophical, racial, and ethnic identity… As a result of the Civil Rights Movement, the passage of the Immigration and Naturalization Act of 1965 allowed for increased immigration from non-Western nations. The INS act incentivized scientists, professors, physicians, and other professionals to immigrate to the US during the Cold War. Subsequently, it was amended in 1986 so that the families of these immigrants could live as permanent legal residents. The high socioeconomic status of these early waves of immigrants, combined with their ambitions to integrate and prosper into the “the land of opportunity” created the perfect storm for Desis to generate animosity toward Blacks. Although colorism was always endogenously prevalent in South Asia, it was more important to assimilate with prejudices that whites had regarding African Americans in order to create a commonality from which to form an intergroup identity. …An examination of the 1990 Census found that 90% of Indian-headed households identified as Indian when in 1970 nearly 75% identified as white. The first wave of South Asian immigrants in the 1960s and 1970s were largely educated professionals, and because of their educational background and the facts that immigrants usually avoid association with Blacks, so some identified as white. The latest wave of South Asian immigrants however, has been working-class and more likely to interact with African Americans and other people of color in urban centers." Volume 3, Number 3 Transforming Ethnic Studies Tokyo Bound: African Americans and Japan Confront White Supremacy Yellow Power: The Formation of Asian-American Nationalism in the Age of Black Power, 1966-1975 Jeffery O.G. Ogbar East of the Sun (West of the Moon): Islam, the Ahmadis, and African America Linking African and Asian in Passing and Passage: The Pagoda and the True History of Paradise B-Boys and Bass Girls: Sex, Style, and Mobility in Indian American Youth Culture Sunaina Marr Maira Building the Antiracist, Anti-Imperalist United Front: Theory and Practice from the L.A. Strategy Center and Bus Riders Union Adding: ‘Left or Right of the Color Line: Asian Americans and the Racial Justice Movement’ from ChangeLab My Uncle James just died of a heart attack. My granddad is sick from gout, hypertension and diabetes. There is no time like the present to get serious about health. All of their ailments are mostly PREVENTABLE. That’s the part that pisses me off. The food we eat is killing us. Fried foods. Red meat in abundance. Hundreds of calories. High in cholesterol. No fiber. No fresh vegetables. No fresh fruits. This is how we eat. It pisses me off because my grandpa is supposed to be at home, growing old with my grandma. Instead, he’s lying in a hospital bed. This makes me want to get serious about my health. I must change my life, for me and to inspire my family to get more healthy.
The historical account on cryptography is incomplete without mentioning the diplomatic intrigue, the Zimmermann telegram and the man responsible, German Foreign Secretary, Arthur Zimmermann. Born in Marggrabova, he studied law, serving as the vice-consul at Shanghai and consul in 1900 at Tientsin and working with the diplomatic branch of the German civil service in 1905 in a subordinate capacity. It was in the year 1916 that he became the foreign secretary on November 25 owing to the support of the military rulers, Paul Von Hindenburg and Erich Ludendorff along with the submarine policy. The date 19 January 1917 was when he sent the Zimmermann telegram to the German minister in Mexico through the German Ambassador in Washington as a repercussion to the threat from America. When this was intercepted by the British, it brutally embarrassed Germany and was crucial in President Woodrow Wilson's involvement of America in the First World War. It revealed German support for the Mexican invasion of the US. After this telegram was deciphered, Zimmermann perplexed political spokesmen everywhere by admitting to the authenticity of the telegram in a speech where he claimed that these were actions to be taken only after the US declared war, and that President Woodrow was the culprit behind loosening ties between America and Germany. Before dying in 1940, he withdrew himself from his post. - Who's Who: Arthur Zimmermann - A profile of Arthur Zimmermann, About.com: European History - http://en.wikipedia.org/wiki/Image:AZIMA.jpg (public domain image)
High levels of Nitrates in water can promote algae blooms and effect turbidity and dissolved oxygen levels. Fertilizers carried to the creek by runoff can raise Nitrate levels. Nitrates, a product of decomposition, can easily be converted into nitrates and also impact pH. A high concentration of nitrates in the creek can lead to a large number of algae blooms in the creek and increase turbidity. Return To Creek Data Home Page
Lead is one the oldest and most prevalent poisons found in the earth's atmosphere. Lead inhalation produces serious heath problems, ranging form lowered IQs to heart disease. Lead inhalation and poisoning can come from many sources. Two of the most common are automotive exhausts and house paint. Other sources include: lead pipes, seams of food cans, ceramic glazes, batteries, and cosmetics, as well as lead mining, smelting, and coal combustion. Exposure to lead can be extremely harmful. Consequences include damage to the brain, kidneys, reproductive system, and cardiovascular system. Lead poisoning, even in small doses, can also significantly reduce the IQs of young children. Still more adverse effects are: aggressive behavior, delinquency, and attention disorders in children, and increased blood pressure and hypertension in adults -- conditions that increase the risk of heart disease. High lead levels are found in people of both developed and developing countries. However, it is poor populations that suffer most from the negative health effects of lead. Exposure is generally higher for these groups and, due to their living conditions, continued exposure is more likely. For example, living by busy roads in urban areas might make the poor more susceptible to automotive fumes; likewise, living in older housing might increase their exposure to lead-based paint. Finally, the human stomach better absorbs lead when it is empty and when a person's diet lacks certain trace elements like iron calcium, and zinc. Chemical profile of lead health effects
Scientists discover how Cordyceps mushrooms fight cancer For thousands of years, the mushroom known as Cordyceps has been used in traditional Chinese medicine to treat problems ranging from coughs and fatigue to impotence and cancer. And once Western scientists started considering that anecdotal reports of the mushroom’s curative powers might be something more than “folk tales”, research started accumulating that backs up many ancient claims about Cordyceps’ health benefits. In much of Asia, particularly China and Japan, mushrooms have been part of the culinary experience but most significantly, they have harnessed the biggest health benefits of mushroom extracts for millenia. With new developments in mushroom cultivation in just the last 20 years, increased availability, research, and cost, use of mushrooms for their health benefits has increased. As an example, the Reishi Mushroom ( Ganoderma Lucidum ) was at first provided solely to Asian royalty and often called mushroom of immortality. Scientists from the University of Nottingham in Great Britain say they’ve documented how Cordiceps can fight cancer — and the new discovery could increase the effectiveness of mushroom-derived cancer treatments. For the study, published in the Journal of Biological Chemistry, Dr. Cornelia de Moor of the University of Nottingham and her colleagues investigated a drug called cordycepin, which was originally extracted from wild growing Cordyceps and is now prepared from a cultivated form of the mushroom. “Because of technical obstacles and people moving on to other subjects, it’s taken a long time to figure out exactly how cordycepin works on cells. With this knowledge, it will be possible to predict what types of cancers might be sensitive and what other cancer drugs it may effectively combine with. It could also lay the groundwork for the design of new cancer drugs that work on the same principle,” Dr. de Moor said in a statement to the media. The University of Nottingham scientists found that the Cordyceps-derived treatment has two important effects on cells that could impact the growth of malignant tumors. At low doses, cordycepin inhibits the uncontrolled growth and division of cells while at high doses it prevents cells from sticking together, essentially blocking the cells from growing. The research team concluded that each of these effects most likely is the result of a single underlying mechanism — cordycepin interferes with how cells make proteins. Specifically, low doses of the mushroom-derived treatment interfere with the production of mRNA (the molecule that directs cells to assemble a protein) and at higher doses cordycepin directly interacts with the making of proteins. “Our discovery will open up the possibility of investigating the range of different cancers that could be treated with cordycepin,” Dr de Moor stated. “We have also developed a very effective method that can be used to test new, more efficient or more stable versions of the drug in the Petri dish.This is a great advantage as it will allow us to rule out any non-runners before anyone considers testing them in animals.” Cordyceps is used in Chinese medicine to make your lung energy stronger. That is related to your immune system in general. If your lung energy is stronger, it is easier to avoid getting sick. Your overall health and energy level determines if you are as likely to catch a cold or flu or suffer from allergies. That is a Chinese medicine theory. Cordyceps has been used to boost the immune system for years. In addition to helping to reduce the likelihood of catching something, once a person catches something, if his overall health and energy were good to start with, it is easier to get over or fight off an ailment. You might think that cordyceps would not help you with insomnia. After all, cordyceps is an energy tonic. However, Cordyceps can help with insomnia. Very often, people wake up in the middle of the night to go to the bathroom. They do not realize that they woke up to go to the bathroom, they think they just woke up and since they were already awake, they take the opportunity to go. In fact, if your kidney energy were strong enough, you should not wake in the middle of the night. I have these discussions with my patients. They say they do wake at night, but it was not DUE to having to go to the bathroom. Unless you drink a bunch of water before you go to bed, you should be able to sleep through the night. Cordyceps can be taken long term to strengthen your kidney energy and reduce or eliminate the urge to go to the bathroom in the middle of the night. More Cordyceps news Scientists at Dalhousie University in Halifax, Nova Scotia, published research in the November 30th edition of the journal Cancer Immunology, Immunotherapy showing that oral Cordyceps can reduce the occurrence of cancer spreading to the lungs in metastatic breast cancer. Although the researchers found that Cordyceps does not reduce the growth of the primary breast tumor, they noted that deaths from breast cancer are primarily due to the development of metastases. That means a treatment that stops the spread of metastatic tumors could save countless lives. Although various medical therapies currently exist that attempt to stop the growth of cancerous metastatic tumors, they have little effect — so this makes the Dalhousie University research into Cordyceps very important. The Canadian researchers stated the evidence they have so far suggests that the mushroom reduces the growth of metastases due to Cordyceps’ effects on the tumors’ cell cycles. Cordyceps is a genus of ascomycete fungi and some of its members including Cordyceps mycelium are commonly used in herbal medicine. The popular species in commercial products are Cordyceps sinensis, Cordyceps ophioglossoides, Cordyceps capita, and Cordyceps. militaris. Cordyceps militaris is a traditional herbal ingredient frequently used for tonic and medicinal purposes in eastern Asia, while, Cordyceps sinensis (dong chong xia cao; 冬蟲夏草) has been suggested to possess anti-tumor, immunostimulant and antioxidant activities. The chemical constituents for most species include cordycepin (3′-de-oxyadenosine) and its derivatives, ergosterol,polysaccharides, a glycoprotein and peptides containing alpha-aminoisobutyric acid. Their benefits are suggested to include anti-tumour, anti-metastatic, immunomodulatory, antioxidant, anti-inflammatory, insecticidal, antimicrobial, hypolipidaemic, hypoglycaemic, anti-ageing, neuroprotective and renoprotective effects. Polysaccharide accounts for the anti-inflammatory, antioxidant, anti-tumour, anti-metastatic, immunomodulatory, hypoglycaemic, steroidogenic and hypolipidaemic effects. Cordycepin contributes to the anti-tumour, insecticidal and antibacterial activity. Ergosterol exhibits anti-tumour and immunomodulatory activity.
If oil revenue is concentrated on infrastructure projects in far-flung regions, such as South Sudan and Darfur, it could prevent those areas from falling back into war. Greater prosperity through an oil-driven economy has the potential to create jobs and generate revenues that would allow both north and south Sudan to mimic the relatively well-managed oil-driven economies of the Persian Gulf states such as Qatar and Dubai. Oil revenues have already made an impact, paying for roads and bridges and hydropower dam projects that the country desperately needs. But most experts believe that the country’s oil may cause more harm than good, increasing government corruption and perhaps even pushing the country back into war. For one thing, the oil wealth is not evenly distributed. The bulk of Sudan’s oil lies in fields along the still-unmarked boundary between the North and the South, between the still-heavily armed Sudanese People’s Liberation Movement and the equally well-armed miltary loyal to President Omar al-Bashir. According to most estimates, more than 70 percent of the oil reserves and more than 70 percent of the current production are inside Southern Sudan.
Telescopes and DSLR cameras available today allow amateur astrophotographers to take very detailed images of the night sky. Combine this technology with advances in acquisition / tracking and processing software allow us to obtain images of nebulae, galaxies and star clusters of similar quality as those taken by professional observatories only a decade ago. Often the targets are very faint, even with extremely sensitive cameras, long exposures ranging from a few minutes to several hours are needed to capture the pleasing detail. Extreme precision is required of the mechanics of telescope, focuser and mount to track these objects as they move along their path across the night sky. A dark site, a bit of experience, and some image processing skills help to achieve pleasing results.
That part of the building wiring system between the final over-current protective device (typically at the electrical panel) and the electrical equipment that consumes power. any one of the complete routes for carrying electric current usually emanating from the main line. That part of a wiring circuit between the final set of fuses and the place where the lighting fixtures or drop cords are attached. An electric circuit with a separate fuse or circuit breaker. A circuit that supplies power to the electrical loads in a building and is terminated at a distribution device (panelboard, loadcenter, etc.). a circuit supplying many outlets BX - see armored cable an electric circuit that supplies current to a limited number of outlets and fixtures A circuit that supplies power to the electrical loads in a building and is terminated at the panelboard or switchboard. Wiring that runs from a service panel or sub-panel to outlets. Branch circuits are protected by fuses or breakers at the panel. An electrical circuit with its own circuit breaker in the service panel. A division of a load circuit with current limited by a fuse or circuit breaker. The circuit conductors between the final overcurrent device protecting the circuit and the outlet(s). Any circuit that comes from the main power lines.
Getting Started with VBA in Word 2010 Summary: Introduces new programmers to using Visual Basic for Applications (VBA) programming to extend Microsoft Word 2010. This topic summarizes the VBA language, instructions about how to access VBA in Word 2010, a detailed solution to a real-world Word VBA programming problem, and tips about how to program and debugging. Applies to: Microsoft Word 2010 | Microsoft Office Word 2007 | Microsoft Office Word 2003 Published: April 2010 Provided by: Shauna Kelly What is a Macro, Why Would I Want One, and How Can I Create One? A macro enables you to put Word on autopilot. It can be frustrating to perform a frequent task in Word if you have to use a long sequence of mouse clicks or key strokes every time that you perform the task. When you write a macro, you can bundle up a collection of commands and instruct Word to start them with only one click or one keystroke. This article assumes that you are a skilled user of Word. It also does not assume that you know anything about computer programming or software development. This article is in two parts: The first part of the article describes an end-to-end example: how to write a simple macro, where you can save it, how to create a button on the Quick Access Toolbar or create a keyboard shortcut to run the macro, and how to copy that macro to another computer (so you can use it at home and at work, or so you can share it with a colleague). That will give you a simple example. The second part of the article examines macros in detail: how to write more complex macros and how to ensure that your macro actually does what you want and does not create errors or problems. Part 1: An End-to-End Macro Example One purpose of macros is to enable you to perform a task with one click or one keystroke. I often create documents that contain hyperlinks to Web sites. By default, Word requires me to follow, or open, a hyperlink, by pressing CTRL while clicking the hyperlink. The following image shows the default option for following a hyperlink. Figure 1. Press CTRL and click the hyperlink to follow it The following image shows the alternative option for following a hyperlink by clicking it. Figure 2. Click to follow hyperlink Sometimes that is useful, but sometimes I prefer to follow a hyperlink by clicking it (Figure 2). I may configure that setting many times a day. To do that, click the File button, and under Help, I click Options. In the Word Options dialog, I click Advanced, and then under Editing options, I select (or clear) Use CTRL+Click to follow hyperlink and then click OK. Doing that sequence of mouse clicks repeatedly is frustrating. I want to be able to click one time to configure this setting. This section describes how to create a simple macro to toggle between using CTRL+Click and only clicking a hyperlink. To create this macro, you will learn how to: Create a Word Macro-Enabled Template file (.dotm) to store your macro and save it in the Word Startup folder. Open and use the Visual Basic Editor, which is part of Word. Write a macro in the Visual Basic Editor. Test your macro. Create a button on the Quick Access Toolbar to run your macro. Create a keyboard shortcut to run your macro. Save your file, load it as an add-in and, if you need to, share it with a colleague, or use it on another computer. This section describes each step. Creating a File to Hold Your Macro Before you start to write a macro, think about how and where you want to save it. Word lets you save macros in two Word file types: a Word Macro-Enabled Document file (.docm) and a Word Macro-Enabled Template file (.dotm). It is not a recommended practice to store macros in a document. Macros are usually stored in a Word Macro-Enabled Template file (.dotm). It is possible to save macros in the Normal Template (Normal.dotm), but that makes it difficult to organize or share your macros. It is usually better to create your own file and store your macros there. Where to save your template file depends on how you want to use the macros. You must decide whether you want to use your macros with all the documents that you might work on or only with some documents. For example, you may create a template for monthly sales reports, and you may have macros to format those reports. In this case, save the template file in the User Templates folder, and use that template as the basis for new documents. (To verify the User Templates folder, click the File tab, and then click Options. In the Word Options dialog, click Advanced, and then click the File Locations button.) If a macro is stored in a template, then only documents attached to that template can access the macro. Other documents will be unable to “see” the macro. In this article, you will create macros that can be used by any document. To do that, you will create a Macro-Enabled Template file. First, close any Word files that you have open. To create the new file to hold your macros, click the File button, click New, and then click My Templates. In the New dialog box, select the Template option button and then click OK. Save the file as a Macro-Enabled Template file that is named MyWordTools.dotm in the Word Startup folder. (To verify the Word Startup folder, click the File tab, and then click Options. In the Word Options dialog, click Advanced, and then click the File Locations button.) Using the Visual Basic Editor Word macros are written in a programming language called Visual Basic for Applications (VBA). You write macros in a part of Word that most users never see: the Visual Basic Editor (the VBE). Open the VBE by using any one of the following methods: Press the keyboard shortcut, Alt+F11 Click the Visual Basic button on the Developer tab. To do that, click the File tab, and then click Options. On the Word Options dialog box, click Customize Ribbon. In the right side of the dialog box, select the Developer tab. Click OK to return to your document, and then on the Developer tab, click the Visual Basic button. Add the Visual Basic command to the Quick Access Toolbar. Before you start to use the VBE, on the Tools menu, click Options. In the Options dialog box, on the Editor tab, ensure that all the check boxes are selected. In the upper-left side of the VBE you will see the Project Explorer. This shows all files that are currently open in Word. Expect to see Normal (which refers to Normal.dotm) and MyWordTools (the new file that you just created). The following image shows the MyWordTools project in the Visual Basic Editor Project Explorer. Figure 3. Visual Basic Editor Project Explorer Within your file, macros are stored in Modules. To add a module to your file, in the Project Explorer, select the MyWordTools file (Figure 3). On the Insert menu, click Module. When you add a module to your file, you will see the module added in the Project Explorer and in the Properties Window under it (Figure 4). You can rename a module using the Properties Window. Leave the name as Module1. The following image shows Module1 in the Visual Basic Editor Project Explorer. Figure 4. Visual Basic Editor Project Explorer Finally, you will see that the entry for your file is named TemplateProject. Although it is not necessary, it is recommended that you give it a more descriptive name. To do that, right-click the entry for the MyWordTools file, and then click TemplateProject Properties on the shortcut menu (Figure 5). The following image shows the TemplateProject Properties… menu item Figure 5. Properties menu item In the Template Project-Project Properties dialog, change the Project name to MyWordTools. Writing Your Macro The large white area on the right in the VBE is where you write the code of your macro. If you cannot see it, on the View menu, click Code to see the code window. Word automatically inserts the statement Option Explicit at the top of the code window. Do not delete this. You want a macro to toggle between the two possible settings, in Word, for following a hyperlink. It will work like other buttons that toggle a setting in Word. The Bold button on the Home tab, for example, will text bold if it is not bold and make text not bold text if it is currently bold. Copy and paste the following macro code example into the code window. Sub ToggleHyperlinkCtrlClick() Options.CtrlClickHyperlinkToOpen = Not Options.CtrlClickHyperlinkToOpen End Sub This is a short macro named ToggleHyperlinkCtrlClick, and it only has one line of code. The one line of code in our macro means "Change the Word option that controls whether I have to use CTRL+Click to open a hyperlink to the opposite of its current setting, that is, to Not its current setting" (Figure 6). The following image shows a line-by-line explanation of the Figure 6. Explanation of the ToggleHyperLinkCtrlClick method Testing Your Macro To test your macro, use the following procedure. To test the macro Arrange the Word and VBE windows so you can see them side by side. Click the main Word window. Type several paragraphs of text into the MyWordTools.dotm document. Include some hyperlinks in your text (Figure 7). The following image shows the MyWordTools document and the ToggleHyperLinkCtrlClick code in the VBE side by side. Figure 7. Document and Visual Basic Editor side by side In the VBE, click anywhere within your macro. To run your macro, on the Run menu, click Run Sub/User Form or press F5. The setting to follow a hyperlink will be changed. Mouse over a hyperlink in the main Word window to see that the tooltip has changed. Re-run the macro to toggle the setting. You can also run your macro within Word itself. On the View tab, in the Macros group, click the Macros button. Your ToggleHyperlinkCtrlClick macro will be listed in the Macros dialog. To run your macro, click the name of the macro, then click the Run button. Creating a Button on the Quick Access Toolbar to Run Your Macro To get one-click access to your macro, you can add a button to the Quick Access Toolbar. To do that, use the following procedure. To create a button on the Quick Access Toolbar Right-click the Quick Access Toolbar and then click Customize Quick Access Toolbar on the shortcut menu. Under Customize the Quick Access Toolbar, in the Choose commands from list, select Macros. In the Customize Quick Access Toolbar list, select MyWordTools.dotm. (You must select MyWordTools.dotm so that Word will save the button on the Quick Access Toolbar in the MyWordTools.dotm file. Only then will the button be available when you copy that file to another computer.) Select the ToggleHyperlinkCtrlClick macro and then click Add. Click the Modify button to select a symbol and change the name to ToggleHyperlinkCtrlClick. You can now run your macro at any time by clicking the new button on the Quick Access Toolbar. Creating a Keyboard Shortcut to Run Your Macro You can also create a keyboard shortcut to run your macro. To do that, use the following procedure. To create a keyboard shortcut to run the macro Right-click the Quick Access Toolbar, and then click Customize the Ribbon on the shortcut menu. Next to Keyboard shortcuts, click the Customize button. The following image shows the Customize Keyboard dialog box Figure 8. Customize Keyboard dialog box In the Customize Keyboard dialog box (Figure 8), you must: In the Categories list, select Macros. In the Macros list, click your macro name. Click in the Press new shortcut key box, and type the keyboard shortcut that you want to use. I used Alt+H, which I can remember because this toggles the setting for Hyperlinks. The dialog box also tells me that this shortcut is currently unassigned, so I will not interfere with an existing keyboard shortcut. In the Save changes in list, select MyWordTools.dotm. (You must select MyWordTools.dotm so that Word will save the keyboard shortcut in the MyWordTools.dotm file. Only then will the button be available when you copy that file to another computer.) To run your macro, press Alt+H. You have now created a new file to hold your macro (MyWordTools.dotm), added a Module (Module1), created your macro (known as ToggleHyperlinkCtrlClick), created a button on the Quick Access Toolbar to start the macro, and created a keyboard shortcut to start the macro. Now, save the MyWordTools.dotm file (you can do this in the VBE or from the main Word window). Because you have finished with the VBE, close it and return to Word. To do that, in the VBE, click the File menu and then click Close and Return to Microsoft Word. To text your macro, click your button on the Quick Access Toolbar. When you confirm that it is working, save and close the file. Managing and Loading Add-ins The plan was that this macro should be available regardless of what document that you are working on. However, if you create a new document (use CTRL+N) you will be unable to see your button on the Quick Access Toolbar. There is no way to start your macro. To make the macro stored in MyWordTools.dotm available to any document that you have open, you must load MyWordTools.dotm as an add-in. (When a .dotm file is used as an add-in it is also known as a ‘global template’.) You can load a .dotm file as an add-in either manually or automatically: To load a .dotm file as an add-in manually, close the.dotm file if it is currently open. On the Developer tab, click the AddIns button. In the Templates and Add-ins dialog, click Add, locate your .dotm file and then click Open. To load a .dotm file as an add-in automatically, the .dotm file must be saved in the Word Startup folder. Quit and restart Word. Word will load your add-in automatically. You can see what add-ins are currently loaded in Word. On the Developer tab, click the Add-Ins button. Because you saved MyWordTools.dotm in the Word Startup folder, close and restart Word. Word will automatically load your MyWordTools.dotm as an add-in. You should see the button on the Quick Access Toolbar, and you can use your macro. When MyWordTools.dotm is loaded as an add-in, you will not see any of the text you may have left in the main Word window in MyWordTools.dotm. Word makes no use of any content on the face of the document itself. Delete any text in the main Word window before you save a .dotm file that is used as an add-in. It is important to differentiate between opening an add-in file (for example, by clicking File and then clicking Open) and loading an add-in (for example, by using the Add-Ins button on the Developer tab). Open a file when you want to edit it and test it. Load the file as an add-in when you want to use it. Never load a file as an add-in if that file is currently open in Word. Never open and edit an add-in file if it is currently loaded as an add-in. This may result in unpredictable behavior. Using Your Macro on Another Computer To use your macro, you must have: The macro code. A button on the Quick Access toolbar that runs the macro. The keyboard shortcut that runs the macro. You saved all three in MyWordTools.dotm. So, to use your macro on another computer, or to share your macro, you just have to to copy MyWordTools.dotm to the Word Startup folder on another computer. When you start Word, Word will load MyWordTools.dotm as an add-in and your macro, the button and the keyboard shortcut will all be available. Part 2: Writing More Complex Macros The end-to-end example in Part 1 includes a simple one-line macro. Part 2 describes how to extend that basic end-to-end process to create more complex macros. Adding Additional Macros to the MyWordTools.dotm File The file that you created in Part 1, MyWordTools.dotm, now contains one module (Module1), and that module contains one macro (ToggleHyperlinkCtrlClick). Do not edit a file while it is loaded as an add-in. To add additional macros to your file, you must: Unload the file as an add-in. To do that, on the Developer tab, click Add-Ins, and then click to clear the check box for your add-in and then click OK. Open the file for editing. To do that, click the File tab. Under Info, click Open. Locate the file, click the file, and then click Open. When your file is open for editing, you can: Add a macro to the existing module, Module1, by typing in the code window, or Add a new module to the file, and type a new macro in the new module. Which method that you use depends on how complex your macros are. If you only have several short macros, then it is common to put them all in one module. If you have many long macros, you may prefer to store them in separate modules and rename each module to indicate what macros the module contains. When you have added a new macro to your file, you can test it, create a button for it on the Quick Access Toolbar, and create a keyboard shortcut for it, exactly as you did for your first macro. Professional developers generally include comments in their code to explain what the code is intended to do. You can add comments to your macros by preceding the text of the comment with a single apostrophe. By default, the VBE displays comments in green. Writing Robust Code: A Macro to Sort Text There are two things to consider if you share a macro with a colleague. First, the other person’s computer is almost definitely not set up identical to yours. Second, the other person is likely to be less forgiving of poor code than you are. So your code must be robust. If a macro is robust then the user will not see inexplicable error messages and the macro will do what it is intended to do—no more and no less. To demonstrate some issues in writing robust code, consider the following macro that is used to sort text. It takes several mouse clicks to sort several paragraphs of text. If you must sort text frequently, it may be useful to have a single button to do a simple sort (and when you must do a more complex sort, the builtin button will still be available to you on the Home tab). This macro action is shown in the following code example. You can test this macro as you tested the previous one. Tile the Word and VBE windows side by side. In the main Word window, type several paragraphs of text and select them. In the VBE, run the macro by clicking in the macro code and pressing F5. Word will sort the paragraphs alphabetically. Our SortText1 macro seems to work well. However, it is not robust. To see what can go wrong, insert a picture into your document. Make it a floating picture. To do that, select the picture. On the Format tab, under Picture Tools, in the Arrange group, click Position. Select any one of the With Text Wrapping page position options. Select your picture, and then run the SortText1 macro. Because it does not make sense for Word to sort one floating picture, you will see an error message from Visual Basic. That is not a robust macro! Our one line of code, Selection.Sort, only works correctly if you have selected ordinary text. Before you find a way to solve that problem, consider another problem. If you click in your document without selecting any text and run the SortText1 macro, the whole document is sorted. At best, that is unexpected; at worst, it might corrupt data. The following code example shows how to restrict the macro to sorting only when there are two or more paragraphs of selected text. Sub SortText2() ' A macro to sort the Selection if the user has selected more than one ' paragraph of text. If Selection.Paragraphs.Count > 1 Then Selection.Sort End If End Sub Experiment running the SortText2 macro. Unlike SortText1, this macro will not display an error message if you select a picture and then run the macro. Also, it will not let you accidentally sort your whole document. To see what happens, use the following procedure to step through the code one line at a time. To step through the VBA code Click in the main Word window, and select your picture. Click anywhere within the SortText2 macro. Press F8. Word will highlight the first line in the macro ( Press F8 repeatedly to step line by line through the code. Using F8 to step through code is a common way to see what the code is doing, and to troubleshoot problem code. In this case, you see that, when a picture is selected, Word processes the If statement, but skips the There are several important things to see in the new SortText2 macro: You can use If and End If to control whether Word processes code or skips it. The If line ends with the keyword Then. Each If must have a corresponding End If. This macro uses Selection.Paragraphs.Countprovides information about what the user has selected. Selection.Sortsorts the selected text. The following code example show how to extend this macro to provide information to the user by using the Else and MsgBox keywords. Sub SortText3() If Selection.Paragraphs.Count > 1 Then ' The user has selected more than one paragraph of text. ' Sort the selection. Selection.Sort Else ' Tell the user what to do. MsgBox "Please select two or more paragraphs and try again." End If End Sub From this macro, you can see that: You can use If…Then, Else and End If to control whether Word processes code or skips it. Using an Else statement is optional. Each If statement may have zero or one Else statements. If you use Else, you still need an End If. The command displays a message to the user on the screen. Type the text of the message after the keyword MsgBox, and enclose it in double quotation marks. If you provide any text in a macro (such as the text of this message), you must enclose it in quotation marks ("). Without quotation marks, Word would try to find "Please" in the Word object model. Because Word does not know what "Please" means, the macro would cause an error message. To see the problem, use the following procedure. To test the SortText3 macro Create a button on the Quick Access Toolbar to start the SortText3 macro. Instruct Word to save the button in MyWordTools.dotm. Test your button to ensure that it runs the SortText3 macro. Save the MyWordTools.dotm file Quit and restart Word If a document is open, close it so that there is no document open. Test your button on the Quick Access Toolbar again. You should see an error message. The error is occurring because the first line of the macro refers to the Selection. The add-in is loaded. If no document is open in the main Word window, then there is no Selection. Without a Selection, Word cannot run the code in your macro. Make one more change to the macro to make it robust. Add a comment to the top of the macro. That is a good way to document your macro so anyone will know what the macro is intended to do. Unload the add-in, and then open the file for editing. Replace your existing macro with the following code example. Sub SortText() ' A macro to sort the selected text, if the user has selected ' more than one paragraph If Documents.Count > 0 Then ' The user has at least one document open. If Selection.Paragraphs.Count > 1 Then ' The user has selected more than one paragraph ' of text, so sort it. Selection.Sort Else ' Tell the user what to do. MsgBox "Please select two or more paragraphs and try again." End If End If End Sub In this final version of SortText, this example used nested If…End If blocks. It is important to indent lines of code. Each If is matched to the correct End If. A macro that is not robust can result in: Error messages (for example, when the Selection object was used, but there was no document open, and therefore no Selection), or Unwanted behavior (for example, when the Selection.Sortcommand unintentionally sorted the whole document). Here are tips for writing robust macros. If your macro refers to the Selection, create a test document and test your macro by running it after you select different parts of the document: just click so you have selected no text, move the curso to the top of the document or the bottom of the document, or select a single word, several paragraphs, a picture, part or all of a table, text in the header or footer, text in a footnote or endnote or comment, and so on. Close your file, load it as an add-in and test your macro to ensure that the macro works if there is no document open. Use If Documents.Count > 0to verify if a document is open. Learning About the Word Object Model The macros created to this point have used three important Word elements: Options. This refers to the Options used to set up Word. Documents. This refers to all the Documents currently open in Word. You can use the keyword ActiveDocument to refer to the currently-active document. Selection. This refers to whatever the user has selected in the ActiveDocument. It might be one word, several paragraphs, a picture, or any other kind of content. For each of these macros, the following lower-level elements were used: These elements (Options, Documents, and Selection) are known as Objects, and together they are part of the Word object model. An object model is a hierarchy. The hierarchy includes all the parts of Word that you might need a macro to control. For example, the ActiveDocument object refers to the currently-active document. If your current document that is contained a table with at least 3 rows and at least 2 columns, then the following would display the text in the second cell in the third row of the first table in the active document. A macro can manipulate an object (such as the Selection object or the ActiveDocument object). Actually, objects have two ways to manipulate them: methods and properties. Methods are like verbs, they describe action. For example: Properties are like adjectives, they describe objects. For example: ActiveDocument.Paragraphs(1).Range.Text = “New text for Paragraph 1” In the first example here the macro is “reading” the Text property and displaying the text in a message box. In the second example, the macro is “writing” (or setting) the Text property. Running that line of code will change the existing text in the first paragraph to “New text for Paragraph 1”. In general, your macro can “read” a property or it can “write” a property. However, some properties are readonly. For example, the Documents object has a property named .Count. Consider the following: MsgBox Documents.Count ' Reports the number of currentlyopen documents. Documents.Count = 4 ' Does not work: .Count is a readonly property. Writing Shortcut Code: A Macro to Toggle the Display of Text Boundaries When I am editing a big document, I like to see the text boundaries in my document. When I am proofreading, I prefer to turn off text boundaries. I may turn the display of text boundaries on and off many times in a day. To do that, I click the File button, and under Help, I click Options. In the Word Options dialog, I click Advanced, and then under Show Document Content, I select (or clear) Show Text Boundaries and then click OK. Doing that sequence of mouse clicks repeatedly is frustrating. I would prefer to click one time to turn text boundaries on or off. The following code example is a macro that is used to toggle the display of text boundaries. Sub ToggleTextBoundaries1() ActiveDocument.ActiveWindow.View.ShowTextBoundaries = Not ActiveDocument.ActiveWindow.View.ShowTextBoundaries End Sub Word does not mind that the single line of code is so long, but it is much to type, and can be difficult for humans to read. To shorten the macro and make it easier to read, use the With keyword as shown in the following code example. Sub ToggleTextBoundaries2() With ActiveDocument.ActiveWindow.View .ShowTextBoundaries = Not .ShowTextBoundaries End With End Sub Within the object model, you can “drill down” from one object to another using a period (“.”) between each item of the hierarchy. Because the period marks off a child object from its parent, you cannot start a line of code with a period—except when using the With keyword. Each With must be paired to an End With statement. Between the With and End With, you can use a period to begin a shortcut. With ActiveDocument.ActiveWindow.View ' Between With and End With you can use ' a period to begin shortcut code. ' The shortcut will refer to ActiveDocument.ActiveWindow.View. End With For example, between this With and End With, the shortcut .ShowTextBoundaries refers to When you have code that refers to the same object several times, it will be less to type, and easier to read, if you use With and End With. This is not yet a robust macro, but you can make a final solid version as shown in the following code example. Sub ToggleTextBoundaries() If Documents.Count > 0 Then With ActiveDocument.ActiveWindow.View .ShowTextBoundaries = Not .ShowTextBoundaries End With End If End Sub In this final version of ToggleTextBoundaries, there is a With…End With block nested in an If…End If block. By indenting the lines, you ensure that each With is matched to its End With, and each If statement is matched to its End If statement. Using With and the Selection Object: A Macro to Insert a Landscape Section Here is a more complete example using the With keyword: a macro to insert a landscape section at the cursor. This macro uses the With keyword with the Selection object (that is, With Selection). As you become more skilled in writing macros, you will find that you use the Selection object less and less, because it is often more effective to use a Range object. Studying this macro is a good way to learn how Word manages the Selection. With Word and the VBE tiled side by side, click within the macro and press F8 repeatedly to step through the macro line by line, to see what happens in the main Word window. This macro also shows that, when you have many lines of text it is recommended that you break them up into logical ‘paragraphs,’ with comments to describe what the code is intended to do. Public Sub InsertLandscapeSectionHere() ' Purpose: Insert a landscape section at the insertion point, ' and insert text to tell the user where the landscape section is. If Documents.Count > 0 Then ' The user has a document open, so insert a ' landscape section. With Selection ' Do not accidentally over-write selected text .Collapse Direction:=wdCollapseStart ' Insert section breaks with blank paragraphs ' in the new section. .TypeParagraph .Style = ActiveDocument.Styles(wdStyleNormal) .InsertBreak Type:=wdSectionBreakNextPage .TypeParagraph .TypeParagraph .TypeParagraph .InsertBreak Type:=wdSectionBreakNextPage .MoveUp Unit:=wdLine, Count:=3 ' Set the orientation of new section to landscape. .PageSetup.Orientation = wdOrientLandscape ' Provide guidance to the user. .TypeText Text:="Your landscape section starts here." End With Else ' Tell the user what to do. MsgBox "Please open a document and try again." End If End Sub You will see two kinds of syntax in this macro: The first syntax is uses an equal sign (=): Selection.PageSetup.Orientation = wdOrientLandscape The second syntax uses a colon and an equal sign (=): For more information about the differences between these syntax types, see the article “Understanding Visual Basic Syntax” in the VBE Help. Learning More About Word Macros and the Visual Basic Editor The Visual Basic Editor (VBE) includes tools to help you write your own macros. The following is a list of things to consider: The Word macro recorder is a powerful tool. If you want to write a macro, but do not know which parts of the Word object model that you must use, the macro recorder is the best way to find out. For more information, see Recording a Macro to Generate Code and How to modify a recorded macro. When you type the name of an object in the Word object model (such as Selection or ActiveDocument), and then type a period (.),Word will show you a list of all the properties and methods of that object. For more information, see Getting to grips with VBA basics in 15 minutes. The Immediate Window in the VBE can help you learn more about the Word object model and test simple macros. For more information, see Extending Word 2007, Excel 2007, and PowerPoint 2007 with a Single Line of VBA Code. You can customize the menu and toolbars in the VBE. One useful customization is to create keyboard shortcuts for comments. To do this, right-click in open space near the menu or toolbars and select Customize. In the Customize dialog, on the Toolbars tab, select the Edit toolbar. Leave the Customize dialog box open. On the Edit toolbar, right-click the Comment block button, change the name to &Comment, then click Image and Text. Do the same to the Uncomment block button, renaming it as &Uncomment. You can now select several lines of code and use the keyboard shortcuts Alt+C and Alt+U to comment, or uncomment, the selected code. The VBE includes a large amount of Help information. This will be your main source of information about the Word object model: For general information, in the VBE, on the Help menu, click Microsoft Visual Basic for Applications Help. Be sure that you display the Table of Contents by clicking the little book icon in the toolbar of the client help viewer. It makes browsing the developer help much more intuitive. In particular, under Concepts, read the topics at Frequently Asked Visual Basic Questions. For specific help about a keyword or an object from the Word object model (such as ActiveDocument), click in the word and press F1 to read help about that keyword or object. Word macros can be long, complex, and powerful. You could write a macro to talk to an external database, perform calculations and return the result to Word. You can also write a macro in Word to control Excel, PowerPoint, or Outlook. Sometimes simple macros can be useful and save you time. If your macro is to format text, it may be better to create a Style instead of writing a macro. If the macro is to control the layout of a document, it may be better to create a new template. Also, there may be a built-in command that you can add to the Quick Access Toolbar (when customizing the Quick Access Toolbar, in the Choose Commands From: list, click All Commands). Have fun creating macros, but do not re-invent the wheel by writing a macro to do what Word can already do on its own. For more information, see the following resources: About the Author Shauna Kelly develops Word add-ins and templates to help people in government and business use Word more efficiently. She has been a Microsoft Word MVP since 2002.
Lakes, Streams & Watersheds In addition to its position at the center of the Great Lakes Basin—shorelines on four of the five Great Lakes—Michigan boasts more than 11,000 inland lakes, thousands of miles of rivers and streams and more than 80 major watersheds. Michigan State University Extension’s focus on lakes, streams and watersheds brings together educators and researchers with local, state, regional and federal agencies and stakeholder organizations to protect and promote the sustainable use of the state’s distinctive aquatic resources. May 22, 2013 | Steve Stewart | Teachers attending the 2013 Great Lakes Conference at Michigan State University shared their best practices, and their needs, regarding Great Lakes Literacy education. May 22, 2013 | Dan O'Keefe | Two programs on Grand River fish habits and habitats are being offered in June. A free public presentation at Spring Lake Library and a weeklong camp for 6th to 9th graders includes fish sampling in Ottawa County waters. May 21, 2013 | Dan O'Keefe | The transition from spring to summer marks some of the best fishing of the year for a variety of species in local waters. With mild weather in the forecast, it should be a pleasant weekend for a family bank fishing trip or an excursion onto the big lake. May 16, 2013 | Brandon Schroeder | A growing Saginaw Bay walleye population contributes to a broader Lake Huron fish community, as well as many anglers’ catch. May 15, 2013 | Jane Herbert | Recent flooding serves as a reminder that upstream stormwater management efforts can reduce impacts at home and downstream. May 2, 2013 | Steve Stewart | The 2013 Summer Discovery Cruises season will offer new opportunities in Southeast Michigan to learn about the Great Lakes by being on the Great Lakes. Registration is open! May 1, 2013 | Ron Kinnunen | As deaths increase on Lake Michigan due to drowning by dangerous currents, three workshops are scheduled to help communities save lives of beachgoers. April 30, 2013 | Brandon Schroeder | Annual spring Lake Huron fisheries workshops bring coastal communities together with fisheries scientists and decision makers, sharing good news regarding the status of this diverse and valuable fishery. April 26, 2013 | Jane Herbert | State law regulates the process of establishing, constructing and maintaining a legal lake level. April 23, 2013 | Steve Stewart | Michigan Sea Grant Extension is working to reduce marine debris in the Great Lakes, and you can too! To find an MSU Extension expert who covers your location or to find an expert in your area of interest, please visit the MSU Extension Expert Search. To reach an MSU Extension county office, call toll-free 1-888-678-3464. Events»View more events » Date: Jun 4, 2013 Location: P.J. Hoffmaster State Park, Norton Shores, MI 49441 Date: Jun 5, 2013 Location: Grand Traverse Civic Center, 1213 W. Civic Center Drive, Traverse City, MI 49686 Date: Jun 6, 2013 Location: USDA Forest Service Office, W1900 West US-2, St Ignace, MI 49781 - Clean Boats Clean Waters - Michigan Lake and Stream Leaders Institute - N.E. Michigan Great Lakes Stewardship Initiative - Summer Discovery Cruises - Michigan Natural Features Inventory - Great Lakes Education Program - Conflict, Collaboration & Consensus in Natural Resource Issues - Be Phosphorus Smart - Michigan Certified Natural Shoreline Professional - Michigan Sea Grant
Birthplace: Oak Park, Illinois Typewriter: Corona No. 3 & No.4*, Underwood Noiseless Portable*, various Royal portables*, Halda portable (recently sold in an online auction) Hemingway started his career as a writer in a newspaper office in Kansas City at the age of seventeen. Before the United States entered the First World War, he joined a volunteer ambulance unit in the Italian army. Serving at the front, he was wounded, was decorated by the Italian Government, and spent considerable time in hospitals. After his return to the United States, he became a reporter for Canadian and American newspapers and was soon sent back to Europe to cover such events as the Greek Revolution. During the twenties, Hemingway became a memb er of the group of expatriate Americans in Paris, which he described in his first important work, The Sun Also Rises (1926). Equally successful was A Farewell to Arms (1929), the study of an American ambulance officer's disillusionment in the war and his role as a deserter. Hemingway used his experiences as a reporter during the civil war in Spain as the background for his most ambitious novel, For Whom the Bell Tolls (1940). Among his later works, the most outstanding is the short novel, The Old Man and the Sea (1952), the story of an old fisherman's journey, his long and lonely struggle with a fish and the sea, and his victory in defeat. The Ernest Hemingway Collection at the John F. Kennedy Library The Hemingway Resource Cent er Picturing Hemingway – A writ er in his time by N ational Portrait Gall ery Hemingway's home in Key West, Florida Ernest Hemingway Home and Museum
The clouds are classified in different categories depending upon the height of the cloud’s base with reference to earth’s surface. There are some types of clouds that are not categorized with respect to height. There are high, middle and low clouds and the most popular cloud is the cumulus cloud. Cumulus clouds have different types, the most common types include: - Cumulus fractus - Cumulus congestus - Cumulus pileus - Cumulus humilis - Cumulus castellanus However, note that this article deals with cumulus castellanus clouds in particular. What are cumulus castellanus clouds? On the whole, cumulus castellanus clouds are different types of cumulus clouds that can be easily identified from the multiple towers they display from their top. The vertical air movements are clearly visible in these clouds. They have got their name because of the resemblance they show with the crenellation of the medieval castles and these clouds are considered to be an indicator of thunderstorms or forthcoming showers. What height are cumulus castellanus clouds found? Cumulus castellanus can sometimes get quite tall. There are times when the top of these clouds are as high as middle level clouds or even high level clouds while maintaining their bases at a low level. They never lose their flat base and only tend to grow vertically. These clouds are formed at a height of 330ft above ground level. The fact that cumulus castellanus has a flat base clearly indicates that water vapor does not condense till the time it reaches a certain height. As water vapors move upwards, these clouds attain the fluffy shape. How are cumulus castellanus formed? Just like any other type of clouds, they are formed when air water vapors move upwards and later condense when they reach a certain height above the earth level. Certain climatic conditions are to be met for the formation of cumulus castellanus clouds. As mentioned earlier they are formed right before a thunderstorm and are also seen under fair weather conditions. Several other types of clouds like cumulonimbus clouds are formed by cumulus castellanus when they experience different weather conditions. Although they are formed very close to the earth’s surface, they can get really huge because of their tendency to rise vertically. If they keep on getting the required amount of rising air, they can then keep on rising and eventually result in thunderstorms. Cumulus clouds can attain the height of up to 20000ft if they receive sufficient moisture content in the air. What do cumulus castellanus clouds look like? Cumulus castellanus clouds are usually white or grey in color with a strong resemblance to floating cotton balls. Their fluffy shape makes them unique and easy to be indentified from other types of clouds. They do not have much width but do have vertical height which is easy to see when viewed from the side. How common are cumulus castellanus clouds? Cumulus clouds are generally known as low level clouds and contain water droplets that cause drizzle. The very first image of clouds when someone mentions clouds is very similar to cumulus castellanus clouds. They have huge projections and this is what makes them most prominent from other types of cumulus clouds and are often considered to be an indicator of thunderstorms and thus can be easily seen at that time.
In a world of plenty, famine is a very serious thing. It shouldn’t happen in the 21st century, but it is happening in Somalia. In simple terms, famine is declared when there is evidence of acute malnutrition affecting more than 30 per cent of children. In some parts of Somalia, levels have reached 50 per cent. Drought conditions tipped vulnerable people into the shocking predicament the world has witnessed over the past few months, forcing them to leave the meagre comfort of their homes to search for help and to face the awful choice of leaving weaker children behind to save the stronger ones. But the reasons that no assistance was to hand were political rather than climatic. The loudest voices blame Al Shabab, the radical anti-Western group that controls much of southern Somalia. Its suspicion of the West and its attacks on aid workers made access for UN and other Western humanitarian agencies well nigh impossible. Others blame the US government for having passed legislation to stop any of its assistance reaching areas held by Al Shabab. This intimidated aid agencies, which feared that any humanitarian assistance could be seen as ‘support for terrorists’. Still others say the famine represents a failure of the UN system brought about by the politicization of aid. The World Food Programme stopped its support to the most vulnerable people in southern Somalia just when they needed it most. As starving people turned up in their thousands, filling makeshift camps in Kenya, Ethiopia and the Somali capital, Mogadishu, those responsible started to squirm and to cover up their part in the crisis. Al Shabab now declared that it would let aid in. Its leaders claimed they had no objection to the delivery of humanitarian assistance provided it came without a political agenda. The US government provided a ‘clarification’ to say that humanitarian agencies had no need to fear prosecution if they delivered assistance to needy people under Al Shabab’s control. The World Food Programme said that all it lacked were the funds to allow it to operate. The famine in Somalia represents a collective failure of the international community to uphold one simple principle. This is that humanitarian assistance – the type of assistance that the poorest people need to avert starvation in a drought – should have no political strings attached. In an environment as fragmented and dangerous and highly politicized as Somalia, the only agencies that were able to carry on their work were those such as the International Red Cross and Red Crescent that insisted on the principle of impartiality and stuck to a mandate of giving help to the needy. The politicization of humanitarian aid, by Al Shabab and by Western governments, is the cause of this famine. The poorest and youngest people in Somalia are paying the price. This first appeared in our award-winning magazine - to read more, subscribe from just £7
COLLEGE STATION, - Forty million years ago, rodents from Africa may have colonized South America by rafting or swimming across the Atlantic, Texas A&M University biologists theorize by studying the evolution of rodents, looking at their genes instead of their fossils - an approach that promises to revolutionize the field of evolutionary biology. "We have good evidence that suggests that South America was founded by a single ancestral stock of caviomorph relatives from Africa," says Rodney Honeycutt, a Texas A&M professor of biology who has been studying the evolution of rodents for the last eight years. "The radiations of these rodents to South America are too young for continental drift to have played a role in their colonization by African ancestors. This presents a dilemma, because either caviomorph ancestors dispersed to South America from Africa over water (e.g., by rafting) or the caviomorph radiations are considerably older than suggested by the paleontological evidence. We're seeking answers by using molecular data to address these questions." Honeycutt, his Ph.D. student, Diane Rowe, and a former student, Ron Adkins from the University of Massachusetts, are using genetics to study the evolution of South American and African rodents. The approach has been to sequence specific genes from several diverse rodent species and then use changes in these genes to reconstruct the evolutionary history of these rodents. Scientists sequenced genes of several rodents and looked at the changes - or mutations - among the genes. Then they compared the number of changes between the genetic sequences of the rodents and constructed the rodents' evolutionary tree. "We found that the South American and African radiations were unique but they do share a common ancestry," Honeycutt says, adding that the time when the two groups diverged may be considerably older (45 million years) than what the fossils are suggesting (36 million years). Contact: Judith Whtie Texas A&M University
There’s a veritable menagerie of alien worlds out there and astronomers may have spotted the weirdest yet: An exoplanet that ‘oozes.’ The exoplanet in question is 55 Cancri e and it orbits a star 40 light-years from Earth. It is also very well known to astronomers — having been detected in 2004. In the last few years, our understanding as to the nature of 55 Cancri e has evolved and due to its rapid orbit (of only 18 hours) — which takes it across the star’s disk slightly dimming some starlight from view — this world is easy to observe. Astronomers have also worked out its mass and physical size arriving at the conclusion that 55 Cancri e belongs to the rocky “super-Earth” class of exoplanets. It is nearly nine-times the mass of Earth and twice its radius. Also, due to the close proximity to its parent star, astronomers have long assumed 55 Cancri e to be a barren and hellishly hot world. But now, due to observations by NASA’s Spitzer Space Telescope, all may not be as it seems. Even though 55 Cancri e has an orbit 26-times closer than Mercury’s orbit around the sun, Spitzer has discovered that a fifth of the planet’s mass must be composed of light elements and compounds, including water — potentially liquid water. But don’t go getting the idea that 55 Cancri e is covered in oceans lapping alien shores, this water is like nothing we experience in nature. Any liquids that do exist on or near the 55 Cancri e surface will be in a supercritical fluid state. Usually (at sea level pressures on Earth) water will turn from a liquid to a gas at 100 degrees Celsius (373 Kelvin) — i.e., the temperature water boils in a kettle. However, if you try to boil water in a high pressure environment, the boiling temperature increases, allowing the water to remain in a liquid state beyond 100 degrees C. Supercritical fluids act as natural solvents in industry — water becomes supercritical in power plant steam turbines and supercritical carbon dioxide is used to remove caffeine from coffee beans, for example. This is what scientists believe is happening on 55 Cancri e; although the surface temperature may top 1,000 degrees C, supercritical solvents are oozing from the rocks on its surface. So 55 Cancri e may not be the barren, hot world it was once assumed to be. But it is a world from hell — supercritical liquid rising to its molten surface, producing a hot, steamy atmosphere detectable from 40 light-years away. Source: NASA Science News Image: A size comparison of 55 Cancri e and Earth. Credit: NASA/JPL-Caltech
Friends may be just the medicine that one needs. A study out of the University of California, San Francisco found older people who feel lonely could find it affects their health. Researchers looked at 1,600 people with an average age of 71. Those who felt lonely are 60% more likely to lose the ability to do the activities needed for daily living. 45% were more likely to die. Carla Perissinotto with UC San Francisco shares how you can help them avoid loneliness, “Examine what is going on in their life. Think about the quality of relationships they have and focus on increasing social connectivity.” Start nurturing those relationships and expanding the horizons early so they are in place when needed. The US Department of Health and Human Services Health Beat has tips on how to manage stress and deal with depression. Listen below as News 92 FM’s Pattie Shieh shares additional information:
MIT reports boosting the efficiency of fuel cells by 100 times by virtue of a cheap yet super-efficient new electrode material. Look for in-home fuel cells that charge during the day from sunlight, then power you home at night by the end of the decade. R.C.J. Massachusetts Institute of Technology (MIT) has licensed to Sun Catalytix a durable and low-cost material for the oxygen-producing electrode in a fuel cell. The cobalt and nickel-borate-based electrode materials enable the creation of more efficient fuel cells that have increased their rate of oxygen production from catalysts by one-hundredfold. The inexpensive cobalt and nickel-borate-based compounds eliminate the current need for expensive platinum catalysts and dangerous chemicals, such as the concentrated alkali used in commercial systems. Sun Catalytix promises safe, super-efficient versions of the system, suitable for homes and small businesses within two years. Full Text: http://bit.ly/NextGenLog-cY38
Each month, the NSF Current newsletter highlights research and education efforts supported by the National Science Foundation. If you would like to automatically receive notifications by e-mail or RSS when future editions of NSF Current are available, please use the links below: The use of prescribed burns to manage western forests may help the U.S. reduce its carbon footprint. Results of a new study, funded by the National Science Foundation (NSF), show that such burns, often used by forest managers to reduce underbrush and protect bigger trees, release substantially less carbon dioxide than wildfires of the same size. The results were published last month in the journal Environmental Science & Technology. "It appears that prescribed burns can be an important piece of a climate change strategy," says Christine Wiedinmyer of the NSF-funded National Center for Atmospheric Research (NCAR) in Boulder, Colo. "If we reintroduce fires into our ecosystems, we may be able to protect larger trees and significantly reduce the amount of carbon released into the atmosphere by major wildfires." Forests have emerged as important factors in climate change. Trees store, or sequester, significant amounts of carbon, thereby helping offset the large amounts of carbon dioxide emitted by vehicles, factories and other human sources. When trees burn, much of that carbon is returned to the atmosphere. It can take decades for forests to sequester the amount of carbon emitted in a single fire. Wiedinmyer and her colleagues found that widespread prescribed burns can reduce fire emissions of carbon dioxide by 18 to 25 percent and by as much as 60 percent in certain forests. Read more here. NSF-funded investigators at UCLA and the University of California (UC), Irvine, have teamed up with police in Los Angeles to find ways to fight the spread of crime. The researchers developed a mathematical model that has revealed the existence of two types of crime hotspots: super-critical hotspots that give rise to a rapid chain reaction of lawlessness, and sub-critical ones that result in large spikes of criminal activity in an otherwise stable area. Super-critical hotspots form from small crime spikes spread over a large area. When police force is applied to this type of hotspot, new spots of crime bud off from existing ones. These buds form around small spikes of activity away from police presence, and the crime continues. Police action applied to sub-critical hotspots, on the other hand, is found to suppress crime, since the area surrounding the hotspot is relatively stable. One of the investigators, Andrea Bertozzi of UCLA, explains that the model can reveal which type of hotspot may be acting in a certain part of the city. "Policing actions directed at one type will have a very different effect than policing actions directed at the other type," she says. George Tita of UC Irvine and Jeff Brantingham and Martin Short of UCLA also participated in the study. "Our model was originally built around a model for residential burglaries," explains Short. "However, based on real data, we believe the ideas in the analysis apply to a variety of crime types." Read more about the work and watch videos of the mathematical simulations here. NSF-funded investigators have recently discovered that the same physical process that gives some diamonds a stunning yellow color can also turn the sparkling gemstones into computing devices that use light, instead of electrons, to store and transmit information. The team of scientists, led by Marko Loncar of NSF's Nanoscale Science and Engineering Center (NSEC) at Harvard, developed the new device based on tiny wires constructed from diamond crystals. The diamond contains nitrogen vacancy (NV) color centers formed when a nitrogen atom, which normally turns diamonds yellow, sits next to a vacant site. NV centers can function as registers in quantum computers, storing information as quantum bits, or qubits, in place of the normal magnetic bits found in standard computing systems. Although quantum computing has been an intriguing possibility for some time, the location of the color centers deep within the diamond crystal prevented transmission of information from one location to the other. Loncar, who is also an NSF CAREER awardee, explains: "What was missing was an interface that connects the nano-world of a color center with the macro-world of optical fibers and lenses." The new diamond nanowires greatly increase the probability that quantum computing may one day become a reality. Read more about this work here. NSF-funded scientists aboard the research vessel Melville are undertaking an expedition to explore the rupture site of the recent 8.8-magnitude earthquake off the coast of Chile. The rapid response expedition is possible because the Melville had been conducting research in the Chilean coastal area when the earthquake struck. An important aspect of the mission involves swath multibeam sonar mapping of the seafloor to produce detailed topographic maps. Data from measurements of the earthquake zone will be made public soon after the research cruise ends, according to Julie Morris, division director for ocean sciences at NSF. The researchers plan to compare the new data with pre-quake data obtained by German scientists at the Leibniz Institute of Marine Sciences, with the goal of finding changes caused by the earthquake rupture. "We'd like to know if the genesis of the resulting tsunami was caused by direct uplift of the seabed along a fault, or by slumping from shaking of sediment-covered slopes," says research team member Dave Chadwell of the Scripps Institution of Oceanography at the University of California, San Diego. Daily updates from Chadwell and others on the research mission are available on the group's blog, SIOSEARCH. Read more about the project here. On Feb. 19, 2010, NSF and the University of Southern California (USC) School of Cinematic Arts announced a new partnership designed to forge collaborations between scientists and entertainment scholars with the goal of producing materials that inspire and inform audiences about science and engineering concepts. The partnership establishes the Creative Science Studio, which will provide NSF-funded researchers with novel opportunities to create engaging and informative expositions of their work for the public. In addition, USC students will gain experience with complex and cutting-edge scientific concepts. "This novel and creative partnership will enlist the power of the entertainment media to inspire audiences to learn more about science and engineering," explains Thomas Kalil, deputy director for policy in the White House Office of Science and Technology Policy. Read more about the innovative partnership here. Bonnie Bassler, NSF-funded investigator at Princeton University, is fascinated by the social lives of bacteria. She has discovered that bacteria communicate using a chemical language. Her realization, that these tiny organisms form a social network and communicate with each other using a process called quorum sensing, has transformed the way biologists think about bacteria and may one day lead to new types of antibiotics. Quorum sensing begins when bacteria release small molecules into the surrounding medium. These molecules are detected by neighboring bacteria that are essentially able to "count" how many of their brethren are present by the type and strength of the chemical signal. Some chemical signals allow bacteria to talk only with neighbors just like them, while other molecules allow communication between bacterial species. The ability of bacteria to count their numbers is crucial to their virulence. In order to infect a host organism, bacteria do not want to begin releasing toxins too soon, and quorum sensing allows them to restrain their attack until enough of their numbers are present, so they have some hope of overcoming the host's immune system. Bassler has shown that this basic communication mechanism is common to some of the world's most virulent microbes, such as those responsible for cholera and plague. Bassler's work in bacterial communication grew from her interest in finding out how information flows between the cells in the human body. Quorum sensing, she says, allows a colony of bacteria to act essentially like a multi-cellular organism. "If we can understand the rules or paradigms governing the process in bacteria," she says, "what we learn could hold true in higher organisms." It seems appropriate that a scientist who is committed to understanding another species' communication methods would be devoted to communicating with her own species. Bassler is very involved in talking with the public about science and has developed a series of presentations for a lay audience, including one highly popular TED lecture. She also presented the 2009 Howard Hughes Holiday Lecture for teachers and educators. In addition, Bassler has worked with the Liz Lerman Dance Exchange to develop a traveling dance piece entitled Ferocious Beauty Genome, with the Baltimore Science Center on the permanent exhibit, "Cells: the Universe Inside Us," and with a whole host of other organizations including NOVA scienceNOW, NPR and others. To read more about Bassler's work, visit her Web site. Climate Change Prediction Project Launched (United Press International and Scientific American) NSF has teamed up with the U. S. Departments of Energy and Agriculture to develop higher-resolution computer models to better predict climate change and guide policy and decision makers. Modern Dogs Likely Descended From Middle Eastern Wolves (New York Times, NPR, ABC News and others) By comparing DNA from dogs and wolves, NSF-funded researchers from UCLA have shown that domesticated dogs can be traced back to wolves living somewhere in the ancient Middle East. Plan to Fertilize Ocean With Iron Could Lead to Ecological Disaster (London Free Press) Research partially funded by NSF has shown that adding iron to the ocean to halt global warming encourages the growth of toxic algae. Iron doubles the amount of toxin produced, threatening other marine life. AIDS Virus Can Hide in Bone Marrow (Associated Press) The AIDS virus can hide in bone marrow, avoiding drugs and surfacing later to cause illness, according to research partially funded by NSF. Language DVDs Don't Promote Learning in Kids (The Guardian) A new study, funded by NSF, has shown that exposing young children to vocabulary-building DVDs does not improve language skills. The Census Bureau estimates that twenty percent of Americans have some type of disability, about half of which are considered severe. With an aging population, that percentage is expected to rise over the coming decades. The Quality of Life Technology (QoLT) Center, an NSF Engineering Research Center jointly managed by Carnegie Mellon University and the University of Pittsburgh, is working to turn cutting-edge discoveries in computing, engineering, robotics and other fields into new devices and tools that can improve disabled peoples' lives. In March, NSF announced a $1.5 million Innovation Award to the QoLT Center to help start-up companies affiliated with the center move these new technologies to the marketplace. The new advances will help millions of Americans continue to be productive and successful members of society. A new exhibit that explores what it means to be human has recently opened at the Museum of Natural History at the Smithsonian Institution in Washington, D.C. The exhibit, partially funded by NSF, provides detailed and up-to-date research findings of the Human Origins Initiative at the Smithsonian. The initiative's goal is to explore the universal human story over a broad time scale and to stimulate new research that will deepen our understanding of what makes us unique as a species and how we came to be. For more information, explore the exhibit's Web site. This winter marked the tenth anniversary of the High Performance Wireless Research and Education Network (HPWREN), a wireless technologies research platform managed by the University of California, San Diego. Centered in California's San Diego, Riverside and Imperial counties and supported by NSF, HPWREN looks into ways wireless technologies can better serve academic researchers and the broader public. In addition to linking up academic institutions ranging from elementary schools to national supercomputer centers, police and firefighters use the network to work across jurisdictions. Native reservations use it to break through geographic isolation. And during the area's historic wildfire season a couple of years ago, evacuated residents used the system's scores of cameras to see if their houses were safe. HPWREN has demonstrated how wireless networks can make communities smarter, safer and more cohesive.
CO2 is so highly persistent and therefore cumulative in the atmosphere that so long as any extra 'unnatural' CO2 is being constantly added to the atmosphere it must build up/accumulate in the atmosphere. The big thing about CO2 is that a lot of lasts in the atmosphere practically 'forever'. The answer is in what happens to all the CO2 being pumped into the atmosphere - constantly. Since 1750, 329 billion tons of carbon dioxide have been poured into the atmosphere from burning fossil fuels. It does not just vanish into thin air. Like methane and nitrous oxide CO2 is called a' long lasting' GHG (i.e i the atmosphere). Most of it is still in the atmosphere and in the oceans where it is harmful to the living planet. Why is this ? Why does 20% of CO2 emissions last in the atmosphere for thousands of years The graph opposite is a computer model run that shows what happens to a single instantaneous pulse of CO2 emitted to the atmosphere. Forty percent of the CO2 pulse disappears from the atmosphere in ten years- mainly being absorbed by the oceans. But after that, the removal of the CO2 pulse from the atmosphere is very slow. After one hundred years a third of the CO2 is still in the atmosphere radiating heat energy. One thousand years after the CO2 is emitted to the atmosphere 20% is still there. The reason why so much of the carbon dioxide emissions stay in the atmosphere 'like for ever' is that our planet can only remove CO2 (really sink it) from the atmosphere in a permanent sense, which means removing from the carbon cycle at an ultra slow rate. It takes thousands to hundreds of thousands of years for the planet to remove and sink carbon from the carbon cycle. There are two ultra long-term/super slow carbon cycles by which the planet takes carbon out of atmospheric carbon dioxide. One is land based and the other is ocean based. They are both ultra slow processes happening over thousands to hundreds of thousands of years. They both operate by dead organisms literally sinking and being compressed into liquid or sold matter. The land sink forms what we call fossil fuels in the depths of the soil and rock. The same sort of thing happens deep below the ocean floor where dead marine organisms form limestone and dolomite rock which in the course of ages of time has formed rock on land. The graph is an illustration of climate change computer model runs. The computer model is run for the equivalent of 1000 years. Starting at today on the graph global carbon dioxide emissions are rapidly reduced to virtual zero. The rate of the carbon dioxide emissions are constantly reduced till they reach virtual zero, which is about 90% on this graph. As emissions reduce the rate of increase of atmospheric CO2 slows to a plateau, but it never falls even with CO2 missions at a sustained virtual zero. The rate of increase of global temperature lagging behind the slowing increase fate of atmospheric CO2. Even at 1000 years the temperature has not stopped increasing. At 1000 years there is no sign of the rate of sea level rise slowing as land ice continues the melt more rapidly. How then can we ever stop global warming and climate disruption? First we have to stop emitting carbon dioxide. Everything else depends on doing that rapidly. However even then there will be some increase in the carbon emissions, making the best that we can achieve virtual zero carbon emissions. To get to zero, or net zero, it is necessary to sink carbon over the very long term or extract some carbon dioxide out of the air. This necessity of 'negative carbon emissions' or the development of 'artificial carbon sinks 'is another definite conclusion of the climate change scientific published research. A 2011 paper by J. Hansen and many other experts suggests that increasing the capacity or forests and agriculture to take up carbon could provide a carbon sink These measures certainly be tried and started right away. Biochar production from wood does have the potential to sink carbon for hundreds of years. This graph published by the IPCC in its 2001 assessment shows all of the essential aspects of the climate change science that all results from the fact that about 20% of all carbon dioxide emitted to the atmosphere lasts there radiating heat for 1000 years. The ocean carbon sink process is initiated and depends on the plankton biological carbon pump. Green phytoplankton take up about half of the total to carbon taken up by the planet's green plants. The carbon captured by the phytoplankon becomes incorporated into shell bearing marine organisms that sinking to the ocean bottom are incorporated into the ocean sediments. only zero carbon only zero carbon only zero carbon only zero carbon only zero carbon only zero carbon only zero only zero A CO2 molecule. CO2 is an extremely long lasting GHG. For this reason, not only do CO2 emissions accumulate in the atmosphere, a large proportion that will persist for 1000 years is accumulating. In a practical sense that increasing proportion is there for ever. A large proportion (over 30%) of all the hundreds of billions of tons of CO2 that has been emitted is in the oceans as carbonic acid and that will remain in the oceans for thousands of years. Only zero carbon emissions can result in the stabilization of atmospheric CO2 Why is CO2 so persistent in the atmosphere? The ultra long term slow carbon cycles that sink carbon from the atmosphere o In the air - atmosphere o On land - plants, soil, rocks o In the oceans The carbon cycle on the land just continually re-cycles carbon. There is no long term land carbon cycle anymore because we are burning fossil carbon. Green plants absorb CO2 from the atmosphere to grow and when they die the CO2 is emitted back to the atmosphere. It is a short term carbon cycle that cannot help us in lowering the atmospheric CO2 concentration. Planting trees is essential for several environmental reasons but won't work for off setting carbon emissions. If there was a massive global reforestation program atmospheric CO2 would be absorbed and could buy us a few decades to get to virtual zero and take carbon out of the air by other means.
Hard to believe, but it was 50 years ago today that Russian cosmonaut Yuri Gagarin became the first man from Earth to go into outer space. At the time, the news of tiny April 12, 1961 Vostok spacecraft which orbited the Earth made Yuri Gagarin an instant world celebrity and set a challenge for the young Kennedy Administration to set a challenge to land man on Earth's moon before the end of the decade. MiG pilot Yuri Gagarin also had a history of faith in God, making him somewhat unique among the other 19 potential cosmonaut candidates. He even had his daughter baptized at the Russian Orthodox church that the family sometimes attended shortly before his flight into space. His family celebrated both Christmas and Easter. Gagarin reportedly asked his wife to remarry if he should die during his flight into space. However, the first manned mission in space proved both successful and safe and Gagarin returned safely. However, seven years after his trip into space Gagarin and a co-pilot were killed when something went wrong in a routine training flight in a MiG-15UTI. Gagarin's body was cremated and buried in the walls of the Kremlin as a national hero of Russia. Today, the crew of the international space station sent words of praise today for Yuri Gagarin's great achievement in Russian, English and Italian. A sign of how much the world has changed since the April 12, 1961 flight which sent Cold War shivers through the U.S. at the time, but is viewed as a pioneering space exploration achievement today. Yuri Gagarin's achievement was the beginning of the future.
The question of why primates, and especially humans, have more strongly developed cognitive skills than other mammals has a long history in science. The most widely accepted notion has been that primates' superior cognitive abilities have evolved in the social realm. Many primate species live in complex societies, and, the argument goes, this favored the evolution of especially developed social skills. Although there is much empirical evidence in favor of the social-intelligence hypothesis, very little work has been conducted to address its alternative, the idea that primate cognition has evolved to deal with problems of an ecological nature, such as foraging for food. With their new work, the researchers sought to address this anomalous gap. By following a group of wild gray-cheeked mangabeys from dawn to dusk over 210 days in their natural rainforest habitat of Kibale Forest, Uganda, the scientists obtained an almost complete record of their foraging decisions in relation to their preferred food, figs. The data showed that the monkeys were more likely to revisit fig trees (in which they had found fruit before) after a period of warm and sunny days than after a period of cold and cloudy days. Temperature and solar radiation are known to accelerate maturation of fruits and insect larvae inside them. The researchers were able to show that past weather conditions--as opposed to sensory cues such as the smell of ripe fruit--accounted for the behavioral trend they observed. These findings are consistent with the idea that monkeys make foraging decisions on the basis of episodic (or "event-based") memories of whether or not a tree previously carried fruit, combined with knowledge of recent and present weather conditions and a more generalized understanding of the relationship between temperature and solar radiation and the maturation rate of fruit and insect larvae. The findings are also consistent with the idea that the evolution of primate cognitive skills has proceeded, at least in part, as a result of ecological challenges associated with foraging for intermittently available food such as ripening fruit. *For expert commentary on this work, be sure to see the related Dispatch in this issue of Current Biology from Dr. Michael Platt: "Animal Cognition: Monkey Meteorology." The researchers include Karline R.L. Janmaat, Richard W. Byrne, and Klaus Zuberbühler of the University of St. Andrews in St. Andrews, Scotland, United Kingdom. This research was funded by the Wenner-Gren and Leakey Foundation, the University of St. Andrews' School of Psychology, the Schure-Bijerinck-Popping foundation of the KNAW, the Stichting Kronendak, the Dobberke Stichting voor Vergelijkende Psychology, the Lucy Burger Stichting, and the Foundation Doctor Catharine van Tussenbroek. In Uganda, the researchers thank the Office of the President, the Uganda National Council for Science and Technology, the Uganda Wildlife Authority, the Makerere University Biological Field Station, and the Kibale Fish and Monkey Project for logistic support and permission to conduct research in Kibale National Park. Janmaat et al.: "Primates Take Weather into Account when Searching for Fruits." Publishing in Current Biology 16, 1232–1237, June 20, 2006, DOI 10.1016/j.cub.2006.04.031. www.current-biology.com Related Dispatch by Platt et al.: "Animal Cognition: Monkey Meteorology." Last reviewed: By John M. Grohol, Psy.D. on 21 Feb 2009 Published on PsychCentral.com. All rights reserved.
Investigating the development of data evaluation: the role of data characteristics. A crucial skill in scientific and everyday reasoning is the ability to interpret data. The present study examined how data features influence data interpretation. In Experiment 1, one hundred and thirty-three 9-year-olds, 12-year-olds, and college students (mean age = 20 years) were shown a series of data sets that varied in the number of observations and the amount of variance between and within observations. Only limited context for the data was provided. In Experiment 2, similar data sets were presented to 101 participants from the same age groups incrementally rather than simultaneously. The results demonstrated that data characteristics affect how children interpret observations, with significant age-related increases in detecting multiple data characteristics, in using them in combination, and in explicit verbal descriptions of data interpretations. Version: za2963e q8zac q8zb5 q8zcd q8zd8 q8ze9 q8zf8 q8zgf
Water quality is a key issue facing the restoration of the Florida Everglades. For example, agriculturally derived phosphorus entering the Everglades has caused eutrophication of 6 to 10 percent of the ecosystem, resulting in changes in the native plant communities. Phosphorus contamination remains an important issue facing Everglades restoration, but recent work by the U.S. Geological Survey (USGS) has pointed to other water-quality issues critical to the overall health of the Everglades ecosystem. Among the most important of these other water-quality issues in the Everglades is sulfate (SO42-) contamination. Sulfate contamination affects as much as one-third of the freshwater Everglades, and sulfate concentrations are 60 to 100 times historical levels in heavily affected areas. The effects of sulfate contamination on ecosystem health have not been fully studied, but links between sulfate contamination and high levels of methylmercury in the ecosystem have been documented. Also, sulfate contamination may be contributing to declines in native plants by altering chemical conditions in the sediments. Ongoing USGS research will further delineate these effects and explore possible options for mitigating sulfate contamination in this unique ecosystem.
TOO DARNED BIG TO TEST Testing large systems is a daunting task, but there are steps we can take to ease the pain. KEITH STOBIE, MICROSOFT The increasing size and complexity of software, coupled with concurrency and distributed systems, has made apparent the ineffectiveness of using only handcrafted tests. The misuse of code coverage and avoidance of random testing has exacerbated the problem. We must start again, beginning with good design (including dependency analysis), good static checking (including model property checking), and good unit testing (including good input selection). Code coverage can help select and prioritize tests to make you more efficient, as can the all-pairs technique for controlling the number of configurations. Finally, testers can use models to generate test coverage and good stochastic tests, and to act as test oracles. HANDCRAFTED TESTS OUTPACED BY HARDWARE AND SOFTWARE Hardware advances have followed Moore’s law for years, giving the capability for running ever-more complex software on the same cost platform. Developers have taken advantage of this by raising their level of abstraction from assembly to first-generation compiled languages to managed code such as C# and Java, and rapid application development environments such as Visual Basic. Although the number of lines of code per day per programmer appears to remain relatively fixed, the power of each line of code has increased, allowing complex systems to be built. Moore’s law provides double the computing power every 18 months and software code size tends to double every seven years, but software testing does not appear to be keeping pace. Unfortunately, the increased power has not, in general, made testing easier. While in a few cases the more powerful hardware means we can do more complete testing, in general the testing problem is getting worse. You can test square root for all the 32-bit values in a reasonable time,1,2 but we are now moving to 64-bit machines with 64-bit values (and even longer for floating point). Assuming a nanosecond per test case, it would take 584 years to test all the values. Sure, you could scale to, say, 1,000 processors, but you would still need six months for just this one test case. Two other issues of complexity to consider are: the number of different execution states software can go through, and concurrency. User interfaces were originally very rigid, letting users progress through a single fixed set of operations—for example, a single set of hierarchical menus and prompts. Now, good user design is event driven with the user in control of the sequence of actions. Further, many actions can be accomplished in multiple manners (e.g., closing a window via the menu [File/Close], a shortcut key [ALT-F4], or the mouse [click on the Close icon]). Multiple representations of items (e.g., file names) also generally mean multiple execution paths. A manifestation of this is the security issue that occurs when an application makes wrong decisions based on a noncanonical representation of a name.3 These explosive combinations make it virtually impossible for a tester to both conceive of and then test a reasonable sampling of all the combinations. Concurrency is becoming more prevalent at many levels. Beyond multiple processes, we see increasing lower-level concurrency, especially as a result of multicore chipsets that encourage parallelism via threading and hyperthreading; soon, everyone will have a dual-processor machine. Machine clustering and Web services are increasing the concurrency across applications. All of this creates greater possibilities for more insidious bugs to be introduced, while making them harder to detect. MISUNDERSTANDING CODE COVERAGE AND STOCHASTIC TESTING Before discussing possible approaches to these problems, let’s clear up some pervasive misconceptions. Myth: Code coverage means quality. Tools for measuring structural code coverage have become increasingly easy to use and deploy and have thus deservedly gained greater acceptance in the testing arsenal. A fundamental flaw made by many organizations (especially by management, which measures by numbers) is to presume that because low code-coverage measures indicate poor testing, or that because good sets of tests have high coverage, high coverage therefore implies good testing (see Logical Fallacies sidebar). Code coverage merely measures that a statement, block, or branch has been exercised. It gives no measure of whether the exercised code behaved correctly. As code gets too big to easily roll up qualitative assessments, people start summarizing code coverage instead of test information. Thus, we know low code coverage is bad, because if we have never exercised the code, it appears impossible to say we have “tested” it. However, while having something that just executes the code may consequently reveal a few possible flaws (e.g., a complete system failure), it doesn’t really indicate if the code has been what most people consider tested. Tested indicates you have verified the results of the code execution in useful ways. The fallacy of high code coverage will become more evident with the next generation of automatic unit testing tools (discussed later) that can generate high code coverage with no results checking. Further, results of real software comparing code coverage and defect density (defects per kilo-lines of code) show that using coverage measures alone as predictors of defect density (software quality/reliability) is not accurate. Denying the Antecedent: Myth: Random testing is bad. One of the big debates in testing is partitioned (typically handcrafted) test design versus operational, profile-based stochastic testing (a method of random testing). Whereas many organizations today still perform testing (such as choosing the partitions) as a craft (either manually or semi-automatically), it is unclear that this is the most effective method, especially with the increasing complexity of software. Current evidence indicates that unless you have reliable knowledge about areas of increased fault likelihood, then random testing can do as well as handcrafted tests.4,5 For example, a recent academic study with fault seeding showed that under some circumstance the all-pairs testing technique (see “Choose configuration interactions with all-pairs” later in this article) applied to function parameters was no better than random testing at detecting faults.6 The real difficulty in doing random testing (like the problem with coverage) is verifying the result. How, for random inputs, can we judge the output? It is easier to have static tests with precomputed outputs. In testing, being able to verify the output is called the test oracle problem.7,8 There are numerous current areas of work on this problem. Another problem is that very low probability sub–domains are likely to be disregarded by random testing, but if the cost of failures in those subdomains is very high, they could have a large impact. This is a major flaw of random testing and a good reason for using it as a complementary rather than the sole testing strategy. Part of the testing discipline is to understand where we have reliable knowledge of increased fault likelihood. The classic boundary value analysis technique succeeds because off-by-one errors increase the fault likelihood at boundaries. Many other analysis techniques have a less tried-and-true underlying fault model to reliably indicate increased fault likelihood. For example, partitioning based on code coverage doesn’t guarantee finding the fault, as the right input must be used to stimulate the fault and sufficient verification must be done to realize the fault has produced a failure. TEST ISSUES: DO OVER It’s clear that increasing structural code coverage doesn’t solve the testing problem and that stochastic testing is frequently as good as or better than handcrafted test cases. Given this state of affairs, how do we address the issue of “it’s too darned big to test”? There are several approaches to consider: unit testing, design, static checking, and concurrency testing. Good unit testing (including good input selection). Getting high-quality units before integration is the first key to making big system testing tractable. The simple way to do that is to follow the decades-old recommendation of starting with good unit testing. The IEEE Standard for Software Unit Testing has been around for years [Std. 1008-1987] and requires 100 percent statement coverage. This has been nicely reincarnated via the test-driven development movement, which also espouses the previously seldom-adhered-to idea of writing the tests before the code. Testing a big system is especially difficult when the underlying components and units are of low quality. Why isn’t good unit testing prevalent? In many shops the problem is partly cultural. Software developers presumed someone else was responsible for testing besides them. Some development groups have spent more effort building out stubbed components and test harnesses than on building the actual components. Standard unit testing harnesses now exist for many environments, and the creation of mock objects is becoming semi-automated. Another reason for poor unit testing is that software was once simpler and reliability expectations lower. In today’s world insecure systems arising from simple problems (for example, buffer overruns) that good unit testing could have caught have really motivated management to drive for better unit testing. A variety of tools exist in this arena. Unit testing frameworks include the XUnit family of tools, Framework for Integrated Test (Ward Cunningham’s Fit), Team Share, etc. Further, IDEs (integrated development environments) such as Eclipse and Rational can generate unit test outlines. Any development shop should be able to find and use applicable ones. In general, today’s unit testing requires developers to handcraft each test case by defining the input and the expected output. In the future, more intelligent tools will be able to create good input coverage test cases and some expected output. We may see 100 percent code coverage, with no results checking. We will need to become more aware of the thoroughness of results checking instead of just code coverage. How to measure results-checking thoroughness hasn’t really been addressed. Current automatic unit-generation test techniques work from the simple known failure modes (such as null parameters as inputs) to analysis of the data structures used. Data-structure generation can be based on constraint solvers, models, and symbolic execution. Other tools can enhance an existing set of tests. For example, Eclat creates an approximate test oracle from a test suite, then uses that oracle to aid in generating test inputs that are likely to reveal bugs or expose new behavior while ignoring those that are invalid inputs or that exercise already-tested functionality. Even without tools, just training people in simple functional test techniques can help them choose better inputs for their unit testing. You can also start with generic checklists of things to test and then make them your own by recording your organization’s history of what makes interesting input. You can find industry checklists in several books under different names such as the following: “Common Software Errors” (Kaner, C., Falk, J. and Nguyen, H.Q. 1993. Testing Computer Software. Van Nostrand Reinhold); “Test Catalog” (Marick, B. 1994. Craft of Software Testing. Prentice Hall PTR. www.testing.com/writings/short-catalog.pdf); or “Attacks” (Whittaker, J.A. 2002. How to Break Software: A Practical Guide to Testing. Addison Wesley); as well as on the Web (for example, http://blogs.msdn.com/jledgard/archive/2003/11/03/53722.aspx). Good design (including dependency analysis). Many books and papers explain how to do good up-front design (the best kind), but what if you already have the code? Dependency analysis can be used to refactor code (rewriting to improve code readability or structure; see http://www.refactoring.com/). Many IDEs are incorporating features to make code refactoring easier. Good unit tests, as agile methods advocate, provide more confidence that refactoring hasn’t broken or regressed the system. Dependency analysis allows testers to choose only those test cases that target a change. Simplistic analysis can be done by looking at structural code dependencies, and it can be very effective. More sophisticated data-flow analysis allows more accurate dependency determination, but at a greater cost. Either technique can be a big improvement over just running everything all the time. A test design implication of this is to create relatively small test cases to reduce extraneous testing or factor big tests into little ones.9 Good static checking (including model property checking). The convergence of static analysis tools with formal methods is now providing powerful tools for ensuring high-quality units and to some extent their integration. For example, PREfix is a simulation tool for C/C++ that simulates the program execution state along a selected set of program paths and queries the execution state to identify programming errors.10 The issue creating the greatest resistance to adoption for most static checking remains the false positive rate. Some tools can only make reliable guesses and sometimes guess wrong. Developers get frustrated if they can’t turn off the tool for the places they know it is wrong. Sometimes developers get carried away and turn off too much, then miss problems the tool could have caught. Some developers don’t believe the issues found are really bugs even when they are. Many of the more advanced checks require additional annotations to describe the intent. You can emulate some of the stricter checks in any language just using compiler or language constructs such as assert(). It is still a challenge to convince many developers that the extra annotations will pay off in higher-quality, more robust, and more maintainable code. Others, however, who have used it and seen nasty issues that would have taken hours, days, or weeks to debug being found effortlessly up front, would never go back. All of this data can also help independent testers later because defects found through static analysis are early indicators of pre-release system defect density. Static analysis defect density can be used to discriminate between components of high and low quality (fault-prone and non-fault-prone components). Concurrency testing. Concurrency issues can be detected both statically (formal method properties such as liveness, livelock/deadlock, etc.) and dynamically (automatic race lockset tracking). Static detection frequently requires extensive annotation of the source code with additional information to allow the necessary reasoning to be done about the code. The advantage of static analysis is its potential to uncover all race conditions. Dynamic race detectors, such as code coverage tools, require a good set of test cases as input because they look only at what the tests force to occur. They can, however, determine race conditions even though the test itself didn’t force it. As long as the test executes the code of the overlapping locks, it is likely to be detected. Several tools, such as Java PathExplorer (JPAX), have been created that extend the Eraser lockset algorithm.11 Besides no guarantee of completeness without complete test cases, the other issues still being resolved with dynamic race detectors include their performance and the number of false positives they generate. More traditional methods of concurrency testing involve controlling the thread scheduler, if possible, to create unusual timings. Fault injection tools, such as Security Innovation’s Holodeck, can be used to create artificial delays and increase the likelihood of race conditions. As a last resort, the old standby of just running lots of activity concurrently under varying conditions has been fruitful in finding defects (although perhaps not as efficiently as many of the newer methods). The previous section helps set the stage at a basic level with units of code and analysis, but most test organizations deal with testing the integrated whole. Testers have to be creative in addressing the overwhelming size of today’s software systems. When it’s too darned big to test, you must be selective in what you test, and use more powerful automation such as modeling to help you test. Fault injection tools again come in handy as a way to tweak things when reaching through the overall big system becomes daunting (which it does quickly!). Use code coverage to help select and prioritize tests. Prioritizing test cases has become increasingly practical in the past two decades. If you know the coverage of each test case, you can prioritize the tests such that you run tests in the least amount of time to get the highest coverage. The major problem here is getting the coverage data and keeping it up to date. Not all tools let you merge coverage data from different builds, and running coverage all the time can slow down testing. The time to calculate the minimization can also be a deterrent. For many collections of test cases, running the minimal set of test cases that give the same coverage as all of the test cases typically finds almost all of the bugs that running all of the test cases would find. Industrial experience at several companies appears to confirm this. For example, for an industrial program of 20,929 blocks, choosing tests only by function call coverage required only 10 percent of the tests while providing 99.2 percent of the same coverage. Reducing by block coverage meant that 34 percent of the tests provided the same block coverage and 99.99 percent of the same branch (segment) coverage. Further, there were no field-reported defects found that would have been caught by the full set of tests but missed by the coverage-reduced set of tests. Best of all, anyone can easily run the experimental comparison. First run the minimal set of tests providing the same coverage as all of the tests, and then run the remaining tests to see how many additional defects are revealed. You can also combine this with dependency analysis to first target only the changed code. Depending on the sophistication of your dependency analysis, you may not have to do any further testing. Use customer usage data. Another great way to target testing is based on actual customer usage. Customers perceive the reliability of a product relative to their usage of it. A feature used by only a few customers on rare occasions will have less impact on customer satisfaction than bread-and-butter features used by virtually all customers all of the time. Although development teams can always make conjectures about product usage, which makes a useful start, actual data improves the targeting. Ideally, you can automatically record and have actual customer usage sent to you. This approach has several problems, including privacy, performance, and security. Each of these can be dealt with, but it is nontrivial. Alternatively, you may have only a select few customers reporting data, you may do random sampling, or you may fall back on case studies or usability studies. You can also work with your customers for cooperative early testing either through beta programs or even earlier pre-release disclosures.12 Customer usage data is especially important in reliability testing where an operational profile improves the meaningfulness of the reliability data. Customer usage data is also critical to prioritizing the configurations that you test. Setting up environments similar to those of your customers helps you find interactions that you might otherwise miss. Choose configuration interactions with all-pairs. A major problem that makes the testing task too darned big is worrying about all the possible configurations. Configurations cover things such as different hardware, different software (operating system, Web browser, application, etc.) or software versions, configuration values (such as network speed or amount of memory), etc. If you understand which interactions of your configurations cause problems, you should explicitly test for them. But when all the software “should” be independent of the configuration, and experience has shown you it isn’t, then you get worried. Setting up a configuration to test is frequently expensive, and running a full set of tests typically isn’t cheap, either. If you are doing performance testing, you need the more formal design-of-experiments techniques associated with robust testing and requiring orthogonal arrays. For simple verification, however, another easy, cheap technique used by many test shops is all-pairs. The most common defect doesn’t require any interaction. Just triggering the defect under any configuration will cause a failure about 80 percent of the time. After basic testing eliminates those defects, the next most common defects turn out to require the interaction of just two aspects of the configuration. Which two aspects? That’s the trick! Using all-pairs, you can cheaply verify any pair of interactions in your configuration. Public domain tools exist (http://tejasconsulting.com/open-testware/feature/allpairs.html), and it takes less than a half hour to download the tool and create a description of your configuration issues. You’ll then get a very quick answer. As a simple example of this, I consulted with a test group that had tested 100 different configurations, but were still getting errors from the field. I taught them the technique and tool, and in less than an hour they had a result showing that only 60 configurations were needed to cover all pairs of interactions, and the interaction most recently reported from the field was one of them. That is, if they tested with fewer configurations, they could have found the bug before shipping. All-pairs testing may be useful in other domains, but the studies are not yet conclusive. USE OF MODELS FOR STOCHASTIC TESTS One way toward stochastic testing is via the use of models. You can begin with the simplest of models, such as “dumb monkey” test tools or “fuzz” testers. You can enhance them toward “smart monkey” tools and formalize them using a variety of modeling languages, such as finite state machines, UML2 (Unified Modeling Language, version 2), and abstract state machines. Testers’ reluctance in the past to embrace formal methods came principally from the state explosion problem that appeared to relegate most formal methods to “toy” problems instead of the big system-level problems that many testers have to deal with. Recent advances in many fields have helped, but especially recent breakthroughs that make SAT (propositional satisfiability) solvers highly efficient and able to handle more than 1 million variables and operations. Models can be used to generate all relevant variations for limited sizes of data structures.13,14 You can also use a stochastic model that defines the structure of how the target system is stimulated by its environment.15 This stochastic testing takes a different approach to sampling than partition testing and simple random testing. Even nonautomated models can provide useful insights for test design. Simple finite state machine models can show states the designers hadn’t thought about or anticipated. They also help clarify the expected behavior for the tester. You can build your own finite state machines in almost any language. A simple example for C# is goldilocks.16 The major issue with model-based testing is changing testers’ mind-sets. Many testers haven’t been trained to think abstractly about what they are testing, and modeling requires the ability to abstract away some detail while concentrating on specific aspects. Modeling can be especially difficult for ad hoc testers not used to approaching a problem systematically. Another reason modeling fails is because people expect it to solve all their problems. Frequently it is better just to add a few manually designed test cases than to extend the model to cover all the details for all cases. ECONOMY IN TESTING The size and complexity of today’s software requires that testers be economical in their test methods. Testers need a good understanding of the fault models of their techniques and when and how they apply in order to make them better than stochastic testing. Code coverage should be used to make testing more efficient in selecting and prioritizing tests, but not necessarily in judging the tests. Data on customer usage is also paramount in selecting tests and configurations with constrained resources. Pairwise testing can be used to control the growth of testing different configurations. Attempting to black-box test integrations of poor-quality components (the traditional “big bang” technique) has always been ineffective, but large systems make it exponentially worse. Test groups must require and product developers must embrace thorough unit testing and preferably tests before code (test-driven development). Dependencies among units must be controlled to make integration quality truly feasible. - Kaner, C. 2000. Architectures of test automation. http://www.kaner.com/testarch.html. - Hoffman, D. So little time, so many cases. www.cs.bsu.edu/homepages/dmz/cs639/So%20little%20time,%20so%20many%20cases.ppt. - Howard, M., and LeBlanc, D. 2002. Writing secure code, 2nd edition. Microsoft Press. - Nair, V. N., et al. 1998. A statistical assessment of some software testing strategies and application of experimental design techniques. Statistica Sinica 8: 165-184. - Ntafos, S. C. 2001. On comparisons of random, partition, and proportional partition testing. IEEE Transactions on Software Engineering 27(10). - Bach, J., and Schroeder, P. 2004. Pairwise testing: A best practice that isn’t. 22nd Annual Pacific Northwest Software Quality Conference, Portland, OR. http://www.pnsqc.org/proceedings/pnsqc2004.pdf. - Hoffman, D. 1999. Heuristic test oracles. Software Testing and Quality Engineering (March/April). http://softwarequalitymethods.com/SQM/Papers/HeuristicPaper.pdf. - Hoffman, D. 1998. A taxonomy for test oracles. Quality Week ’98. http://softwarequalitymethods.com/SQM/Slides/ATaxonomyslide.pdf. - Saff, D., and Ernst, M. 2004. Automatic mock object creation for test factoring. ACM SIGPLAN/SIGSOFT Workshop on Program Analysis for Software Tools and Engineering (PASTE’04), Washington, DC (June 7-8): 49-51. - Larus, J., et al. 2004. Righting Software. IEEE Software 21(3): 92-100. - Savage, S., Burrows, M., Nelson, G., Sobalvarro, P., and Anderson, T. 1997. Eraser: A dynamic data race detector for multithreaded programs. ACM Transactions on Computer Systems 15(4): 391-411. - Tate, A. 2003. The best testers are free. Software Testing Analysis and Review (STAR West). - Marinov, D., and Khurshid, S. 2001. TestEra: A novel framework for automated testing of Java programs. Proceedings of the 16th IEEE Conference on Automated Software Engineering (November): 22-31. - Ball, T., et al. 2000. State generation and automated class testing. Software Testing, Verification and Reliability 10(3): 149-170. - Whittaker, J. A. 1997. Stochastic software testing. Annals of Software Engineering 4: 115-131. http://www.geocities.com/harry_robinson_testing/whittaker97.doc. - Robinson, H., and Corning, M. Model-based testing in the key of C#. http://www.qasig.org/presentations/QASIG_Goldilocks2.pdf. KEITH STOBIE is a test architect in Microsoft’s XML Web Services group (Indigo), where he directs and instructs in QA and test process and strategy. He also plans, designs, and reviews software architecture and tests. With 20 years of distributed systems testing experience, Stobie’s interests are in testing methodology, tools technology, and quality process. He is also active in the Web Services Interoperability organization’s (WS-I.org) Test Working Group creating test tools for analysis and conformance of WS-I profiles. He has a B.S. in computer science from Cornell University. Originally published in Queue vol. 3, no. 1— see this item in the ACM Digital Library
It’s too easy for blogging to end up as nothing more than a series of rants, so here’s something positive. It’s that time of year when millions of people (leastwise, those in the northern hemisphere) are starting new school years. As someone who tutors (other) students with ADD and learning disabilities, I thought I’d share a bevy of helpful ideas I’ve scraped together over the years. GETTING READY TO READ Put the material into the Big Picture. Before starting a chapter/ module/ unit, review your syllabus to see how the content of this one fits within the logical flow of the previous unit, and how it might be important to the next unit. This helps the material make more sense and seem less like a giant pile of loose facts. Read the textbook backwards. Start with the Summary in the back of the chapter; this is the “TV Guide” version to what the chapter is about, so you know what you’re heading into before you dive into all the excruciating details. Read over the new terms in the Glossary, so when you encounter them in the text you won’t have those unintelligible speed-bumps that interrupt your understanding of the reading. This is helpful if this subject is entirely new to you and you have little or no background in the concepts and terminology of this particular field of study. From the first day of class, create a personal glossary of new terms and their definitions. This is imperative if you are starting a new field of study because you will soon find yourself in possession of a swarm of new words for which you are responsible. Trying to look up a word for its definition by flipping through masses of notes, handouts and textbooks only slows you down and makes you frustrated. Staring into space, pacing, rocking or banging your head do not aid in remembering new terms, so having that personal glossary will give you a ready list to access. Don’t forget to add helpful tips to your definitions, such as cautions about similar-sounding words that you might confuse, or terms with complementary or opposite meanings. This is especially helpful if you are slow at recalling words, or have difficulties with spelling. Block off distracting printed material with a mask. Use a half sheet of thin cardboard, a 3/4 sheet cut into an “L” shape, or two blank index cards to mask off distracting graphics, or simply to block off everything but the single question, objective, or paragraph you need to focus upon. This is helpful if you are someone who is easily distracted by fascinating pictures, or if you have reading difficulties. TAME THE PAPER TIGER Assign a particular color to each class. I like to have the binder match the textbook color, so when I’m getting things together for class I only have to grab “two red things”. After the test, keep the notes and handouts in the colored binder or manilla folder. Use that color of ink to mark due dates for assignments and test dates on your calendar. Use that color of manilla or pocket folder to keep all the stray bits of useful stuff you are collecting for a report/project – having that special “parking place” will help organize and reduce the “file by pile” mess on your desk, floor, table, window ledge and other random surfaces… Buy a hole punch with a trap. The trap collects all the “dots” so they don’t litter the floor. A 3- or 4-hole punch (depending on whether you use 8.5″ x 11″ or A4 paper) is vastly easier than a single-hole punch, as it not only reduces the number of clenches you have to perform, but also because it makes hole spacing that is perfectly even for the binder. Hole-punch all of your handouts and put them into the binder with your notes, so the two can live in wedded bliss. Buy several packages of index dividers so you can separate the different chapters/units in your binder and more quickly flip through them for studying. Make liberal use of colored sticky-notes. These are the greatest invention since the microwave oven! They will save tremendous amounts of time from having to endlessly flip through textbook, lab manual, handout, and note pages to track down important information. Use colored sticky-notes to mark where important graphs, lists, charts, and diagrams are located in the textbook – write a key word on the external, flagging end of the sticky. Use different colors of sticky-note for different chapters/modules/units, to make studying easier when you have tests that come after you have begun the next chapter/module/unit. Use sticky-notes to mark chapter sections for those classes that skip around a lot within a textbook. If you are only using section 3.2 of a chapter, then you may begin by reading the summary for just section 3.2 of that chapter, but it might also be helpful to briefly review what the rest of the chapter summary has to say, to understand how the ideas in this section are connected to other ideas. NOTABLE TIPS FOR NOTE-TAKING & STUDYING Always take notes in black ink. There is nothing more horrifying during Midterm or Final Exams than discovering that a semester’s worth of pencil-written lecture notes has turned into a smeary, unreadable mess. Oh, the horror… Also, some kinds of blue ink are close to “non-photo / non-repro” blue, a color that’s nearly invisible to many photocopiers; this is usually not a problem unless you need to photocopy those notes for any reason. Always date and/or number your note pages. Of course, if you live a charmed life and never have sudden “gravity fluctuations” in your part of the planet that cause you to drop or spill note papers, or you never own binders that lose their “bite”, then don’t bother. Otherwise, dating the pages lets you keep track of what was lectured on at a particular time (handy if someone asks to borrow your notes from last Tuesday). If you take more than one page of notes per day (which is nearly always) then numbering the pages instead of or in addition to dating them makes it even easier to put spilled pages back to rights. Title each page. Even if it’s just by abbreviation, describe the page of notes by the lecture topic, the unit or chapter title. This not only makes it easier to find the right notes when studying for tests, but it also helps you remember what the overall pattern of ideas is during the course of the class across the semester. MITOSIS WED 2 FEB p.1 Take notes in two columns: the left side for listing the main idea titles, important names, terms, dates or formulae, and the right side for all the regular details and sentences. If there is a page in your textbook, lab manual or whatever that has a particular graph, chart or listing, write down that page number on the left side as well, as well as a word or two to title why that page number is important. This speeds up your test studying because you can glance through pages of notes to find the one that has the specific information you’re looking for. Use the Objectives listed in the chapter/unit/module as your study guide for the test, and write out a full answer to each one as though it were a question. Pay attention to key verbs such as Describe, Compare, List, Define or Identify – these can give you an idea of what kind of test question could be asked. Writing these out does two things: it not only helps you self-test your own understanding before you get to the class test, but it also changes your answers from something you have to invent during the test (which is time-consuming) into something you just have to recall during the test (which is much quicker and easier). Writing out answers to the objectives in full sentences is especially helpful if English is not your first language, and/or if you are slow at remembering words,and/or otherwise have difficulty expressing the knowledge that’s stuck in your head. DECIMATED BY NUMBERS Turn lined paper sideways to have ready-made columns for keeping your place-values straight in big arithmetic calculations. Another option is to use green “engineer’s paper” that has graph squares on one side and is blank on the other side, but the graph grid is still somewhat visible on the blank side, and the green tint is more restful on the eyes. This is especially helpful if your handwriting tends to wander around or slope down a page, and will keep your numbers and decimals in order. If you are doing mathematical equations or other things that are processes, write out your own set of numbered directions describing how to do the process. For instance, it may not be as obvious to you as it was to the author of the formula that you need to determine the value of “C” before you put the other values into the formula. So in your own directions, you should note “Find the value for “C” by ~ ~ ~” as one of the earlier steps. Whenever you solve an equation or do a statistical analysis, write out in a complete sentence what the answer to the calculations MEANS in regards to the original problem/story/question given. These are especially helpful if you are more of a “words” person than a “numbers” person. If you have several different formulae , make yourself a flow-chart (meaning, a series of decisions) that helps you figure out which one you use for different kinds of circumstances. When you are studying a chapter or doing that day’s homework, it’s obvious which one you need to use – it’s the one you’re learning that day! But come test time, you will need to be able to understand which one you use for each kind of situation. This is especially helpful if you are one of those people for whom “all the formulae look the same”. Use name and address labels on everything, and add your phone number or email as well. Put them on your textbooks, lab manuals, various notebooks, calculator, data CD, flash/keychain drive, assorted binders, notepads, calendar-organizer, each piece of art & drafting equipment plus the carrying case, and all the other things that you need to survive as a student, to help guarantee that the person who finds them can help get them back home to you. This is especially helpful if you are forgetful, distractible, prone to leaving things in various places, and/or are juggling a variety of classes and jobs. (You can imagine why I know this.)
Good progress is being made with the return to use of the derelict 784km (497 mile) rail network in the Central American state of Guatemala. This is being undertaken under a 50-year concession to the Compañia Desarrolladora Ferroviaria SA consortium, an affiliate of the Railroad Development Corporation (RDC) of Pittsburgh, USA. Observers believe that the take over will result in the revitalization of a once moribund stateowned rail network, and result in the provision of a rail system to lead the country out of difficult economic conditions. Guatemala has the largest population and economy in Central America with low inflation, a tradition of maintaining a stable currency, democratic government, and a policy open to domestic and foreign investment. At its height, FEGUA enjoyed a status as the most important rail network in Central America, despite operating on a nonstandard 1067mm gauge, and linking with neighboring countries’ 1435mm systems. More importantly, it provided links between the country’s interior and the Atlantic and Pacific oceans, which, in reality, are its economic lifelines, providing the main routes by which its principal exports of coffee, bananas and sugar can be transported to their main markets. Containers, steel, cement and paper are also carried. Redevelopment of the country’s railways faced the familiar stumbling block of the constraints imposed by being under state control, which had allowed unregulated road transport to gain a decisive upper hand in winning new traffic. The would-be operators of the new system were keen to stress that they brought to the table valuable experience from the operation of two rail route concessions in Argentina, and their main business, the Iowa Interstate Railroad, from Chicago to Omaha in the United States. RDC started to rebuild the old main line between Puerto Barrios on the Atlantic Coast and Guatemala City early in 1998. Plans called for the main line to be re-opened by the end of 1998 between Puerto Barrios and Guatemala City, but massive damage from Hurricane Mitch delayed this until spring 1999. Several bridges were knocked off their abutments and needed to be replaced. One of the FEGUA-built diesels, dating from 1982, which is now in the hands of the RDC. The country’s rail network is 1,067mm gauge line. This links with the Mexican system, ,435mm gauge, and the El Salvadorian network, also 1,067mm. The country’s network consists of 784km of 1067mm gauge line. As previously mentioned, this links with the Mexican system, which has a 1435mm gauge, and the El Salvadorian network, also with a 1067mm gauge. RDC believes the Guatemalan operation had strengths far beyond simple rail operations - its ownership of ports and real estate in and around the main stations gives it considerable potential. However, physical deterioration reached such a stage that passenger traffic ceased in 1994, and despite the possibility of privatization, the country’s entire rail system closed two years later. RDC vigorously opposed this move, saying it would be far more difficult to rehabilitate the network once closed, rather than keep it running in its present state. But the scale of the investment needed meant it was forced initially to concentrate on the corridor offering the maximum return - the 300km corridor from Guatemala City to the Atlantic ports of Puerto Barrios and San Tomás, where road traffic carried goods as diverse as coffee to scrap metal, and bananas to cement. Railroad Development Corporation faced a number of unusual problems beyond the obvious obstacles of a lack of capital investment. Theft, vandalism and the widespread occupation of former railway premises by squatters all made the RDC’s task more difficult, but road construction had also meant that connections to several former freight customers’ premises had been severed. On 15 April 1999 a goods train ran between a cement plant in El Chile to the Guatemala City, and in the first full operational year 63,000t of goods were carried, rising to 110,000t in 2001. Under normal conditions one train per day each operates in each per direction and day, taking 12 hours to travel 300km. Several links to private sidings will need to be restored, but the most important new project is a new terminal to handle container and transshipment traffic. At a new location, it will move this important hub from the former Central station to a site more convenient for major industrial customers. The operator owns a small fleet of 15 modern diesel locomotives, which date from 1982. Only seven engines were operable on the take-over. General Electric U10 locomotives were found to be in relatively good condition, having seen little use as traffic declined before closure. FEGUA now has 200 freight vehicles in working order, mainly of low-capacity box and flat cars. The poor state of bridge structures is a major worry. This damaged bridge, at Villa Canales, gives some idea of the scale of the task facing the RDC in reviving railways in Guatemala. One of the four operational General Electric diesel locomotives which it is hoped will be returned to traffic. Signaling / Communications The project has been described by RDC as ’bootstrap capitalism’, by which his company intends to get the network running with the minimum investment, and then see what is needed to meet any demands created by new traffic flows. By this, he means that signaling will be restored where needed, and further investment will be dependent on any growth in traffic. A simple ’one train in section’ system will, therefore, be adequate for most needs, until the growth in traffic demands a more sophisticated method. The rail network of this Central American state has begun its revitalization from an inimaginable state of decay, but the new American operators are under no illusions as to the scale of the task which they have begun. The 200km Pacific corridor from Tecun Uman on the Mexican border to Puerto Quetzal and Esquintla - from where lorry loads could be transshipped to Guatemala City - was seen as the next priority, while Phase three is seen as the construction of a new line to Cementos Progreso, and Phase four, the restoration of the a link between the Pacific and Atlantic corridors with the reinstatement of the Esquintla- Guatemala City line. RDC has its sights set on the reinstatement of the long-closed link between Guatemala and El Salvador. Reinstatement of passenger traffic is a possibility if it does not interfere with freight FEGUA Rail Network – Specification System Ownership Compañia Desarrollodera Ferroviaria SA (a subsidiary of US-based Railroad Development Corp.) Date opened 1904 Route length 784km (490 miles) Maximum line speed 80km/h (50mph) Phase 3 To Cementos Progreso Phase 4 Zacapa-Anguiatu-El Salvador Opening date To be determined Fleet total 15 locomotives serviceable, steam in reserve Builder Alco/General Electric New stock on order Nil At its height, FEGUA enjoyed a status as the most important rail network in Central America.
The old ram stands looking down over rock slides, stupidly triumphant. by John Gardner Grendel is Mr. Gardner's re-telling of the classic Anglo-Saxon epic, Beowulf which tells the tale of a young, heroic warrior who travels to Denmark to do battle with the monster who has been raiding the great hall of the local Danish king. Beowulf is a wonderful work, I'm very fond of Seamus Heaney's recent translation which everyone should read. I guess I had forgotten just how small a role Grendel plays in Beowulf. Grendel really just gets the action going. After several attacks on the Danish king, Beowulf shows up and saves the day, ripping off Grendel's arm in a epic battle that ends with the monster retreating to the sea to die in his mother's arms. Grendel's mother then comes to attack the Danes. Beowulf defeats her, becomes king, and then, very late in life, faces a terrible dragon who finally brings Beowulf to a fitting warrior's end. John Gardner gives Grendel his due, exploring the character of the monster in a re-telling that long preceded the current trend of re-telling classics. One thing that really struck me about Grendel was how closely tied it is to Mary Shelley's classic Frankenstein. Much of Frankenstein, the best parts of it if you ask me, is focused on the monster's point of view. Created by a man, the monster longs to become a human, but he is forever cast outside the realm of humanity because he was created by a human. The monster, who has been assembled from parts of several human corpses, is also cursed with an appearance that everyone he meets finds horrific. In one of the books more touching sections, the monster finds a little house, occupied by a small family. He spends several weeks observing the family, watching them from afar, then from outside the window, envious of everything this rather poor family has even though all they really have is each other. The monster is truly alone. Not only does he have no one, he knows that there is no one in the world he could ever have. When the monster makes himself known to the family's blind grandfather, he hopes that by befriending one person, he can befriend the entire family and thereby connect with humanity. Of course, we all know how this turns out. As soon as someone who can see sees the monster, he is chased away by the terrified family. People who don't know the novel expect Frankenstein to be a horror story, but it's really a long meditation on the challenges of being human, mainly of connecting with other humans. Even Victor Frankenstein faces this dilemma in his own search to find a true friend. Grendel covers much of the same territory. Before he first encounters humanity, Grendel lives as a wild creature, a monster, but he is something more than his un-named mother. He has a facility for language, for thought, that she clearly lacks. We do not know who Grendel's father was, but clearly something more was added to his genetic make-up, something that turned out to be no- so-good for him. Once he sees people, he can't help but follow them. He wants to know more. He wants to find 'people' equal to his own capacity for thought and for wonder. He is able to spy on the Danes un-noticed for quite a few pages, but eventually they see him and attack. While he tries to speak to them, his words come out as intelligible moans' he is trapped by his own monstrous body much like the monster in Frankenstein was. "You're all crazy!" I tried to yell, but it came out a moan. I bellowed for my mother. "Surround him!" the king yelled, "Save the horses!" --and suddenly I know I was dealing with no dull mechanical bull but with thinking creatures, pattern makers, the most dangerous things I'd ever met. Both Grendel and the monster stand outside the homes of humanity looking in at what they cannot have. Grendel stands outside the hall of the Danish king listening to the singer inside singing songs about him and about the creation of the world. The harp turned solemn. He told of an ancient feud between two brothers which split all the world between darkness and light. And I, Grendel, was the dark side, he said in effect. The terrible race God cursed. I believed him. Such was the power of the Shaper's harp! Even if he is meant to be the villain in the tales the humans tell, Grendel still wants a place in their lives much the same way the monster in Frankenstein does. The monster will also take on the role of villain by the novels end. Both try to connect, both are seen as monsters when they do so, and both must finally wrestle with their 'creators.' In both stories there is also a real sense that they are condemned to isolation because neither was made by 'God'. Frankenstein closes with an account of the monster chasing his creator across the frozen arctic ice while a group of men aboard a ship watch from a distance. How close is this to the way Grendel's life ends? Chased by a group of men, across a frozen northern landscape, to a body of water where he is dealt a death blow by Beowulf. By the end of John Gardner's novel, I began to wonder just how much Mary Shelley had been influenced by Beowulf when she wrote Frankenstein. Of course, Beowulf did not create Grendel. John Gardner plays with the question of who created Grendel in his novel. The original epic poem was written by an unknown source. This unknown poet is present in Mr. Gardner's novel as a singer who entertains the Danes with songs and tales of epic battles. The character has a clear historical basis though I don't recall him acting as much of a force in Beowulf. However, in Grendel, the Shaper, as Grendel calls him, is the main antagonist. Grendel fights and kills many characters, but the one he really wrestles with, the way the monster wrestled Victor Frankenstein, is the Shaper. Grendel after listening to one of the Shapers songs: It was a cold-blooded lie that a god had lovingly made the world and set out the sun and moon as lights to land-dwellers, that brothers had fought, that one of the races was saved, the other cursed. Yet he, the old Shaper, might make it true, by the sweetness of his harp, his cunning trickery. It came to me with a fierce jolt that I wanted it. As they did too, though vicious animals, cunning, cracked with theories. I wanted it, yes! Even if I must be the outcast, cursed by the rules of his hideous fable. The god this Shaper creates in his tale is not exactly the Judeo-Christian god, but close enough. The Bible has its share of warring brothers and cursed races. Is Grendel to be seen as Cain? as Judas? It's interesting to me that even if his only role is to be the villain in the story, he wants a part so badly that he will take it. He complains that the Shaper is not telling the story correctly, but he still wants a part in the tale. The character of the Shaper can't help but call the idea of religion and religious explanations into question on a more fundamental level. Has the teller come before the tale? Did humanity's natural inclination towards finding patterns bring about the notion of god in the first place? If Grendel is what really happened in Beowulf, then what really formed the basis for the other tales we use to explain our world. Is there anything behind the tale at all? At one point Grendel meets the dragon who will eventually bring about Beowulf's end. The dragon is eternal in the sense that he knows the entire story from start to finish all at once. For the dragon, what will happen, has already happened. The dragon tells Grendel about humanity: 'They sense that, of course, from time to time; have uneasy feelings that all they live by is nonsense. They have dim apprehensions that such propositions as "God does not exist" are somewhat dubious at least in comparison with statements like "All carnivorous cows eat meat." That's where the Shaper saves them. Provides an illusion of reality--puts together all their facts with a gluey whine of connectedness. Mere tripe, believe me. Mere sleight-of-wits, if anything: works with the same old clutter of atoms, the givens of time and place and tongue. But he spins it all together with harp runs and hoots and they think what they think is alive, think Heaven loves them. It keeps them going -- for what that's worth. As for myself, I can hardly bear to look.' Dragons really don't make for happy company. However Grendel makes for fascinating, unexpected reading. While I thought I would get an exciting adventure with lots of battle scenes, I got an intriguing view of humanity's habits, foibles perhaps, along with a new sense of what it is to create a character and how important it is to be a part of a story in some fashion. Grendel! Grendel! You make the world by whispers, second by second. Are you blind to that? Whether you make it a grave or a garden of roses is not the point. Like the monster wrestles with Frankenstein and Grendel wrestles with the Shaper, I have been wrestling with John Gardner's little book for some time now. I think this is one that I'll come back to again some day. And it's my new favorite book.
Wolfram Education Training M310: Digital Image Processing This course presents the theory and practice of digital image processing with Mathematica and focuses on the Digital Image Processing package. The features and capabilities of the package are demonstrated, and numerous examples and practical hands-on exercises are included. The material is presented as a sequence of eight one-hour lectures. Each lecture covers a major image processing topic, typically consisting of a discussion of the basic theoretical concepts and including examples that illustrate relevant, practical imaging problems. The lectures are followed by exercise sessions to help students understand the material and to provide a focused and practical learning experience. The course is designed primarily for people who need to analyze and process imaging data with Mathematica. Students typically have wide-ranging backgrounds and include engineers and professionals in the physical, life, and medical sciences. Introduction: lists, matrices, and images; basics of programming with Mathematica; image representation and display Image Histogram and Point Operations: first- and second-order image histograms, contrast enhancement via point transformations, histogram equalization, and color transformations Geometric Operations: image resizing, interpolation and decimation, affine spatial transformations (rotation), and higher-order spatial transformations (warp) Linear Processing: convolution and correlation, linear filtering, FIR filters, blurring, sharpening, and edge detection Selective Processing: block processing, region-of-interest processing, and line profiles Nonlinear Processing: nonlinear noise reducing filters (e.g., median, outlier, and adaptive) and image morphology Frequency Domain Processing: fourier analysis and unitary image transforms: DFT, DCT, DHT, and DWT Performance and Extensions: packed arrays, Compile, J/Link, and Java Advanced Programming (JAI) All Wolfram Education Group offerings are "hands-on" interactive courses and are taught by certified instructors in a professional computer classroom environment. For more information: www.wolfram.com/weg/.
• Flat is an attribute that can be assigned to a symbol f to indicate that all expressions involving nested functions f should be flattened out. This property is accounted for in pattern matching. • Flat corresponds to the mathematical property of associativity. • For a symbol f with attribute Flat, f[f[a, b], f[c]] is automatically reduced to f[a, b, c]. • For a Flat function f, the variables x and y in the pattern f[x_, y_] can correspond to any sequence of arguments. • The Flat attribute must be assigned before defining any values for a Flat function. • New in Version 1.
In the 2008 election, 110,270 Denton residents voted for the McCain/Palin ticket, while 70,388 resident voted Obama/Biden. Denton is largely a Republican county, according to 2008 data. But If you break down and map the data by voting precinct, it’s easy to see which areas of the county Democrats and Republicans are likely to dominate the vote. Based off the map below, city of Denton residents are largely Democrat, while in the more rural areas of the county, Republicans are dominate. The grey patches on the map represent precincts created after the 2008 election and there is not any data available – yet. Denton seems to follow the same trend as the entire state, but on a much smaller level. In my last blog, The political landscape of Texas, I created a map that shows which Texas Counties are likely to vote Republican or Democrat. And according to the map, urban areas like El Paso, Dallas and Houston all voted Democrat, while more rural cities and counties voted Republican. What environmental factors may influence political loyalty? For the interactive map, click here - http://goo.gl/JN8Yj Source: Denton County Elections Administration Map Created By: John D. Harden
CNET GLOSSARY: Terms for the techie Typically used to connect a television to cable TV services, coaxial cable consists of a small copper tube or wire surrounded by an insulating material and another conductor with a larger diameter, usually in the form of a tube or a copper braid. This cable is then encased in a rubberized protective material. Can't find your term here? Have a question about the glossary? E-mail us.
The segments of our DNA that are commonly called "genes" consist of protein-coding exons and non-protein-coding introns. Initially, the entire DNA segment is transcribed into RNA, but between ninety and ninety-five percent of the initial RNAs are "alternatively spliced."It's up to you, dear readers, to figure out all the things wrong with this explanation. You can start with the math. Arithmetic isn't one of their strong points. Or maybe it's an understanding of biology that's the real weak point? What is alternative splicing? Imagine that the initial RNA derived from its DNA template has the organization A—B—C—D—E—F, where the letters represent blocks that specify amino acid sequences and the dashes in between the letters stand for introns. Alternative splicing enables multiple proteins to be constructed given the same RNA precursor, say, ABCDF, ACDEF, BCDEF, and so forth. In this way, hundreds or thousands of proteins can be derived from a single gene. There’s more. The messenger RNAs that are produced by this process—and therefore the proteins that are made in a cell—are generated in a way that depends on the stage of development as well as the cell and tissue type. In the above example, a nerve cell may express the ACDEF version of a messenger RNA whereas a pancreatic cell may produce only the BCDE version. The differences are biologically essential. What does this have to with introns? Everything. It is the presence of introns that makes this permutative expansion of messenger RNAs possible in the first place. So let’s do the math. At least ninety percent of gene transcripts undergo alternative splicing, and there are at least 190,000 introns in the human genome. That means we have at least 0.90 x 190,000 = 171,000 introns that participate in the alternative-splicing pathway(s) available to a cell. Thursday, June 03, 2010 Creationists, Introns, and Fairly Tales Richard Sternberg thinks that introns are important. He has to think that way because he's an Intelligent Design Creationist and the idea that introns could be mostly junk and not have a function isn't part of his faith [Matheson’s Intron Fairy Tale].
Who was the first to ride the surf at San Onofre? No one can say for certain. The late, great George “Peanuts” Larson of Laguna would have claimed the title for himself; but others contend that it was Orange, California’s Matt Brown and Lorrin “Whitey” Harrison who first stumbled onto ‘Nofre’s forgiving waves while heading south from Corona Del mar in search of surf. In the early 1920’s, motorists began noticing a few cars driving up and down the old Coast Highway with surfboards tied on top or sticking out the rear window. Also, along about this time, Dutch Miller, the life-long Chief of Lifeguards in Long Beach observed Ted Sizemore making a “paddleboard” on the beach over on the west side. Since the advent of waterproof glue in the manufacture of plywood, anyone could now make his own board. Ted said he got the idea from Barney Wilkes who had gotten the idea from either Pete Peterson or Lorrin Harrison. Dutch borrowed the concept, and introduced the use of paddleboards for life saving in the surf along this part of the coast. By the late 1920’s there were probably as many as 50 to 75 regular surfers in Southern California. These hearty nomads would explore every inlet, cove and point in search of new and better places to ride waves. One of the spots they discovered was an obscure and unknown beach about 60 miles south of Los Angeles near a railroad siding identified as “San Onofre”. Over the years, there has been much debate about the origin of San Onofre’s name. Some believe it was a Native American word, bastardized by the Spanish to describe the area’s creek and surrounding valley. Others think that Spanish missionaries named the area after the sixth century Egyptian saint known as Onofreas. Whatever the case, the name appears as far back as the Capistrano Mission era, and in the Santa Margarita Land Grant documents from 1836. When the Santa Fe Railroad Company put in a coastal line from Los Angeles to Oceanside in the 1880’s, a train stop was erected at San Onofre to help local farmers get their crops to market. The area’s namesake, clearly scripted on the sign hanging from the train station wall at that time, has remained unchanged ever since. For a few years San Onofre would remain a secret, frequented only by handful of discoverers who would be counted later as the best watermen in California. Sometime during the mid-’30s, rumors began buzzing of a “fish camp,” with incredible surf, located on a beach in an obscure; desolate place called San Onofre. As surfing’s popularity had increased at spots like Long Beach flood control, Corona Del Mar and Palos Verdes’ Bluff Cove, so had the idea of searching for new breaks. In those days, riding 11-14 foot planks and paddleboards at a dumping beachbreak like Hermosa was a challenging feat. So when word got back to the surfers at Corona Del Mar and Palos Verdes, about a gently sloping, long-peeling wave with size and consistency, they were immediately intrigued. Depending where you lived in SoCal, the drive averaged one to three hours, down the two-lane coast highway to get to ‘Nofre. It was an inland cruise by way of San Juan Capistrano through the citrus groves from which Orange County got its name– there was no highway along the Laguna coast at the time. The chaparral-covered valleys and canyons surrounding San Onofre’s beach were owned by the Santa Margarita Ranch, which dates back to the Mexican Land Grant era of the 1830s. During the 1930s, the Haven Ranch leased the farm property across the railroad tracks from the beach. The Santa Margarita Ranch also leased out the small piece of sea level, beachfront property at San Onofre to a not-so-friendly (read: gun-toting) chap who first started the “San Onofre Beach Fishing Camp.” The camp was quite popular with anglers, who reportedly used to pull in massive catches of sizable corbina, sea bass and halibut. The knee-deep tide pools produced an abundant seabed of clams as well. By ’37, a local San Clemente resident, Frank Ulrich, took over the lease for the camp and promptly added a Texaco gas station and café’ on the east side of the coastal highway. A handful of consistent waveriders were allowed to frequented the beach - as long as they coughed up their “two-bits” (25 cents), of course. And, for 50 cents, one could stay for the whole weekend. “Times were tough in those days, money–wise,” remembers 86-year-old E.J Oshier, who still frequents ‘Nofre to this day. “It was still in the depression days, see? Money was hard to come by, and some of us guys were coming all the way down from the Palos Verdes area. So we’d get a few of us to pitch in enough for some gas and jug of wine.” Oshier adds with a smile. “That’s why we didn’t spend money on surfboards or trunks back then. If any of us had [money] we were off to ‘Nofre. Pulling into Ulrich’s camp, you’d turn off the highway at the train depot and cross the tracks by way of a rutted dirt road. On the other side of them was a small shack with a meager selection of food and camping supplies. Ulrich would take your money and hand you a parking pass. “It was a pleasure to beat him out of 25 cents when you could,” laughs 80-year-old Art Beard. More often than not, they slipped in for free, eluding a distracted Ulrich, who was usually preoccupied by his gas station or café’. From there, the road dropped down to the beach, past the dirt lot with a few shabby, wooden cabins Ulrich had built to rent out to the weekend fishermen. Early on, the surf casters there despised the growing number of waveriders and wanted them barred from the camp, claiming they were nuisances and were scaring the fish away. They also pointed out to Ulrich that the rocks along the sea floor there were terribly dangerous and made for unsafe swimming conditions. But the camp’s proprietor paid no attention - times were slow and he needed any sort of business he could drum up. The surfers, of course, weren’t going anywhere. They already knew that those very rocks made for one of the longest, smoothest rides in Southern California. Over the next few years, the number of surfers increased while the presence of fishermen slowly dwindled. “There used to be these pilings there, right about where the entrance to The Point is today. You couldn’t drive all the way to Old Man’s,” remembers then lifeguard Dorian “Doc” Paskowitz, who (along with “Storm Surf” Taylor and Lloyd Baker) was among the first San Diego surfers to head up to San Onofre. “That was in about ’36, but I think by ’37 or so, Lorrin Harrison and some of the fellas had cut [the wooden posts] down so many times that Ulrich finally gave up on denying surfers vehicle access to the beach down by Kukae Canyon.” Kukae Canyon, that well-known arroyo (which no longer exists since the Edison nuclear power plant filled it in for a parking lot) was located just a 100 feet or so north of where the shack at Old Man’s is today. Named after the Hawaiian work for “shit,” Kukae Canyon was where everybody took care of “business” until outhouses were installed some 15 years later. The small canyon, which was the mouth of a dried-up creekbed, was what helped spit out some of the cobblestone rocks that formed the mellow, sloped rock bottom. “I was glad when they cut down the pilings at ‘Nofre. Before that, it was a heck of a walk with those old, 90-pound redwood planks,” says 84-year-old Leroy Grannis, who used to come down with guys like “Hoppy” Schwartz from the Palos Verdes Surfing Club. “We’d camp out there for days at a time and just sleep on the beach. During the week you would get it all to yourself; but on the weekend, all sorts of characters showed up. Pete Peterson used to come down, what a paddler he was.” “In ’37 and ’38, you had guys coming from all over to surf ‘Nofre – it was paradise,” says Oshier. “Someone always had a guitar or ukulele, and after we surfed our brains out, someone would pass the bottle and we’d play Hawaiian music on into the night. We’d wake the next day and do it all over again. There was never really any friction. Nobody cared if you shoulder-hopped; there was never any question of whose wave it was. We just all went…I tell ya, we had an out and out ball down there – it was the best time of our lives.” “We were all just surfers,” explains 84-year-old Paul “Willard” Luton. “We barely ate anything all day and stayed out in the water for hours. We’d come in and someone always had a fire going – either a tire or some bits of dried shrubs. I also remember a lot of us wearing trench coats we got from Goodwill for 50 cents. It would be pretty cold there in the morning in those days.” By 1939, the “Nofre lifestyle was well established, and so were the group of guys that were considered regulars. Some of them included: Eddie “The Mayor” McBride, Don Smith, Barney Wilkes, George Brignell, Lorrin “Whitey” Harrison, George “Peanuts” Larson, Dexter Woods, E.J. Oshier, “Laho Lio,” Paul “Willard Luton, John Waters and more. Surfing Clubs became popular in the mid-‘30’s, and organizations such as Palos Verdes and the Kekala surfriders (from Orange) regularly frequented ‘Nofre. However, there was never really an official San Onofre Club at that time. “The real ‘Nofre guys didn’t care about a Club,” explains Oshier. “They went there to get away from that. They didn’t want organization, rules, bylaws, meetings and such. I mean, you knew who the real “Nofre guys were. There, everybody was a member, or nobody was a member, because nobody cared. You ust wanted to catch another wave, play another Hawaiian song, and, for some, have another sip of wine.” In the late ‘30s, few would argue that there was no better place on the planet to live the surfing lifestyle. But the approaching storm clouds of world war would temporarily tear these wave riders from their idyllic paradise for years to come. Surf Waves, Radio Waves, Shock Waves It was a beautiful, clear morning at San Onofre on December 7th, 1941. The sun was hot, the crowds were thin, and the surf was running about five feet. Bud “Augie” Anderson had been down in the area for a few days. Saturday night he and other surfers had gone to the movies at the old Miramar Theater, a few miles north in sleepy San Clemente. “Larry Roth came paddling out” Augie remembers, “and shouted, ‘Hey, the Japs just bombed Pearl harbor!’ We didn’t believe him, so we went in and turned on our car radios – which is when we heard President Roosevelt declare war against Japan.” “I joined the Navy in June of ’42, Augie continued, “Another San ‘O regular, Stanford King and I were stationed at Los Alamitos Navel Air Station. I surfed the whole summer of 1942 with my buddies at San Onofre. That winter I shipped out. I stayed in the Navy through World War II and the Korean Conflict. “I didn’t surf much during those years.” Neither did many others. In early 1943, the Santa Margarita Ranch property fell victim to eminent Domain by the U.S. Department of Interior, and was subsequently leased to the Marine corps becoming a training base named Camp Pendleton. Ulrich lost his camp, but he was able to hang onto his petrol station and café across the street. At the same time, the new USMC commanding officers lowered the boom on San Onofre and shut down all public access to its beach. Some years later, the old camping area and train station was turned into the Enlisted Men’s recreational area. San Onofre was nearly empty during the War years; surfers, like everyone else, went into military service in the far corners of the Pacific or European Theaters. Civilians weren’t legally permitted to surf on what had become a strategic military base, but local guys who weren’t enlisted still found ways to sneak in. Frequently they could get past the Military Police, but eventually daily beach sweeps would force their removal. With gas rationing and surfboard materials in scarce supply, the short, sweet surfing life of San Onofre almost ceased to be. In the post-War 1940’s, surfing began to come to life again at San Onofre. The Beach was still closed, but Camp Pendleton’s various commanding Officers allowed a sort of loose arrangement, essentially turning a blind eye to the surfers, most of who were WWII vets. That same quirky bunch of characters from the late Thirties began returning to ‘Nofre, this time, though, with a new social flavoring. Most of them had gotten married and the Baby Boom years were on. Notable among the service men returning to civilian life (and their first love of surfing) was Dr. A.H. “Barney” Wilkes. After his discharge from the Army, Barney opened his dental practice in San Clemente and spent every available moment on the beach and waves at San Onofre. As a successful and respected businessman, and one of San ’Os most frequent beach-goers, it was only natural that he would become a spokesman for the surfing group when communicating with the Marine corps. The Marine Corps was not enforcing any “off-limits” restriction at the beach, and the various changing commands at Camp Pendleton usually considered it to be good public relations to permit public usage of the beach area for which they had only occasional use. The kids that grew up during the war and the men returning from the service wanted the very most from their available recreational time. And surfing was “big” with this particular segment of sportsman. As the surfing population increased, so did the interest and concern of the Marine corps. Each new succeeding General who took command at Camp Pendleton became more and more concerned about “all those civilians” having unregulated admittance to military property. Civilian access to the beach had always been tenuous at best. Understandably, most of the surfers had a desire to formalize the arrangements with the military. Likewise, the Commanding General had an equally strong desire to achieve some order and control among the fringe element of unruly surfers whose behavior was beginning to become untenable: Vandalism and disregard for sanitation was such a serious problem it was difficult for the C.O. to ignore. To say the least, he wasn’t happy with what was taking place on his turf at San Onofre Beach. Continuos brush fires and other incidents at San Onofre had gotten ugly, and Wilkes got wind that if Pendleton’s Commanding General did not see conditions improve immediately, he was going to lock the whole beach down indefinitely. To avoid this, Wilkes, along with Andre “Frenchy” Jahan, took action with a scheme to establish a surf club to regulate the flow of traffic, as well as take responsibility for keeping the place clean. It was a win-win situation: The General would be relieved of the thorn in his side and enjoy better PR for his base. And Barney, would save the beach for all of the longtime ‘Nofre regulars. In 1950, Andre “Frenchy” Jahan, an old-time surfer of foresight, enlisted the aid of friends on the staff at Camp Pendleton in pursuing this goal. The resulting correspondence between Barney and the Commanding General produced a letter of permit granting the San Onofre Surfing Club access to the beach. At this time the San Onofre surfing Club was not a formally organized Club. Barney would accept token dues at the beach, and keep your name on an “official” list of Club members. Organizational attempts had been made in the past few years but without a pressing need, nothing had come of it. Barney would simply convey the general concerns in his communications with the Marine Corps. The San Onofre Surfing Club was loosely formed in 1951. Membership included a key to the swinging iron gate that was established for entry control. That system was short-lived, though, as copies of the keys made their way up and down the coast within months. In early ’52, the Club tightened up its program by hiring paid gate guards, who checked the official Club windshield decal (the logo of which was fashioned by artist Don Smith, a machinist for Disney) and membership ID card against the Club’s roster list. The SOSC was officially off and running with its exclusive beach and soon-to-be, well-entrenched “big family” lifestyle. This era was the beginning of a tradition that had its roots firmly planted by the early-'50s, but in many respects would continue to represent the overall style of life at San‘O for decades to come. “It was a kid’s Nirvana,” says 53 –year-old Don Craig. “Every Friday, after school, we’d load our converted panel van with all the weekend’s supplies, except the boards, because we were too little to get them on the roof. We’d wait for Dad to get home then head down to San ‘O from the south Bay. We’d camp at the San Clemente State Park and be out of there before sunrise to get a prime parking spot. If we weren’t in the water, we were up hiking and exploring the finger canyons behind the bluffs. We’d come back and Mom would have lunch ready. I couldn’t imagine a more ideal way to grow up.” “The simplest way to describe San Onofre is ‘a way of life,’” says pre-WWII surfer, and longtime ‘Nofre observer, Art Beard. “We were all just raising our families, and it was a cheap, easy and fun way to do it.” While some Club members affectionately refer to the ‘50s as the “golden years for raising families, there was still a handful of kids that caused some trouble, like a teen-aged Miklos Dora, who went by the surname “Chapin” at the time. His stepfather (a hell of a surfer) Gard Chapin, used to bring Miki and the family down for weekends at ‘Nofre. “We called him ‘Meatball,’” remembers Paul Luton. “And if we didn’t call him that, it was ‘Kamikaze Kid,’ because he would just go out and run everybody over. He was just the little meatball out there, knocking people off their boards. Rumor had it he was one of the kids who burned down one of the grass shacks in the ‘50s.” There were also a handful of notable women who surfed ‘Nofre, starting with Mary Ann Hawkins back in the ‘30s. When the ‘50s and ‘60s rolled around, Benny Merrill’s daughter, Linda Merrill put in a lot of water time there, as well as Paul and M.J. Luton’s daughter, Pauline Luton, who became Mike Doyle’s tandem partner for a while in the mid-‘60s. And, of course, you can’t forget 70-something Eve Fletcher, who has been consistently surfing San Onofre longer than just about any female surfer around today. In late 1951 the Marine Corps tightened their requirements. They insisted on a gate for controlled access to the beach, with membership cards for personal identification, windshield decals, and responsible representation for the surfing Club. The first formal meeting of the San Onofre Surfing Club was called to order by Dr. Barney Wilkes on the beach on April 24, 1952. This meeting established the following: adoption of Club by-laws as prepared by Andre Jahan, dues budget and banking procedures, membership cards, decals, beach access control, beach maintenance and rules of conduct and access as agreed upon with the Marine corps. The Club was now formally organized. Things went relatively smoothly until 1955. At this time the Club President, Al Dowden was notified that the Club was no longer to control access to the beach. All marine personnel were to be allowed access, and all other civilians were to be allowed access by simply going to the San Mateo M.P. station and getting a day pass. In early 1956 it was added that the trestle area would be open under the same conditions, and that the Marines would provide trash removal service for the area. Things went to hell in a hurry! Surfers with no past, present or future concern set fire to the brush in the San Mateo creek estuary and nearly burned down the railroad trestle. They threw burning wood at the commuter trains and piled debris on the tracks, causing a passenger train to grind to an emergency stop. They tore down parking and other regulatory signs and used them as firewood. There were incidents wherein the M.P.’s were so provoked that they fired rifles and pistols at the trestle surfers. The disease spread to some of the less responsible members of the regular surfing community at San Onofre. The big grass shack was torched into a raging inferno that would rival a college homecoming fire. A cave on the cliffs was filled with old tires and gasoline and turned into an 8 foot diameter blow-torch so intense in jetting flame that the Marines couldn’t get the fire truck anywhere close. The 4th of July fireworks display, in violation of Camp Pendleton rules was wild enough to almost cause the Base Fire Marshall to lapse into a state of coma. Some surfers set up camp overnight on the beach in defiance of patrolling M.P.’s. The Marine Corps demanded that the Club maintain order or all civilians would be restricted from the beach. The Club however was able to disclaim any responsibility in the matter since the Marines had allowed free and uncontrolled usage of the area to the public and Marine Corps personnel. All these conditions had inevitably resulted in serious damage to the Club’s structure. The paid memberships of 490 members in 1955 had dropped off to 288 memberships in 1957 due to the open beach policy. In late 1957 the Club was notified in writing that the new commanding General of Camp Pendleton would consider a request from the Club to re-establish controlled access to the surfing beach. Subsequent correspondence resulted in issuance of a lease, with additional and eventual renewal of the lease without modification for the next five years. From 1958, with the support of the Commanding General and a license to control access, the dignity of the Club were regained. With the Club by-laws and the Camp Pendleton Base rules being faithfully observed, the appearance of the beach and the conduct of the surfers improved. At this point the Club was running smoothly, and it had a well-established relationship with CampPendleton. The one-dollar-a-year lease from the Marine Corps further solidified the arrangement with the Club. Operating with nearly a $25,000 yearly budget, the Club kept the military happy by hiring full-time maintenance men, E.J. Oshier and John Huckins, to clean the beach and maintain the heads. San Onofre was without the cleanest, most organized beach in Southern California. “I was getting three bucks an hour working for E.J., almost double what minimum wage was back then,” says 45-year-old Bobby Lombard, current President of the SOSC and former teenage beach sweeper. “E.J.’s limit was a match stick. If he found anything bigger than that he was gonna get on you about it.” In 1958 all 288 active members renewed. The 68 inactive past members also renewed, and 150 new members brought the total to 506 memberships. As in earlier years, there had been talk of limiting the Club membership. The Marine corps had been pressing for this limitation. It was voted and passed at the annual Club meeting in 1958 that the membership would be limited, and a waiting list would be established. New memberships would be accepted only as old members dropped out, and with Marine corps concurrence, and as facilities were expanded. From 1959 through 1961 the Club functioned with Barney Wilkes again as President. In the past, officers had been elected at the annual meetings on the beach. It was the general feeling that this procedure had become impractical and non-representative. At the annual beach meeting on August 27, 1961, it was proposed and so voted that the bylaws, as prepared by Andre Jahan, be changed to provide for a “mail vote” election of a Board of Directors. In turn, the elected Board would then elect the Club officers. In April of 1962, the newly elected Board of Directors elected Al Dowden to again serve as President of the Club. Of the many accomplishments of Al Dowden in the improvement in Club management and efficiency, it should be noted that he was the instrumental force in establishing the position of a hired professional Club Manager. By this time the Club had expanded to 800 paid members. It was evident to President Dowden that the part time management of SOSC’s affairs was no longer effective. In addition to the guards for the gate, the Club now employed John Huckins and E.J. Oshier as beach maintenance men, and a young but efficient corresponding secretary named Pat Hazard to handle Club commitments and activity. In 1963 the newly elected Club President. Gene Hornbeck, hired Lex Stout as our first manager. Lex was active and effective in his job, but subsequently suffered a stroke and passed away after his first year. Jerry Gaffney, who remained as manager until 1970, replaced him as interim manager. The mid-‘60s saw the beginnings of troubled years ahead for the Club, starting with the coming of the nuclear reactor at the south end of the beach. “It was an unknown entity then,” remembers John Waters. “Some people thought the warm water would be good, you wouldn’t have to wear wetsuits, you know? Others feared it would suck all the fish away.” Although most Club members didn’t know what to make of it, one of the more outspoken and controversial members, Tom “Opai” Wert led many of the anti-nuke demonstrations against the plant. Some were supportive of his efforts, either because they feared the results of the Nuclear power or because they thought the plant would bring more crowds. But others felt it brought unnecessary media attention to San Onofre Beach, and contravened Club tradition – the “no publicity” policy that dates back to before the Club’s official inauguration. An article, “The ski Bums: Revisited,” in the 1950 August issue of LIFE had blown the lid off the informal attitude Pendleton had about the surfers. It added to the already existing military embarrassment about the area, not to mention attracting those who weren’t already in the know about surfing at San Onofre. Since then, and even to this day, a number of Club members refuse to talk with any media about “Nofre. In March 1965 Andre Jahan was elected President. Andre one of the earliest pioneers of the Club was an avid surfer and committed member. He was a popular and respected president, but his term was to be cut short. In September of 1966, he passed away doing what he loved most: surfing in front of his CyprusShores home. Andre was replaced by Jack McManus who was vice-president at the time. In February of 1968, the sad news came that our founding father, Barney Wilkes, had died in Mexico. Barney had served six terms as President in our Club’s first ten years of formal organization and had served as consultant and adviser for many more years. Jack McManus passed away in July, 1969. In the spring of 1969, exceptionally severe winter rains caused unprecedented damage to the road, parking areas, and the beach. Serious repair work was required, and the SOSC stepped up to the plate with financial investment to solve the problem. With the support of the Marine Corps, Club member Art Beard, a lifelong surfer and civil engineer, directed heavy equipment to clear sand and dirt, making the beach accessible again. Without the Club’s resources, that disaster could have marked the end of the San Onofre Surfing Club, as we know it today. In April 1969, the Board of Directors elected the long time surfer and diplomat, Doug Craig, to lead the Club through the troubled years ahead. The controversial nuclear power plant had been built on our immediate south and had, for a time, threatened the public usage of the beach. President Nixon had established the Western White House just up the beach to the north. The obscure little fishing and surfing camp of the 1920’s now had unwanted national prominence. Everybody, surfer and non-surfer alike, had heard the name San Onofre. Headline-hungry politicians proclaimed they had uncovered an outrageous condition, whereby a select group of privileged individuals had isolated and gained control of a magnificent public beach area for their own private and exclusive use. To further complicate matters, the federal government was in the process of turning over certain portions of the old Rancho Santa Margarita, on a 50-year lease, to the State of California for development of a new state beach and park. This lease included all of the beach area traditionally leased and maintained by the San Onofre Surfing Club. It would be the culminating crisis of the Club’s history, putting it to the ultimate test. Like every crisis facing a community, strong leaders are always essential to the solution. With the looming likelihood of a State Parks take over, the Club was facing the potential loss of all it had created. Great leadership was needed for its very survival. Some leaders are born, some become great through their efforts, and some have greatness thrust upon them. In the case of the San Onofre Surfing Club, it was lucky enough to get all three. In the spring of 1970, Sam Conroy relieved Jerry Gaffney as Club manager and pursued his duties with unprecedented vitality. An excellent skier and surfer, Sam assisted Club President Doug Craig and the Board of Directors. An ex-Marine officer and popular San Clemente High School teacher, he effectively handled the liaison work with both military and State personnel, in addition to the regular duty with Club members. Probably Sam’s most noteworthy quality has been his ability to take any small insignificant incident out of his past and turn it into the damnedest story you ever heard. But even Sam could not stop the inevitable march of national politics and military bureaucracy. By 1971, a tentative lease had been prepared whereby the U.S. Navy would lease certain parcels of CampPendleton land to the State of California to be used as a State Park and Beach. This included the entire San Onofre surf beach. It was generally believed that President Nixon wanted a California State Park and Beach to be named after him. On Saturday, August 31, 11:00 AM, Lease NF®13233 was signed into effect at the Western White House. President Nixon was present and all smiles at the signing. However, at no mention in the lease of the actual name of the park or beach. The lease took effect at midnight, August 31, 1971, for a period of 50 years, for a total consideration of $1 to be paid to the U.S. Government. It was the end of an era for many Club members. Without the valuable benefit of restricted beach access, membership renewals dropped from 1,000 memberships at $50 annually, with 2,000 names on the waiting list, to less than 300 renewals at $5 that following year. There seemed little reason to pay dues if you could just stroll in for only $1 a day. Talk of disbanding the Club was even proposed, but President Doug Craig insisted it stay together, remain strong, and do what it could to work with the State to preserve the beach for future generations. The State Parks Division announced immediate plans to pave and curb the entire dirt road and lot that had been untouched for almost 35 years. There was even talk of the state just shutting the lower beach lot altogether, making everyone park on the bluff above and walk down a flight of stairs. Fortunately, that never happened. At the spring meeting in 1973 the question was raised as to whether or not the Club should disband. The question was overwhelmingly rejected by all present. Since the Club no longer had to carry the annual expenses of gate guards and maintenance crews, the annual membership dues were reduced by unanimous voice vote, and the paid services of corresponding secretary Pat Hazard and Club manager Sam Conroy were placed in an “as needed” status. Foreseeing the previously unthinkable possibility that this venerable surfing Club could cease to be, Club President Doug Craig, the officers and Board of Directors (in concurrence with the general membership) authorized the funds to have a yearbook published. All Club members were encouraged to contribute photos and other memorabilia. Babs Fitzgerald spent the entire winter with layouts all over every piece of furniture in her home. Val Hying spent months of research, correspondence, and interviews with old timers such Ted Sizemore, Pete Peterson, Loren Harrison, Pop Proctor, Duke Kahanamoku and Dutch Miller. Published in 1974, the yearbook was mailed to all 1,000 members of record. Club members who had the foresight to hang on to their copies saw them become treasured collector’s item. The Club was facing the biggest challenge in its history and members knew it. The loss of the Club’s lease forced it to reinvent itself and hope it would be strong enough to survive. “But what doesn’t kill you makes you stronger” says Doug Craig, “You have to accept life’s changes.” The Seventies was the decade where the radical cultural concepts of the Sixties shifted into the mainstream. Change was everywhere, and the SOSC made the biggest change in it life – from a club of privileged power to a community of protective pride. The social life at the beach had its share of changes in the Seventies as well: Volleyball became a major activity in the lives of many of our surfers, with an annual tournament resulting in participation that would compare to our annual surfing contest. The marbles contest, the Bocci Ball contest, the table-folding contest, all were received with overwhelmingly appeal. The 1973 Club Olympic Games, with competition in horseshoes, darts, chug-a-lug, tug-o-war, etc. brought out the best of participation and competitive spirit in hundreds of Club members, young and old alike. The pie throwing contest, to raise a donation for the church in San Clemente, resulted in over $200.00 in contributions. The annual Costume Promenade produced an astonishing array of prize-winning costumes. As the beach opened up to a more public place, the need for more formalized experiences, including surf schools, citizen advisory boards, and events. In the mid 70’s, Dorian “Doc” Paskowitz opened his Surf School at ‘Nofre and began to teach visiting surfers from all around the globe about the physical, spiritual and scientific value of surfing and its surrounding lifestyle. Still a fixture at the beach, the entire family has at one time or another been involved with the School which has now introduced the wonders of wave riding to hundreds of surf stoked kids. By 1978, California was experincing the sixth year of a drought that produced water shortages, incomparable vintages of wine and a record number of sunny days at the beach. It would also be a year of great south swells, remembered by some as one of the best. Around 1979, the Beach Citizens Advisory Committee was formed, for the purpose of providing advice and assistance to the State arks System regarding the policies and requirements of the beach and its users. A 12 member panel (with 6 from the Club and 6 from the State) It included former San Clemente mayor Donna Wilkerson, Benny Merrill, Sam Conroy, Les Williams, Tom Turner, Lynn Hicks Art Beard, and Ruth Yielding. Later the group would expand to include Steve Pezman, and Jack Stowe from the Parks Department, as well as many other guest committee members who gave perspective and expertise to the board. It was (as Charles Dickens said) the best of times and the worst of times. The fabric of the Surfing Club had been stretched to the breaking point, battered by national politics, burned by President Nixon, torn by desertion, and frayed by the winds of change. But thanks to the remarkable solidarity of the membership, the dedication of Club President Doug Craig, and the leadership of the Board, it had held together. The decade of the Eighties was marked by efforts to preserve the physical elements of the beach and begin to commemorate the traditions of the past. While challenges continued the SOSC met them with continued vigilance. From 1971 to 1982, the State did very little to improve conditions at the surf beach. They built a different entrance with a toll gate, began charging $2 per day, slipped in an occasional life guard tower and placed logs all along the road to separate the road and parking area from the beach. The most notable change was to restrict maximum occupancy to 350 vehicles at any given time. In 1982, the State built five restroom buildings on the bluff side of the road with an outdoor, cold-water shower at each. By 1983, the run-off water from those showers was literally destroying the road with ever deepening trenches. SOSC President Bob Fitzgerald had pleaded and reasoned for several months with State Parks District Superintendent to let SOSC correct the problem. The answer was always the same: “It has to wait for approval from Sacramento.” Finally, Bob took the bull by the horns and work parties were formed and the water lines were piped under the road and new showers were constructed on the beach side of the road. As a result, bamboo flourished with the water run-off, and the road was once again saved by your SOSC. Thirty years after SCOC’s inception, the Club officially designated June 30th as Founders Day, in 1982. Designing a special invitation from an old photo of San Onofre, the Club commemorated the date with one of its most memorable parties, bringing members from far and wide across various decades. It was a poignant celebration, where many of the old timers would get together for the last time. Some years are remembered for the memorable events, some for the great weather, some for the epic surf. At San Onofre, the south swells are the most impressive, when the south facing beach and cobblestone shelf create a picture of point perfection. For San Onofre the summer of 1984 was one of those. Big, deep-fetch Southern Hemi’s from New Zealand, combined with back–to-back Mexican Chabascos, created weeks of solid overhead swell. As word traveled up and down the coast, surfers traveled from both north and south of San ‘O to sample the long reeling waves. And there was so much surf nobody seemed to even notice the extra crowds. Although everyone had always celebrated it, in 1985 a group of people at San ‘O began to formally celebrate Memorial Day with a formal flag raising ceremony, complete with Kate Smith belting “God Bless America”, and Mike Gleason on the bugle. The festivities include some very innovative imbibing, courtesy of the Green family. And even though no one can remember the precise year, the New Year’s Day Polar Bear Dip (a dawn patrol go out with no wetsuits) is a tradition that is firmly entrenched from the mid-eighties. The Nineties was an era of growth, celebration and revival. While the previous decade had kept the flame alive, this one was marked by a new found energy that produced a resurgent pride and productivity. The explosion of long boards brought a renewed set of members, while the economic affluence of the era created a welcome sense of comfort and collective optimism. The Hawaiian Surf Club of San Onofre was founded in January of 1990 by a group of transplanted Hawaiian surfers that wanted to share the Hawaiian culture and the Aloha spirit in surfing. “The surf Club was founded in honor of one of Hawaii’s legendary surfers, Mr. Raymond Leialoha Patterson, one of the Patterson Brothers,” says Club President Paul Strauch. The 1989 Desert Storm Parade was started as a lark by Don Craig and his large, extended gang at the north end of the beach. Responding to complaints by Viet Nam era vets that there was never a parade to honor military men from the era, Don announced that there would be a full military revue complete with General Shwartzkoff. Then with the help of the Point Crew, they threw together a 1971 Cadallac convertible, some vintage jeeps and other military vehicles and dressed up in combat uniforms of every branch of the service. With Doug’s brother Rod played General Shwartzkoff and Joe Weaver’s dad handled the role of General Colin Powell, they started down the beach at the announced hour. The parade was a sensation. By the time they reached the south end of the beach, a crowd of several thousand had lined the roadside, creating one of the most well attended and successful spontaneous events in the history of the Club. The Point crew, a sub-tribe of individuals who generally surf their point break exclusively, have always considered their area as a somewhat independent territory. It was not surprising that after a number of spots along the State Beach have become identified by their architectural structures (Pink Pole, the Shacks at old Man’s and Dog Patch, the Volleyball Courts, the Showers at Four Doors) that the Point would want a statement of their own. In 1995, the Point crew created their own unique sculpture sign denoting “The Point”, using recessed concrete and glazed blue tile. While the Club has no particular interest in promoting contests at San Onofre, the State Park has allowed a small number of events to take place on an annual basis. The Hobie San Onofre Classic is one of these, and it is often attended by Hobie Alter himself. A frequent competitor in the early years at Nofre, and a long-time Club member, Hobie is still a local hometown favorite. The event, started in 1995, has continued to flourish, adding a “Concours De Elegance” classic board display featuring some of the most outstanding examples of rare boards found outside an auction room or private collection. The Roxy Wahine Classic, first run in 1996, has been a very popular girl’s contest that is also one of the few allowed by the State. Produced by Alan Seymour (whose Club membership dates back to the early 70’s) it attracts a huge contingency of younger girl surfers who use the event as an introduction to contests, or to cut their teeth for other more intense competition. In 1997 Steve Gibby produced the documentary surf video Surfer’s Mecca – San Onofre in collaboration with LongBoard Magazine. A fairly comprehensive documentary, it caught a good deal of the flavor and feeling of the Club as well as some unique water footage rare for San Onofre. Perhaps the most impressive aspect of the film was the hot footage of so many well known surfers – from Hawaiian legend Rabbit Kekai, and former World Champ Joyce Hoffman to top pros Steven Slater and Jeff Kramer, to current longboard maestro Joel Tudor and his mentor, David Nuuhiwa. Colin McPhillips, a long time member, began training for the World Longboard Championships, which he eventually won. He would be the fourth Champ to have used San Onofre for his training grounds, joining Joyce Hoffman, Corky Carroll and Rolf Arness, three of the most dominant and impressive surfers of their respective eras. Colin has followed suit, capturing the 1998 World Longboard title and then winning again in 2000 and 2001 for two titles in a row. About the same time that the Club was producing the 1972 yearbook, it also produced another smaller classic: the San Onofre Surfing Club Cookbook. In 1999, the Club produced a new edition with updated recipes from many of the younger members as well as the tried-and-true standards of their elders. Illustrated by surf artist and Surfer Magazine creator John Severson, it contains such classics as Polly Buckingham’s Emergency Potato Salad, Dianne Jappe’s Evening Glass Off Guacamole, and Egg Foo Young by Pop Proctor (who at 91 was considered the oldest surfer in the world). It also includes recipes unique to San Onofre such as the world-famous Middle Beach Pit Barbeque by Chuck Joyce (used for large events in every decade) the legendary Shark Steak and Subgum Stew by James Arness (passed down from a recipe of Whitey Harrison in the Thirties) and the and Guard Gate Special contributed by Steve Pezman (cooked on your surf mobile’s engine manifold on the drive to the beach.) The New Millennium Huell Howser, the National Public Broadcasting host of the critically acclaimed “California's Gold” series about notable people and places in the Golden State, produced an episode on San Onofre. It aired in 2000, and was well received by the general television audience. Unfortunately (or maybe not so) the camera crew arrived on one of the flattest days of the year. Although the show producers did capture some great interviews with many of the Clubs leading lights, Hoole was pretty clue-less when it came to wave knowledge. And while the film crew did shoot a few of the musicians playing the Bamboo Room, they didn’t hang around to shoot the Wednesday evening sessions, when the real magic of the music emerges. But that’s showbiz, and San Onofre really isn’t about that at all. After years of outstanding performances at Old Man’s, (which probably rivaled any tandem talent pool in the world) Steve and Barrie Boehne, produced a video chronicling the fine art practiced at San Onofre. Titled The Art Of Tandem Surfing, it is an elegant testimony to the grace and beauty of tandem riding which has been kept strong and vibrant (like some lost art) through the years here, thanks to Steve and Barrie’s efforts and a host of excellent tandem partners. Although Club members have exhibited their musical talents for over seven decades, no official document of the myriad music sessions had ever been archived. But in 2002, a select group of the Bamboo Room Philharmonic went into the studio and recorded a disc of 14 tracks. It is available through the Club or________. Another session is expected. In the year 2002, The SOSC membership roster again topped 1,000 in the year 2002. The Club remains solid, with strong leadership in the board of directors, both young and old. Excellent relations prevail between the Club representatives and the top officials at the District Office of the State Parks and Beaches. “The Club butted heads [with California State Parks] at first, but over the years we have been able to work together and stop proposed projects like paving the lot,” says current SOSC President Bobby Lombard. “That’s why the beach is still relatively like it was 50 years ago, except that now there is running water. If the state had their way, that place would have already been ruined. A lot of our own money is spent to preserve it. We have guys with tractors that grade the road at no expense to the state. We buy ‘em car counters, and we suggested the turnabout they use to deal with the heavy overflow of traffic trying to get in. The average wait to get in these days is three hours. It’s ridiculous. But as long as they use that car counter and keep it to 350 cars, that’s the best the Club can hope for.” San Onofre was never about who you were or what you did; it had always been, and would always be, equal ground for anybody, whether you were a ditch digger, an astrophysicist or a movie star. It was all about being there, living a carefree surfing lifestyle that began with a handful of colorable bachelors hanging out along a narrow strip of sand below the crumbling, yellow sandstone bluffs just south of the San Mateo Creek opening. “For me, the cross section of people we had in those days was amazing,” says 80-year-old Doug Craig. “From Otis Chandler, the publisher of the LA Times, to Don Cram, a Nobel Prize winner in chemistry, to James Arness, world-famous actor, to bricklayers, to carpenters to teachers - we had every walk of life at our beach.” “Whether you were a north-ender, middle-ender or south-ender, we lived our lives there as a group, as a family, really,” recalls John Waters. “And when we went home we went back to our private lives. You might have been a democrat or republican or somebody famous. It didn’t matter; we didn’t treat anyone different than the next down there. Over the past five decades, some things have inevitably changed. Crowds have increased, loved ones have passed on, and many of today’s beach-goers are not the diehard, tight-knit clan that once ruled its shore. Most certainly, more changes are in store. It would be naïve to expect otherwise. In the case of our Surfing Club and the San Onofre Beach, some changes may not necessarily denote progress. But whatever the future brings, the members of San Onofre Surfing Club will always take pride in the lives they have lived and the families they have raised in this uniquely wholesome environment. The outstanding safety record of our beach and the deep social camaraderie of our Club members will live indefinitely in the collective memory of all who have experienced this remarkable way of life. Because of this rare place called San Onofre, this tiny island of beach life in a sea of global chaos, our children past and present, are the finest group of boys and girls and young men and women in the world. And let no one say otherwise. Despite all odds, the “Nofre experience is still intact, and the SOSC members have been successful in preserving the traditions their folks and grand-folks established almost 70 years ago. Through half a century of constant challenge, the people and families that call San Onofre home never let their life style be diminished. Today, the culture continues with the next generation of surfers who create and continue their own family traditions, and their own San Onofre way of life into the new millennium.
SAT Skills Insight Select a score band Skills needed to score in this band SKILL 1: Use vocabulary clues to determine the tone of a text (the author’s attitude toward the subject of the text and toward the audience) SKILL 2: Recognize elements of figurative language (such as metaphor) in a text Use vocabulary clues to determine the tone of a text (the author’s attitude toward the subject of the text and toward the audience)Example A reading passage Line Number Text Duke Ellington considered himself “the world’s greatest listener.” In music, hearing is all. Judging by the two or three thousand pieces of music Ellington wrote, he could probably hear a flea scratching itself and put that rhythm Line 5 into one of his compositions. For him the sounds of the world were the ingredients he mixed into appetizers, main courses, and desserts to satisfy the appetite of his worldwide audience. He wasn’t averse to going out in a boat to catch the fish himself. He would raise the fowl Line 10 himself. But when that musical meal appeared before you none of the drudgery showed. The author most likely refers to the “flea” in line 4 in order to Skills needed to score in the next band As you read a text, think about the author’s purpose for writing (for example, to express his or her feelings about a subject, to inform the reader, to present an argument). As you read a text, look for literary devices such as personification, metaphor, and hyperbole. As you identify them, think about what effect they have on the text.
In considering the history of science, it is easy to think of scientific discovery and debate as distinct from the rest of culture. Academics picked away at fossils and squinted at the eyepieces of microscopes in isolation, and only in exceptional cases did science jump the cultural barrier to challenge cherished beliefs. This is nonsense, and even though arguments over scientific minutiae may have been restricted to journals and learned societies, there have been many times when scientific discoveries have stirred great public interest. One way to get at the interaction between science & popular culture is to look for cartoons, paintings, and poetry. It would seem absurd that anyone should write poetry about a giant sloth, for example, yet the 19th century German poet Joseph Scheffel did just that. In 1789 the skeleton of a giant sloth, Megatherium, was found in the vicinity of Buenos Aires and was shipped to Madrid where it was displayed. (This individual is depicted in the illustration to the left.) It was a bizarre and fascinating representative of a lost world, and here is Scheffel’s poem about the beast, translated into English by Charles Leland; What hangs there like a frozen pig, Or knot all twisted rude? So giant lazy, giant big, In the prim-rim-aeval wood? Thrice bigger than a bull–at least Thrice heavier, and dumb– A climbing and a clawing beast, The Megatherium ! All dreamily it opes its jaws And glares so lazily. Then digs with might its cutting claws In the Embahuba tree. It eats the fruit, it eats the leaf, Soft, happy, grunting ‘ Ai!’ And when they re gone, as if with grief, Occasionally goes ‘ Wai!’ But from the tree it never crawls. It knows a shorter way ; For like a gourd adown it falls, And will not hence away. With owly eyes awhile it hums. Smiles wondrously and deep ; For after good long feeding comes Its main hard work–to sleep. Oh, sceptic mortal–brassy, bold. Wilt thou, my words deride ? Go to Madrid and there behold His bones all petrified. And if thou hast before them stood, Remember these my rhymes, Such laziness held only good In ant’diluvian times. Thou art no Megatherium, Thy soul has aims divine, Then mind your studies, all and some, And eat not like a swine. Use well your time–’tis money worth, Yea, work till death you see. And should you yield to sloth and mirth. Do it not sloath–somely !
Most PCBs are patterned using a ferric chloride or copper chloride based etch which is almost perfectly isotropic - it etches equally in all directions. This results in the mask being undercut by approximately the thickness of the copper layer (35 μm for 1oz copper). |PCB cross section showing sloped sidewalls from undercut (homemade board)| Ever since I did my first BGA board and noticed some of the pads looked a bit small I've wanted to do a formal study of mask undercut on this process so that I can design future boards with this in mind. While the undercut is small and insignificant for many applications, it can become a significant issue when working with fine-pitch BGAs, 0201 sized passives, and other very small components! This was my first time imaging PCBs at high magnifications (more than the 30x of my inspection microscope) so I fooled around a little bit with techniques. Somewhat surprisingly I found that darkfield illumination gave better results than brightfield - brightfield images tended to be far too bright due to reflections from the soldermask. My first test was one of the leftover boards from my BGA test in January. In darkfield the pads stood out very well and were easy to measure to within a few μm. All dimensions were taken with the focal plane at the top of the trace. At some point in the future I plan to cross-section a board and measure edge profiles. |200 μm nominal diameter BGA pads (darkfield image). The circles below and to the left of the pads are damage to the soldermask from a misaligned BGA reflow attempt.| Further inspection showed that one of the isolated pads (not connected to a trace) had been etched away entirely. |Brightfield image showing empty hole in soldermask where a pad was etched away. Note strong glare from soldermask in brightfield mode.| A higher magnification image allowed me to measure trace and pad sizes more precisely. |Higher magnification image of same area| In order to verify that the undercut was not specific to this one board I tried a more recent one - my 0201 passive test panel from last week. |0201 capacitor footprint| |Higher magnification view of a single pad| The same board also contained an 0.8mm BGA footprint. Pads are nominally 400 μm on 800 μm centers. |0.8mm BGA footprint| In conclusion, the fabricator used by the DorkbotPDX batch order exhibits nontrivial undercut on small features. Observed undercut values ranged from 68 to 133 μm and were fairly consistent within a single board but varied from board to board. The 0.5mm BGA test board exhibited an average undercut of 113.5 ± 19.5 μm and the 0201 test board exhibited an average undercut of 67 ± 4 μm.
Neologisms are the meaningless words used by psychotics — or so Stedman’s Medical Dictionary describes. A more contemporary description, neologism is itself a coined term that is applied to new concepts, to synthesize pre-existing concepts, or to make older terminology sound or containe a more contemporary meaning. Neologisms are especially useful in identifying inventions, new phenomena, or old ideas that have taken on a new cultural context. One of the most popular examples of the latter is Vladimir Nabokov’s creation of the word nymphet, a term used liberally wherever there is “a need to invoke the raw and radiant power of youthful sexuality, be it in the realms of the arts, of cultural commentary, psychology, or pornography.” From a monologue by Humbert in Nabokov’s Lolita: Now I wish to introduce the following idea. Between the age limits of nine and fourteen there occur maidens who, to certain bewitched travelers, twice or many times older than they, reveal their true nature which is not human, but nymphic (that is, demoniac); and these chosen creatures I propose to designate as “nymphets”. It will be marked that I substitute time terms for spatial ones. In fact, I would have the reader see “nine” and “fourteen” as the boundaries – the mirrory beaches and rosy rocks – of an enchanted island haunted by those nymphets of mine and surrounded by a vast, misty sea. Between those age limits, are all girl-children nymphets? Of course not. Otherwise, we who are in the know, we lone voyagers, we nympholepts, would have long gone insane. Neither are good looks any criterion; and vulgarity, or at least what a given community terms so, does not necessarily impair certain mysterious characteristics, the fey grace, the elusive, shifty, soul-shattering, insidious charm that separates the nymphet from such coevals of hers as are incomparably more dependent on the spatial world of synchronous phenomena than on that intangible island of entranced time where Lolita plays with her likes. The etymology of Nabokov’s neologism nymphet, most simply, draws on the modern English meaning of “nymph”, that of a young girl. Furthermore, it more poetically draws on the Greek myth of the nymph — beutiful creatures with the power to capture men’s attention, bringing about dumbness, besotted infatuation, madness, and even stroke. Humbert refers to himself as a nympholept, a man entranced by a nymph — and another of Nabokov’s neologisms. “This is more than mere infatuation — according to legend, nymphs could reciprocally grant their nympholeptic worshippers powers of prophecy and great poetic skill, which goes some ways towards accounting for the manner in which Nabokov expresses such complex [word] play in the narrative voice of Humberts, an otherwise fairly ordinary man.” “Lolita” has also become one of Nabokov’s great literary neologisms, to which it has largely eclipsed even the popularity of his “nymphet”. “Surely it is one of those great ironies of life that Nabokov, one of the greatest wordsmiths of modern times, is most widely recognized for the coinage of a term he wasn’t even trying to introduce.”
Notre Dame de Paris - For other uses, see Notre Dame Notre Dame is a very old cathedral in Paris, France. It is one of France's most famous landmarks and many people visit it each year. The cathedral is on a small island on the River Seine. It is the cathedral of the Catholic Archdiocese of Paris, one of 23 archdioceses in France. The cathedral was built during two centuries. It is a very good example of Gothic Architecture in France. At the time it was finished, the cathedral was the biggest such building in Western Europe. Between 1844 and 1864, the building was renovated under the direction of Eugène Viollet-le-Duc, who incorporated elements into the design that the medieval building had not possessed before. The two rose windows which ornate the transept are among the biggest in Europe: each has a diameter of 13m. The bronze plate in the parvis before the cathedral is the origin of all road distances in France calculated from Paris. The national geographic institute in France uses the building as a benchmark. The coordinates and altitude for this point are known precisely. Victor Hugo wrote a famous story. The story takes place around the cathedral. The English title is 'The Hunchback of Notre Dame', but in French it is called 'Notre-Dame de Paris'. Tradition has it that Notre-Dame’s first stone was laid in 1163 in the presence of Pope Alexander III. On May 21, 2013 French historian Dominique Venner committed suicide in at Notre Dame de Paris because of his beliefs for gay rights. About 1,500 people were evacuated from the cathedral.
Sarah Yorke Jackson Sarah Yorke was born in Philadelphia, Pennsylvania. Her parents were very wealthy. The exact date of her birth is not known. But, it was in July 1805. Her father’s name was Peter Yorke. Her father had worked as a sea captain, and then became a merchant. Her mother’s name was Mary Haines York. Her father died in 1805. While going to New Orleans, her mother died in 1820. Sarah had two other sisters. Her two aunts raised all the three children. On November 24 1831, Sarah married Andrew Jackson, Jr. He was the adopted son of president Andrew Jackson. The couple lived in the family plantation farm in Tennessee. There in 1834, a fire destroyed the main house. They came to Washington, and started living in the White House. By that time, they had become parents of two children. The couple and their children reached the White House on November 26 1834. She immediately started to act as the First Lady. At the same time, the President’s niece Emily Donelson was also acting as the First Lady. This is the single example of two ladies acting as the First Lady in the history of the United States. In 1836, Emily died of tuberculosis. After that, Sarah continued to act as the First Lady. President Jackson’s term as president expired in 1837. Until then Sarah remained in the White House, and continued to act as the First lady. By that time, a new house was constructed at Tennessee. The couple and the former president lived there for about eight years. In 1845, Andrew Jackson died. She continued to live there with her husband. When the American Civil War broke out, the couple moved to Mississippi. She died on August 23 1887.
Lately the flu, specifically the swine variety, has been on everyone’s minds. While the H1N1 swine flu doesn’t pose much of a threat to our pets, there is a risk of contracting the H3N8 dog flu, a highly contagious respiratory illness. Last week, the U.S. Department of Agriculture announced that it had approved the first vaccine for the canine influenza . The vaccine does not always prevent infection, but it decreases the severity and duration of the illness. Canine influenza was first identified in 2004 when pneumonia killed a third of racing greyhounds in Florida. Today, the flu has been found in 30 states and the District of Columbia, mostly in close quarters like animal shelters and pet stores. The flu has hit Florida, New York, Philadelphia, and Colorado the hardest. Dr. Cynda Crawford, one of the early researchers of the virus , estimates that over 10,000 dogs have been infected in the United States, a hefty number but small compared to the nation’s 70 million dogs. At one time there was a fear that the flu would kill one to ten percent of the nation’s dogs. Most dogs have no immunity to the dog flu, so it has the potential to spread quickly. The virus can be passed through a water bowl or person’s clothes, no direct dog-dog contact is necessary. According to the U.S. Centers for Disease Control and Prevention, there is no evidence that dogs can transmit the virus to humans. Symptoms, which may not show up until several days after contracting the virus, include coughing, sneezing, and nasal discharge. However, a laboratory test is the only conclusive way to know if you dog has canine influenza. Most pups make a full recovery, but if it progresses to pneumonia, it can be life-threatening. Unlike human influenza, the dog flu infections occur year round. I try to minimize the number of vaccines that my pets receive. For instance, my crew gets blood titers instead of boosters at their annual checkup. My dogs seem to be in the moderate risk group -- I never board them, but they do have a lot of contact with other canines at obedience school and agility trials. I’m leaning towards not vaccinating since the overall risk doesn’t seem high, but I plan on consulting with my veterinarian to get his opinion before making a final decision.
A local effort is under way to keep plastic from farms out of the waste stream and put it to good use. The Otsego County Conservation Association signed on with Cornell University to represent Cornell’s Recycling Agricultural Plastics Project in February, and is ready to reach out to local farmers to help them participate in this statewide effort. OCCA Program Director Travis Sauerwald will lead the local project in conjunction with RAPP coordinators, the Otsego County Soil & Water Conservation District and the state Department of Environmental Conservation. “RAPP provides an exciting, environmentally responsible method of disposing of agricultural plastics by recycling old products into new,” OCCA Executive Director Darla M. Youngs said. “We’re eager to start working with the Otsego County agricultural community through this statewide recycling initiative,” she said. According to the media release, farms use a lot of plastic in the growing, processing and storage of crops, and up to the point, the large plastics have been difficult to get recycled, in part because they are cumbersome to manage. The RAPP program will take agricultural plastics and convert them for use in sidewalk pavers, plastic lumber, oil and other products. The program aims to provide mobile compaction equipment for use on farms and in rural areas to make storage easier. Since 2009, RAPP has coordinated the collection of nearly 1 million pounds of used plastic that would otherwise have gone into landfills, been burned in open fires, or left behind in fields. Sauerwald will be working directly with Otsego County farmers to develop a recycling plan that meets their needs and the requirements of the program to reroute agricultural plastics from the waste stream. “When a farm collects enough plastic to make a full 1,000-pound bale, RAPP will arrange to have a plastic baler brought to the site,” Sauerwald explained. “For those handling smaller volumes, their plastic can be baled along with plastic from other farms.” Otsego County users of film or rigid plastics in agriculture — including livestock farmers, landscapers, producers of maple syrup and horticultural crops — can call Sauerwald at 282-4087 to learn more about the RAPP program and how to participate locally. Site visits will be scheduled, after which Sauerwald will assist program participants in developing individualized best management plans to prepare their plastics for recycling. More information about the initiative is available onlinen at http://environmentalrisk.cornell.edu/AgPlastics.
Area of Parallelograms and Triangles Free Geometry Worksheet - Area of Parallelograms and Triangles 5. To find the area of parallelograms and triangles Problems in this worksheet require the use of the formulas for the area of parallelograms (A = bh) and triangles (A = 1/2bh) to find the area of different polygons. # of Problems: 7 # of Pages: 2 Answer Key?: Yes Sample Problems from Worksheet (Worksheet pictures and diagrams not shown) Area of Parallelograms & Triangles 1-3) Find the area of each triangle given the base b and height h. 4-7) Find the area of each figure.
Bendigo ) is a major regional city in the state of Victoria, Australia, located very close to the geographical centre of the state and approximately 150 km north west of the state capital Melbourne. It is the second largest inland city and fourth most populous city in the state. The estimated urban population is 91,713. It is the administrative centre for the City of Greater Bendigo which encompasses both the urban area and outlying towns spanning an area of approximately 3,000 square kilometres (1,158 sq mi). Inhabitants of Bendigo are known as Bendigonians.Originally taking its name from the Bendigo Creek, the name became official in 1891.Bendigo is one of the most significant Victorian era boomtowns in Australia. Gold was discovered in 1851 at The Rocks on Bendigo Creek and the Bendigo Valley was found to be a rich alluvial field where gold could easily be extracted. News of the finds intensified the Victorian gold rush bringing an influx of migrants to the city from around the world within a year and transforming it from a station to a major settlement in the newly proclaimed Colony of Victoria.Bendigo is notable for its Victorian architectural heritage and gold mining history. Since 1851 over 22 million ounces of gold have been extracted from its goldmines, making it the highest producing 19th Century goldfield in Australia in the nineteenth century and the largest gold mining economy in Eastern Australia. It is the largest finance centre in Victoria outside of Melbourne as home to Australia's only provincially headquartered retail bank, the Bendigo Bank, and the Bendigo Stock Exchange (BSX).
In any given year, 1/3 of the uninsured go without needed medical care. Eight million uninsured Americans fail to take medication their doctors prescribe--because they cannot afford to fill the prescription. Two hundred and seventy thousand children suffering from asthma never see a doctor. Three hundred and fifty thousand children with recurrent earaches never see a doctor. More than three hundred and fifty thousand children with severe sore throats never see a doctor. 27,000 uninsured women are diagnosed with breast cancer each year. They are twice as likely as insured women not to receive medical treatment until their cancer has already spread in their bodies. As a result, they are 50% more likely to die of the disease. 32,000 Americans with heart disease go without life-saving and life-enhancing bypass surgery or angioplasty because they are uninsured. Whether the disease is AIDS or mental illness or cancer or heart disease or diabetes, the uninsured are left out and left behind. In hospital and out, young or old, black or white, they receive less care, suffer more, and are 25% more likely to die than those who are insured. Medical costs account for about half of all bankruptcies, affecting over 2 million people annually. More than 9 million families spend more than 1/5 of their total income on medical costs.
Rovers roll on, horizon moons loom Q: Are the Mars' Rovers still active or did they finally say goodbye? (George, Somewhere, World) A: Opportunity and Spirit roll on. NASA extended the original 3-month mission to September (if they last that long). The early lander, Spirit, tootling in northern climes half a planet from Opportunity, has racked up a record 1.9 miles (3 km) — 25 times farther than the last (1997) mission rover and 5 times farther than planned this time. Meanwhile, down south, Opportunity finished the rovers' primary mission in March when it found evidence of an ancient shallow salt lake. Water once flowed on Mars!!! But, who knows what else they'll find? More geological clues to Mars' watery past? Divining the seasons on Mars from atmospheric hints? Evidence of life!? By the way, on May 25, mission controllers told Opportunity to creep 4 inches (10 cm) closer to the edge of Endurance Crater. Opportunity "saw" the crater's edge in its navigational camera and stopped ("Good grief! No!"). Since then we have sent commands to bypass such "over conservatism." On June 10, she tentatively dipped over the edge and then returned successfully to brim safety. A few days later she ventured about 13 feet (3.9 m) deep into the crater, got comfortable, and began drilling rocks. •Mars Exploration rover mission: latest news •NinePlanets.org: Mars by Bill Arnett Q: Recently in the early West Texas morning the moon was setting through a dusty horizon. Why did the moon seem so incredibly large? I am guessing it is the same reason the sun can look so large. (Scott, Tuscon, Arizona) A: It is the same reason that the setting Sun looks huge. Also constellations near the horizon look large for that same reason. The reason, though, is a mystery. We can rule out some hypotheses. Others make some sense but don't explain the whole phenomenon. The illusion: Some people think the Moon looks twice as large near the horizon as up in the sky. Most think it's about 50% to 75% larger. A few don't think it looks larger at all. For most people, however, the effect is large. The Moon doesnít change size as it rises. The atmosphere doesnít make it look bigger. Weíve checked this out by comparing photographs. The Full Moonís apparent size is constant 0.5° as it moves from horizon to horizon across the sky — about the size of a pea held at armís length. The bending of the Sunís rays through the atmosphere does distort the low Moon: a slight 1.7%. Thatís too small to notice with the naked eye and the effect is in the wrong direction: refraction makes the horizon Moon appear flatter (smaller), not larger. We can refute a common explanation — reference cues, or the idea that we compare the moon to trees and such when it's closer to them — simply by looking at the horizon Moon between our legs or standing on our heads. The Moon looks normal size then (for most people). Also airline pilots above the clouds see a "big" Moon when it rises or sets. The horizon Moon looks big to sailors on a featureless ocean or to farmers on a flat plain. So, the explanation can't be that the presence of nearby objects (for example, trees, mountains, or roads) makes the Moon look larger than it does when high in the sky where we have no reference objects. It's something about how our brains process information — filling in details from memories or allocating the brain's resources. "Many visual illusions result from the brain's attempt to reconcile conflicting information into a sensible picture," says Donald Simanek, emeritus physics professor at Lock Haven University of Pennsylvania. "Multiple depth and distance cues provide consistent, unambiguous information for everyday nearby scenes that we can observe from various vantage points and even touch. But the perception mechanisms that work so well in ordinary situations are insufficient to deal unambiguously with the unusual setting of the horizon moon in comparison with the moon seen high in the sky. For many people, this distortion looms large." But how? As early as 7th century BC, Chinese and Greeks wondered about the illusion. Aristotle pondered the matter in 350 BC. And the reason still escapes us. Good question. • Lock Haven University: The Moon illusion, an unsolved mystery by Donald E. Simanek (Answered June 25, 2004) April Holladay, science journalist for USATODAY.com, lives in Albuquerque, New Mexico. A few years ago Holladay retired early from computer engineering to canoe the flood-swollen Mackenzie, Canada's largest river. Now she writes a column about nature and science, which appears Fridays at USATODAY.com. If you have a question for April, please e-mail us, including your hometown and your state (or, if you're writing from outside the U.S., your country). To read April's past WonderQuest columns, please check out her site.
Marriage is clearly a big deal for Catholics. Even many non-Catholics know that, for instance, the Catholic Church doesn’t recognize divorce and that being married in the Church is important to Catholics. Delving into Catholic teaching itself, Scripture is filled with references to marriage, and the Church presents it as a vocation and as one of the Sacraments, a visible sign of God’s gift of grace. What might be more surprising is that, for Catholics, marriage is also a key public policy issue, in fact one of six raised by the U.S. bishops when they reissued Forming Consciences for Faithful Citizenship, their call to political responsibility. This means marriage is not only something that matters to the doctrine of the Church and the private lives of the people entering into it. It matters to all society. The reason it matters is because marriage affects the common good. In fact, the two are inseparably intertwined. As the Second Vatican Council put it, “The well-being of the individual person and of human and Christian society is intimately linked with the healthy condition of that community produced by marriage and the family.” In fact, because the union of husband and wife is uniquely capable of welcoming new life into the world, the Church describes marriage as the very “condition” for society’s existence. The family founded on marriage plays a profound educational role in society. Pope John Paul II taught that the family is “the first and irreplaceable school of social life,” where each person “learns what it means to love and to be loved, and thus what it actually means to be a person.” This “dynamic of love” emanates from the total self-giving union between husband and wife. Because of marriage’s unique contribution to society, all people should be concerned with its well-being. In Forming Consciences for Faithful Citizenship, the bishops write that, in light of the tragic consequences of marriage’s breakdown or disappearance, especially for children, “policies on taxes, work, divorce, immigration, and welfare should help families stay together.” They also advocate for wages that “allow workers to support their families” and for public assistance for struggling families. In addition to urging policies that strengthen marriages and families, the bishops are deeply concerned with “intensifying efforts” to redefine marriage, namely proposals to remove sexual difference from marriage. This is not “expanding” marriage, as the bishops see it, but rather redefining it and in effect dismantling it. Sexual difference is not an optional component of marriage but rather an essential element, rooted in the nature of the human person created male and female. Both the bishops of the United States and Pope Benedict XVI have stated that defending marriage as the union of one man and one woman is, as the Pope taught in one ad limina talk, “ultimately a question of justice, since it entails safeguarding the good of the entire human community and the rights of parents and children alike.” Defending marriage does justice to the child by providing him or her with the best possibility of knowing and being loved by both mother and father together. In contrast, redefining marriage asserts that mothers and fathers are interchangeable and denies a child the right to know both a father and a mother. It also obscures the core of marriage, namely the union of husband and wife founded on sexual difference. Standing up for marriage can be difficult and often uncomfortable in today’s cultural climate, but it’s essential for the good of marriage itself and for the common good, especially for the welfare of children. The USCCB offers resources to help Catholics understand and articulate what marriage is and why it matters. Preserving, strengthening and defending marriage are matters of justice that should matter to everyone. Bethany Meola is program specialist for the USCCB’s Subcommittee for the Promotion and Defense of Marriage.
Central Heating Operation This article is about operating Central heating systems using hot water fed radiators. It assumes the system is functional. There are also articles about: - Central Heating Design which may help understanding what sort of system you have. - Central Heating Repair for help diagnosing and fixing a faulty system - Other heating topics are found by clicking the heating category link at the bottom. Fanned warm-air ducted systems are occasionally found in the UK. These operate quite differently and their use is not covered here. (There is a discussion on updating existing warm-air systems here) Time & temperature controls - Programmer and room thermostat - check there is a room stat - check no other sources of heat or valves turned down in stat area - Programmable thermostat - fully programmable or set-back - optimum start (Honeywell issue) - combis with pre-heat Setting boiler temperature control As well as a room thermostat, there is also a thermostat on the boiler itself. This thermostat adjusts the temperature of the circulating heating water. The lower the temp of this water, the more efficient the boiler operates, but the lower is the system heat output. If a heating system is failing to keep the house warm enough, turning this thermostat up often resolves the problem. In milder weather it can be turned down to improve efficiency. Some systems experience kettling when this thermostat is adjusted too high. Kettling is water boiling in the boiler exchanger when the pump stops. It sounds like banging and thumping, and can prematurely damage some boilers. Run 24/7 or turn up and down Total energy use is always less when the system is turned down when not needed. At any moment, rate of heat loss is proportional to the temperature difference between the house interior and outdoors, so total heat loss is reduced when house temp is reduced when heat ins't needed. Frost protection when house empty Set the room thermostat to minimum or 5 degrees C. Switching the system fully off instead can sometimes result in water or heating pipes freezing and bursting, potentially causing thousands of pounds of damage. Setting Thermostatic Radiator Valves (TRVs) If there is a TRV on the radiators in the room with the room thermostat, these TRVs should be set to max temp setting. Radiators in other rooms that have TRVs should have their lockshield (manual) valves set full on, and the TRV adjusted to give the desired room temp. Setting manual radiator valves Check the valves on the radiators in the room with the room thermostat are fully open. To turn these valves full, screwing the caps anticlockwise. The temperature in this room is now controlled by the room stat rather than the rad valve. Adjust rad valves in other rooms to achieve even temperature through the house. With some houses the room temps take a day to stabilise. Little used rooms will generally be adjusted to provide a lower room temperature. With a new or very maladjusted system, all valves fully open is a good place to start from. Radiators are balanced by adjusting the manual lockshield valves or TRVs. See the Radiator controls section above. Bleeding a CH system Bleeding radiators is needed when some of the radiator is hot, but the top is not. This problem is caused by air buildup in the radiators, which is caused by - sucking in air from the header tank - or pumping over in the header tank, which oxygenates the water To bleed a radiator, loosen the little bleed screw at one end of the top of the radiator using a radiator key or pliers. Air will escape with a hiss, when the hissing stops and just water comes out, retighten the screw. Dribbles can be anything from clean water to black, so have a towel ready to catch a little spillage. The need to bleed any more than very occasionally indicates some sort of problem. Topping up the system corrosion inhibitor solves this in a lot of cases. Topping up a sealed system - why? check for how pressure being lost
Secondary Air Injection A secondary air injection system was offered as an emissions control device in later revisions of cars with the MMS411A ECU (starting in late 1997). The engine requires a relatively rich mixture for smooth operation on cold start. For this reason, the ECU operates in open loop mode with a fixed fuel map for the first 20 to 120 seconds of engine operation until the oxygen sensors have heated to operating temperature. Because of this, exhaust gases contain high levels of carbon monoxide and hydrocarbons after cold starts. The unburned hydrocarbons could be further oxidized, except there is no oxygen left after combustion. By feeding air into the exhaust manifold (secondary air), CO and HC are oxidized through afterburning at temperatures over 600°C to form water and carbon dioxide. The exothermal reaction also increases the exhaust gas temperature, which warms the catalyist more quickly. To achieve efficient warmup operation, a high secondary air flow rate must be achieved within the first few seconds of engine startup, and the air flow rate must be maintained until oxygen sensor control is in operation. Air flow is maintained by an electric air pump (6). Once the oxygen sensors and catalytic converters have reached their operating temperatures, valves (4, 7) cut off the secondary air flow. What the ECU monitors An activated secondary air injection system leads to an increase in oxygen content in the exhaust (air flow from the secondary air pump). This increase is noted by the ECU by reduced pre-cat oxygen sensor voltage. If the engine management system gives the opening signal to the secondary air injection vaolve and switches on the pump, an extremely lean mixture should be found at the pre-cat sensors if the secondary air system is operating correctly.
App DescriptionThe Renaissance was a cultural movement that spanned the period roughly from the 14th to the 17th century, beginning in Italy in the Late Middle Ages and later spreading to the rest of Europe. Though availability of paper and the invention of metal movable type sped the dissemination of ideas from the later 15th century, the changes of the Renaissance were not uniformly experienced across Europe. - Paintings and sculptures - no internet connection required - Hi Quality photos - Slideshow mode - Favorite list - Quiz game - Universal app Renaissance Pro is a great collection with paintings and sculpture from renaissance movement. January 22, 2013 Initial Release
Phytochemicals: What Are They and What Benefits Can They Give to Our Bodies?Published:July 18th, 2011 Technically speaking, “phytochemical” is just another term that means plant chemical. Some of these plant chemicals provide beautiful colors to fruits and vegetables. With the aid of modern technology, scientists discovered hundreds of phytochemicals from different plant sources. Phytochemicals have different common names. The most common of which are polyphenols, antioxidants, isoflavones, flavanoids, catechines and carotenoids. Many people believe that phytochemicals have the ability to improve human health. Though, the results pertaining to the effects of phytochemicals are still not proven beyond reasonable doubt and most of the claimed benefits of these wonder plant chemicals are still left unproven, we all know that these foodstuffs are truly healthful. They aid in the proper digestive process of out foods, improve our bowel movement, decrease our risk for chronic diseases such as hypertension, cardiovascular diseases and type 2 diabetes. What are the most popular phytochemicals? Polyphenols and flavanoids Polyphenols and flavanoids are two of the healthiest known phytochemicals in this planet, says an article published by the University of Maryland Medical Center. Polyphenol is a general term used to signify a type of phytophemical and flavanoid is a family member of the polyphenol family. Laboratory studies show that certain flavanoids have the ability to suppress the growth of tumors, prevent blood clot formation, interfere with the effects of sex hormones and suppress the occurrence of inflammation. The most common sources of flavanoids include onions, cranberries, celery, kale, apples, cherries, red wine, soybeans, tomatoes, tea, and thyme. Berries are particularly rich in antioxidant, a type of phytochemical that suppress the activity of free radicals, the chemicals that destroys the integrity of cell walls and gene production resulting in early aging and development of cancer. Flavanoids are further divided into subtypes. The popular of which are: Resveratrol is a flavanoid found in wine, grapes, peanuts and berries. It is particularly rich in red wine, a beverage known to decrease coronary artery disease development by 20 to 30 percent, according to Linus Pauling Institute. Other sources of resveratrol include berries and peanuts. In animal models, it was shown that high intake of resveratrol results in the suppression of tumor growth. However, this fact still needs to be seen in human subjects. In fruit flies, fish and mice, it was shown that daily consumption of resveratrol resulted in increased lifespan. The only problem with resveratrol is that it appears to have a relatively low bioavailability when ingested by human subjects because the body tends to metabolize it fast and effeciently. Quercetin is a popular antioxidant that protects cells from the damaging effects of free radicals, says the University of Maryland Medical Center. It may also act like an antihistamine or antiallergy. Furthermore, it can protect the cardiovascular system against inflammatory processes that lead to cardiovascular diseases. Green tea extract is the most common source of catechins. It acts like an antioxidant. Interestingly, it can also reduce the body fat composition of a patient that ultimately leads to weight loss, according to a study conducted by a group of scientists from Health Care Products Research Laboratories, Kao Corporation, Tokyo and published by the January 2005 issue of The American Journal of Clinical Nutrition. Phytoestrogens or isoflavanes Isoflavones are known to act like estrogen, a female hormone. According to the University of Maryland Medical Center, the consumption of soy, a food rich in isoflavones, results in the reduction of symptoms associated with the menopausal period. However, further studies are still needed to establish this fact. Lignan is another type of phytoestrogen. Flaxseed is the richest source of this type of phytochemical. As mentioned by Linus Pauling Institute, lignans prevent cardiovascular diseases as well as hormone-dependent types of cancer such as breast cancer, endometrial cancer, ovarian cancer and prostate cancer. Lignan also appears to improve the bone mineral density in post-menopausal patients suffering from osteoporosis. Brocolli, cauliflowers, kale, mustard, brussel sprouts and cabbage are the best sources of isothiocyanates, a type of phytochemical that converts the action of estrogen hormones into benign forms. Isotiocyanates also have the ability to block steroid hormones that trigger the development of prostate and breast cancers. Already heard about monoterpenes? Monoterpenes are basically essential plant oils found in rich proportions in fruits, veggies and herbs. According to a study published by Environmental Health Perspectives in June 1997, monoterpenes have the ability to prevent cancer formation process and may be effective in the treatment of early and late stages of cancer. It may also be effective in treating patients with leukemia. However, further studies are still needed to establish this fact. Organosulfur compounds appear to improve the immune system function and help the liver detoxify carcinogenic substances taken into the body. In fact, according to the University of Maryland Medical Center, it can also reduce the liver’s production of cholesterol. Garlic, onions, shallots, scallions and leeks are the best sources of organosulfur compounds. Do you know that peppers contain a unique type of phytochemical not seen elsewhere? According to the University of Maryland Medical Center, peppers contain rich amount of capsaicin, a type of phytochemical that appears reduce the levels of substance P, a chemical substance within the body that worsen inflammation and pain signals from the central nervous system. Sterols are special phytochemicals found abundantly in vegetable oils. It has different subtypes: stigmasterol, squalene, campesterol, and sitosterol. Sitosterol is the most popular among these sterols because it has a cholesterol-lowering mechanism that appears to benefit those who have problems with their blood cholesterol levels. Furthermore, many people claim that it can also help improve urine flow and excretion, therefore, improving urinary symptoms association with enlarged prostate glands in m ale patients with benign prostatic hyperplasia. The South African star grass, scientific name Hypoxis rooperi, is a plant known for its rich content of sitosterols. - Cancer.org: Phytochemicals - Linus Pauling Institute: Phytochemicals - Linus Pauling Institute: Resveratrol - Linus Pauling Institute: Lignans - University of Maryland Medical Center: Phytochemicals - University of Maryland Medical Center: Quercetin - The American Journal of Clinical Nutrition; Ingestion of a tea rich in catechins leads to a reduction in body fat and malondialdehyde-modified LDL in men; Nagao, T. et al.; January 2005 - Environmental Health Perspectives; Cancer chemoprevention and therapy by monoterpenes; Gould, M.N.; June 1997
A Blue-Ground and Green-Enamel 'Dragon' Bowl, Mark and Period of Kangxi - Photo Sotheby's the deep rounded sides rising steeply from a short straight foot to a flared rim, brightly painted in green enamel against an underglaze-blue ground, the interior with a central medallion enclosing a five-clawed dragon writhing fiercely among flames within a double-line border, the inner walls left white, the exterior similarly decorated with two five-clawed dragons striding among flames and clouds in pursuit of 'flaming pearls', above three tiers of pointed overlapping petals skirting the foot, the base inscribed with a six-character reign mark within a double ring in underglaze blue, the fifth claw of each dragon later hidden beneath black enamel; 13.8 cm., 5 1/2 in. Estimation: 500,000 - 700,000 HKD PROVENANCE: Spink & Son Ltd, London. Christie’s Hong Kong, 7th July 2003, lot 573. LITTERATURE: Regina Krahl, Chinese Ceramics from the Meiyintang Collection, London, 1994-2010, vol. 4, no. 1721. NOTE DE CATALOGUE: Two-coloured bowls of this type decorated with five-clawed dragons were the customary food vessels used in the Qing palace during large banquets, where the colours were regulated according to rank. Similar bowls were in use throughout the dynasty and the same pattern can be found with many different reign marks, those of the Kangxi period (AD 1662-1722) being the earliest. On the present bowl, the three dragons all have the fifth claw on each leg hidden under flame-like motifs in black enamel that were added at a later stage, probably to hide palace provenance of the bowl. Another Kangxi bowl of this design in the Shanghai Museum is published in Wang Qingzheng, ed., Kangxi Porcelain Wares from the Shanghai Museum Collection, Hong Kong, 1998, pl. 194. A similar pair of bowls in the Roemer Museum, Hildesheim, is published in Ulrich Wiesner, Chinesisches Porzellan. Die Ohlmersche Sammlung im RoemerMuseum, Hildesheim, Mainz am Rhein, 1981, cat. no. 111 and col. pl. 209. Compare also a Kangxi dish decorated with this design in this colour scheme illustrated in John Ayers, The Baur Collection Geneva: Chinese Ceramics, Geneva, 1968-74, vol. IV, no. A 555. Sotheby's. The Meiyintang Collection, Part IV - An Important Selection of Imperial Chinese Porcelains. Hong Kong | 09 oct. 2012 www.sothebys.com
Science Fair Project Encyclopedia Arthur Griffith (Árt Ó Gríobhtha in Irish) (31 March, 1871 - 12 August, 1922) was the founder and first leader of Sinn Féin. He served as President of Dáil Éireann from January to August 1922, and was head of the Irish delegation at the negotiations that produced the Anglo-Irish Treaty of 1921. Arthur Griffith was born in Dublin in March 1871, and educated by the Christian Brothers there. He worked for a time as a printer before joining the Gaelic League, which was aimed at promotion the restoration of the Irish language. His father had been a printer on The Nation newspaper - Griffith was one of several employees locked out in the early 1890s due to a dispute with a new owner of the paper. The young Griffith was a member of the Irish Republican Brotherhood (IRB). He visited South Africa from 1897-1898 when convalescing from tuberculosis; there he supported the Afrikaner republics against British expansionism and was a strong admirer of Paul Kruger. In 1899, on returning to Dublin, he co-founded the weekly United Irishman newspaper with his associate William Rooney (d.1901). Griffith's critique of the Irish Parliamentary Party's alliance with British Liberalism was heavily influenced by the anti-liberal rhetoric of John Mitchel, and he combined fierce hostility to snobbery and deference, and a "producerist" attitude based on skilled craft unionism , with some strongly illiberal attitudes. He defended anti-semitic rioters in Limerick, denounced socialists and pacifists as conscious tools of the British Empire, and successively praised Tsarist Russia and Wilhelmine Germany as morally superior to Great Britain. In 1904, he established an organisation called 'Cumann na nGaedhael ' to campaign against the visit to Ireland of King Edward VII and Queen Alexandra. In 1905, this organisation merged with a number of others to form Sinn Féin (Irish for "We Ourselves"). In 1906, after the United Irishman collapsed because of a libel suit, Griffith refounded it under the title Sinn Féin; it briefly became a daily in 1909 and survived until its suppression by the British government in 1914, after which it was sporadically revived as Nationality . Foundation of Sinn Féin The fundamental principles on which Sinn Féin was founded were outlined in a book published in 1904 by Griffith called the Resurrection of Hungary , in which, noting how in 1867 Hungary went from being part of the Austrian Empire to a separate co-equal kingdom in Austria-Hungary. Though not a monarchist himself, Griffith advocated such an approach for the Anglo-Irish relationship, namely that Ireland should become a separate kingdom alongside Great Britain, the two forming a dual monarchy with a shared monarch but separate governments. Griffith sought to combine elements of Parnellism with the traditional separatist approach; he saw himself not as a leader but as providing a strategy which a new leader might follow. Central to his strategy was parliamentary abstention: the belief that Irish MPs should refuse to attend the Parliament of the United Kingdom in Westminster but should instead establish a separate Irish parliament (with an administrative system based on local government). In 1907 Sinn Féin unsuccessfully contested a by-election in North Leitrim, where the sitting MP had defected to Sinn Féin. At this time Sinn Féin functioned as an IRB front organisation; it had several local councillors (mostly in Dublin, including WT Cosgrave) and contained a dissident wing grouped from 1910 around the monthly Irish Freedom . The IRB argued that dual monarchism should be replaced by republicanism, and that Griffith was excessively inclined to compromise with conservative elements (notably in his pro-employer position during the 1913-14 Dublin lockout, when he saw the syndicalism of James Larkin as aimed at crippling Irish industry for Britain's benefit). Griffith was also sexually puritanical; despite initially supporting WB Yeats' National Theatre he attacked Synge's The Shadow of the Glen and The Playboy of the Western World as slandering Irish womanhood and morality, and was extremely critical of Yeats's acceptance of a literary pension from the British Crown. In 1916 rebels seized and took over a number of key locations in Dublin, in what became known as the Easter Rising. After its defeat, it was widely described both by British politicians and the Irish and British media as the "Sinn Féin rebellion", even though Sinn Féin had no involvement. When in 1917, surviving leaders of the rebellion were released from gaol (or escaped) they joined Sinn Féin en masse, using it as a vehicle for the advancement of their demanded 'republic'. The result was a bitter clash between those original members who backed Griffith's concept of an Anglo-Irish dual monarchy and the new members, under Eamon de Valera, who wanted to achieve a republic. Matters almost led to a split at the party's Árd Fhéis (conference) in October, 1917. In a compromise, it was decided to seek to establish a republic initially, then allow the people to decide if they wanted a republic or a monarchy, subject to the condition that no member of Britain's royal house could sit on any prospective Irish throne. Griffith resigned the party leadership and presidency at that Árd Fhéis, and was replaced by de Valera. War of Independence Griffith was elected a Sinn Féin MP in the East Cavan by-election of mid-1918, and Sinn Féin routed the Irish Parliamentary Party at the 1918 general election. Sinn Féin's MPs decided not to take their seats in the British House of Commons but instead set up their own Irish parliament, Dáil Éireann; the War of Independence followed almost immediately. The dominant leaders in the new unilaterally declared Irish Republic were figures like Eamon de Valera, President of Dáil Éireann (1919-21), President of the Republic (1921-22), and Michael Collins, Minister for Finance, head of the IRB and the Irish Republican Army's major strategist. During de Valera's absence in the United States (1919-21) Griffith served as Acting President and gave regular press interviews. He was imprisoned in 1921 but subsequently released. Griffith became central to the Republic again when, in late 1921, President de Valera asked him to head the delegation of Irish plenipotentiaries to negotiate with the British government. Griffith was the member of the treaty delegation most supportive of its eventual outcome (a compromise based on dominion status, rather than a republic); after the narrow ratification of the Anglo-Irish Treaty by the Dáil in January 1922 he replaced de Valera as acting head of government. He was, however, to a great extent merely a figurehead and his relations with Michael Collins were somewhat tense. Under increasing strain because of quarrels with many old friends, and faced with a nation sliding into chaos, Griffith's health deteriorated and he died of a stroke soon after the outbreak of the Irish Civil War. In 1910 he had married Maud after a fifteen-year engagement; they had a son and a daughter. - "In Arthur Griffith there is a mighty force in Ireland. He has none of the wildness of some I could name. Instead there is an abundance of wisdom and an awareness of things which are Ireland." - Michael Collins. - "A braver man than Arthur Griffith, I never met" - comment by one of the British delegates at the end of the Treaty negotiations. - Patrick Maume , The Long Gestation (Gill & Macmillan, 1999). - There is a 2003 reprint of The Resurrection of Hungary with an introduction by Patrick Murray (University College Dublin Press). |- style="text-align: center;" | width="30%" |Preceded by: Count Plunkett | width="40%" style="text-align: center;" |Minister for Foreign Affairs 1921-1922 | width="30%" |Succeeded by: George Gavan Duffy |- style="text-align: center;" | width="30%" |Preceded by: Eamon de Valera | width="40%" style="text-align: center;" |President of the Irish Republic 1922 | width="30%" |Succeeded by: the office of President of the Executive Council The contents of this article is licensed from www.wikipedia.org under the GNU Free Documentation License. Click here to see the transparent copy and copyright details
Log In to Anagrammer Scrabble word: SCHIST In which Scrabble dictionary does SCHIST exist? Definitions of SCHIST in dictionaries: - noun - any metamorphic rock that can be split into thin layers - Any of various medium-grained to coarse-grained metamorphic rocks composed of laminated, often flaky parallel layers of chiefly micaceous minerals. - n - a rock that readily splits into parallel layers There are 6 letters in SCHIST: C H I S S T Scrabble words that can be created with an extra letter added to SCHIST All anagrams that could be made from letters of word SCHIST plus a wildcard: SCHIST? Scrabble words that can be created with letters from word SCHIST 6 letter words 5 letter words 4 letter words 3 letter words 2 letter words Images for SCHIST - Footwall fault schist microscope view. Ribbons of quartz run through ... - Rocks Amethyst Brain Rock Collection. Chadds Ford PA - iKatelin.com - Kyanite with Garnet in Mica schist 7 - Black Biotite Mica schist Rock Texture - Free high resolution photo ... - Figure 10g-2 schist. - FileChalcopyrite-Chlorite-schist Metamorphic rock near Copperopolis ... - Great forgotten metalrock. - AR15.COM - Staurolite-Garnet-Mica schist - talc schist-A - No obstante si lo que queris es gozar un rato de las preguntas ...
To use 1.1, you need Visual Studio 2008 Beta 2 Standard or Pro (NOT Express), the , and the Silverlight 1.1 SDK. The best place to get started is at Microsoft’s Silverlight page, in the Getting Started section (imagine that). They start out with a simple clock for the first example. Each hand is a XAML Paths which is rotated with a simple repeating animation. I copy the SecondHand object to create a msHand and edit its shape. The example uses inline path markup syntax, something I hadn’t come across yet. Data attribute here defines a path that starts ( (-1.5, 16) (in the Silverlight canvas, (0, 0) is in the top-left), and moves ( l) down by (0, 70), then left (3, 0), then up (0, -70), and finally closes ( z). That’s the second hand. I want the millisecond hand to be thinner, longer, and darker. Each path also contains a Path.RenderTransform element which defines its angle, and is needed for the path to rotate. Now to animate it. This is the second-hand animation, which just changes the object’s angle from 180 to 540 over 1 minute and repeats forever. Adapting it to the millisecond hand is simple: The only differences are the target and the duration. The final step is the set up the initial condition. The path definitions create all hands pointing down, towards 6, but we want them to represent the current time. This is where the C# codebehind comes in. But first, how do we get to the code-behind? Well, in version 1.0 the Canvas (root) of the XAML file had a x:Class tells Silverlight that we’re using the ClockCanvas class in Canvas_Loaded method contains C# code to find the canvas ( Canvas can = sender as Canvas), find the animations ( DoubleAnimation secAnim = (DoubleAnimation)can.FindName("secondAnimation")), calculate the current time, and move the hands into position: From properties of the animation object correspond to the respective attributes in the XAML. Adding 180 is done so the “default” is up instead of down (if date.Second is 0, the angle would be 0 and the hand would point down because of the Path definition, but we want it to point up). Duplicate this code for the new millisecond hand and replace .Millisecond, and we’re done. We now have a 4-hand clock.
The Maccabiah Games |Home - Arza Tours > Israel Travel Guide > Israeli Sports >The Maccabiah Games The establishment of the Maccabiah Games in 1932 marked an astounding achievement for Yosef Yekutieli, a 15-year-old in pre-state Palestine who was caught up in the excitement of the 1912 Olympic games. As the ‘new Jews,’ who were interested in being strong and capable, as well as more physical than the European Jews, the Jews of the Yishuv period had an interest in sports. Yekutieli ultimately received the sponsorship of the British Palestine High Commissioner Sir Arthur “Andy” Wauchope in 1931. The games were scheduled as part of a commemoration of the Bar Kochba rebellion against the Roman empire 1,800 years earlier. The British High Commissioner made the request that Arab as well as British Mandate athletes be allowed to compete along with Jewish athletes. Following a 15-year break between 1935 and 1950, the post-war games included a team of survivors from Germany. In recent years, the Maccabiah movement has expanded to games outside of Israel, including the Pan American, European, Australian and US/JCC Maccabi games and has grown to host more than 50 countries and over 7,000 athletes. As a celebration of Jewish awareness, Zionism, identity and sportsmanship, it has helped to redefine the meaning of sportsmanship for Jews everywhere. Come to Israel during a Maccabiah year. * Israel Tours * Bar / Bat Mitzvah Tours * Jewish Heritage Tours
Conservation International Takes 52,000 Photos In Camera Trap Mammal Study By Juliana Chan | Featured Research August 16, 2011 Conservation International’s first global camera trap mammal study documented 105 species in nearly 52,000 images across seven different wildlife preserves, including Indonesia and Laos. AsianScientist (Aug. 16, 2011) – The first global camera trap mammal study, announced today by a group of international scientists, has documented 105 species in nearly 52,000 images across seven protected wildlife preserves in Suriname, Costa Rica, Indonesia, Tanzania, Brazil, Uganda, and Laos. Suriname had the most diversity and Laos the least. The study, led by Dr. Jorge Ahumada, ecologist with the Tropical Ecology Assessment and Monitoring (TEAM) Network at Conservation International, has been published in Philosophical Transactions of the Royal Society B. The pictures reveal an amazing variety of animals in their most candid moments, from a minute mouse to the enormous African elephant, plus gorillas, cougars, giant anteaters and – surprisingly – even tourists and poachers. To gather data, 420 cameras were placed around the world, with 60 camera traps set up in each site at a density of one per every two square kilometers for a month in each site. After photos were collected from 2008-2010, scientists categorized animals by species, body size and diet, among other things. “We take away two key findings from this research. First, protected areas matter: the bigger the forest they live in, the higher the number and diversity of species, body sizes and diet types,” said Dr. Ahumada. “Second, some mammals seem more vulnerable to habitat loss than others: insect-eating mammals — like anteaters, armadillos and some primates, are the first to disappear — while other groups, like herbivores, seem to be less sensitive,” he added. Of the sites researched, the Central Suriname Nature Reserve presented the highest number of species diversity (28) and the Nam Kading National Protected Area in Laos presented the lowest number of species diversity (13). The body size of species photographed ranged from 26 g (Linnaeu’s Mouse Opossum, Marmosa murina) to 3,940 kg (African elephant, Loxodonta africana). With around 25 percent of all mammal species under threat and little global quantitative information available, this study fills a very important gap in what scientists know about how mammals are being affected by local, regional and global threats such as overhunting, conversion of land to agriculture, and climate change. Since 2010, more cameras have been installed in 17 new sites, including sites in Malaysia and India. Source: Conservation International. Disclaimer: This article does not necessarily reflect the views of AsianScientist or its staff.
Myths are always common when new things are developed. Technology has seen its fair share of myths and misunderstandings over the years. One of the greatest technology myths is that teenagers and younger adults are the ones that benefit most from technology and that they know better how to use it than the older generation. Many feel that the technology age has completely bypassed Generation X’ers and Baby Boomers This could not be further from the truth. In fact, while teenagers and young adults are mostly tech savvy in this day and age, they are not the only ones that know how to use current technology. A few years ago, many believed that older people could not program a VCR or that women did not know how to properly use their microwaves. As technology grows, so does the knowledge that consumers develop as to how they should use this technology. Teenagers may be able to text their friends faster than their parents or grandparents but studies have shown a stiff increase in the number of older consumers who are learning how to use various levels of technology. Some older consumers find that having a cell phone is a must in this day and age. Many use their cell phones to call for assistance when they need it or to simply keep up with family and friends. Some use various health apps that are designed to help them to take better care of themselves. There are a number of smartphone apps that are designed for the older generation and although obsession with cell phones do not normally reach into older age groups, many have begun to find joy in playing games and performing other tasks on their smartphones. Studies have also shown that the number of older people, generally those over 60, who are using various computer applications has also increased. In today’s internet age, it simply makes sense that consumers will learn how to use basic computer applications. Studies confirm that many of the older generation have taken classes or have watched video tutorials online that give them instruction on how to do certain computer related tasks. Technology is a growing industry and as it grows, so do the numbers of people who are learning how to use these newly developed apps, devices and gadgets. Several years ago, many had no idea how to use a video cassette recorder or VCR. Today, there are Blu-ray players that make television viewing nearly as profound as watching movies on the big screen. The myth that only the younger generation knows how to enjoy technology has been dispelled many times. As technology grows, so will the ages of those who are using this technology. Many of the middle class generation remembers certain technologies as they were introduced while they were teenagers. Teenagers today will know how to use certain technologies when their own children become teenagers. Technology is a growing field and consumers who learn to grow with it will find it much easier to understand newer developments as they are released. The diagnosis and treatment of ADHD has changed significantly over the past few decades. Today, many technologies are used to better diagnose as well as treat this condition. In the past few years alone, great strides have been made in the field of biometric measurement on a personal level. Technological tools are available today that allow for better monitoring of many health concerns from fatigue and food sensitivities to calorie ingestion and brain monitoring. New technology in the field of Attention Deficit Hyperactivity Disorder today even includes smartphone applications that can help to detect and even improve many of the problems associated with worsening ADHD symptoms. Lack of sleep, food sensitivities and stress can all trigger symptoms of this disorder and these new technologies offer better ways of dealing with these symptom triggers. Single channel EEG bands can now be used to send wireless signals to computers that allow users to learn more about their current state of attention. This neurological feedback tool is now available for home use by patients and allows them to better train themselves to overcome certain symptom triggers. Future technology could see applications designed to consistently monitor the brain and send updated feedback directly to smart phones that contains detailed information about fatigue and individual attention levels. The smartphone applications would notify the user by beeps or chimes that will instantly alert them of periods when they may be falling asleep or losing their attention grasp. Emotiv Life Sciences as well as NeuroSky currently provide EEG monitors that are available for home use. The products are inexpensive and require no special skills to set up and use. These two companies are also currently working on attention and brain health applications that can be used on smartphones for alerts of certain symptom triggers or effects. Hoffmann-La Roche has developed a monitoring device that is called iBrain and is a personal brain monitor that is designed to be used for monitoring and diagnosis on a professional level. It is a single channel EEG device that is very similar to the headsets provided by Emotiv Life Sciences and NeuroSky. The device is designed to monitor brain activity and provide healthcare professionals with wave data and feedback regarding how certain ADHD medications are working in patients. The company has announced that the device can be used for diagnosing and monitoring sleep disorders and the use of iBrain for ADHD diagnosis and monitoring is still currently being studied. Technology in the healthcare industry has provided a number of great strides toward better patient care and diagnosis. Personal monitoring tools are being developed at a rapid pace, encouraging those who suffer from ADHD and other disorders that better means of monitoring and treatment are soon to be available. Devices in the future will be made much smaller than what is currently available and be a bit more affordable so that patients who need them can easily access them. As technological advances in the healthcare industry increase, so will the technology behind ADHD monitoring devices. If you are unsure of whether or not your email is private, now is the time to find out. CIA Director David Patraeus thought that his email was protected and he was proven wrong. You certainly do not want just anyone having access to your mail and your privacy settings are essential in protecting you and those that you communicate with through your email account. The best and easiest way to begin is to hide your location. Every email that you send out includes a header that has your IP address. This IP address can be used to reveal who you are and where you are. Using a VPN or Virtual Private Network is a good way to mask your location. When you connect using a Virtual Private Network, the IP address that is associated with the email that you send out is the IP address of the server and not your location. You can also simply hide the communication that you are engaged in. You can find a number of free services that encrypt email communication. Some send your messages in plain text while others completely encrypt and decrypt your messages on local computers so the only person who is able to read those messages is the person that you send them to. If you are using encryption, certain people can obtain and then decrypt those messages if they gain access to your inbox. There are services available for a small charge that will encrypt your messages and then eliminate them completely after one week so they are eliminated from your computer altogether. Various communication tools will also give you communication that leaves absolutely no trace so just as soon as your recipient reads the message that you send, it is completely eliminated from your system. Higher prices systems will eliminate your messages and make it impossible for your recipients to forward, print or even copy those messages. Why you want to keep your email messages private is your own business. If you are afraid that someone may be able to hack your system and read your messages or if you are avoiding subpoenas or other legal issues, there are things that you can do to ensure that the messages you send are never seen by anyone other than who you send them to. One way to ensure that you never get mixed up in a legal battle is to simply never send emails that you should not send. If you have sensitive data that should never be shared, then simply do not share it. Even the best service in the world is not going to stop someone who is really set on getting into your mail so if you are sharing private information that should not be shared, there is no way for you to really hide it. If you are simply looking to ensure that someone does not make off with private data that may have to do with your business or personal life, these steps can help you to better protect your email privacy. The internet allows us to open the door to new and exciting worlds. We can reach out and touch almost anywhere in the universe with a few clicks of the keyboard. For all of the benefits it offers, there are dangers too. Entering unsafe websites can cause viruses or malware to affect your files and damage your entire computer system. Opening attachments that accompany emails can also allow dangerous files to enter your computer. How can you prevent unwanted intruders from entering your computer? People store all kinds of important and confidential information on their computers. A virus can destroy every file on your computer in less than 5 minutes if allowed to run unchecked through your system. Installing virus protection programs will help prevent this kind of damage and keep your system safe from crashes and corrupted files. Installing virus protection is only a small part of protecting your files and computer system. Another key way to protect your files is to limit the amount of personal information you share with others online. Never reveal passwords or information concerning online email or cloud accounts. Keeping passwords confidential protects hackers from getting into your email and including attachments that could infect your email recipient’s systems. Never reply or open emails you believe to be spam. Once you acknowledge your email account as valid by opening an email, it allows for more emails to be sent from various other spam accounts. While spam may be just a nuisance, some spam emails may contain harmless files that can corrupt files and harm your system. High speed internet companies report an increasing number of spam emails sent over personal and professional emails. With a high speed internet connection, a hacker can break into an email account and compromise your files in a short amount of time. Be aware of a website’s security. Financial websites take great pains to make sure their site is secure. Many websites have copies of their security certificate posted on their website so customers can view it for themselves. A few of the major search engines have indicators showing the level of security of each website. Most anti-virus programs will flash a warning if a dangerous or questionable site is about to be accessed. Search engines can spot dangerous sites no matter how fast the high speed internet connection. Depending on where you access your computer files will determine the level of security you can be assured of. At home your security level is high. You are using your own system and you know what your anti-virus program is capable of. If you access your personal files at work, it is wise to remember that some companies monitor the sites their employees frequent while on the job. Accessing files from a public internet provider, like that used at a coffee shop or public library, can leave your computer open to hackers and other internet predators. High speed internet connections that are rated as public are commonly known to have limited security at best. Protecting your computer and the confidential files it contains is easy if you take necessary precautions. With an effective anti-virus program and a common sense approach as to who you provide your personal information, you can keep your files safe and your computer running smoothly. An EPOS or Electronic Point of Sale is essentially equipment that a business has in place to process checkout tasks such as credit and debit card transactions. Many retailers now use POS or Point of Sale systems which are simply the location where the process takes place. The EPOS system can secure various types of different transactions as well as printed sales reports and any other data that is provided by the company. EPOS touch screens are the visual display screens that allow employees to navigate the system. This technology is not used in a number of industries from hotels and restaurants to retail stores, libraries and many other establishments. The system allows businesses to perform many tasks with more efficiency and accuracy. The touch screen technology that is included in EPOS systems will allow businesses to quickly and easily keep track of their financial flow. Many of the processes required are fully automated which helps to provide for higher levels of productivity. When using an EPOS system, businesses can quickly create sales reports and other documents needed to make quick decisions about business dealings. The system can assist management in knowing what specific products need to be ordered or which ones are not selling and should be dropped from the inventory altogether. The technology itself is very easy to use. Most touch screen systems offer customizable interfaces so the system can be completely simplified and it can provide data that is specific to the business’s needs. Employees can easily operate the EPOS system where the older POS system versions may have been a bit more difficult to maneuver due to complicated key bars and screens. EPOS technology allows for much more accuracy because stores that have consumable products can utilize the barcodes and keys to input information into registers for a much more accurate transaction. This virtually eliminates human error which ultimately stops losses altogether and offers more security over products. Businesses that have large warehousing needs can check the availability of individual products easily. The EPOS system will provide automatic updates of ledgers, data and other information to provide timely information and the touch screen technology offers real time accounting that will provide a much more accurate sales and inventory projection. The actual type of EPOS system that a business will use depends on the type of operation that the business performs. Retail stores will have different needs than restaurants. Smaller businesses or vendors will use systems that offer a more basic range of features simply because their operations are not as complicated as those of larger retail establishments. These smaller systems are just as efficient and very effective for the business that has it but larger businesses will need systems that can handle a higher range of tasks. Those interested in EPOS systems can do a bit of research on the topic and learn more about the benefits and operation of these systems as well as what specific system may be best for their individual business needs.
Leon Walras (1834-1910) transformed economics from a literary discipline into a mathematical, deterministic science. For the first time, Walras expressed rigorously the view that all markets are related, and that their relationships can be described and analyzed mathematically. These interrelated markets tend toward a "general equilibrium" position, undergoing a constant interactive adjustment process that Walras called a "tatonnement". This conception of economics led to important new insights about the stability of markets and about the capitalist economic system. A follower of Walras, Vilfredo Pareto (1848-1923), viewed economics as part of the broader science of sociology, extending Walrasian analysis to say that society at large is an equilibrium system. This view profoundly influenced the modern course of the "social sciences", in which quantitative techniques have become standard analytical tools. Many later scholars have been awarded the Nobel Prize for developing the Walrasian analysis of capitalism. © and (P)1988 Knowledge Products, Inc.
Results Around the Web for Africa Back to the Africa name page Africa is the world's second-largest and second-most-populous continent. At about 30.2 million km² (11.7 million sq mi) including adjacent islands, it covers six percent of the Earth's total surface area and 20.4 percent of the total land area. Race and ethnicity in the United States Census, defined by the federal Office of Management and Budget (OMB) and the United States Census Bureau, are self-identification data items in which residents choose the race or races with which they most closely identify, and indicate whether or not they are of Hispanic or Latino origin (the only categories for ethnicity). African Americans (also referred to as Black Americans or less commonly Afro-Americans, and formerly as American Negroes) are citizens or residents of the United States who have total or partial ancestry from any of the native populations of Sub-Saharan Africa. African people are natives or inhabitants of Africa and people of African descent. The African-American Civil Rights Movement were social movements in the United States aimed at outlawing racial discrimination against black Americans and restoring voting rights to them.
Music — especially blues and jazz — has little bits of all of us in it according to celebrated jazz musician Wynton Marsalis. Last Thursday, Marsalis’ lecture and performance “Music as Metaphor” at Sanders Theatre was the kick-off for a two year initiative “Hidden in Plain View: Meanings in American Music” with Harvard University. The series is the latest example of Harvard’s commitment to integrating arts and education. The nine-time Grammy award-winning trumpeter, documentarian and author is also a gifted griot that expertly stitched together quotes from Bessie Smith, Louie Armstrong and Nietzsche that attempted to capture in words the essence of music. With more than 70 albums to his credit, Marsalis delivered a poetic soliloquy peppered with musical vignettes for more than three hours to make audiences feel its significance. He taught concertgoers that the rhythm section is the foundation of music and the epitome of sacrifice. The rhythm section sets the tone and the pace but is always relegated to the back of the stage, never getting the glory. Coupled with the delicacy of strings and the charisma of soloists, each section comes together in a melody or riff to say something. He talked about the evolution of certain genres of music such as Negro spirituals and the blues borne out of the hearts of slaves bearing heavy burdens. Injustices like slavery, internment camps, wars and other cruelties have spawned beautifully painful melodic contributions from all over the world. But it’s up to music lovers to be in tune with where the artists are coming from and receive their messages. Throughout the ages, artists have been truth tellers for civilization; they speak about the essence of their society in ways that others cannot or will not. “You need to listen to what’s being said,” he explained. The New Orleans native is the artistic director at the Lincoln Center, and holds an honorary doctorate from Harvard. Marsalis’ motifs for his lecture were love, listening and the connectedness of the human spirit. Learning the “math of music” or music theory is not enough to move people. All the notes can be in the right place and the music can be technically sound, but if the person’s soul is not in it, then nothing is being said. Humbly, he urged audience members to listen ever so carefully to what a song is saying. In 1995, Marsalis worked on a documentary called “Marsalis on Music” where he asked numerous musicians what music means to them, and every musician talked about some form of love. Whether it’s blues, jazz, techno or rock, Marsalis claims that all music is essentially the same because it comes from the same place — the spirit. He said he believes it can help us live our nation’s creed and transcend our differences. “This music cost us a lot,” Marsalis said. “But not knowing what it means, costs us a lot more.” Grammy-winning jazz musician Herbie Hancock has been named the 2008 Artist of the Year by the Harvard Foundation of Harvard University, the foundation announced Tuesday. Hancock, the unanimous choice of the selection committee, will be awarded the foundation's most prestigious medal, which bears the signature of the university's president, at Harvard's annual Cultural Rhythms ceremony on Saturday, March 1. "The students and faculty of Harvard are delighted to present the 2008 Artist of the Year award to the legendary jazz pianist and Grammy award winner Herbie Hancock," said Dr. S. Allen Counter, director of the Harvard Foundation, in a statement announcing the award. "He is a masterful artist, a brilliant composer of modern jazz who has shared his musical gift with humanity in ways that inspire and unify." Tour mate and fellow NEA Jazz Masters Award recipient Paquito D'Rivera has achieved great acclaim in his own right as both a performer and composer of jazz, Latin and classical music. The Cuban-born D'Rivera has earned numerous Grammy and Latin Grammy awards, and was honored with the Kennedy Center for the Performing Arts' Living Jazz Legend Award in 2007. With four members - keyboardist James, bassist Nathan East, drummer Harvey Mason and guitarist Larry Carlton - who have each achieved critical acclaim individually, Fourplay is something of a contemporary jazz all-star team. The group's sound, a unique mixture of jazz, R&B and pop sensibilities, has been synonymous with smooth jazz for over 17 years. About 60 members of the toddler and preschool set were treated last week to a lesson in the sounds of jazz at the Boston Public Library. "Riffs & Raps - Jazz for the Very Young" was part of Jazz Week, a series of concerts and programs presented by JazzBoston, a nonprofit group with a mission to expose audiences of all ages to jazz and to promote Boston as a great jazz city. A group of 3- to 5-year-olds from South End Head Start, along with a handful of other children accompanied by their parents or caregivers, perched on auditorium seats or sat right up onstage as local jazz musicians Arni Cheatham and Bill Lowe gave an interactive presentation.
seems odd to say that Kents Cavern's caves in Torquay are 100 years old in 2003 - because we all know that they've been there for hundreds of thousands of years. In fact, the cavern is the oldest recognisable human dwelling in Britain, with artefacts and remains dating back half a million years. The caves are a scheduled ancient monument and historical site. The surrounding woodlands and the inside of the cavern are also a site of special scientific interest (SSSI). But the cavern's life as a visitor attraction didn't begin until the early 20th century - 1903 to be exact. The official centenary day is 23rd August, when special events are being held to mark the anniversary. caves were first lived in 700,000 years ago was when the Haldon estate - including the caves - was sold to the Francis Powe initially used the caves as a workplace to make beach huts for nearby Meadfoot Beach. It was his son, Leslie, who spent his life turning the caves into a major attraction. He laid concrete paths, put in electric lighting, and added all the usual visitor facilities. Then his son John Powe carried on the work and added modern facilities. John died in 2000 and his father died in 2001. So now, John's eldest son Nick Powe is at the helm. Excavation work is still as important as ever at the caves, and exciting new finds are regularly unearthed. When the digging began... Excavation work really got under way in earnest at Kents Cavern in the 19th century. of the exhibits on show at Torquay Museum and flints were dug up in the 1820s, before archaeologist William Pengelly set to work in the second half of the century. He found hand axes dating back 450,000 years. Importantly, he kept meticulous records of his discoveries and made a note of which layer of rock they were found. Then, between 1925 and 1941, Arthur Ogilvy unearthed many ancient Among them was a piece of bone, which was in recent years dated as 31,000 years old - making it the oldest piece of bone from modern man in Britain. In fact, the advent of modern dating methods - such as carbon dating - is helping us to build a better picture of who and what lived in the caves and when. Kents Cavern has been inhabited for around 700,000 years, and the digging will continue in the hope that remains and artefacts from the very early days can be found. For those wanting to see the finds, the biggest display is at Torquay Museum, which is just half a mile away from the caves. Article first published: 31st July 2003
Mars' entire surface was shaped by water The whole of Mars' surface was shaped by liquid water around four billion years ago, say scientists. Signs of liquid water had been seen on southern Mars, but the latest findings reveal similar signals in craters in the north of the Red Planet. The team made their discovery by examining data from instruments on board Europe's Mars Express and Nasa's Mars Reconnaissance Orbiter. They report the findings in the latest issue of the journal Science. John Carter, of the University of Paris, led the team of France- and US-based scientists. "Until now, we had no idea what half Mars was made of in terms of mineral composition," he told BBC News. End Quote John Carter University of Paris We're seeing signals of what were once river beds, small seas and lakes” "Now, with the Esa and Nasa probes, we have been able to get a mixture of images and spectral information about the composition of the rock." He explained that these instruments had revealed clay-type minerals called phyllosilicates - "the stuff you would find in mud and in river beds." "It's not the species of mineral itself that's important," said Dr Carter, "it's more the fact that the minerals contain water. "This enhances the picture of liquid water on Mars."Liquid rock Previously, researchers have seen similar signs of water in the highlands of southern Mars in rocks that are up to four billion years old. But in the northern part of the planet, more recently formed rocks have buried the older geology. The prevailing theory for why this is, is that a giant object slammed into northern Mars, turning nearly half of the planet's surface into the Solar System's largest impact crater. Dr Carter explained that this meant a thick veneer of younger rock covered the older geology, "so the craters are the only way of accessing the older stuff". But the craters are relatively small and more difficult for the orbiting probes to take measurements from. "There's also ice and dust coverage in the north of the planet, making it harder to get signals from these craters," said Dr Carter. The new findings suggest that at least part of the wet period on Mars, that could have been favourable to life, extended into the time between that giant impact and when volcanic and other rocks formed an overlying mantle. This indicates that, 4.2 billion years ago, the planet was probably altered by liquid water on a global scale. But Dr Carter said that the findings did not paint a picture of huge Martian oceans. "It was probably a very dry place," he said. "But we're seeing signals of what were once river beds, small seas and lakes."
antibiotics given to bees before there is a problem can cause a problem Symbionts as Major Modulators of Insect Health: Lactic Acid Bacteria and Honeybees Honeybees possess an abundant, diverse and ancient LAB microbiota in their honey crop with beneficial effects for bee health, defending them against microbial threats. Beekeeping management practices can negatively impact this microbiota. The crop microbiota of A. mellifera is composed of 13 bacterial species within the genera Lactobacillus and Bifidobacterium , , and it plays a key role in the production of honey and bee-bread , long term stored food for both adult honeybees and larvae. We have demonstrated by both in vitro and in vivo studies that the LAB microbiota in A. mellifera inhibit one important honeybee pathogen, the bacterial brood pathogen Paenibacillus larvae that is the cause of the brood disease American foulbrood (AFB) To dumb it down – antibiotics given to bees before there is a problem can cause a problem! Tags for this Thread
Jeffrey Ecker, M.D. is an Associate Professor of obstetrics, gynecology and reproductive biology at Harvard Medical School and a member of the Department of Obstetrics and Gynecology at Massachusetts General Hospital, where he practices maternal-fetal medicine. If you have genital herpes, is it possible to have children without them being infected? Although the herpes virus can be devastating to newborns, most women with genital herpes will have uninfected children. This happy reality is the result of both biology and careful medical care. Women infected with herpes virus "shed" the virus from their genital tract from time to time. Some women have attacks many times a year, others just once or twice in a lifetime. Herpes outbreaks are marked by small ulcers or sores and symptoms of burning, itching, tingling or pain. Sometimes, however, the virus can be detected even when symptoms and sores are absent. Babies can be infected as they pass through the birth canal when the virus is present. Usually, however, the virus is dormant and not being actively shed. So there is little risk of a baby becoming infected during labor and delivery. Also, women with a history of herpes usually have developed antibodies to the virus. These antibodies pass through the placenta to the baby, providing some degree of protection from infection during delivery. The active virus is much more likely to be shed if the mother has symptoms or sores. Therefore, to minimize exposure to the newborn, most obstetricians recommend Caesarean delivery if: There is evidence of herpes at the time of labor The membranes have ruptured, or The Caesarean delivery was previously planned. Some doctors and pregnant women will choose to use medications in the weeks before the due date to reduce the likelihood of an outbreak when it is time to deliver. If you have a history of genital herpes tell your doctor or midwife. It is especially important that you report any symptoms that make you think you might be having an outbreak at the time you come to deliver. If you have a history of genital herpes your provider will ask you about symptoms when it comes time to deliver and will do a careful examination of the vagina, cervix and surrounding tissues to be sure there is no evidence of an outbreak. Suspicious sores or symptoms may lead you to choose a Caesarean birth. The greatest risk a baby faces in being infected is associated with a primary (first) herpes infection at the time of labor and delivery. The amount of virus shed is highest during the initial infections, partly because the mother has not yet made antibodies to protect the baby. If you have concerns about new sores or pain in the vagina or surrounding skin, let your doctor know. Women who believe they are uninfected should be especially careful about having sex with men who have or have had herpes in order to avoid a primary infection during pregnancy.
December 15, 2003 | When the license restricts modifications and distribution, that's when. Read the fine print. Unfortunately, there is no single, official definition of open- source software (OSS). Rather, there are variations on a licensing theme. In other words, OSS can best be defined and understood by a list of required licensing terms that apply to the use, modification, and distribution of the software. The most widely recognized licensing terms have three core requirements: - Mandatory availability of source code - The right to use and redistribute the source code freely by anybody to anybody for any purpose - A requirement to make derivative works available on like terms While the Open Source Initiative, a nonprofit organization, promotes one widely accepted definition of "open source," even the OSI Web site offers a menu of alternatives, with the GPL (Gnu Public License), for example, as one substitute definition and a commonly used form of license. The OSI has developed its own certification program for OSS, and there are now more than 40 types of certified OSS licenses, with many more, in varying degrees, based on the consensus OSS requirements. Many licenses include some but not all of the above OSS requirements, however, and some software is made available under more than one form of license — for example, an OSS license and a proprietary license. Examples include "freeware," which often allows for unrestricted use of the software with no further restrictions on the user apart from copyright notices, or the Perl "Artistic License" and the Sun licenses governing Java, which include restrictions on modifications and distribution of the software. "Public domain" software, on the other hand, is freely available with no restrictions on use, modification, or distribution, by declaration of the author(s). Back to The Hidden Costs of 'Free' Software
Head injuries--let's face the facts. Health risk assessment Sports injuries (Diagnosis) Sports injuries (Care and treatment) Sports injuries (Complications and side effects) Brain (Care and treatment) Brain (Complications and side effects) |Publication:||Name: South African Journal of Sports Medicine Publisher: South African Medical Association Audience: Academic Format: Magazine/Journal Subject: Health Copyright: COPYRIGHT 2011 South African Medical Association ISSN: 1015-5163| |Issue:||Date: Dec, 2011 Source Volume: 23 Source Issue: 4| |Geographic:||Geographic Scope: South Africa Geographic Code: 6SOUT South Africa| I have just read an interesting article that describes how concussion and other head injuries are becoming a national epidemic in America, where 1.7 million people are diagnosed with a traumatic brain injury every year. (1) This number is increasing, not because the sporting environment is becoming more dangerous, but because there is a greater awareness of the consequences of concussion and therefore more cases are reported. To some extent this can be attributed to Drs Ellenbogen and Batjer, who in March 2010 were appointed to co-chair a medical committee of the NFL, which specialises in head neck and spine injuries. (2) The turnabout has come because the two doctors are well-respected neurosurgeons who faced up to the problem of head injuries and concussion. They are quick to point out that concussion is a brain injury and all brain injuries are serious. They further acknowledge that although some cases are resolved in a few weeks, some children and adults show symptoms of the concussion that last for several months or even longer. This honest, evidence-based approach was in contrast to their predecessor, a 'denialist' who was more conservative in making the link between concussion and negative long-term consequences. We have seen how the HIV problem in South Africa has been tackled full-on, with dramatic positive results, once the denialists were ousted. We are witnessing a similar scenario with the way the problem of concussion is being confronted in America. There is mounting evidence that people who have repeated bouts of brain injury or concussion have a greater risk of getting Alzheimer-like dementia, Parkinson's disease and motor neurone disease, and this message is being well communicated. More resources are being devoted to studies on concussion and the techniques for measurement are becoming more sophisticated. The enlightened doctors have also encouraged everyone dealing with people who may have concussion to adopt the mantra 'when in doubt, sit it out' to ensure maximal recovery and reduce the risk of repeated trauma. This approach raises new issues for participants in sport. For example, now that the risks are becoming better understood, is it not the right of the player to know the risks, so that he/she can make an informed decision about participating? If truth be told, every sport has some risk associated with participation. Are we heading to a point where these risks are stated clearly and unambiguously--like the mandatory statement 'smoking causes cancer' displayed on cigarette packets? Perhaps we are not too far off, particularly since the NFL has taken a much more concerned approach by producing posters for display in the locker rooms making the statement that 'concussion can lead to the early onset of dementia'. (2) Some of these points are discussed further in two commentaries in this edition of the journal. One of the commentaries, 'Ethical issues in return to sport decisions' by Theresa Burgess) picks up on the ethical issues associated with sport, in particular the decision about returning to play after injury, and the other commentary by Ann Shuttleworth-Edwards debates the controversy around neuropsychological testing in managing concussion. Both papers provide insight into complex areas and will certainly stimulate lively discussion. Before closing I would just like to thank all the reviewers who assisted this year. The difficulty of the editor's job is defined by the community of reviewers who are prepared to contribute their time and expertise to the journal. My job is made much easier by the fact that I always seem to be able to find willing and competent reviewers--with their continued support the journal can grow from strength to strength. (1.) Daneshvar DH, Riley DO, Nowinski CJ, McKee AC, Stern RA, Cantu RC. Long-term consequences: effects on normal development profile after concussion. Phys Med Rehabil Clin N Am 2011;22(4):683-700. (2.) Rather D. Kids, head injuries and the NFL. www.huffingtonpost.com (accessed 9 August 2010). |Gale Copyright:||Copyright 2011 Gale, Cengage Learning. All rights reserved.|
Panting is a fast, sometimes spasmodic and noisy breathing characterized by open-mouth respiration. In birds, it can be the result of a variety of etiologies. Panting can be a response to respiratory distress, overheating or sexual arousal. It can be characterized by abnormal tail motion (“tail pumping”) when associated with respiratory distress. If associated with overheating, the bird will often hold its wings away from its body to increase heat exchange from thinly feathered areas under the wings. If it is a response to sexual arousal, it often occurs when a female pet parrot is masturbating or being petted by the owner in a sexual manner. If the bird is having trouble breathing, take it to a veterinarian immediately. If the bird is overheated, reduce the environmental temperature and/or mist the bird and/or its feet with cool, not frigid, water. If the person’s caresses cause the bird to become sexually aroused, the person should immediately stop what he/she is doing and move away from the bird. Disclaimer: BirdChannel.com’s Bird Behavior Index is intended for educational purposes only. It is not meant to replace the expertise and experience of a professional veterinarian. Do not use the information presented here to make decisions about your bird’s health if you suspect your pet is sick. If your pet is showing signs of illness or you notice changes in your bird’s behavior, take your pet to the nearest veterinarian or an emergency pet clinic as soon as possible.
The roots of the American barbecue craze were probably influenced by all meat cookery dating back to the stone age. But the direct inspiration for the current barbecue phenomenon in America is likely the whole pig cookings that were done in what is now known as Eastern North Carolina in the 1600's. Hernando De Soto is said to have introduced the pig into the United States in 1539. Subsequent introductions took place along the eastern coast. These pigs were brought by explorers as a mobile source of food. Many escaped and ran wild. By the mid 1600's, these escaped pigs had multiplied and settled in great numbers in the Carolina region where they fed on the abundant supply of acorns and other mast indigenous to the area. The pigs became a convenient supply of food for settlers because they did not need to be fed or tended. They were cooked whole to avoid the task of preserving the meat because salt was often a precious commodity. Community barbecues served as social events. This is likely when barbecue began its subliminal association with that of an enjoyable experience. The 1800's featured large festive events held on plantations with the cooking of a whole hog being the main culinary attraction. Pigs were cooked at church picnics and pig pickin's became synonymous with political rallies. Roadside barbecue shacks or stands appeared in the 1900's throughout the southeast. These take-out only shacks normally opened for business only on weekends. These were followed by full-service restaurantís, that featured barbecue as their main theme. In the latter part of the century, suburbanites began to flock to barbecue as a pastime with the advent of contraptions known as smokers. Somewhere along this short but destructive path of the twentieth century, barbecue became almost completely unrecognizable in virtually every aspect when compared to the food that it evolved from. Although cooking one-eighth of a pig is now widely accepted as barbecue, it should be obvious to anyone that a perfect rendition of this food could never be achieved without cooking a whole pig. On this note, we now have a basis to continue on with our search..... Will just any pig do? No. In the early days, hogs ran wild and foraged for their food. These hogs likely ate everything from acorns to carcasses. They converted food into fat at a higher rate than muscle in order to survive winters. Being constantly on the move without a steady food supply meant that they grew very slowly and were quite well exercised. All of this adds up to a pig that would be significantly more flavorful than the pork produced in modern operations. So, in keeping with our pursuit of perfection, we can now conclude that the pig must be older, very well exercised and of a lineage that develops high fat to meat ratio. It also must be allowed to forage for food. Read our "Pseudo Swine" web page about today's hog raising. Does the pig need to be any particular size From everything we can gather, yes. From a practical standpoint, small pigs would have been much easier to handle in both the butchering and cooking process. We once speculated along with fellow contributor Dan Gill that small pigs might have been used because they could be killed and cooked quickly enough to prevent the onset of rigor mortis. Meat cooked while passing through rigor mortis will have a stringy and poor texture. In a conversation with America's foremost pitmaster, whose family has been cooking barbecue in Eastern North Carolina since the 1830's, we were told that only pigs weighing around 40 pounds or less were traditionally used for barbecue. We must remember that a pig of that size would have been quite old and flavorful. This is in contrast to modern commercial pigs that are likely to weigh 40 pounds when weaned at only several weeks of age. Based upon this, we now have a general idea of the pig that would be required. How would this pig be cooked? While much meat was cooked over a fire, pigs were not suited to this because of their fat content and the resulting flare-ups that would have occurred. Instead, the hardwood was burned down to coals. The wood used likely varied according to locale and availability, although oak and hickory were quite abundant. The coals were then placed either around the pig, directly underneath or a combination of both. This also would have likely varied between pitmaster and local custom. The faint smoke emitted by the coals would have left no smoke ring in the pig. The result would have been a much cleaner taste than is generally considered optimal by today's standards. In fact, modern barbecue would likely be considered tainted or flawed by those who inspired it. We now determine that the pig should be cooked with hardwood that has been burned to coals. And furthermore that the coals must be placed in a manner so that they have a direct line of light to the pig. Today's off-set smokers, regardless of price, are not designed to produce true barbecue. At what temperature should the pig cook? One thing you'll notice as you navigate through Bob in Ga.com is that specific temperatures are never mentioned in anything that we write. This is because temperatures were simply not measured and we respect your intelligence enough that we will not attempt to make any wild baseless speculations on this most ridiculous topic. The bottom line is that the pigs were cooked by skilled pitmasters who used their experience and judgement to cook them. Dave Lineback, founder of the Society for the Preservation of Traditional Southern Barbecue (SPTSB), is fond of saying that barbecue cannot be made using a thermometer. We concur! At Bob in Ga.com, we are artisans who would never advocate tarnishing the purity of this antique process with modern temperature measuring devices. How long would the pig need to be cooked? Most of what is written above regarding temperature is also applicable to time. The pigs were done when they were judged to be ready by their respective pitmasters. Nothing more can be written on these two topics of an intelligent nature because too many variables were likely involved. Factors such as weather conditions, texture preferences and differences in cooking set-ups would have governed the cooking times. It would have been highly unlikely that any standards were in place as it pertained to pig cooking or their results. Therefore we consider the popular practice of focusing on time and temperature in efforts to replicate barbecue to be downright OK, we have our pig cooked, anything As far as pigs were concerned, the tendency of our forefathers was to use everything but the "oink". Therefore, it is likely that the technique of "blistering" or "popping" the skin into a fried pork rind type texture was practiced. This was usually done at the end of cooking. This technique dates back at least to the 1830's, according to our source. Our guess is that it is likely much older than that. While almost non-existent in modern barbecue, we don't feel that a perfect rendition would be achievable without the inclusion of this blistered What type of sauce or seasoning would be required? True American barbecue sauce is basically seasoned vinegar. Again, the choices of seasoning agents would have likely varied according to availability as well as taste preferences. As a result, nothing concrete can be concluded as far as ingredients are concerned. However, capsicum peppers were plentiful and it's believed that they were the most commonly used seasoning agent along with salt and possibly some form of sweetener. One thing that we cannot stress in strong enough terms is that tomato would have never been included. Tomatoes were thought to be poisonous. Therefore, no sauce or seasoning containing a tomato product could ever be considered legitimate, regardless of how heavily they are currently marketed or rampantly used. Read the paragraph in Dan Gill's article on "Barbecue History 101" at our "Thoughts from Others" web page. Have we finally arrived at the end of the barbecue rainbow? We're very close now! At this point we want all of the meat to be pulled from the pig. That means the highly flavorful shoulders. The prized tenderloins. The sinfully-delicious rib and bacon meat. The well exercised hams, the stately center loin, the delectable jowl meat. And last, but not least, that heavenly crispy blistered skin. We want all of these parts to be lightly chopped or mixed together so that a perfect reconstruction of the pig is achieved every bite. Everything in its proper proportion. All ratios perfect. is the beauty of whole hog barbecue. Simply cooking a single part is one dimensional. It is NOT comparable or interchangeable. As we add a bit of our seasoning and take our first bite, our eyes instantly reflect vivid gold, which lies at the end of We know why food writers have traveled millions of miles in pursuit of the perfect barbecue. We know why the hobbyist's stay up all night after a hard week at work. We can now understand why barbecue, in its purest form, laid the groundwork for our national culinary pastime. Where could one possibly find such Barbecue? Unfortunately, most people are likely to never find or produce such barbecue. The rarity of a traditional pig, the cost of wood and other factors have caused such a perfect rendition to become practically extinct, regardless of venue. Modern taste preferences are now geared toward the various tastes imparted by years of massive attempts to simulate the barbecue described above. Along with most every trendy attempt at simulation, tastes were lost and flaws were imparted. Currently, the flaws are so ingrained that attempts are now more likely to be geared toward simulating the flaws, rather than the true dish.
Experimental Breeder Reactor I Simply begin typing or use the editing tools above to add to this article. Once you are finished and click submit, your modifications will be sent to our editors for review. history of nuclear power The first LMR was the Experimental Breeder Reactor, EBR-I, which was designed at Argonne National Laboratory and constructed at what is now the Idaho National Laboratory near Idaho Falls, Idaho. EBR-I was an early experiment to demonstrate breeding, and in 1951 it produced the first electricity from nuclear heat. A much larger experimental breeder, EBR-II, was developed and put into service... What made you want to look up "Experimental Breeder Reactor I"? Please share what surprised you most...
Dudley Shelton Senanayake Simply begin typing or use the editing tools above to add to this article. Once you are finished and click submit, your modifications will be sent to our editors for review. hartal of 1953 ...subsidies to keep the price of rice stable in the face of a fluctuating world market. By 1952 the subsidies accounted for 20 percent of government expenditure. In July 1953, Prime Minister Dudley Senanayake of the United National Party drastically reduced the subsidies, causing the price of rice to triple. history of Sri Lanka ...consumer goods, and the failure of state enterprise in industry and trade—made people look back to the UNP. This party gained the support of minorities, and in 1965 it returned to power under Dudley Shelton Senanayake, who, as the son of Don Stephen Senanayake, had served as prime minister (1952–53) after his father’s death and briefly in 1960. Senanayake’s government enjoyed a... What made you want to look up "Dudley Shelton Senanayake"? Please share what surprised you most...
AdrarArticle Free Pass Adrar, traditional region of central Mauritania in western Africa. It consists of a low central massif with noticeable cliffs that rise to about 800 feet (240 m). The terrain is arid and almost totally unsuitable for cropping. There is, however, sufficient water at the base of the uplands to support date-palm groves, and during the wetter part of the year there is cultivation of millet, sorghum, melons, and vegetables in gorges. The population of the Adrar (Berber for “mountain”) formerly was nomadic. The major town in the region is Atar. Historic sites include Ouadane, formerly a caravan and gold-trading centre, and Chinguetti, an ancient centre of learning and of Islām. What made you want to look up "Adrar"? Please share what surprised you most...
On Bodies Carried Down by Their Weight Simply begin typing or use the editing tools above to add to this article. Once you are finished and click submit, your modifications will be sent to our editors for review. discussed in biography ...23), was the accurate determination of terrestrial locations. Ancient authors preserved only a few tantalizing allusions to Hipparchus’s other scientific work. For instance, On Bodies Carried Down by Their Weight speculated on the principles of weight and motion, and a work on optics adhered to Euclid’s theory from the Optics that vision... What made you want to look up "On Bodies Carried Down by Their Weight"? Please share what surprised you most...
|CAL Resources Archive The CAL Resources Archive was created to provide our visitors with access to older pages and content from our Web site that they may find useful. Please be aware that information within the CAL Resources Archive is historical in nature and will not be maintained or updated by CAL. CAL Resource Guides Online Several factors influence the identification of English language learners with special needs. Limited prior schooling, lack of proficiency in English, native language background, cultural expectations, and personal or family concerns can all influence a learnerÁs academic progress. It can be difficult to distinguish between a learning disability and learning problems caused by a variety of other factors. English language learners whose difficulties in school stem from these factors may be misidentified as having a learning disability. In other cases, English language learners who have a learning disability may not be properly identified on the assumption that their learning problems stem from linguistic and cultural differences. To further complicate matters, special needs may not have been identified in the learner's native language but may become evident as the student is learning English. Traditional identification instruments designed for English speakers may not be valid with English language learners. Identification of English language learners with special needs should include consideration of the following factors: Learning in any language is affected by learning disabilities, but second language learners with special needs present additional educational challenges. According to the British Columbia Ministry of Education, Skills, and Training (Fowler & Hooper, 1998), instructors of English language learners with special needs should consider the cultural, developmental, and first language background of the learner. They should also do the following: The use of standardized testing to identify and assess the progress of English language learners with special needs is problematic. Normally designed for native English speakers, many assessment instruments do not reliably assess speakers of other languages because they ignore differences among linguistic and cultural groups (Schwarz & Burt, 1995). Assessment of English language learners with special needs should include the following: Fowler, J., & Hooper, H.R. (1998). ESL Learners with Special Needs in British Columbia: Identification, Assessment and Programming. British Columbia: The British Columbia Ministry for Education, Skills, and Training. Schwarz, R., & Burt, M. (1995). ESL Instruction for Learning Disabled Adults. ERIC Digest. Washington, DC: National Center for ESL Literacy Education. (EDRS No. ED 379 966) ERIC/CLL is grateful to Bernadette Knoblauch of the ERIC Clearinghouse on Disabilities and Gifted Education for her valuable assistance in compiling this Resource Guide Online. The following publications, Web sites, and listservs offer additional information about English language learners with special needs. This Resource Guide concludes with an annotated bibliography of ERIC documents related to this topic. Digests provide brief overviews of topics in education. British Journal of Special Education This well-established and respected journal covers the full range of learning difficulties of students in mainstream and special schools. The journal is concerned with a wide range of special educational needs and covers all levels of education, including pre-school, K–12, and post-secondary education. This peer-reviewed journal publishes original research on the education and development of toddlers, infants, children, and youth with special needs, and articles on professional issues of concern to special educators. Teaching Exceptional Children TEC is published specifically for teachers and administrators of children with disabilities and children who are gifted. TEC features practical articles that present methods and materials for classroom use as well as current issues in special education. Topics in Early Childhood Special Education This journal deals with timely, important issues and trends in early childhood special education. TECSE translates theory and research into effective practice; presents articles by professionals in early childhood education, special education, and related fields; and provides a forum for ideas, knowledge, research findings, questions, and answers dealing with the special nature of preschool education. Artiles, A., & Ortiz, A. (in press). English Language Learners with Special Needs: Identification, Placement, and Instruction. Washington, DC: Center for Applied Linguistics. Baca, L., & de Valenzuela, J. S. (1994). Reconstructing the Bilingual Special Education Interface. NCBE Program Information Guide Series No. 20. Washington, DC: National Clearinghouse for Bilingual Education. Available: http://www.ncela.gwu.edu/pubs/pigs/pig20.htm Chapman, J.B., Vaillancourt, B., & Dobbs, C.S. (1980). Learning Disabilities and the Adult Student of English as a Second Language. Palatine, IL: William Rainey Harper College. Genesee, F. H., Paradis, J., & Crago, M. B. (2004). Dual Language Development and Disorders. Baltimore, MD: Paul H. Brookes Publishing Co. Gersten, R., Baker, S., & Marks, S. (1998). Teaching English-Language Learners with Learning Difficulties: Guiding Principles and Examples from Research-Based Practice. Reston, VA: Council for Exceptional Children. Root, C. (1994). A Guide to Learning Disabilities for the ESL Classroom Practitioner. TESL Electronic Journal, 1. Smith, T., Polloway, E., Patton, J., & Dowdy, C. (3rd. Edition). (2001). Teaching Students with Special Needs in Inclusive Settings (3rd ed.). Boston: Allyn & Bacon. Terrill, L., & Flores, M. (2001). NCLE Resource Collection: Learning Disabilities and Adult ESL. The Council for Exceptional Children The Council for Exceptional Children (CEC) is the largest international professional organization dedicated to improving educational outcomes for individuals with exceptionalities, students with disabilities, and the gifted. A service of The Learning Project at WETA in WAshington, DC, LD Online is an interactive guide to learning disabilities for parents, teachers, and children. Learning Disabilities Association of America (LDA) LDA's purpose is to advance the education and general welfare of children and adults of normal or potentially normal intelligence who manifest disabilities of a perceptual, conceptual, or coordinative nature. The Migrant/OLE Project The Migrant/Optimal Learning Environment (Migrant/OLE) Project disseminates a program of effective literacy instruction for migrant students, and culturally and linguistically diverse (CLD) students. National Center for Learning Disabilities (NCLD) NCLD provides national leadership in support of adults and children with learning disabilities by offering information, resources, and referral services. NCELA Resources About Special Education This resource page is comprised of articles and links regarding the education of linguistically and culturally diverse (LCD) students with special educational needs. Learning Disabilities/Learning Challenges: Unlocking the Mysteries of Learning Problems with ESL Adult Students This online PowerPoint presentation by Judith Rance-Roney, Ed.D. Lehigh University, is about ESL learners with learning disabilities. Special Needs Education Discussion Group The Special Needs Education (SNE) project offers five e-mail discussion groups: ESL-SNE is a discussion group for teachers and learners of English as a second language. SNETEACHTALK-L is for educators involved in special needs education. SNEPARENTALK-L is for parents and care providers of children with special needs. SNE-NEWTEACHER-L brings together novice and experienced educators to discuss topics related to novice teaching. SNE-EATALK-L was created for educational assistants and other support personnel to discuss their role in educating students with special needs. An assortment of e-mail discussion lists on many issues of interest to learners with special needs is offered by the ERIC Clearinghouse on Disabilities and Gifted Education. Information on obtaining the documents listed below can be found at the end of this section. These documents were identified by searching the ERIC database using the following combination of ERIC descriptors and keywords: Special Needs Students or Disabilities/DE or Disorders/DE English (Second Language) or Limited English Speaking The Development of English Oral Communication in Learning Disabled Spanish Speaking Students in Puerto Rico. Reyes-Bonilla, Maria A.; Carrasquillo, Angela L. April 26, 1993 In a study that sought to identify the gains in English oral communication skills of Spanish speaking learning disabled students in the elementary schools, the Basic Inventory of Natural Language (BINL) and the Woodcock Language Proficiency Battery (WLPB) were administered to two groups of students aged 8 to 12 years: an experimental group of 20 students in a self-contained class and a control group of 20 mainstreamed students. Students in the experimental group received English as a Second Language (ESL) instruction using the natural approach (NAT) and the control group followed the audiolingual approach (AAT). The BINL was used as a measure of English oral production skills and the WLPB was used to measure English vocabulary comprehension skills and listening comprehension. Results reveal that elementary Spanish speaking learning- disabled students showed gains in English oral communication skills. The NAT and AAT showed instructional strengths for the acquisition of ESL in learning-disabled students. Students under the NAT treatment showed a significant difference in the Woodcock Analogies subtest by age group. The Care and Education of Young Bilinguals: An Introduction for Professionals. Availability: Multilingual Matters Ltd., UTP, 2250 Military Road, Tonawanda, NY 14150 (hardback: ISBN-0- 85359-466-0, $49.95; paperback: ISBN-1-85359-465-2, $15.95). This book is a comprehensive introduction for all professionals working with bilingual children. For speech therapists, physicians, psychologists, counselors, teachers, special needs personnel, and many others, this book addresses the most important issues at a practical level. It is written in simple, nontechnical terms accessible to the layman and provides a brief but comprehensive introduction. Areas addressed include the following: the nature of bilingual children; the everyday language use of bilinguals; the advantages of the bilingual child; the personality and social development of bilinguals; identity issues and solutions; children as interpreters; code-switching; bilinguals and their families; childhood trilingualism; home and school relationships; language assessment and speech therapy in the bilingual context; migrants and refugee bilinguals; the assessment of bilingual children; language delays and disorders; the development of biliteracy; prejudice reduction in school; and bilingual classrooms. The book is divided into 13 chapters and has an index, glossary, and bibliography. Each chapter concludes with suggestions for further reading. Scholarly references appear throughout the text. Accuracy of Teacher Assessments of Second-Language Students at Risk for Reading Disability. Limbos, Marjolaine M.; Geva, Esther Journal of Learning Disabilities, v34 n2 p136-51 Mar-Apr 2001 This study examined the accuracy of teacher (N=51) assessments in screening for reading disabilities among Grade 1 students of English as a second language (N=249) and as a first language (N=120). Results indicated that teacher rating scales and nominations had low sensitivity in identifying reading disability in either group as determined by a standardized reading score. Bilingual Multicultural Special Education: An Integrated Personnel Preparation Program. Gallegos, Anne; McCarty, Laurie L. Teacher Education and Special Education, v23 n4 p264-70 Fall 2000 This article presents guidelines for the development and implementation of a graduate program in bilingual multicultural special education using the program at New Mexico State University as a model. It identifies the competencies of bilingual multicultural special educators, identifies steps in creating a new program, offers a philosophical framework, and reports on outcomes of the New Mexico program. Assessment of Culturally and Linguistically Diverse Students for Special Education Eligibility. In addressing the problem of assessing culturally and linguistically diverse students for special education eligibility, this brief paper organizes suggested strategies according to four principles: (1) convening a full, multidisciplinary assessment team (to include parents, educators, assessors, and a person familiar with the student's culture and language); (2) using pre-referral strategies and interventions to determine whether difficulties stem from language or cultural differences, a lack of educational opportunity, or from a disability); (3) determining the language to be used in testing (assessment of language dominance and proficiency); and (4) conducting a tailored, appropriate assessment of the child and environment that combines nonbiased, appropriate instruments with other sources of information (observations, interviews) from a variety of environments to produce a multidimensional assessment. Learning Disabilities and Spanish-Speaking Adult Populations: The Beginning of a Process. Office of Vocational and Adult Education, U.S. Department of Education This report is designed for use by policymakers and practitioners involved in issues of adult education, welfare reform, and employment training programs. The driving concerns of this report are issues associated with Spanish-speaking adults who are at risk for having learning disabilities. The focus on adults is critical, because such issues as diagnostic requirements and services available are very different for adults with disabilities than for children with disabilities. The ultimate goal of this project is to provide a means to enable persons with disabilities to successfully achieve self-sufficiency through education and work. To develop the needed tools for this effort, a 6-phase project was designed, and this report focuses on the first two phases: reaching a consensus among experts on diagnostic processes for Spanish-speaking adults to determine the presence of learning disabilities (LD); and determining from the existing screen instruments in Spanish which ones should be field-tested to for validity for predicting LD in Spanish-speaking adults. Several specific tests and types of tests are discussed at length. The report is divided into seven sections providing acknowledgements, an executive summary, an introduction, a conference overview, the notes of concern of some participants, findings on the diagnostic process, and screening tool recommendations for field tests. Seven appendices provide details on learning disability definitions, current diagnostic procedures, GED testing requirements, lists of conference participants and project staff. Do My ESOL Students Have Learning Disabilities? A Practical Manual for ESOL Instructors Concerned about Learning Disabilities and the ESOL Learner. Shewcraft, Dianne F.; Witkop, Eileen L. This booklet, conceived, researched, and produced by teachers of English to speakers of other languages (ESOL) practitioners, is designed to help ESOL teachers identify and assess students who may have learning disabilities. Two groups of ESOL learners are the subject of this study: those ESOL adult learners who already have some formal education and are attempting further formal education and knowledge in English, who may also have a learning disability; and the ESOL adult learner seeking the same but has no or very little previous formal education. The book is divided into several sections covering the following topics: the definition of a learning disability, suspecting a learning disability in ESOL learners, approaching the learner, and classroom strategies. Extensive lists of resources are provided, as well as four appendices including a sample hands-on screening kit, a list of common acronyms pertinent to learning disabilities, a copy of the Americans With Disabilities Act, and a practitioner questionnaire. Profiles of Asian American Students with LD at Initial Referral, Assessment, and Placement in Poon-McBrayer, Kim Fong; Garcia, Shernaz B. Journal of Learning Disabilities, v33 n1 p61-71 Jan-Feb 2000 This study examined the characteristics of 26 Asian American elementary students with learning disabilities including demographic characteristics, factors associated with referral, assessment practices, student profiles, and instructional recommendations for special education. Results suggest similarities with Mexican American students with LD and other language minority students. Assessing LEP Migrant Students for Special Education Services. Lozano-Rodriguez, Jose R.; Castellano, Jaime A. Many migrant students are not identified for needed special education services in a timely manner. This digest describes the obligations of schools to provide such services, and discusses approaches for student referral, assessment, and placement and working with migrant families. Federal mandates concerning special education are summarized, and the unmet needs of migrant children are briefly discussed. The referral process takes time and includes requirements that vary among agencies, districts, and states. General information is provided on the referral process for the school-aged child. Because most migrant students are from Hispanic backgrounds, multidisciplinary assessment teams should include bilingual educators and be sensitive to children's cultural backgrounds. Evaluators must consider as many aspects of the student's life circumstances as possible in order to accurately interpret results of educational testing. Schools should give testing priority to migrant students, assess students fairly in their native language or provide a skilled interpreter, and interpret standardized tests with caution. Cultural and experiential differences between school personnel and parents may create difficulties. Suggestions for working with parents are listed. Because migrant students may move frequently, schools must pursue interventions quickly and provide copies of documents to be carried to the student's next school. Effective Instruction for Language Minority Children with Mild Disabilities. Ruiz, Nadine T. This digest describes a model curriculum for children from language minority groups, called the "Optimal Learning Environment Curriculum--A Resource for Teachers of Spanish Speaking Children in Learning Handicapped Programs." This bilingual special education class model is governed by the following principles: (1) take into account the student's sociocultural background and its effect on oral language, reading and writing, and second language learning; (2) take into account the student's learning handicaps and how they may affect oral language, reading, writing, and second language learning; (3) follow developmental processes in literacy acquisition; (4) locate curriculum in a meaningful context where the communicative purpose is clear and authentic; (5) connect curriculum with the students' personal experiences; (6) incorporate children's literature into reading, writing, and English-as-a-Second Language lessons; and (7) involve parents as active partners in the instruction of their children. Turning Frustration into Success for English Language Learners. Brice, Alejandro; Roseberry-McKibben, Celeste Educational Leadership, v56 n7 p53-55 Apr 1999 To teach culturally and linguistically diverse learners with language-learning disabilities, teachers should ensure that students understand assigned tasks, seat students from similar linguistic backgrounds together, start lessons with lead statements, use varied questioning strategies, ask for summaries, use multimodal approaches, and relate information to students' background. Conversations with a Latina Teacher about Education for Language-Minority Students with Special Needs. Bos, Candace S.; Reyes, Elba I. Elementary School Journal, v96 n3 p343-51 Jan 1996 Describes the beliefs, knowledge, and classroom practices of a successful bilingual special educator, noting factors that shaped her teaching, including her own experiences as a second-language learner. Notes the teacher's emphasis on interactive teaching that weaves students' first language and culture with direct instruction, practice, and transfer. The Language-Minority Student and Special Education: Issues, Trends, and Paradoxes. Gersten, Russell; Woodward, John Exceptional Children, v60 n4 p310-22 Feb 1994 This article discusses issues in referral and special education instruction for students from language-minority groups. It discusses inadequacies in assessment and placement; development of effective and feasible instructional strategies through collaboration between bilingual education and special education; and potential solutions involving both academic skills instruction and a natural approach to language instruction. The full text of most materials in the ERIC database with an "ED" followed by six digits is available through the ERIC Document Reproduction Service (EDRS) in microfiche, by email, or in paper copy. Approximately 80% of ERIC documents from 1993 to the present are available for online ordering and electronic delivery through the EDRS Web site. You can read ERIC documents on microfiche for free at many libraries with monthly subscriptions or specialized collections. To find an ERIC center near you, contact our User Services staff. The full text of journal articles may be available from one or more of the following sources: To obtain journals that do not permit reprints and are not available from your library, write directly to the publisher. Addresses of publishers are listed in the front of each issue of Current Index to Journals in Education and can now be accessed online through the CIJE Source Journal Index. If you would like additional information about this or any topic related to language education or linguistics, contact our User Services Staff. Back to RGOs Top of Page
By William Golding "'We've got to have rules and obey them. After all, we're not savages. We're English, and the English are best at everything.'" - from The Lord of the Flies What's it about?: The story is set in the war-ravaged 1940s and begins when a plane, carrying a group of British school boys, crashes on a remote island. With the youngsters left to fend for themselves until help arrives, Ralph, one of the older boys, attempts to organize a democratic system of order. Over time, Jack, a rebellious boy, breaks away from the pack and forms a tribe of his own. Soon most of the children defect to Jack's tribe, partly for protection from what they believe is a monster lurking in the wilderness. What ensues is a battle of ideologies — democracy versus tribal law — as well as the struggle to maintain civility amongst the inevitable uprising of man's most primal instincts. The children themselves become the monsters, the weak are systematically eliminated and those left are forced to not only find a way to survive the elements, but the savages they've become. Why your child should read it: To start, it's a terrific and compelling read. As such, Golding's best known work was another title that received repeated endorsements from those who suggested books for our list. Many young readers are first exposed to The Lord of the Flies in school, where they identify with the characters who are of similar age. Some, no doubt, question whether they would side with Ralph or Jack if put on the island. But more importantly, this novel illustrates the dark side of the human spirit that often emerges alongside dire situations. The book teaches the importance of standing by your convictions in the face of opposition. And if nothing else, The Lord of the Flies was used as the basis for an entire, hilarious episode of The Simpsons. To a teenager, that's one mighty powerful endorsement. |20 Bedtime Books Every Parent Should Have in Their Library|
CBSE Guess > Papers > Important Questions > Class XII > 2007 > Chemistry > Chemistry In Everyday Life Chemistry In Everyday Life - Define Carbon Fibers - What are the main causes for developing carbon fiber? - What is carbon fiber reinforced carbon? - In how many categories made regarding the application of carbon fibers. - What is meant by ceramics? - What is meant by propellant? Classify them? - Describe the following: (a). Double base propellant (b). Hybrid propellant (c). Monoliquid Propellant - Discuss the role of Redox phenomenon in the context of rocket propellants. - What is meant by pheromones? - What are the two groups found in detergents? - Give the structure of sodiumalkyulbenzenesulphonates - What is meant by soap less soap? - Write a short notes on detergent. - What is the major drawback with chemical insecticides? - Differentiate between anionic and cationic detergents. - Why the cleansing action of synthetic detergents not disturbed by hard water - Describe cleansing action of detergents. - Solid State - Chemical Kinetics - P Blocks Element - D and F Block Elements - Coordination Compound - Nuclear Chemistry - Functional Group I - Chemistry In Everyday Life - Practice Paper 1 - Practice Paper 2 - Practice Paper 3
The following Web sites help guide site-based RtI2 team members through the problem-solving process as they discuss data, students’ progress, materials, and interventions. - Creating Shared Language for Collaboration in RtI Creates shared language for communication, so participants can successfully participate in the RtI process. By Barbara J. Ehren, Barbara Laster, and Susan Watts-Taffe. International Reading Association's Commission on RtI. - Create Your Implementation Blueprint Stage 4: Full Implementation Describes the problem solving process and other activities, after teachers have administered the assessments and begin to learn how to provide targeted and effective intervention instruction. By Susan L. Hall. RTI Action Network. - Intervention Documentation Worksheet Supports the site-based problem solving process. By the Florida Problem-Solving/Response to Intervention Project. RTI Action Network. - Preparing for a Progress Monitoring Review Process: Engaging in Reflection for Productive Decision Making Provides an overview and steps, including a “story script” form to guide participants through the progress monitoring process. RTI Action Network. - RTI Action Network Problem Solving Checklist and Critical Components Checklist Provides downloadable Word documents useful for guiding problem-solving meetings. RTI Action Network.
Botswana's wildebeest populations have dropped by nearly 90 percent due to multiple causes including land use, habitat fragmentation, vegetation changes, drought, veterinary fences, fires and poaching. Wildebeest, Botswana [File photo] A recent study conducted by Elephants Without Borders (EWB), a charitable organization dedicated to conserving wildlife and natural resources, showed Botswana's wildebeest is on the brink of extinction and will be difficult to rejuvenate. The wildebeest populations, especially in the northern districts Ngamiland and Chobe, have dropped by nearly 90 percent over the last 15 years. EWB founder Michael Chase was quoted by local daily Mmegi on Tuesday as saying other animals whose declining numbers were of grave concern included giraffe, kudu, lechwe, ostrich, roan, tsessebe and warthog. "Particularly troubling is the almost 90 percent drop in the numbers of wildebeest sighted by the survey," Chase said. "For example, in Chobe, we sighted only 500 wildebeest." According to the survey, the decline in the animals' populations largely due to land use, habitat fragmentation, vegetation changes, drought, veterinary fences, fires and poaching. "Bush fires tend to kill tens of animals at a go," Chase said. "For example, in a national park, we saw 43 carcasses of zebra which had obviously been killed by wildfires." He said increasing rate in poaching further exacerbates the situation. "In one case, poachers killed a lactating lion that had six cubs." While the population of elephant in northern Botswana remained stable at about 130,000 individuals. The survey, which was commissioned by the Department of Wildlife and National Parks, took nearly 250 hours of flying time and flight lines totalling 25,598 km over 73,478 square km in the country's most national parks. It analyzed comparative data from nine similar surveys that were conducted between 1993 and 2004.
|« Back to Article| Time is of the essence in treating strokes By James Grotta, M.D. | June 21, 2012 | Updated: June 21, 2012 7:01pm Stroke is a disease that everyone fears because it is so common (in Houston, almost 100 people suffer a new stroke each day) and because it is one of the leading causes of severe disability leading to the need for chronic care. No one wants to be a burden to their family or to live out their days bedridden in a nursing home. While the chances of stroke rapidly increase as we age above 65, stroke also can occur in children and young adults, particularly African-Americans with hypertension. Stroke also disproportionately affects women, since elderly women outnumber men and are less likely to have a spouse available as a caregiver if they suffer a stroke. Twice as many women die of stroke as from breast cancer. The most important way to avoid the ravages of a stroke is to prevent it. A stroke is due to a blocked or ruptured artery in the brain, and recent studies have shown that aggressive medical management of its well known risk factors is much more effective than surgery or invasive treatments. These measures include cessation of smoking, weight loss, exercise, control of blood pressure, diabetes and high cholesterol, blood thinners for certain cardiac conditions (especially atrial fibrillation), and in some cases surgery or stenting of narrowed carotid arteries. But most patients are not compliant with these measures, even though we have new and affordable medications that have few if any side effects. A new era of stroke prevention is upon us - we know that if patients take their medications, keep their blood pressure and cholesterol low, maintain the proper level of blood thinning, and lead reasonably healthy life styles, their chance of being disabled by a stroke is markedly reduced. But what if a stroke occurs? The first step is to recognize that a stroke is happening and not to ignore the signs. Most patients are not aware of the gravity of the situation, and it is up to a spouse, family member, co-worker or bystander to do the right thing. But fewer than one in four Houstonians knows the symptoms of a stroke. If you remember only one thing from this article, remember the acronym word FAST, for face, arm, speech and time. It is one of those things that should be on your refrigerator door. Look at the person; is one side of the face drooping? Have him hold out his arms; does one drop down faster than the other? Get the person to talk; is his speech slurred or does he have trouble getting his words out? If any of these conditions are present, you must call 911 immediately. Don't call your doctor, don't call your son or daughter or spouse, and don't just drive to the clinic or hospital. Call 911. Why is it so important to call 911? Because time is brain. Brain cells cannot live without the oxygen and glucose that is delivered by the blocked artery. It is estimated that 2 million nerve cells die every minute after a stroke begins. That means that someone having a stroke when you began reading this has already lost more than 10 million nerve cells. We have therapies that can help reverse the damage caused by stroke but they are only effective if given quickly. The most effective and well-established treatment is intravenous tissue plasminogen activator, or TPA. This is a natural clot busting substance that we all have coating our arteries. After a quick CT scan to be certain the stroke is not caused by bleeding, TPA is injected into the vein, circulates through the bloodstream, hones in on the clot and begins to dissolve its components. The treatment has risk, but all studies, including one just completed in Europe with more than 3,000 patients, confirms its benefit if given within three hours of symptom onset. Fifty percent of patients treated within two hours of onset will recover with no disability. The other reason to call 911 is that only some emergency departments in Houston can deliver TPA quickly and predictably. Houston fortunately has a highly trained paramedic force with the Houston Fire Department-Emergency Medical Services (HFD-EMS). HFD-EMS will answer your 911 call and will be able to recognize if the person is having a stroke and will know where to take them. That decision requires balancing the importance of time and the availability of expertise. It doesn't do any good to go to the closest hospital or the one where your doctor works, if they are not prepared to treat you. Severe trauma patients go to one of Houston's excellent Level 1 trauma centers. Similarly, a stroke patient should go to one of Houston's outstanding stroke centers. The Stroke Team at UT Medical School and Memorial Hermann Hospital at the Texas Medical Center (MHH-TMC), in collaboration with HFD-EMS, were one of the first in the world to test TPA when it was first developed in the 1980s. Since then, several other hospitals, but not all, have become stroke centers where you can be assured to receive TPA if you arrive within the three-hour time window and meet other treatment criteria. Currently, 25 hospitals are certified by our local emergency care agency as primary (Level 2) or comprehensive (Level 1) stroke centers. Together, they treated with TPA roughly 10 percent of the stroke patients taken to one of their emergency rooms last year, more than twice the national average. This fall, some of these hospitals will receive national certification as comprehensive stroke centers, which will identify their expertise in delivering more complex stroke therapies and conducting research to find better therapies. After all, a treatment that is used in only 10 percent of stroke patients and that only works up to 50 percent of the time is not good enough. At MHH-TMC, we are testing newer treatments (more powerful clot-busting drug combinations, ultrasonic energy, catheters, brain cooling, to name just a few) that might improve on the benefit of TPA, but all have to be given quickly. We are also testing exciting new approaches to helping patients recover from their strokes. But for the most part, the battle for complete recovery from stroke is won or lost in the emergency department in those first few hours. Not all stroke care is equal. Take ownership of all aspects of your health. In particular, Houstonians need to be informed and active participants in decisions about both stroke treatment and prevention. For any questions or further resource information about stroke, there are a number of excellent online resources: * www.strokeassociation.org/STROKEORG/General/Patient-Infor mation-Sheets_UCM_310731_Article.jsp Grotta is professor and chairman of the Department of Neurology, University of Texas Medical School, and chief of neurology and co-director of the Mischer Neuroscience Institute, Memorial Hermann Hospital-Texas Medical Center. These opinions are the author's and do not necessarily reflect those of the UT Medical School or Memorial Hermann Hospital.
Here's a shocking fact (maybe not to those of you already scientifically minded): Trees are electric. It's true. Scientists know that each individual tree generates a faint (we're talking 100 to 200 millivolts) electrical current. Recently, however, researchers have begun to figure out why and entrepreneurs have begun to harness that slight power in each tree -- multiplied by millions. The why: MIT scientists say the small electrical current comes from the pH imbalance between the tree and the soil it occupies, according to Onearth, a National Resources Defense Council magazine. And a company called Voltree, working with the U.S. Forest Service, is developing an "electricity generation system that harvests metabolic energy from the trees themselves and converts it to usable power." The electricity would power sensors -- now run by battery -- that alert officials to a forest fire. So, what could this have to do with Ohio? Hard to say for sure, but we have a lot of trees and we're getting greener all the time. Remember we told you last year that forest coverage in Ohio is now more than 31 percent, more than double from the 1940s. (Go here to see a nifty graphical depiction of that 70-year growth.) Now, think about harnessing all that tree power . . . and let your eco-imagination go wild. Ohio has most 'Tree Cities': Think we're exaggerating the thing Ohio has for trees? Not a chance: The Buckeye State (named after a tree, see?) now has a 28-year streak as the state with the most Tree City USA communities, according to the Ohio Department of Natural Resources. We had 249 Tree Cities last year, well ahead of states chasing us like Illinois (191) and Wisconsin (177). Go to arborday.org/programs/treeCityUSA/map.cfm to see all the Tree City USA communities in Ohio and elsewhere. Click and plant: We're trailing in another tree race, but there's a way to increase the number of trees in Ohio -- one mouse click at a time. Odwalla, the California fruit juice and food bar company has put up $100,000 to buy and plant new trees in state parks in 11 states -- and you can decide where they go. Unfortunately, 80,000 trees have already been spoken for and most of the other states are leaving us behind: Only 1,420 trees (including one by this columnist) have been set aside for Ohio, while states like Michigan (41,000) and Virginia (22,000) lead the way. Still, you can go to parkvisitor.com/odwalla/ to plant a tree in Ohio (only 1 per Web site visitor while they last), to see the state rankings or just to learn more about the Plant a Tree program (and some interesting tree facts). What tree is that? Speaking of trees, the Arbor Day Foundation is offering a free online (or $5 for the printed, full-color version) tree identification tool. "What Tree Is That?" features hand-drawn botanical illustrations highlighting the specific characteristics of tree species throughout the continental United States, including Ohio. To order the print version, go to arborday.org or mail in your name and address and $5 for each guide to What Tree Is That?, Arbor Day Foundation, Nebraska City, NE 68410.
Adherence With Iron Sprinkles Among High-Risk Infants Compared with iron drops, iron sprinkles supplied for 3 months to high-risk children beginning at age 5-7 months will increase adherence and reduce the rates of anemia and iron deficiency. Drug: Ferrous sulphate drops with vitamins A, D, and C Drug: Ferrous fumarate sprinkles with vitamins and minerals |Study Design:||Allocation: Randomized Endpoint Classification: Efficacy Study Intervention Model: Parallel Assignment Masking: Open Label Primary Purpose: Prevention |Official Title:||Adherence With Iron Sprinkles Among High-Risk Infants| - full adherence, use of iron supplements 6-7 days/week for 3 months - iron deficiency at 9 months of age - anemia at 9 months of age |Study Start Date:||March 2005| |Study Completion Date:||December 2005| Iron deficiency is the most common known nutrient deficiency and cause of anemia in childhood. It is associated with numerous adverse health effects, particularly delayed mental and motor development, that may be irreversible. Despite advances of iron nutrition, the prevalence of iron deficiency remains high among low-income infants and toddlers. Previous studies suggest adherence with iron containing drops is low. Adherence to iron sprinkles among children as tested in studies in less developed countries appears high. Comparison: Children randomized to ferrous sulfate drops will be compared with children randomized to ferrous fumarate sprinkles. |United States, Massachusetts| |Boston Medical Center Pediatric Primary Care Clinic| |Boston, Massachusetts, United States, 02118| |Whittier Street Health Center| |Boston, Massachusetts, United States, 02108| |Principal Investigator:||Paul L. Geltman, MD, MPH||Boston University|
- Component of a computer processor that adds two numbers sent from the processing instructions. - An adder is a circuit that sums the amplitudes of two input signals. A half adder is a group of connected logic gates that create a logic circuit, incapable of handling addition for two numbers. A full adder is a circuit that adds two binary numbers. Also see: Processor definitions
Giovanni Sidotti Batlista (1668 - 1714) was an Sicilian Jesuit missionary, who entered Japan in July 1708, despite the Tokugawa Shōgunate's ban on Christianity. He was arrested upon landing, and taken to Edo for interrogation. He was later imprisoned and spent the remainder of his life in captivity. His arrest resulted in the ban on Christianity being strengthened and anybody who denounced Christians was rewarded with large sums of money. It was only after the downfall of the Tokugawa Shogunate in 1867 and the rapid modernisation of Japan during the Meiji Restoration that the ban on Christianity was lifted. One other result of his arrest was the later publication of the 3-volume work Seiyo kibun (西洋奇聞 A Report on the Occident) which was written by Arai Hakuseki. Hakuseki, a Confucian scholar and high-ranking official of the shōgunate, who was curious about the West (he had read the works of Matteo Ricci and other Jesuits) had interviewed Sidotti, via a Latin interpreter, during his imprisonment.
Explicit Solutions provided for Navier-Stokes Type Equations and their Relation to the Heat Equation, Burger's Equation, and Euler's Equation It is well known that linear differential equations with real coefficients lead to real dispersion relations only if they consist entirely of even derivatives or they consist entirely of odd derivatives or they consist of imaginary odd derivatives and even derivatives. This occurs because we have a direct correspondence between the equation and the dispersion relation through the correspondence Examples are the Klein-Gordon equation (even derivatives), the equation for flexural vibrations of a beam (odd and even derivatives), the linearized Kortweg- deVries equation (odd derivatives), and Schrodinger's equation (imaginary odd derivative and even derivatives).1 Such equations involve elementary solutions where is the wavenumber, is the frequency, and A is the amplitude, a constant in simple cases. The Navier-Stokes Equation Consider differential equations with even and odd derivatives, for example the Navier-Stokes equations are (3) .......... u=u(x,y,z,t) in which u, p are unknown velocity and pressure vectors, letter subscripts denote partial differentials, f is externally applied force (e.g. gravity), is a positive coefficient (the viscosity), is the gradient operator, and is the Laplacian operator. The first of equations (3) is Newton's law f=ma for a fluid element subject to the external force f and to forces of pressure and friction, and is the condition for an incompressible fluid. The x-component of equations (3) reduce to (4) ... u=u(x,y,z,t) and similarly for the y and z components of velocity v and w. Equation (4) assumes that force f has a gradient. Equations (3) and (4) apply in principle both to laminar and turbulent flows although, because of the impossibility of following all the minor fluctuations in velocity associated with turbulence and because of the difficulty and lack of solutions in turbulence problems, they cannot be used directly to solve problems in turbulent flow.2 There is no satisfactory theory for turbulent flow. For these reasons, more efficient solutions to the Navier-Stokes equations are needed. Indeed, a prize is available to provide a mathematical theory which provides a better understanding of the Navier-Stokes equations.3 And, since we have two unknowns u and p and one equation, solutions of equations (3) and (4) simply means finding u as a function of p as well as x,y,z,t. Equation (4) can be replaced by three equations and three unknowns so that if a solution u is found, solutions for v and w are available. The justification for equations (5) assumes that u, the x component of velocity, is mainly a function of x so that its y and z derivatives tend to zero. This assumption says that u=u(x,y,z,t) does not vary greatly when y and z change. Certainly, the last two equations (5) approach zero as . They also approach zero if the vector velocity is mainly in the x direction since then . With these assumptions, adding the last two second order equations to first equation (5) results in equation (4). If the last two terms on the right hand side of the first of equations (5) are dropped, the resulting equation is Burger's equation which, by using the nonlinear transformation may be reduced to the linear heat equation The general solution of the linear heat equation (8) is well known in which is the initial temperature, a known function from initial condition t=0. The solution for u is obtained from equation (7) in which is a known function from initial condition t=0.4 In summary, nonlinear transformation (7) eliminates the nonlinear term in Burger's equation (6) and reduces it to the linear heat equation (8) which has an explicit solution (9) thus providing an explicit solution (10) for Burger's equation. Explicit Solutions for the Navier-Stokes Equation I now propose to obtain an explicit solution for the Navier-Stokes equation, first of equations (5), following Whitham's method of obtaining an explicit solution for Burger's equation (6).5 Whitham's method consists of setting so that first equation (5) may be integrated to in which is evaluated between 0 and x, and then introducing to obtain a linear heat equation Solution With Constant Coefficient A constant coefficient means that -px+fx=0 so that the first equation (5) reduces to Burger's equation with solution provided by equation (10). Independently, substitution of in equation (13) reduces it to the heat equation whose solution is provided in equation (9). The solution for is equation (14) and the solution for is equation (12) which can now be written as and since the explicit solution for the Napier-Stokes equation, first of equations (5), when is constant is and using the last two of equations (5) to obtain solutions for v and w. Solution (17) is identical to solution (10) for Burger's equation. In summary, the explicit solution of the Navier-Stokes first equation (5), equation (17) with constant, requires that the velocity u is proportional to a function in which is provided in equation (9) and leads to solution (10) of Burger's equation. Solution With Variable Coefficient in equation (13) produces the linear nonhomogeneous equation Note that when =constant or when t=0, equation (19) reduces to the heat equation (15) whose solution is provided in equation (9). The corresponding linear homogeneous equation can be represented by an operator provides two roots D1,D2 of equation (20). The solution of the linear homogenous equation is, therefore (22) ....... .......... ........... in which y1,y2 are known functions of x,t and c1,c2 are arbitrary constants. By the method of variation of constants, we write the particular solution of equation (19) (23) ....... .......... .......... in which y1,y2 are known functions of x,t and A,B are undetermined functions of x,t. The method of finding A,B is straightforward.6 Set equation (23) in equation (19) and obtain two equations which can be solved for Ax,Bx (25) ...... ........ ........ and integrating to obtain (26) ..... ....... ........ Solution (23) then takes the form (27) ...... .......... The solution of equation (19) can be written as The solution for is equation (18) and the solution for is equation (12) which can now be written as and, since , the explicit solution for the Napier-Stokes first of equations (5) is and using the last two of equations (5) to obtain solutions for v and w. In summary, the explicit solution of the Navier-Stokes first of equations (5), equation (30) with variable, requires that the velocity u is proportional to a function in which is provided in equation (28). However, the Navier-Stokes first equation (5) is an exception to the general rule that solutions can be defined by initial conditions. Since , equations (25)-(30) do not exist when constant or when t=0 or when x=0 since then W=0, solutions y1,y2 are no longer independent, and integrals do not converge. For this reason, solution (30) does not reduce to solution (17) when constant, and we cannot give the velocity u(x,t) an arbitrary initial distribution u(x,0)=f(x) or u(0,t)=f(t) and expect to find u(x,t) as a function of f(x) or f(t). Nevertheless, solutions to many physical problems can still be defined by boundary conditions. In the present theory, solutions u,v,w are obtained using each component of vector velocity equation (3), i.e., we have three sets of solutions u,v,w at each point x,y,z,t. The solutions, therefore, must be identical. ................ x component of .................. y component of .............. z component of ............................. ................... equation (5) .......................... equation (5) ..................... equation (5) ........................ in which the descending diagonals are solutions u,v,w in terms of x,y,z,t and the remaining second order non-diagonal terms are solutions of u,v,w in terms of second to first derivative velocity ratios. There are a number of equations obtained by omitting, or adding, a term(s) on the right hand side of the Napier-Stokes first equation (5). With omissions, it would be hoped-for that the Napier-Stokes solution (30) would lead to solutions for some related equations. Unfortunately, since , this does not occur. The Burger equation is the first of equations (5) with -px+fx=0 which means -p+f= constant and thus constant. However, while setting -px+fx=0 in the Napier-Stokes equation produces Burger's equation, this same setting in the Napier-Stokes solution (30) does not produce Burger's solution (10). The reason for this is because equations (26)-(30) do not exist when and thus W=0 and solutions y1,y2 are no longer independent and, therefore, cannot predict the solution to Burger's equation. The Euler equation is the first of equations (5) with set equal to zero. However, while setting in the Napier-Stokes equation produces Euler's equation, this same setting in the Napier-Stokes solution (30) does not produce an Euler solution. The reason for this is because equations (12)-(30) do not exist when and, therefore, cannot predict the solution to Euler's equation. If the right hand side of equation (5) is set to zero, the result is which, since , gives u=constant. However, while setting the right hand side of the Napier-Stokes equation to zero produces equation (34), this same setting in the Napier-Stokes solution (30) does not produce solution u=constant. The reason for this is because equations (12)-(30) do not exist, since and constant and thus W=0, and, therefore, cannot predict the solution u=constant to equation (34). The Navier-Stokes equations (5) cover the case of turbulent flow regarded as unstable motion. Unfortunately, rapidly fluctuating instantaneous velocity u is impractical to compute in turbulence problems but, for many engineering purposes, we can replace u by where is the mean and is the turbulent rapidly fluctuating part of velocity. This requires introducing the Reynolds stress into equations leading to the derivation of equations (5). However, while a relationship exists between the viscous stress p and the derivatives of u leading to the derivation of equations (5), no relationship is known to exist between the Reynolds stress r and the derivatives of with result that adding a Reynolds term to equation (5) does not produce a more tractable equation.7 The modified equation is in which is the divergence of r. Equation (35) requires replacing second equation (11) by in which p,f and are known in many practical situations. Note that since the y and z components of stress r are second order effects, the integral in (36) reduces to r, the x component of stress. The solution to equation (35) is provided by equation (30) provided equation (36) replaces equation (11). 1 G. Whitham, Linear and Nonlinear Waves, John Wiley & Sons, 1974 pages 366-367. 2 J. Kay, Fluid Mechanics & Heat Transfer, 2d ed. Cambridge University Press, 2d ed. 1962 pages 150-151. 3 Clay Mathematics Institute at http://www.claymath.org/ 4 Whitham above pages 97-98; Lester Ford, Differential Equations, McGraw-Hill Book Co., 1955 pages 272-274. 6 Earl Coddington, Ordinary Differential Equations, Dover Publications, Inc. 1961. pages 66-68. 7 Kay above Chapter 14. Copyright © 2003 by James Constant By the same author: http://www.coolissues.com/mathematics/sameauthor.htm
The Virtues: The Seven Habits of Champions Ė Show #5: Fortitude By: Dr. Marcellino DíAmbrosio A. Peter Kreeft, Back to Virtue. Society was good at the hard virtues a century ago. Now we are good at the soft ones. B. Epidemic of fear and weakness. We are overly tolerant of evil C. Fortitude & courage synonyms. Deal with 2 diff emotions Fear & Anger D. Definition of Fortitude: willingness to suffer personal injury, even death, for the sake of truth and justice.Ē The virtue that enables us to take risks. Virtue of irascible (fear & anger) passions. Preserves the good. Clears the way for doing it. (CCC 1808) A. We are vulnerable, fragile, creatures. We can use life & health. Given to animals & us. Impulse to cause us to flee from danger, death, hurt. B. But we often fear the wrong thing. Peer pressure causes us to do dangerous things: drunkenness, premarital sex, drugs. Fear disapproval of friends more than C. Prudence helps you know if a fear is rational, whether to flee when fear D. Fear of failure: ridicule does not kill you. Genuine fear: Neglecting a serious responsibility. Failure to take initiative. E. Fortitude is the strength that makes fear subordinate to what is right & true. Keeps you from acquiescing, giving in to, cooperating with evil A. Godís wrath: B. Given to us to provide us with the energy to go after and attack evil and injustice. C. Love the sinner and hate the sin. Anger gives you the energy to go after the sinner and restrain him from doing harm. D. Duty to attack evil: In Old Testament. God gets angry when people fail to go after the enemy, dislodging the Canaanites. Davidís response to Goliath a good example of righteous anger and aggressive fortitude. E. Anger is the proper response to injustice. A. Passive: Keeps us from running away from a confrontation B. Attack: use anger right to go after evil and attack it. C. Fortitude is not foolhardy and wasteful: mean between cowardice and rash, foolhardy daring D. Thomas More as example of passive fortitude & the balance involved 1. Balance: prudence 2. Fool throws self against something that is evil with no chance of victory. Waste. 3. Coward: shrinks from duty due to fear. 4. Friend & Chancellor under Henry VIII. Did not rant & rave against the king. Family. Did not need to speak. Strong signal to those in Europe. No legal grounds. Preserve his life while protesting. 5. Thomas would not compromise with evil. 6. Caving in to the king and lying would injure his own integrity, greater injury than the king could inflict upon him. 7. Fortitude understands that the loss of life is not the greatest tragedy, but rather the loss of oneís soul V. Doing the Good involves injury and hardship: Annexed Virtues of fortitude 1. Be careful what you 2. Dealing with hardships with serenity and cheerfulness 3. Children a gift 1. Go out start doing something. Didnít know it was going to be so hard. 2. Sticking with your call as multiple obstacles crop up 3. Thomas Edison: 10,000 failures. Turns misfortune into a learning experience. 4. Maintains vision VI. Dimensions of Fortitude A. Lack of Fortitude one of the biggest problems in the church B. Bishops fearful of confronting priests; priests of people, people of others C. Attack: not necessarily violent Aggressive D. Catholics are not pacifists. Couldnít have a police force. Sniper in Wash DC: must be stopped, even if it takes deadly force. Just war theory. E. Only the state can do this: not take justice into ones own hands F. Taking initiative to confront and restrain evil G. Evil: the lack of a good that ought to be there: 1. Example of loss of 18-35 yrs olds. Lack of evangelization 2. Poverty: an objective evil 3. Sickness & disease H. In Acts: boldness is one of the most common words I. Gift of the Spirit of Fortitude. Peter proves that fearful people can change! VII. 4 Cardinal Virtues A. When you have all 4, you are a leader B. All of us are called to be leaders. C. Most are called to be parents; parents must be leaders; therefore most are called to rise to leadership and must acquire the character necessary D. Sheep who become Shepherds E. Canít happen without grace VIII. Discussion Questions A. Name some irrational fears that paralyze people. How about you? B. Name some things we really should fear. C. Which annexed virtue under fortitude do you need most? Patience or Perseverance? D. Discuss examples of fortitude in Scripture. This outline appears in the Virtues & Christian Character, Christian Lifestyle and Prayer & Spirituality sections of The Crossroads Initiative Library. EWTN - The Virtues: Seven Habits of Champions Dr. Marcellino D'Ambrosio Dr. Marcellino D'Ambrosio offers a spirited 8-part mediation in this new EWTN Home Video on the four moral virtues - Fortitude, Prudence, Justice, and Temperance - as well as the three theological virtues of Faith, Hope and Charity. Seven Habits of Champions lays out how God's inspiration for Christian living, contained in the Scriptures and made known through the Catechism and Church documents, is accessible to everyone and demonstrated for us through the lives of the saints. We are all called to holiness. And though the practice of these seven "habits," heroic virtue can be attained! Each DVD tape contains 4 episodes aired on EWTN! Introduction to the Virtues The Cardinal Virtue of Prudence The Cardinal Virtue of Justice The Cardinal Virtue of Temperance The Cardinal Virtue of Fortitude The Theological Virtue of Faith The Theological Virtue of Hope The Theological Virtue of Charity Set of Two DVD with 8 shows - $49.95
Grate, Slice, or Chop: How to Prepare Your Vegetables for Fermentation When you're first getting started with culturing vegetables there are a lot of things to learn, like how to know when they are truly cultured and how to be okay with allowing them to just sit out at room temperature for a while without refrigeration. One thing to learn is how to cut your vegetables, for how you prepare your vegetables will affect the outcome of your ferment and can determine what other ingredients and preparations you need to include in order to create a good finished product. You have a number of options on whether to grate, slice, or chop your vegetables and that decision will affect your next step. Grating a vegetable for fermentation is often done either in a food processor or on a large box grater. You will most often see instructions to grate your vegetable when it is a hard or crunchy vegetable. Good candidates for grating include: Grating your vegetables for fermentation creates the most surface area of any of the preparation techniques. This increased surface area allows salt to penetrate the vegetable more quickly, which draws the moisture out and in turn creates a brine for the vegetable. This means that when grating you are preparing a self-brining cultured vegetable. These are often the texture of a relish when finished. Some people make sauerkraut by grating the cabbage in a food processor. While this does facilitate the drawing out of the brine, it isn’t necessary to achieve a self-brining sauerkraut or other ferment. Grated vegetables are generally the quickest to culture because of their increased surface area. A chopped vegetable is simply one that has been cut to small bite-size pieces. When you cook vegetables and are told to “chop” an onion and then sauté it you most likely end up with a piece of vegetable that is small enough to have several bits in each mouthful. When you chop a vegetable for fermentation it can be that small or much larger, as in the case with a carrot stick. Most often, if the ferment requires something other than a 1” or smaller chop, the recipe will specifically tell you. Chopped vegetables almost always require a brine of salt and water. Chopping requires much less time than grating (or slicing). You can very roughly chop up chunks of summer squash, cucumber, or carrot and throw some salt in water and you will have a very quickly prepared cultured vegetable. Good candidates for chopping include: The main consideration for what size to chop your vegetables is the length of time it will take for the vegetable to culture. A hard carrot stick will take about twice as long to culture as a grated carrot because the brine has to completely penetrate the hard carrot. Chopped vegetables, depending on their size, usually take much longer to culture than grated or thinly sliced vegetables. Slicing thinly is much like grating in that it increases the surface area. Long, thinly sliced vegetables may create their own brine — like sauerkraut — but other vegetables may have a cellular structure that still requires you to create a brine. So while you might simply throw thinly sliced cabbage, sea salt, and some juniper berries together to create a simple sauerkraut, a thick cucumber slice would need a brine in order to culture properly and maintain its crunch. Good candidates for slicing include: Because most recipes call for grating or chopping vegetables, you will have to simply take into account the nature of the vegetable when deciding whether or not to create a brine. Sliced vegetables are generally right in the middle in terms of a culturing time. This may give you a wider window of culturing time in case you need to get your ferment in cold storage when you are leaving town, or accidentally leave it to culture one too many days.
Parts of the Sentence - Verbals Review Instructions: Using all the knowledge learned in the previous lessons, find the verbs (v), subjects (subj), predicate nominatives (pn), direct objects (do), appositives (app), nouns of address (na), adjectives (adj), predicate adjectives (pa), adverbs (adv), prepositions (prep), objects of the preposition (op), prepositional phrases (p ph), indirect objects (io), and objective complements (oc) in the following sentences. If the word is a verbal, tell whether it is a gerund, participle, noun infinitive, adjective infinitive, or adverb infinitive. Tell which word the adjective, adverb, prepositional phrase, verbal, or verbal phrase modify. Example: The actors performed there to entertain and to be seen. (performed = verb, actors = subject, the = adjective modifying actors, there = adverb modifying performed, to entertain/to be seen = adv. infinitives modifying performed, and = conjunction) 1. I finally bought me a hearing aid to hear better. 2. Sometimes I just need to try again. 3. Having decided definitely, he stepped onto the train to leave home. 4. The person winning the lottery will have a different life. 5. You can only reach our place by crossing the river. --For answers scroll down. 1. bought = verb, I = subject, finally = adverb modifying bought, me = indirect object, hearing aid = direct object, a = adjective modifying hearing aid, to hear better = adverb infinitive phrase modifying bought, better = adverb modifying to hear 2. need = verb, I = subject, sometimes/just = adverbs modifying need, to try again = noun infinitive phrase used as the direct object, again = adverb modifying to try 3. stepped = verb, he = subject, having decided definitely = participial phrase modifying he, definitely = adverb modifying having decided, onto the train = prepositional phase modifying stepped, onto = preposition, train = object of the preposition, the = adjective modifying train, to leave home = adverb infinitive phrase modifying stepped, home = adverb modifying to leave 4. will have = verb, person = subject, the = adjective modifying person, winning the lottery = participial phrase modifying person, lottery = direct object to the verbal winning, the = adjective modifying lottery, life = direct object, a/different = adjectives modifying life 5. can reach = verb, you = subject, only = adverb modifying can reach, place = direct object, our = adjective modifying place, by crossing the river = prepositional phrase modifying can reach, by = preposition, crossing the river = gerund phrase used as the object of the preposition, river = direct object to the verbal crossing, the = adjective modifying river DAILY GRAMMAR - - - - by Mr. Johanson Copyright 2012 Word Place, Inc - - All Rights Reserved. For your convenience, all of our lessons are available on our website in our lesson archive at http://www.dailygrammar.com/archive.html. Our lessons are Daily Grammar Lessons Search
Poetry can be fairly considered as one of the most primitive forms of literal arts. Poets have been affecting the cultures and civilizations of the world since the time immemorial. Even today, there is no such society which is devoid of the pride of having world class poets. All languages, all societies and all cultures have nourished poets of their own and it is the rhythmic and spell bound effect of their poetic expressions which have always left long lasting marks on history. Even in the current era, where science and technology seems to have shadowed all other fields of life, the culture of holding such events where poetry and poets are promoted is quite common. Poetry slams are such functions where this practice is carried out. These competitions are help all around the world almost throughout the year. There are different organizations which arrange these events and there are some set patterns for these events. All poets and other contestants compete in such events and there are proper judges which are of course some of the leading artists and poets. In some communities, these poetry slams are carried out once a year, in some, once a month! There are proper schedules for getting registered in such organization which organize these events. Once you are registered, there are certain terms and conditions under which you can compete in such events. In most of the cases, there are lucrative prize amounts and awards for the winners of the poetry slams. There are many websites conveying information on these poetry slams. If you are one of those ambitious young poets who would like to be a part of these events, it would be rather advisable to you to search out the coming dates and schedules for these events and ensure your presence in the next competition in your locality. All performing artists enjoy the limelight, attention and awe from the public audience. Any live performance in turn gets more energetic with audience participation. Musicians today encourage audience participation by motivating them to sing along and clap their hands to the beat. This kind of behavior has also seeped into stand up comedy shows and open mic poetry recitations. Poetry slams events remain a success to date only because of the public audience that is drawn to the charisma of clannish eras – eras wherein almost everyone in the community took part in tribal folk dances, singing sessions and story narration. Poetry slam first started somewhere in the year 1984 by Marc Smith – a construction worker and took place in a jazz club in Chicago. A couple of years later Smith proposed a plan to another club, the owner of the club accepted Smith’s idea. Smith’s idea was to host a live poetry competition on Saturday nights for enthusiastic performance poets, slam poetry was thus introduced. These poetries slam events have pushed poetry to a more energetic level of performance, making it a passionate experience for poets and participants alike. Initially slam poetry was all about certain subjects that revolved around public concerns such as politics, social issues etc. But the poetry slams of today have expanded immensely with a wide variety of themes and subjects. These poetries slam events are not something that will fade away and are surely here to stay. A form of entertainment that is based on imagination and rooted in performance art will always receive not just an enthusiastic audience but also high brow criticism. Poetry slams in the true sense are not merely based on the words or lines that are recited and as a matter of fact it is emphatic narration, voice modulation, eye contact and body language that go into this art of performance poetry. In simple words poetry slam refers to a form of competitive art by way of poetry performance. It lays emphasis on performance as well as writing skills of the poet, thereby encouraging poets pay attention to what is being recited and how it is being recited. In general it is a competition like event that gives aspiring poets a platform to showcase their work and talent to an audience. Poetry slam events are generally held weekly, bi-monthly or perhaps just once in a month, it is all based on the community of the given area. If the community has the strong support of the audience and a lot of enthusiastic participants coming in to perform in the area it is likely that slam event will take place more often in the area. The judges are selected by the co-organizers or hosts from amongst the audience attending the event. These randomly selected judges are then instructed to rate the poet’s work on a scale of one to ten based on the performance and content of the poetry recited. Those who wish to participate can get their names enrolled even on the day of the event at the venue. Prior enrollment would always be recommended since it is an open event based on a first come first served basis. Generally every person who enrolls himself is given an opportunity to perform in round one. The opportunity to perform in the subsequent rounds is purely based on the score of the judges and it is the judge’s vote that determines which poet they would like to hear more work from. • Slam events are generally organized by enthusiastic poets with the motivation to cultivate the art of poetry in their respective communities. All the work is usually done voluntarily and the entry fee if any is used to keep the show going or to pay for any special guest performance held at the slam event. Berkeley slam is one of those many festivities held in appreciation of poetry. It is all about appreciating spontaneous poetry. Berkeley slam is a great platform for amateur poetry enthusiasts who wish to pour out their poetry to an audience in public. You do not need to have any of your poetry in print or any such kind of formal eligibility to take part in the different kinds of poetry related competitions at the Berkeley slam. In fact, it’s an open platform where by one can walk into this open mic event and jet across your poetic verses to the public gathered out there. And who knows you may even be voted towards winning by the randomly chosen judges – from among the audience. Yes, friends you will be surprised to know that the judges here are actually randomly picked up from among the audience gathered, these judges are then made to vote on poetry that is recited. The poet with the maximum votes from the judges is declared winner. The best part being that everyone who is part of the slam gets to have a share in the fun. The audience is treated to interesting new poetry depicting the views of the new world, upcoming poets and poetry enthusiasts get an opportunity to be heard and appreciated by the public and those lucky to be randomly picked to judge are proud for being part of the slam. People who are not fond of poetry also make up a large crowd among the audience, in view of the socialization at the gathering and since socializing makes most people happy they do not mind, in fact all people happily attend the gathering even though it represents poetry. The Berkeley slam has musical guests as part of the show, pulling the audience with their performances between the poetry sessions. The title might seem strange to some, but this article will help us understand what is it that makes a poetry – call poetry. Basically it is the way the expression of the words, confession, meter arrangement, symbolism, irony, resemblance, spontaneity etc. At the end of the poem, the poet who had just recited the poem, will get a feeling that whatever he had in his mind has been revealed in just a few couples of lines. A poem is meant to bring some sort of chill in the opposite person who is listening to it. Reciting a poem also entirely depends upon the poet. An ordinary person might not be able to recite a poem in the best possible manner but a professional poet will ensure that his poem has touched the heart of the listeners. The poet will have a strange expression in his face which automatically touches the heart of the listeners. His movement of the eyes also plays a very important role while he recites a poem. There are chances that a poet does not capitalize or win the heart of his listeners but that is similar to the way how each and every song that we hear might not necessarily touch our hearts in any way. But some songs will mesmerize our minds in such a manner that we do not get bored to hear the song again and again and again. Each poem has got a short story in it. There will be a start and an end to the poem. Even the shortest of poetry will have a meaning and that is exactly why poets are given their due respect. All of us might not be able to become poets but professional poets will make sure that their poems are loved, enjoyed and remembered by their fans for a very long period of time. Incase you are romantically involved with someone and you wish to express your love in one of the best possible way, then expressing your love through a poem written by you to your loved one can be one of the best choices. To write a romantic poem, you do not require any artistic skills. Here are some basic tips on writing a romantic poem to your loved one. 1- Think about the person you love – First and foremost, think about the person with whom you are romantically involved. Think about all the good memories that you both shared together. What feelings do you have when you think about him/ her? Ask yourself, while thinking about your partner, do you get warmth inside your body or do you get butterflies feeling in your stomach or do you visualize angels singing around? 2- Compare the beauty of your loved one – When you think about your loved one, think about what you like in your person and compare that with some scenic beauty. For example – if your loved one has beautiful gray eyes, compare it with beautiful mountains or deep lake. Remember, a poem consists of such imaginary words. 3- Check website for rhyming words – In case you are a bit confused pertaining to some of the words that you need to pen down and you wish getting a rhyming word for your poem, the best way is to go through a website. There are many websites that provide rhyming words for each and every word. So, you can get loads of help here. 4- Make it short and sweet – Try making your love poem short and sweet. You do not need to write a poem like an essay. Just jot down important points and a little bit of help from the internet and you will be able to impress your loved one. Have you ever noticed that funny poems are like a magic which can help transform our lives completely? Well, here are 5 different methods which will help transform our lives. 1- Children love funny poems – Children love hearing funny poems. They are often seen laughing and enjoying themselves all through the program. 2- Late Shel Silverstein is known to have sold millions of copies of funny poems. His poems are found in library’s homes, bookstores, schools etc. What makes his poems so famous? Well, most of his books have sketches painted in black and white color but the main secret is that Shel was actually a kid advocate. He knew what kids love and how to help kids become responsible citizens of their country. 3- Some teachers are known to share a funny poem with the class every time their lecture ends or are about to begin. This helps students to attend the lectures whenever teachers take such an initiative. The students loved hearing the poems because such funny poems made them laugh and also help them see the other side of real life. 4- While narrating a funny poem it is equally important that we enact similarly so that children can relate to the poem easily. Just reading out a poem from a book might not help students love poems. The presenter must also have the ability to act while performing. 5- Laughter is a great exercise and enjoying a natural laugh through funny poetry can help in getting your body all geared up. Laughter helps in boosting a human body’s immune system, improves our mood, relaxes our muscles, helps in defusing conflict and helps others attract towards us. And this is what happens when we listen to funny poems. Try listening to funny poems today and see the difference it can make to your mind, body and soul. Becoming a poet is something which many are not able to achieve. Though poetry is hardly made from few sentences, the entire poem will have a great beginning and an exciting end like a short story. It is surprising to see how poets can create such unique poems. There are different terms of poetry such as metaphor, acrostic, rhyme, sonnet, free verse, prose, haiku are few examples of the various types of poetry forms. Poetry is not something that needs to be feared of. Rather it is an educational tool that helps students in learning of writing and reading which no other genre is capable of. Poetry is one of the best beauties of a human heart that expresses emotions leaving the listener with a complete feeling of contentment. There are no other forms of writing that expresses the ability to inspire which poetry can create. One of the ideal forms of poetry which most of the newcomers can try their hands on is known as ‘Free Verse’. This type of poems helps students gain the required strength and confidence. In a Free verse type of poetry, the poem does not have any particular style or pattern. You can set your own rules to the poem. You can end each line of the poem with a rhythm. Incase you are using rhythm in your poetry, ensure not to stress the rhythm too much and do not add the rhythm just for the sake of adding a rhythm. The rhythm has to flow well with the poem. You can also write a poem without the use of rhythm. That is the reason why this particular type of poetry is known as Free verse. Many people are also of the opinion that poetry means it must include rhythm. Well, that is absolutely not correct and there are plenty of famous poems without rhythms. People love poetry. Poetry is something which is very short and ends with a great meaning. Many of us tend to get confused between poetry and a verse, though both of them resemble to be the same. This article will explain what exactly a verse is. Verse is a musical instrument using words for a metrical or melodious effect. All of us love to use rhythm in words. Let us take a simple example. On your best friend’s birthday, you take a popular tune and convert it by writing your own lyrics to the tune and recite it on your friend’s birthday. This is not called poetry, it is called verse. Similarly, whenever you open a greeting card and read out its wordings, it is called verse and not poetry. And almost 99.9% of greeting cards contain only verses. So what is poetry? Poetry too includes everything that is included in a verse. But poetry can cause a lot of emotions and provides a deep in-depth meaning of each and every word that a poet recites in the poetry. Many of us have the ability to write verses but very few of them can actually write a poem and that is the reason why poets are always given a lot of respect because each and every word that they recite will bring out emotions and feelings. Only an experienced poet can write a true poem. A poem is a short story. Within the few verses, an entire story is hidden. There are many genres of poetry. Some of the famous forms of poetry are Narrative poetry, Epic poetry, Dramatic poetry, Lyric poetry, Satirical poetry, Prose poetry, Elegy, Verse fable etc. Nobody knows when the poetry officially generated but it is said that poetry started more than a thousand years in the past. From the very beginning of time poetry was always considered a means to help people express their deep emotions creating a healing impression. In the Greek mythology, it has been considered that Asclepius, famously known as the God of Healing was the greatest son of Apollo. Apollo was known to be the God of poetry. Similar to writers, poets are also given the same amount of respect. There are many poetry competitions held across the world. As compared to a writer’s job of writing a novel which can turn on for hundreds of pages, in case of poetry, it can end within just a few lines. In spite of that, there are plenty of poetry competitions across the globe to know the poet’s ability to discuss a particular situation within the poem. A poet reciting poetry can be very soothing to the ears, mind and heart of the listeners. There are poems that end with rhymes. Here are some of the myths that people have about poetry – 1- All poetry does not require ending in rhymes. Poetry is completely flexible and it all depends upon how a poet tends to put forth his poems. 2- It is a clear myth that most of the poets writing poetry are all sad and depressed poets. This is absolutely not the case. Almost all of them writing poetry are doing so for the love of it and wish to express their emotions. 3- All poems do not require proper grammar as often discussed. Even if the grammar is not proper, it can be considered as poetry since the words might have been mismatched to flow well with the poem. Another myth many people think is that a poem is mostly based on the life of the poet writing the poem. This again is not right. Poetry can be written by anyone, anytime, anywhere.
Progress on air pollution reduction plan among Utah's environmental highlights in 2012 Scott G Winterton, Deseret News SALT LAKE CITY — A number of developments happened on Utah's environmental front in 2012, chief among them the state's progress to come up with an air pollution reduction plan that will pass federal muster. Utah air quality regulators are not quite there for the Wasatch Front, but the implementation of a tailpipe inspections program in Cache County may get that area in compliance by 2014. Environmental scientists will spend early 2013 crafting the hard limits on what industry can emit into Utah's skies, trying to shave levels of PM2.5 — tiny airborne particles — that clog lungs and impact public health. While components of that plan are expected to be tough and even painfully expensive to adopt, the state risks invoking the ire of the Environmental Protection Agency if clean air standards are not met. The good news on the pollution battle is that the Wasatch Front has prevailed in its efforts to zap ozone problems in the summer, with 14 air-monitoring stations registering no unacceptable limits over the past three years. Some of these successes, detailed in the state Department of Environmental Quality's year-end report, include a multiyear, multiagency effort to dissect the cause of the ozone pollution that persists in the Uinta Basin. A mild winter last year somewhat derailed data collection efforts by scientists to understand what kind of chemistry is at play, but preliminary information gathered will provide a roadmap for more analysis to continue in the coming months. Air quality regulators, too, are taking part in a regional study that looks at air pollution that impacts the eastern section of Utah, national parks or monuments, and western Colorado. Perhaps one of the most notable milestones of 2012 was the successful elimination at Deseret Chemical Depot of 45 percent of the nation's chemical weapons stockpile. Destruction of the last nerve agent was completed well ahead of an international treaty deadline, and 95 percent of the secondary waste has been destroyed as well. The Tooele County facility will undergo an environmental makeover before it takes on a new purpose under the purview of the U.S. Army. With an eye toward protecting water quality this year, the Utah Department of Environmental Quality directed improvements to the Weber River watershed area, including upper reaches of the Weber River, Chalk Creek and Summit County's East Canyon Creek. The department also completed an intensive survey of the Jordan River in the central Salt Lake area in an attempt to better understand sources of pollution in the impaired waterway and what can be done to improve that river's health. Joining forces with the federal government, counties, cities and advocacy groups, the state launched a cooperative monitoring agreement for the Great Salt Lake to enhance knowledge of its ecology and identify ways to best manage it for the multiple uses it supports, including waterfowl populations and industry. To boost delivery of clean water to Utah homes, the department oversaw the allocation of nearly $26 million to 21 different water systems via system improvement projects. Critics, however, blasted the state's issuance of a permit for the U.S Oil Sands PR Spring Project in eastern Utah, saying the oversight was not protective enough of potential groundwater in the area. The state countered that an extensive network of wells revealed an absence of groundwater, and the tar sands operation had sufficient environmental controls in place. - Kennecott lays off roughly 100 workers Thursday - US companies challenging contraception mandate - The future of food? 3D printing moves beyond... - Utah ranks No. 1 for economic outlook for... - 'Mantiques' could be a ticket to more cash - S.L. draws up airport plans - Couples registry gets preliminary nod from... - XanGo seeks ouster of co-founder in new lawsuit - S.L. draws up airport plans 33 - Couples registry gets preliminary nod... 29 - Should we let wunderkinds drop out of... 13 - Obama opposes GOP bill on Keystone XL... 11 - IRS official to take the 5th at hearing 8 - Apple's Cook to face Senate questions... 6 - The future of food? 3D printing moves... 6 - Obama threatens veto of Republican... 6
This mini-lesson consists of two parts. In the first part, I teach you how to create a “crayon rubbing” on a piece of cardboard. In the second part, you will learn how to use the displace filter to make text take the shape of the layer beneath it. You will learn how to: - Use tolerance settings to fine-tune selections - load and use a brush - recolor a brush - create a displacement map - use the displace filter to add dimension to text This lesson includes: - A 9-page, full-color pdf with screen shots (in PS CS3) and step-by-step instructions for Photoshop and Photoshop Elements users. - Four cardboard shapes (K. Pertiet) - One oil crayon brush and png image (K. Pertiet) You will need: Adobe Photoshop or Photoshop ElementsAdobe Acrobat While I do provide every step necessary to complete the techniques, it is helpful to have some experience with opening and closing files and a basic knowledge of the Photoshop or Photoshop Elements workspace. This information was previously distributed in a “How’d They Do That?” Chat. If you participated in that chat, you already have this product. File Size: 18MB LP834205 Designer: Cassie Jones
XML vs SGML XML stands for EXtensible Markup Language. It is defined in the XML 1.0 specification, which is developed by the W3C (World Wide Web Consortium). XML provides a standard way, which is also simple, to encode data and text such that the content could be exchanged across driver hardware, operating systems and applications with little human intervention. SGML (Standard Generalized Markup Language) is an ISO (International Organization for Standardization) standard for specifying a document markup language or a set of tags. SGML is not a document language but a Document Type Definition (DTD). XML is a markup language that is used to transfer data and text between driver hardware, operating systems and applications with little human intervention. XML provides tags, attributes and element structures that can be used to provide context information. This context information can be used to decode the meaning of the content. This makes it possible to develop efficient search engines and perform data mining on the data. Furthermore, traditional relational databases are suitable as XML data because they could be organized in rows and columns but XML provides less support for data with rich content such as audio, video, complex documents, etc. XML databases store data in a structured, hierarchical form which allows queries to be processed more efficiently. XML tags are not predefined and the users can define new tags and document structures. Also, new internet languages such as RSS, Atom, SOAP, and XHTM were created using XML. SGML is based on the idea that even though a document can be displayed with different appearances depending on the output medium used, it contains some structural and semantic elements that do not change with reference to how it is displayed. SGML based documents can be created without concerning about the document’s appearance which can change overtime, but concerning about the document structure. Further, SGML compiler can interpret any document using its DTD, therefore these documents provide more portability. Also, documents based on SGML could be easily re-adapted to different media (for example, document intended for print medium could be readapted for a display screen). What is the difference between XML and SGML? While XML is a markup language that is used to transfer data and text between driver hardware, operating systems and applications, SGML is an ISO standard for specifying a document markup language or a set of tags. XML is actually a markup language that is based on SGML. But XML imposes some restrictions that are not in SGML. For example, XML imposes following restrictions: entity references must be closed with a REFC delimiter, references to external data entities in content are not allowed, character references must be closed with a REFC delimiter, named character references are not allowed, etc. Furthermore, some constructs such as unclosed start-tags, unclosed end-tags, empty start-tags, empty end-tags that are permitted in SGML when SHORTTAG is YES, are not allowed in XML. Additionally, some SGML declarations such as DATATAG, OMITTAG, RANK, LINK (SIMPLE, IMPLICIT and EXPLICIT), etc. are not allowed in XML.
When you are traveling across Europe to its own train leave, it can be very tempting to concentrate around you the comfort of the train, and spend some magnificent scenery of rolling your window. Dramatic landscape that simply can not miss the region, however, is that the Carpathian Mountains. This spectacular mountain range can be seen from many different cities and countries in Europe, and it's really powerful and unique event for any train journey. Carpathians stretch from the Czech Republic, Romania, Slovakia, Poland, Hungary and Ukraine through the formation of a bow across Central and Eastern Europe, about 1,500 km long second-longest mountain chain in Europe. What's his name? Real name "Carpathian" is derived from ancient Greek "Karpaty Oros', meaning" Rocky Mountain ". Old Polish word "carp" meaning "aggressive violations, rugged roots or trunks. 'More common word" batter "- sharp rocks or other vertical landscapes. Otherwise, the name may come not from the Indo-European" Kwerps "or" convert " wart-like Greek, meaning "wrist" - perhaps a reference to the mountain bends or veers, L-shaped. car also means "king" and the same "road", so carpati "perhaps the king of way - apt comparison trains leave the region greatness. Although commonly referred to as mountain ranges, the Carpathian actually not a continuous chain of mountains. Instead, they consist of several distinctive groups in the mountains, with a wide structural diversity. It keeps a lot of travelers to their trains leave comment on the similarities in the Alps - But there are many differences in the Carpathian Mountains which is unique. While they may share a common look and climate of the region's average Alps, the Carpathian Mountains feature a little more gentle terrain, without the extensive snowfields, large glaciers, high waterfalls or shared among the Alps. Carpathian Mountains are separated from the Alps, the Danube, marking them as unique in their own right. Highest among the Carpathian Mountains, Gerlach Peak (Gerlach? Titus in its original form) has a symbolic role in several Central European nations and rulers in the eyes of the population, making it a perfect look on your train during peak holidays. Between 19 and mid-20th century, it had four different names and six name changes and was able to be the highest mountain in the Kingdom of Hungary and Czechoslovakia and Slovakia - in a space only about two decades of the 20th century. - Author Articles - More From This AuthorThe hotel offers modern amenities and warm hospitality at the entrance to historic Košice city center. Our newly renovated hotel is located downtown within walking tours, shopping, business centers and...The area usually is Snina Ukraine to the east and the north border of Poland. These villages are mostly Rusyn and declining population. Before any changes are more permanent, I'm trying to portray them as they...Thus, starting from the capital Bratislava, you will find a beautiful medieval old town, surrounded by a flat block and interesting suburbs. The city is unusual in that it borders two other countries, Austria...One thing that most people who are into digital photography will never learn how to use a digital camera and its functions and settings. They pay for pictures that are just OK, as soon as the camera's...Slovakia region is one of the most important tourist destinations in Central Europe, Slovakia and the region area is divided into types. The first is a mountainous area, which is the Carpathian Mountains and... - Related Articles - Related ArticlesNo doubt there are other fabulous thrift stores in Tampa that deserve a 'Best of' title. strapless prom dresses the perfect wedding gown is chosen comes the key task of choosing the right bridesmaid dress...The online marketing is getting more popular with the increasing users of the internet. This is true wherever the internet reached the online market reaches there. Moreover, the shoppers have become smart as...There must be a question that is raising in your mind right now, and that is “How can I select the best Photography studio from a huge bunch of photographers?” Well I would like to help you in that case....Digital techniques have eased the production of business cards. Although offset printing is the most sought after in terms of quality, digitalization increases the production with very less turnaround time....Mange is a range of dog skin problems that affect the overall well being of your canine. What causes dog mange is several species of parasitic mites that are transferred from dog to dog upon close contact with... - Latest Articles - Latest ArticlesDerma Rolleris extremely useful in fighting against premature aging by encouraging the production of collagen and by helping the skin in better absorption of skin care products. We also discussed whether it...Today, when anything and everything is available online; translation services also turn out to be no major exception. In fact, online translation services are pretty lucrative and profit generating as well....No doubt there are other fabulous thrift stores in Tampa that deserve a 'Best of' title. strapless prom dresses the perfect wedding gown is chosen comes the key task of choosing the right bridesmaid dress...The online marketing is getting more popular with the increasing users of the internet. This is true wherever the internet reached the online market reaches there. Moreover, the shoppers have become smart as...There must be a question that is raising in your mind right now, and that is “How can I select the best Photography studio from a huge bunch of photographers?” Well I would like to help you in that case....