ESSAI Created in 2002 by a cross-disciplinary faculty, ESSAI, The College of DuPage Anthology of Academic Writing Across the Curriculum is an annual academic journal that represents a unique but inclusive discourse community of student writers; it fosters student success through writing across the curriculum and in all levels of learning at College of DuPage. The resulting published work is their expanded cultural and epistemological self-endowed with talent, sophistication, and fresh scholarship, all expressed with clarity, eloquence, and grace. Join with us in honoring the exceptional talents of our student authors. Volume XII It rests on our belief that writing empowers and prepares our students for meaningful careers and advanced academic pursuits and that, to realize such goals, the entire college community is responsible for incorporating writing into their courses as a vital part of the academic experience at our college. That we practice the philosophy of student success through good writing will become even more apparent in the journal’s name. It harkens back and pays homage to Michel de Montaigne, whose seminal style of writing “essais” connoted one of trials and attempts. The College of DuPage Anthology of Academic Writing Across the Curriculum What is Essai? ESSAI 2013-2014 The College of DuPage Anthology of Academic Writing Across the Curriculum Volume XII 2013-2014 ESSAI The College of DuPage Anthology Of Academic Writing Across the Curriculum (Volume XII 2013-2014) The College of DuPage, Glen Ellyn, Illinois The contents of this anthology copyright, 2014 College of DuPage Awards and Recognitions ESSAI is a recipient of the prestigious recognitions of The 2005 Distinguished Achievement Award of the Community College Humanities Association The Society of Midland Author’s Qwertyuiop Award (Founded in 1915, the Society of Midland Authors has included for its members Hamlin Garland, Harriet Monroe, Vachel Lindsay, Edna Ferber, Carl Sandburg, Jane Addams, Clarence Darrow, and other literary luminaries.) and The 2009 Diana Hacker Two-Year College English Association (TYCA) Award for Outstanding Program for Two-Year Colleges and Teachers In the Category of Fostering Student Success (The TYCA is a national organization that belongs to the National Council of the Teachers of English-NCTE). II Acknowledgments We warmly salute all submitting instructors for their dedicated and inspired teaching which has contributed to the successful writing of their students and make it possible for ESSAI to continue. We also offer our special thanks to Dr. Robert Breuder, President, Dr. Joe Collins, Executive Vice President, Dr. Jean Kartje, Vice President of Academic Affairs, Catherine Brod, Vice President of Development/Executive Director, Foundation Dr. Daniel Lloyd, Dean, Liberal Arts, Dr. Sheldon Walcherl, Associate Dean, Liberal Arts, Mr. Tom Cameron, Dean, Health and Sciences, Dr. Laura Ortiz, Associate Dean, Humanities, For their constant, gracious support and encouragement. Our gratitude also goes to Karyin Boulom. Her meticulous and careful word processing expertise has been indispensable. Pearson Publishing’s continuing sponsorship of the celebratory reception of the student writers is deeply appreciated. Our special and heartfelt appreciation goes to Professor Blevins’ student, Elizabeth Brousil, who designed the book cover. The others who submitted designs were Rae Crawford, Lucas Gregulak, and Derick Magante. The publication and associated recognition event are underwritten, in part, by a grant from the College of DuPage Foundation III “In this work, when it shall be found that much is omitted, let it not be forgotten that much likewise is performed.” Samuel Johnson, the Preface to his Dictionary IV Foreword Created in 2002 by a cross-disciplinary faculty, ESSAI, The College of DuPage Anthology of Academic Writing Across the Curriculum is an annual academic journal that represents a unique but inclusive discourse community of student writers; it fosters student success through writing across the curriculum and in all levels of learning at College of DuPage. It rests on our belief that writing empowers and prepares our students for meaningful careers and advanced academic pursuits and that, to realize such goals, the entire college community is responsible for incorporating writing into their courses as a vital part of the academic experience at our college. Like all the past volumes, this Volume XII continues to harvest our belief in fertile cooperation with our cross-disciplinary faculty who implement writing to learn and promote discipline in their teaching. Some of the published essays, therefore, have resulted because our students made use of journals, logs, microthemes, and other informal writing assignments encouraged in the classroom. In this way, students wrote frequently and improved their writing skills enabling them to comprehend better and to retain information longer. Others – expository work such as reports, article reviews, criticisms, research papers – attest to students’ successful participation in academic discourse according to the conventions of language use and style expected in specific disciplines. That we practice the philosophy of student success through good writing will become even more apparent in the journal’s name. It harkens back and pays homage to Michel de Montaigne whose seminal style of writing “essais” connoted one of trials and attempts. Thus ESSAI chronicles our students’ intellectual trials wherein they dialogically encounter diverse culturescapes, critically weigh various issues at hand and then engage in analytical exercises, while each evolving from a personal to a public writing self. The resulting published work is their expanded cultural and epistemological self endowed with talent, sophistication, and fresh scholarship, all expressed with clarity, eloquence, and grace. We continue our editorial tradition in making the best selections for publication. The most exemplary of the year’s submissions are those that: • • • • • • • • Are completed according to assignments’ purposes Have mechanics, grammar, and other technical points in place Are imaginative, creative, logical, and risk-taking with respect to assignments’ purposes Exhibit clarity of writing with respect to purposes Follow the disciplinary format Include relevant literature reviews where required Demonstrate analytical abilities Generate interest Join with us in honoring the exceptional talents of our student authors and our student artist whose graphic design adorns our cover. And watch for Volume XIII of ESSAI which will be published in 2016. We appreciate your readership. Through your interest and support, our anthology will keep growing. The ESSAI Editorial Board: Brian Blevins, Graphic Design Franz Burnier, English Richard Jarman, Chemistry Keith Krasemann, Philosophy Sarah Magin, English Chris Petersen, Biology Lois Stanciak, Education Ben Whisenhunt, History V Table of Contents Awards and Recognitions ......................................................................................................... II Acknowledgments..................................................................................................................... III Foreword ................................................................................................................................... V Julia Andersen So You Want to Write YA: Social Responsibility and Young Adult Literature ...................... English 1102 1 Nicholas Bashqawi Kill to Conserve: Ethical Implications of Trophy Hunting Conservation Measures ................ Honors Philosophy 1110 19 Harry Bodell The Role of Discontent in Communism's Downfall: An Analysis of How We Survived Communism and Even Laughed................................................................................................ History 2235 24 Kellie Brennan Today’s Registered Nurse ......................................................................................................... English 1102 27 Kevin Brewton Sankana Essay........................................................................................................................... Philosophy 1100 33 Elizabeth Cole The Effects of Tetracycline and Ibuprofen on Common Duckweed, Lemna minor L.............. Honors Biology 1152 40 Neil Davis Role of Women in Service ......................................................................................................... History 1120 44 Kelly Donlan The Usefulness of Rubrics ........................................................................................................ Education 1100 46 VI Randy Dziak Civil Engineering: A Surprisingly Noble Career ...................................................................... English 1102 50 Katrina Favis The Legend of Starved Rock..................................................................................................... English 1102 57 Emily Garvanovic Miasmatic Theory ..................................................................................................................... Chemistry 1152 60 Frank Giuliani The Composition, Structure, Sources, and Applications of Eugenol........................................ Honors Chemistry 1552 63 Hayley Hickson Saving the Tallgrass Prairie ...................................................................................................... English 1102 67 Hope Javier The Four Humours Theory........................................................................................................ Honors Chemistry 1552 72 Jonathan Kaye Communicative Action via Internet Technologies ................................................................... Philosophy 1800 75 Syeda Khalid Pediatrics: Because We Care About the ‘Little World’ ............................................................ English 1102 82 Marie Labbe Physical Therapy: A Profession with a Promising Future ........................................................ English 1102 88 Amanda Manton Term Limits in the United States .............................................................................................. Honors English 1102 94 VII Matt Mazur Is there an Animal Consciousness? A Phenomenological Approach ...................................... Philosophy 1800 98 Virginia Meglio Life, Lottery, for the Pursuit of Organs..................................................................................... Philosophy 1100 107 Bryan Mundt A Fountain Very Exhaustible: The Pollution and Restoration of Lake Michigan .................... English 1102 113 David O’Sullivan Hospitality ................................................................................................................................. Philosophy 1110 117 Ashley Padavonia Effect of Resource Indulgence on Aquatic Vascular Plants: A Closer Study of Lemna minor ............................................................................................................................. Biology 1152 119 Becky Reichel A Humorous Lesson.................................................................................................................. Honors Chemistry 1551 123 Michael Robles Video Games as an Art Form.................................................................................................... English 1102 126 Dina Rudenya The 1968 Riots: Protesting the Democratic Convention........................................................... English 1102 130 Jake Sumeraj Changing Baseball Forever ....................................................................................................... Honors English 1102 134 Emily Trca Being a woman in the time of Stalin: Becoming the Other ...................................................... History 2225 139 VIII Alexandra Wickersham Mothers, Martyrs, Damsels, and Demons: Women in Western Horror from Romanticism to the Modern Age......................................................................................................................... English 1102 143 Ryan Woitas Naturally Occurring Persistent Organic Pollutant..................................................................... Chemistry 1552 156 Tribute to Faculty...................................................................................................................... 159 Cumulative Index of Student Authors....................................................................................... 160 IX So You Want to Write YA: Social Responsibility and Young Adult Literature by Julia Andersen (English 1102) D espite what the uninitiated might assume, there is much more to Young Adult literature than supernatural love triangles; writing quality YA is a difficult and time-consuming task. Although no one can deny the popularity of titles like Stephenie Meyer’s best-selling Twilight, for example, many a critic has lampooned her flimsy characters and flat writing, and the borderline-abusive relationship between the novel’s main character and her love interest has been thoroughly decried. Meyer’s novel, written in three months, took off in the Young Adult market, but the message she sends to teens about gender roles and relationships is considered by many to be a troubling one. YA authors must consider their impressionable adolescent audience while crafting their story, and take into account young adults’ many and varied needs during a time of intense life changes. As a relatively new genre, Young Adult literature is still coming into its own; therefore, it has great potential to address the needs of all its readers, including members of minority groups traditionally ignored by the publishing industry. As an aspiring YA author, you must educate yourself on the concerns and needs of your audience, the current state of the Young Adult writing and publishing industry, and the potential for that industry to change over time. Is YA truly the genre for you? Of course, to answer that question, you first have to figure out exactly what YA lit is. Unfortunately, there’s no easy answer to that question; because it’s a relatively new genre, Young Adult literature’s definition and conventions are still being determined. As Michael Cart notes on behalf of the Young Adult Library Services Association, a branch of the American Library Association, it is difficult to give such a dynamic genre a static definition. He says, “The term ‘young adult literature’ is inherently amorphous, for its constituent terms ‘young adult’ and ‘literature’ are dynamic, changing as culture and society—which provide their context—change.” Currently, YALSA defines a “young adult” as someone between the ages of twelve and eighteen. One of the few things on which most experts seem to agree is that Young Adult novels typically address coming-of-age issues, and feature a main character between the ages of twelve and sixteen. This age window is conducive to discussions of difficult topics that are often associated with adolescence, ranging from drug abuse to sexuality to race to self-determination to social acceptance. Some novels approach such issues directly, while others deliver guidance in the form of metaphor or analogy. Some are concerned with the suburban teenager’s everyday concerns—dating drama, changing or disintegrating friendships, familial pressures or parental troubles—while others take place in fantastic or far-removed settings, such as J.K. Rowling’s Hogwarts or J. R. R. Tolkien’s Middle Earth. Yet, more often than not, these fantastic and far-removed stories still deal with many of the same issues as their more immediately realistic fellows. For example, Harry Potter finds in the Weasleys the family he was denied by the Dursleys, struggles to separate himself from the evil he fights (and which, he finds, literally lurks inside of him), and deals with his share of romantic frustrations. Joanne Brown and Nancy St. Clair, authors of Declarations of Independence: Empowered Girls in Young Adult Literature, 1990-2001, explain: 1 [F]antasy and science fiction, the two main forms of literature of the fantastic, are usually set in either an unspecified past or future. Either setting allows for a world populated by beings or technology not found in a contemporary setting. Yet while the world of the novel is significantly different from the world of the reader, it frequently is plagued by the same anxieties about the same conflicts as the world the reader inhabits. (140) Evidently, addressing common adolescent concerns is a major part of what classifies a book as Young Adult literature. According to Pam B. Cole, author of the McGraw-Hill textbook Young Adult Literature in the 21st Century, other “characteristics that have historically defined the genre [of YA lit]” include: 1. The protagonist is a teenager. 2. Events revolve around the protagonist and his/her struggle to resolve conflict. 3. The story is told from the viewpoint and in the voice of a young adult. 4. Literature is written by and for young adults. 5. Literature is marketed to the young adult audience. 6. Story doesn’t have a ‘storybook’ or ‘happily-ever-after’ ending—a characteristic of children’s books. 7. Parents are noticeably absent or at odds with young adults. 8. Themes address coming-of-age issues (e.g., maturity, sexuality, relationships, drugs). 9. Books contain under 300 pages, closer to 200. (49) However, Cole immediately goes on to warn that “reliance upon a laundry list of characteristics… results in a narrow and misleading definition of young adult literature.” She mentions, for example, that “YA books range from thin novellas to books with 800+ pages,” as in the case of Rowling’s Harry Potter series, mentioned above (49). So, while the protagonist is almost always a teenager, and events are often related from their viewpoint and in their voice, not all aspects of Cole’s list are or remain accurate. A notable example of the changing face of YA lit, and one that has already begun to void Cole’s list, is the increasing interest of adults in the Young Adult market. As Cole herself has noted, “In [recent] years the young adult genre has evolved, becoming more sophisticated, more inclusive, has gained more widespread popularity” (50). In fact, even more adults than teens are beginning to flock to YA lit. According to Ashley Strickland’s recent article, “A Brief History of Young Adult Literature” on CNN.com, “55% of young adult books purchased in 2012 were bought by adults between 18 and 44 years old.” Many experts cite the inherent turmoil of adolescence as a draw for audiences of all ages, whether they be teens eager to see their real-life struggles dealt with in some form, or adults looking to revisit the transformative teen years in print. As one young adult author said in an interview with Pepperdine University’s Pepperdine Magazine last year, “‘I think a lot of adults like the theme of finding yourself. You get to go back to that time in your life and relive the wonder of your teenage years, and all of those not so wonderful firsts, too’” (Suzanne Lazear qtd. in Fisher). Fellow YA author Cynthia Struloeff agrees, “‘Teenagers’ lives are about transitions and defining moments. That’s very exciting in a narrative sense, so it makes sense to me that adults are drawn to YA’” (Struloeff qtd. in Fisher). As Kat Falls, author of YA novel Inhuman, told Bob Sirott and Marianne Murciano during her WGN radio interview: “I read YA, and so much is being done in that genre that’s exciting and new and edgy. And you don’t get boxed into a genre the way you do with adult fiction. Like, if you’re sci-fi, you’re in this one section of the bookstore, and only sci-fi lovers head over to it. Well, YA, all those books are shelved together. They’re just lumped as YA because it has a teenaged protagonist.” 2 The variety within the YA genre is something readers of all ages may find attractive—the genre’s inclusion of titles addressing everything from death and life-after-divorce to first dates and after-school jobs, making Falls’s definition seem particularly apt. However, not all authorities in the field agree. In a personal interview, Young Adult author Trina Sotirakopulos, argues: Adult books with young protagonists are still Adult books, such as Lovely Bones. … Lovely Bones is [narrated by] a character, Suzy Salmon, who’s twelve… but she’s writing from heaven, and she’s writing about her brutal murder. …So that’s Adult. …I don’t think you can look at the age of a character and say that this is Adult or Young Adult. Young Adult is more responsible. …The Adult book, you can write however you want. You could end with horrible, horrible sadness, and you’re not held to any other standard…. But in the Young Adult novel, it must end with hope. Something must give the teenager hope. …I think that’s the difference. However, one of the key attributes of Young Adult literature, which differentiates it from Children’s literature, is the absence of “storybook endings”—so while Young Adult novels do generally end on a hopeful note, this characteristic does not necessarily prevent the author from addressing difficult or delicate topics. After all, YA includes titles like Speak, The Hunger Games trilogy, and The Fault in Our Stars, each of with conclude with the sort of practical optimism that has become a characteristic of the genre. The first is a novel by Laurie Halse Anderson that centers around a rape victim’s struggle to acknowledge and overcome the trauma she suffered; the second, by Suzanne Collins, features protagonist Katniss’s struggle to first survive in, and later bring down the tyrannical government of, her dystopian near-future setting; and John Green’s The Fault in Our Stars deals with issues of mortality and grief through the experiences of teenaged cancer patients. In each, the protagonist works to overcome, or at least come to terms with, their difficulties. Rather than curing Hazel, the main character of The Fault in Our Stars, Green writes of her appreciation for life even as she fears dying and comes to terms with the death of a friend (also a cancer victim). Rather than promising a perfect new world when Katniss succeeds in escaping the previous government regime, Collins offers the potential for large-scale healing in an improved (but still imperfect) world—hope over a trite happily-ever-after. However, not all YA novels accomplish this balance of realism and optimism. Cole claims, “The societal belief that stories should have hope impacts the way some writers end their stories.” She explains, “[w]hile good writers shy away from the happily-ever-after endings, many still tend to ‘lighten’ events by story’s end,” despite the fact that, typically, the time frame of a YA novel “is not conducive to deep character development and, most importantly, it is not long enough realistically for characters to work through personal and family issues” (65). These endings may be the basis for some criticism on the general quality of the YA genre, although such criticism is generally considered surface and uninformed, as there is great variety in any other category of literature. As Cole explains in Young Adult Literature in the 21st Century: 3 Grounding their criticism in erroneous and misleading media coverage and in limited familiarity with young adult literature, some critics assert young adult literature is worthless and inferior to classical texts and adult literature. Those who study the young adult genre, however, understand multiple factors contribute to this negative pigeonholing. They also know that young adult literature offers some of the best in contemporary literature. … A number of issues contribute to a stigmatic view of the genre. First, the location of young adult literature in bookstores is problematic. Some bookstores combine young adult literature with children’s literature, and by doing so, send an unwritten message that young adult literature is not sophisticated enough for teens, especially older readers. Other bookstores place young adult literature in close proximity to children’s literature, sending a similar message. (57-60) So, while some would consider Young Adult literature childish or simple, it is actually a relatively sophisticated and often challenged genre. In fact, because so many YA novels tackle important and sensitive subjects, covering everything from abuse to poverty to sexuality to race to religion, the genre is a favorite target of censorship groups across the U.S. Cole explains, “[w]ouldbe censors genuinely believe they have a definitive take on ‘morality’ and… criticize young adult authors for ‘corrupting the young,’ ‘being evil,’ or ‘writing shocking, violent, and obscene stories to make a buck.’” She condemns their censorship, which “violates intellectual freedom, the individual right to access information from all viewpoints free of restriction,” and connects their behavior to that of oppressive regimes: Censors of young adult literature fail to see the parallels between their own desires to suppress information about such topics as sexuality, religion, violence, vulgarity, and the successful attempts by dictators, governments, and military groups to control and even brainwash their countries’ population. They see their actions as ‘helpful’ and ‘decent.’ (70) Cole also provides a list of classical titles challenged and the censors’ reasons for objecting, including Diary of a Young Girl by Anne Frank for being “obscene and blasphemous,” J.D. Salinger’s The Catcher in the Rye—“A dreadful, dreary, recital of sickness, sordidness, and sadism”—and Harper Lee’s To Kill a Mockingbird, because “The word ‘rape’ is used several times. Children should not see this in any literature book” (70-71). Of course, contemporary YA titles are also challenged fairly regularly; now, however, people are fighting back very publicly. The American Library Association, for example, runs a Banned Book Week each September encouraging people to read books that have been banned or challenged. According to the ALA, the top three reasons materials are banned are that “the material was considered ‘sexually explicit,’ …contained ‘offensive language,’ … [or] was ‘unsuited to any age group.’” The ALA even keeps lists of the most challenged authors by year, and the top one hundred banned books by decade. Recent years have featured names like Sherman Alexie, John Steinbeck, Judy Blume, Ellen Hopkins, Suzanne Collins, J.K. Rowling, Stephen Chbosky, Chris Crutcher, Lois Lowry, Toni Morrison, and John Green as most challenged authors. The list of most challenged titles in recent years include Rowling’s Harry Potter series, the novel I Know Why the Caged Bird Sings by Maya Angelou, and even the Captain Underpants series by Dav Pilkey. Individual authors have also been seen taking a stand against book banning. When her book, The Bermudez Triangle, was challenged in Barlesville, Oklahoma, author Maureen Johnson mobilized her fan base by posting multiple blog entries about the attempted censorship and her response. The parent in question challenged Johnson’s book on the grounds that it contains “[h]omosexual content, unprotected sex, underage drinking, and reckless promiscuity” and “wrongly 4 promotes a ‘do whomever you want to discover yourself’ mentality. There’s no mention of the myriad of diseases, pregnancy, destruction of friendships and lives that are very real consequences of a ‘sexual free-for-all’ decision” (Angela Rader qtd. in I Am A Very Dangerous Person, Johnson). While there is a homosexual relationship in The Bermudez Triangle, there is no sexual activity beyond kissing, invalidating Rader’s charges of “reckless promiscuity” and eliminating any worry of pregnancy or sexually transmitted infections. Furthermore, Johnson writes, “The book is entirely about what happens to friendships when relationships enter the picture,” and therefore devotes much attention to the possibility of ruined friendship as a consequence of romantic involvement (Johnson, “I Am A Very Dangerous Person”). When the school board, who admitted they had not read the book before voting, recommended removing the book from the high school library, Johnson took action. It was not simply because she thought her book did not deserve to be banned, she explained in multiple entries, but because book-banning sets a dangerous precedent, especially when school committees do it so quietly that the majority of their towns’ residents don’t even realize it’s happening. She emailed the committee members, alerted her readers, and even involved the town’s local newspaper. She continued to chronicle her struggle against the book banning until the case’s conclusion, almost two months after the committee originally voted. They decided that Johnson’s book would be kept on a “special shelf,” and only be checked out to students with a parental permission slip (Johnson, “News Flash: Victory in Bartlesville (Sorta, Kinda)”). Johnson credited this partial victory in large part to the hundreds of letters and emails her readers and fans sent protesting the banning. Johnson is not the only individual author who has taken a stand against book banning, either. Chris Crutcher is known for his arguments against censorship, and other authors with a notable social media presence have used their online clout to help anti-book banning movements gain momentum. Authors like Johnson or John Green may provide the professional email addresses of the public officials involved so that concerned citizens can share their views on literary censorship directly with the people in charge of determining whether or not a challenged book will remain on the shelves. Of course, social media can be used for publicity outside of anti-book banning movements; the vast majority of authors today have some online presence, and some have particularly large and devoted followings. Of these, few can rival Internet and YA giant John Green. Green has certainly had his fame augmented thanks to his social media presence as half of the wildly popular Vlogbrothers YouTube channel, and co-founder of the DFTBA record label and product line (named after a Vlogbrother community catchphrase, “Don’t Forget to Be Awesome”). The Vlogbrothers “brand” has expanded to include multiple YouTube channels like CrashCourse, the Art Assignment, and even an Emmy Award-winning web series. Of course, Green’s literary work can also stand on its own; as Cole writes in Young Adult Literature in the 21st Century: [Y]oung adult literature offers some of the best in contemporary literature. It’s almost impossible to read a book like John Green’s Looking for Alaska (2005), for instance, and not develop huge respect for young adult literature and those who write for this audience. Deeply philosophical and richly textured, Looking for Alaska is an intense coming-of-age story about a group of teens, a boarding school, and one boy’s search for the ‘Great Perhaps.’ First-time novelist Green depicts the intensity of adolescence seldom seen in contemporary literature. (57-58) Green has since gone on to write several more books, including The Fault in Our Stars, which was recently turned into a movie. Another novel of his, Paper Towns, is also being adapted into a film. While his novels are certainly high-quality, it’s not unreasonable to assume that the enthusiasm of his widespread online fanbase was encouraging to filmmakers considering adapting his work for the screen. Of course, Green is an outstanding case; while some authors also gain large followings on 5 other social media sites—Johnson currently has 100,000 Twitter followers, for example—few have yet to reach Green’s level of online prominence. Even so, almost every professional author also has a blog with biographical and bibliographical information, and many run blogs similar to Johnson’s. Official websites, like that of the Authors Guild, also demonstrates the online presence of authors; a list of members on the site includes the likes of Laurie Halse Anderson and Judy Blume. But these are public concerns of authors that have already been published and are looking to maintain or augment their recognition; what kind of training is required to produce and sell a book in the first place? According to Sotirakopulos, education plays a key role in a writer’s ability to produce a high-quality novel. In a personal interview, she explained: I think it’s a combination of experience in writing/education, or just the education. But [the author] Linda Sue Park is a great one who will say that to write a novel you need to read one hundred novels, to write a picture book, …[read] one thousand picture books. … So I don’t know if you need a degree, or certificate, but you certainly need to study the craft. So where in the United States can aspiring YA authors go to study the craft and get a degree or certificate along the way? Unfortunately, traditional sources on the topic of career training have little information tailored to Young Adult authors. However, by taking a sample of fifty critically acclaimed and popularly successful American Young Adult authors, one might evaluate these authors’ history of higher education.1 The results might be surprising—not only did 18% of the authors surveyed for this paper not complete any formal higher education at all, the remaining 82% showed remarkably little consistency in their schooling choices. Six attended Ivy League schools, but thirty-five did not. Four authors attended schools in the South, fifteen in the West, eighteen in the East, and eleven in the Midwest. The survey also found that seven attended colleges with less than 10,000 students, while the remaining thirty-four went to large universities ranging from 20,000 to 50,000 students. It seems, then, that there is no great regularity to the colleges these authors have attended, only that the majority—82%—completed some formal higher education, and most of them did so with a focus in an English- or Communications-related field. If only 18% of these authors succeeded without a degree or certificate, how did the vast majority’s degree or certificate contribute to their career success? According to Sotirakopulos, level of education is generally reflected in the quality of an author’s writing. She explains that authors like “John Green and his editor… Julia Strauss-Gabel… know literary work, so they can still write for teens, they still produce work for teens, but they follow a nice literary pattern… [with] a complex plot and beautiful prose.” She goes on to explain that, in her mind, the important thing is a balance between education and experience: I think there is something to be said about education, or at least recognition in the field. …I can’t say whether or not people look at [a writer’s] education and publish their [work] because of the education [they have]. Although I am editing an anthology on women’s life shifts, and when we were looking at submissions we did take into consideration …not only the quality but [also] the author’s interest in the field, submersion in the field, you know. So it wasn’t necessarily where did they go to college or what did they teach, that kind of thing, but it was more or less did they know the field, …are they attending conferences, are they active in some kind of organization. That was the focus. 1 SeeAppendix for full list and findings. 6 So while it is unlikely that an author’s education will be directly responsible for the success of their work, the indirect effect—namely, the improved writing quality due to education—will most likely contribute to career success. But how does a YA author’s level of education affect their salary? Unlike many professional careers, a high-level degree does not guarantee an increase in pay, as demonstrated by the fact that 18% of the YA authors surveyed earlier never completed any form of higher education. But while a degree or certificate may not be absolutely required to succeed as an author, knowledge of the publishing industry is—at least traditionally. The recent self-publishing craze seems to be leveling the playing field for the less-experienced writer, although an increasing number of expert voices deny that self-publishing gives novice authors any real advantage. Certainly, self-publishing requires much less knowledge of the publishing world and all its intricate relationships and maneuverings than navigating the traditional publishing house process. Unfortunately, as article after article will tell you, this ease does not always lead to success. Perhaps even more unfortunately, all the effort that goes into having your book published in the traditional manner doesn’t always pay off, either. Let’s start with traditional publishing—a process that begins with polishing your own manuscript, and never seems to end. According to author Nathan Bransford, once you have a full manuscript that you’re relatively happy with, you have to continue working on it. Polish it “so much it’s shining like the top of the Chrysler Building. Now it’s time to find an agent, right? Nuh uh. Time to learn about the publishing business.” He advises, “Before you embark on your quest for a literary agent, you should devote many, many hours to familiarizing yourself with the business, literary agents, editors, and anything else you can possibly do to discern how this unique industry operates.” This advice is echoed by Jerry Simmons of Writers Digest, who explains, “Knowing how [production, marketing, sale and distribution of your book] comes together doesn’t just increase your odds of crafting a submission that will get you a deal—it also gives you a better chance of impacting the decisions that can make or break your book’s success.” These important decisions, of course, include things like the book’s cover, which will be discussed later in this paper. Simmons adds: Writing is a craft, but publishing is a business. …Being able to intelligently discuss books and authors with professionals who make their living in publishing builds your credibility as a market-savvy author—especially if you can demonstrate an understanding of where your own book will fit in. Once you know exactly where your work belongs on the shelf, be sure that knowledge is reflected in the query or cover letter accompanying your submissions to agents or editors. It will increase your chances of getting one. A query letter is a brief letter that describes your work, which you send to prospective agents in the hopes they’ll be interested and want to take you on as a client. Bransford also recommends networking as much as possible in order to increase your chances of a referral by an established author, explaining that agents are “more likely to [give your manuscript] a thorough look” if it’s been endorsed by someone that agent respects. Considering the amount of work that crosses their desk, this is a definite advantage. But, no matter how desperate you are to have an agent look at your work, author Maureen Johnson advises wariness: “Anyone that offers a quick-fix, a sure deal, a new method of getting to editors… are lying to you. Money up front is a 100% sign that an agent is a fake. Real agents only take a percentage from something they sell.” Johnson recommends contacting the Author’s Guild if you are unsure of an agent’s credentials or a contract you’ve been offered, whether by an agent or publishing company.2 2 See Appendix for more information on the Authors Guild, including services, dues, and membership types. 7 Even once you’ve snagged a bona fide agent with your Chrysler-Building-shiny manuscript, referrals, and/or query letters, which can take months or longer, there’s quite a bit of work to be done. The purpose of having an agent, of course, is that they know the ins and outs of the publishing world; as Johnson puts it on her professional website and blog: Editors tend to specialize in one type of book. …Writers out in the world won’t know what these editors specialize in. …[Agents] know which editors are publishing which books. They will know that Editor Suzy LOVES books with a little romance, but doesn’t want to hear about your vampire. Editor Grace, who works at the same company, is gasping for a vampire. She’s got a long list of vampire books. The editor will know to send the romance to Suzy, the vampire to Grace. Once you have an agent to send your manuscript, Johnson warns there will be long periods of waiting for publishing companies to respond. If you’re lucky enough to have more than one interested publisher, your agent takes care of auctioning off your work. When your agent has negotiated a deal with a publishing company, they will come to you for the final sign-off. After that, Simmons of Writer’s Digest cautions, it’s out of your hands: “The moment that contract was signed, you lost rights and control over how your manuscript will be published.” This includes decisions like cover art, marketing, and even the book title. Simmons recommends meeting with the staff in charge of putting your book out into the world and developing relationships with them to increase the chances that your suggestions will at least be taken under consideration. He insists that maintaining these relationships are key to ensuring a final product everyone is more or less happy with: “[E]very smart writer knows his editor is also his main contact (and advocate) for all things relating to the publication of his manuscript. Conduct yourself accordingly.” So, when does the money start rolling in? Well, not until after your book makes a profit. The publishing company first has to make back what it spent on your advance—the amount the publishing company paid for the rights to your work. Advances can range “anywhere from ten grand, which is very low-end, to six figures,” according to Sotirakopulos. Once the book has begun making a profit, authors can begin receiving their royalties, or percentage of earnings from book sales. According to Johnson, this amount is usually between 8% and 15%. Of course, this only happens if your books make it off the shelves and into the hands of readers. Which brings us to one of the great lures of self-publishing: it can be very, very cheap, and very, very lucrative—if the author hits mainstream fame. According to Leslie Kaufman’s article in The New York Times, “New Publisher Authors Trust: Themselves”: While self-published authors get no advance, they typically receive 70 percent of sales. A standard contract with a traditional house gives an author an advance, and only pays royalties—the standard is 25 percent of digital sales and 7 to 12 percent of the list price for bound books—after the advance is earned back in sales. However, according to Sarah Fay’s article, “After ‘Fifty Shades of Grey,’ What’s Next for Self-Publishing?” in The Atlantic, “The success rate [of self-published books] is less than one percent.” In a personal interview with recently self-published fantasy author and fellow College Of DuPage student Nathan Kiehn, he explained to me: 8 I basically wrote it with no intention of publishing it, and then my dad mentioned, “Hey, why don’t you, you know, try to publish it?” And I thought, “OK.” So my mom had read it over, and then I read it over. We made a couple changes, mostly notes, to catch spelling and grammatical things, and that was pretty much all the editing we did. He added that they did not come up with cover art or spend any money to have a company format it for him, and did not spend money to print any hard-copies; instead, he simply formatted it as an HTML document and uploaded it to Amazon KDP. How much did it cost him to publish? I wanted to know. “Nothing.” And how many copies had he sold? Nine or ten, for $2.99 each—technically a profit, though certainly not on the scale of the e-published Twilight fanfiction that became an international best-seller. (As of April 18th, 2013, Fifty Shades of Grey had already sold over 70 million copies worldwide, according to the Huffington Post.) Certainly Kiehn’s experience was not as complicated as the traditional publishing route described earlier: no agents or publishing houses are required, and the author retains all rights to their work.3 Perhaps he hadn’t sold as many copies as E.L. James’s hit, but he’d still come out in the black. Naturally, there are also some downsides to self-publishing your work. For one, the selfpublishing field is so new that the market it still flooded with low-quality, formulaic genre books that would normally never leave a casual writer’s attic or closet floor or similarly disused section of their computer’s hard drive. So, if you hope to catch the attention of a big publishing house this way, your chances are low. Even books of decent quality are not always successful in the Wild West of selfpublishing, as it “now accounts for more than 235,000 books annually, according to Bowker” (Kaufman). Of course, there are a few tales of moderate success amidst the disappointments and the dizzying glories of self-published authors; Sotirakopulos explained, for example, that her colleague Ophelia Julien “does very well” with her self-published work. However, Julien has also been writing for about twenty years, and has had previous experience and success in the publishing world. It seems, then, that experience may be a key factor in the success of self-publishing, barring phenomena such as Fifty Shades of Grey, especially as more established authors begin exploring selfpublishing options. Such established authors include the likes of Pulitzer Prize-winning playwright and author David Mamet; according to Kaufman’s article in The New York Times, Mamet was attracted in part by the idea that self-publishing allows “well-known figures like Mr. Mamet to look after their own publicity,” as “traditional publishers have cut back on marketing” in recent years. So what’s Sotirakopulos’s take on self-publishing—was it something she had ever considered? In my interview with her, she explained: I did not consider that, only because I work with friends [who] have a press, [and] they kind of grandfathered me into it. But my situation is quite rare. … I know people who do quite well with [self-publishing]. [A] couple [of my colleagues] ended up self-publishing—…They opened a press and did it. So I think if you have a press name, you’re fine. But I think it’s the people who go off on their own and go do it [who might get in trouble]. …I recommend it if the book is edited, and clean, and a good representation. The publishing market right now, I don’t know, it’s tough. Although traditional publishing may include more paperwork and rejection letters, and does not guarantee success in the end, its success rate is certainly higher than self-publishing’s one percent. As of now, self-publishing is still a virtual free-for-all with no guarantees beyond ease of 3 For more information on copyrights, see Appendix. 9 production, should you take the simplest route. And, be warned—the cost increases substantially if you bring in professional editors, printers, formatters, packagers, and so on. Even choosing to publish electronically, while definitely the cheaper option, may negatively impact your success. According to Voices of Youth Advocates, or VOYA, a publication targeted at Young Adult librarians, the vast majority of teens still prefer reading physical copies of books. In her February 2014 article, “I Am Not a Robot: Teens and Technology,” librarian Geri Diorio reports that her teen advisory group’s preferred reading platform is still print: “Only one TAG [teen advisory group] member preferred ebooks [to print], but not by much: ‘…[I prefer] ebooks because of convenience, but it’s 60 to 40 (ebooks to paper)’” (22-23). How else might the self-publishing and e-publishing revolution affect the industry? Regina Sierra Carter, Ph.D., suggests that customizable cover-art could significantly alter a reader’s experience with literature. Her paper, “YA Literature: The Inside and Cover Story,” published in YALSA’s official publication The Journal of Research on Libraries and Young Adults, “acknowledges a need for cover art that reflects the ‘changing faces of today’s teen’ and offers suggestions for how this may be realized on a small scale.” She explains: As of now, there is little that lovers of YA literature can do about cover art that grossly misses the mark, which is unfortunate. However, in the future, readers may have more of a say in how cover art is constructed. When that time comes, there may be a move toward a more creative and reader-oriented approach to producing cover art. For example, in the future, cover art may be personally selected via pre-order services or created using kiosks. For, as much as most people hate to admit it, a book’s cover largely informs potential readers whether or not a particular book is something in which they’d like to invest their time. Thus, the book’s cover has a large amount of power in determining what will be read and what won’t, the expectations of its readers, and sometimes the mental image the reader has of the main character before reading. A very serious concern, therefore, is the “whitewashing” of book covers featuring people of color as their protagonists—something that could very well be addressed by the innovative methods of cover art development Carter suggests in “YA Literature: The Inside and Cover Story.” “Whitewashing,” or misrepresenting the race of the book’s main character on the cover of the story—usually a person of color incorrectly portrayed as White, hence the term “whitewashing”—is an enduring controversy. (“People of color” is a term that includes anyone who identifies as Black, Asian, Indian, Native American, multiracial, and so on.) Even before the covers can be whitewashed, of course, the protagonist must be a person of color. As Carter notes, “Racial minorities have not traditionally experienced widespread coverage in YA literature [in the first place]”. With such limited representation already, whitewashed covers are particularly concerning. Carter expresses this concern in her paper, quoting from the Cooperative Children’s Book Center’s publication, Choices: For quite a few years we have been commenting on how few books by and about people of color are published in the United States in relation to the overall number of books produced annually. As the population of the United States continues to not only increase but become more diverse, the output of publishing houses has not been a mirror of society, at least in terms of the numbers. Carter and the Cooperative Children’s Book Center (CCBC) insist that “‘[i]n our ever-more-diverse nation, we need books that provide all children the opportunity to see themselves and the world in which they live reflected.’” This sentiment is echoed by Young Adult librarian Robin Brenner in her interview with Lisa Weidenfeld of Metro. She says, “[Y]ou want books to be either mirrors or 10 windows. Mirrors should show you yourself in some way, and windows should show you other people so that you can see what it’s like to not be you, and that’s where the empathy starts to grow in terms of reading fiction.” Unfortunately, there is still a distinct lack of minority voices in YA—not only in terms of characters, but also writers, leaving much to be desired in the way of “mirrors and windows.” In a recent article for BuzzFeed, “Diversity Is Not Enough: Race, Power, Publishing,” author Daniel José Older describes his struggle to be heard as an author of color. In a majority-White industry, he was confronted with agents “that said they loved [his] writing but didn’t connect with the character, … didn’t think [his] book would be marketable even though it was already accepted at a major publishing house, … [and] ones that wanted [him] to delete moments when a character of color gets mean loks from white people.” He even encountered a “white magazine editor who lectured [him] on how [he’d] gotten [his] own culture wrong.” And his experience is not unique. Older explains, “My friends all have the same stories of whitewashed covers and constant sparring with the many micro and mega-aggressions of the publishing industry.” Older goes on to quote publishing industry expert Sarah McCarry, who explains that even if a non-traditional book is published, the publishing house’s advertising “money and attention overwhelmingly goes to what the industry has already decided is ‘marketable’—heterosexual narratives featuring white characters [instead of those books featuring people of color or LGBTQ+ members].” Indeed, issues in representation are not limited to young people of color; perhaps more pressing, suggests Talya Sokoll, is the dearth of transgender youth representation in Young Adult literature.4 She laments that “[w]hile in the past four years publications of these [trans*-oriented] titles has slightly improved, …the number of books being published is still very low, and some of these titles are published by small press, with little or no marketing, which makes them difficult for library staff, or teens, to locate” (Sokoll 23). She emphasizes that “[t]he lack of titles does not mean there is a lack of need for these books,” citing findings that “‘Transgender students experienced more hostile school climates than their non-transgender peers—80 percent of transgender students reported feeling unsafe at school because of their gender expression’” (“2011 National School Climate Survey” qtd. in Sokoll 24). She also referred to Parrotfish author Ellen Wittlinger’s goal to “normalize homosexuality and transexuality and make gender and sexual orientation just two of the many ways in which we are different from each other. She hoped that her book would not only appeal to adolescents who saw themselves in the characters, but to straight adolescents who did not think they knew any gay teens’” (24). Unfortunately, some authors’ attempts to include trans*, genderqueer, multiracial, or other minority group members may backfire, as these characters are too often reduced to stereotypical token characters, and/or over-exoticized. So, how does one fairly present characters who are also members of one or more minority groups? According to Sotirakopulos, author of trans*-oriented YA novel In Her Skin: I’ve read books where people have just put [in] a… token character… and it does a disservice, it’s an insult, I think to that group of people [they’re trying to write about]. So, I would say, base it on reality. …If you, or another writer, is interested in writing about a specific group of people, submerge yourself and be that person for a while. 4 Transgender is one identity included in the trans* umbrella; other identities represented by the trans* prefix include transsexual, transvestite, genderfluid, bi-gender, agender, and non-binary, all of which are severely underrepresented in YA literature. 11 She explains that her own experiences with a close friend since childhood, who went through the gender-transitioning process, is the only reason she felt comfortable writing about the topic of a transgender character herself, and agrees that “there isn’t much fiction [for trans*-youth], and it’s not fair.” As for over-exoticization, it can occur both within a text itself and, of course, on the cover. As author Mitali Perkins notes in her article for the School Library Journal, “Straight Talk on Race”: Sometimes books may be packaged with covers depicting a character as more foreign than he or she is described in the story. Nowhere in Cynthia Kadohata's novel Weedflower (S & S/Atheneum, 2006), for example, does 12-year-old Sumiko wear a kimono. But she does on the cover. For a story about the Japanese internment in North America during World War II, why did the powers-that-be make an American protagonist appear more culturally Japanese than American, especially when a girl in jeans behind barbed wire would have been more historically accurate? Perhaps they were trying to tap into a fascination with all things foreign, amping up the exotic factor so that those looking for a book about "faraway cultures" might buy this one. She goes on to explain that “[o]verexoticizing a nonwhite character to appeal to white readers can happen inside a story as well as on a cover,” and uses her own book as example. Despite her own Bengali heritage and awareness of the overexoticization issue, she had, almost unconsciously, inflicted it upon her own main character. She explains that she didn’t even notice what she’d done until a reviewer pointed out the “‘unnecessary exoticization of [her] progatonist,” particularly the description of the character looking like “some exotic Indian princess” at the end of the novel: “I fumed, but, dang it, the reviewer was right. Exotic Indian princess? What was I thinking? Enduring a twinge of shame, I moved on and tried to learn from my mistake.” Unfortunately, misrepresentative covers don’t begin and end with whitewashing and overexoticization; sexism is still a very real issue faced by female authors and their readers. Because a book cover largely informs the reader on the apparent worth and subject of a book before reading, even issues seemingly as insignificant as cover art that has little to no connection to the plot can leave a reader unprepared for the story they’ve just picked up. For example, librarian Darcy Lohmiller picked up Chris Crutcher’s The Crazy Horse Electric Game “because on its cover were three kids celebrating apparent victory on a baseball field, and [she] needed a sport book in [her] repertoire,” only to find that “this was a scene from the first chapter,” after which “[t]he hero… gets into a water-skiing accident and is partially paralyzed, his parents’ marriage crumbles over their guilt and grief, and he is quickly alienated from his friends, family, and past life” (13). Not what she was expecting, and not what your average reader would predict from the cover. While this is a relatively neutral example, there is an epidemic of misrepresentative covers for fiction written by women; due to sexist marketing, the perceived value of books by women is often very low. That is something people like Maureen Johnson want to change. Several months ago, Johnson, who often Tweets about her experiences in the writing and publishing industries, made an off-hand comment about how vexing it is to receive messages from prospective readers who find her covers too embarrassingly “girly” to actually read. She came up with the idea of a “coverflip”—reimagining the cover of a famous book as if its author’s gender were “flipped,” both to expose the excessive gendering of book covers and to get a good laugh. When followers’ responses to her idea of a “cover-flip” took off, she even partnered up with Huffington Post to run a short competition, putting the top cover-flipped entries of famous and not-so-famous books on the Huffington Post’s website. Johnson argued that pointlessly gendered book covers are so common that most people do not even realize they are inferring the quality of a book based on its cover. Unfortunately, the general 12 consensus is that “chick lit”—books by women, perceived to be “for” women—is of an inferior quality. Emily Matchar’s article in The Atlantic, “Chick Lit Remixed: The Simple Brilliance of Gender-Flipping,” discusses the general trend of people using gender-flipping to expose the absurdity of many media’s hyper-feminine and/or hyper-masculine treatment of certain subjects. She then moves specifically onto cover-flipping: “Wherein ‘guy’ books are given the ‘women’s fiction’ treatment, with pastel colors and misty photos of pensive-looking girls, and girly books get bro-ed up with graphic black-and-white. … Highlights [include] John le Carré’s Cold War spy classic Tinker, Tailor, Soldier, Spy done up in pink and baby blue with four model-like men lounging provocatively on the jacket, [and] Freedom by ‘Jane Franzen’ featuring a manic pixie dream girl raising her arms to the wind, Titanic-style.” Matchar goes on to suggest that “if more men read ‘women’s fiction’ it would cease to be ‘women’s fiction’ at all, and simply become what male-written novels have long been called: ‘fiction.’” It wasn’t the first time Johnson called out the stigma attached to women writers and so-called “chick-lit”; she also started the hashtag trend “#IsItChickLit,” in which Twitter users summarized a great male-authored literary work as if it were being promoted as “chick lit.” A few outstanding examples include The Odyssey—“When the hubby’s away, will Penelope play? She has to learn to rely on herself—and wait for true love,” from user @turning_Beth; a summary of Hamlet from @elmify: “One big castle, one Danish invasion, and one ZANY boyfriend—what’s a girl to do?”; and even The Great Gatsby, from @maureenjohnson herself: “Sure the parities [sic] are amazing and the clothes are fabulous, but how will Daisy’s torrid love affair with Gatsby end?” As anyone who has taken high school English can tell you, that is certainly not the focus of F. Scott Fitzgerald’s Jazz Age novel—which is exactly Johnson’s point. As she writes in her Huffington Post article on the subject, “The Gender Coverup”: [T]he simple fact of the matter is, if you are a female author, you are much more likely to get the package that suggests the book is of a lower perceived quality. Because it’s ‘girly,’ which is somehow inherently different and easier on the palate. A man and a woman can write books about the same subject matter, at the same level of quality, and that woman is simple [sic] more likely to get the soft-sell cover with the warm glow and the feeling of smooth jazz blowing off of it. Frustratingly, as Johnson notes on her blog, authors seldom have control over their publisher’s marketing decisions, up to and including the book cover. So, while it is somewhat discouraging to acknowledge the biases people have regarding female writers, it may be bolstering to know that there are those like Johnson and the countless #IsItChickLit and CoverFlip participants actively trying to change such gendered stereotypes. As Johnson notes in her article on The Huffington Post’s website, “This idea that there are ‘girl books’ and ‘boy books’ and ‘chick lit’ and ‘whatever is the guy equivalent of chick lit’ gives credit to absolutely no one, especially not the boys who will happily read stories by women, about women.” In a literary world inundated by Manic Pixie Dream Girls and damsels in distress, it is important that books by women about women are taken seriously, especially if they provide a more realistic view of the world as it is, with full and flawed female characters. Obviously, representation is an issue that starts at the cover and continues to the core of much of today’s Young Adult literature. Covers misrepresenting the content of their books, whether that misrepresentation is as serious as whitewashing and sexist marketing, or as seemingly mundane as illustrating an insignificant scene from the story, are rampant. Thankfully, professionals in the field are bringing these issues to public attention—whether it be through articles, “coverflipping”, or hashtag trends. As seen in public responses to the social media movements instigated by Maureen Johnson, readers and writers agree that trends in misrepresentative cover art must be corrected. There 13 is a dire need for more realistic representation of minority youth in young adult literature—not merely token characters, or white-washed covers, but true-to-life “mirrors and windows” that allow young adults to see themselves and their peers truthfully portrayed, as Young Adult librarian Robin Brenner suggests in her interview with Lisa Weidenfeld of Metro. After all, adolescence is an important and difficult time for many, and YA literature often provides adolescents with a method of understanding and navigating the world in which we live. Now that you know the basic process of becoming a published YA author, as well as some of the current controversies in the field, you may be questioning your decision to write Young Adult literature. After all, there is a lot of hard work and risk involved in this profession. Not only must you take into account your impressionable audience and strive to produce a quality piece of literature that fairly represents our world, you must learn the ins and outs of publishing to first become published, and then try to make your published work a financial success. Is it worth the struggle? That is something you have to answer for yourself, bearing in mind that Young Adult literature has an important social and cultural impact, especially because it speaks to a rapidly maturing audience. As Michael Cart explains in his YALSA whitepaper “The Value of Young Adult Fiction”: Whether one defines young adult literature narrowly or broadly, much of its value cannot be quantified but is to be found in how it addresses the needs of its readers. …[T]hese needs recognize that young adults are beings in evolution, in search of self and identity; beings who are constantly growing and changing, morphing from the condition of childhood to that of adulthood. That period of passage [is]… distinguished by unique needs that are—at minimum—physical, intellectual, emotional, and societal in nature. By addressing these needs, young adult literature is made valuable not only by its artistry but also by its relevance to the lives of its readers. YA authors have a responsibility provide a realistic view of the world we inhabit to these “growing and changing” readers. By arming young adults with an understanding of the world they are coming into, both good and bad, Young Adult authors are arming their readers with an ability to shape and change that world. Whether you choose to write Young Adult literature or simply carry on as a consumer, you play an important role in shaping this evolving genre; as YA authors continue writing for and about adolescents, addressing any number of coming-of-age issues, readers continue to show their support for this literary form and its perpetual progress. Perhaps publishers will take note of consumer reactions to their products, particularly in regards to gendered and/or misrepresentative cover art, and adjust their packaging and marketing accordingly. Or perhaps the future of YA literature lies in self-published authors, or will turn entirely to e-books; perhaps changes in platforms will also contribute to clearer and more accurate cover art, and provide opportunities for voices that might have otherwise been stifled to provide new and innovative perspectives. However Young Adult literature evolves, it is an important, useful, and enjoyable resource for young adults as they struggle to understand themselves and the world around them during a period of intense life changes. 14 Appendix 1. Survey of fifty successful YA writers’ histories of higher education: Rainbow Rowell—unreported. Possibly no formal higher education John Green—Kenyon College Maureen Johnson—University of Delaware, Columbia University Suzanne Collins—Indiana University, New York University Rick Riordan—University of Texas (Austin) Judy Blume—New York University Beverly Cleary—Chaffey College, UC Berkeley, University of Washington S.E. Hinton—University of Tulsa Veronica Roth—Carleton College, Northwestern University Libba Bray—University of Texas (Austin) Lois Lowry—Brown University (incomplete), University of Southern Maine Amanda Hocking—self-published, no formal higher education Meg Cabot—Indiana University (Bloomington) Sherman Alexie—Gonzaga University (incomplete), Washington State U Carl Hiaasen—Emory University (incomplete), University of Florida Stephanie Perkins—unreported. Possibly no formal higher education Ransom Riggs—Kenyon College, University of Southern California Maggie Stiefvater—University of Mary Washington Cassandra Clare—unreported. Possibly no formal higher education Holly Black—College of New Jersey, Rutgers University Marissa Meyer—Stanford University David Levithan—Brown University Scott Westerfield—Vassar College Kami Garcia—has an MA in education, school unspecified Laini Taylor—UC Berkeley Richelle Mead—University of Washington, University of Michigan, Western Michigan University Kiera Cass—Radford University Margaret Stohl—Amherst College, Stanford University Ally Carter—Oklahoma State University, Cornell University Marie Lu—University of Southern California Rick Yancey—Roosevelt University Sarah Dessen—University of North Carolina (Chapel Hill), Greensboro College Neal Shusterman—University of California (Irvine) James Dashner—Brigham Young University Michael Grant—unreported. Possibly no formal higher education. Jodi Picoult—Princeton University, Harvard University Simone Elkeles—Purdue University, Loyola University (Chicago), University of Illinois (Champaign-Urbana) Laurie Halse Anderson—Onondaga Community College, Georgetown University Ellen Hopkins—University of California (Santa Barbara) Gayle Forman—unreported. Possibly no formal higher education. Lauren Myracle—University of North Carolina (Chapel Hill), Colorado State University Rachel Cohn—Barnard College 15 Gordon Korman—New York University Margaret Peterson Haddix—Miami University (Ohio) Patrick Carman—Willamette University Linda Sue Park—Stanford University Ann Brashares—Columbia University, Barnard College Lauren Kate—University of California (Davis) Michele Jaffe—Harvard University Patrick Ness—University of Southern California 2. The Author’s Guild is a not-for-profit organization that represents book authors (including coauthors, ghostwriters, translators, and self-published authors) and freelance writers. Dues are typically $90 per year for all three categories of membership (Regular, Associate, or Member-AtLarge), although after the first year the dues are technically on a sliding scale. The following services are the benefits of membership, according to the organization’s official website: Members of the Authors Guild receive free reviews of U.S. book contracts from experienced legal staff, low-cost website services including website-building, e-mail, and domain name registration, access to [the Guild’s] free Back in Print service, [their] quarterly print Bulletin, and invitations to panels and programs throughout the year. Regular membership is available to those who qualify as “an author published by an established U.S. book publisher, a freelance writer published by periodicals of general circulation in the U.S., or as a book author or freelance writer earning writing income” (Authors Guild). The Guild specifically includes self-published authors in the category of freelance writers. A writer may qualify as an Associate Member if they have “been offered a contract with an American book publisher, …an established literary agency has offered to represent [that writer], or if [the writer] has earned… at least $500 in writing income… in the 18 months prior to applying for membership” (Authors Guild). The Guild grants Associate Members the same rights and privileges as Regular Members, with the exception of voting in Guild elections, according to the Guild’s official website. They publish a quarterly bulletin that covers “publishing, copyright, tax, legal and legislative news” (Authors Guild). 3. According to the U.S. Copyright Office, a copyright is “a form of protection grounded in the U.S. Constitution and granted by law for original works of authorship fixed in a tangible medium of expression [and] covers both published and unpublished works.” There is a $35 fee to register your copyright online, according to their page on Registering Your Work, although the Copyright in General FAQ explains, “Your work is under copyright protection the moment it is created and fixed in a tangible form that it is perceptible either directly or with the aid of a machine or device.” However, you must register your work “if you wish to bring a lawsuit for infringement of a U.S. work,” according to the official website of the U.S. Copyright Office. They add, “Registered works may be eligible for statutory damages and attorney's fees in successful litigation. …[I]f registration occurs within 5 years of publication, it is considered prima facie evidence in a court of law.” 16 Works Cited American Library Association. The American Library Association, 2014. Web. 22 Apr. 2014. Authors Guild. The Authors Guild, 22 Apr. 2014. Web. 22 Apr. 2014. Bransford, Nathan. “How to Find a Literary Agent.” Blog.NathanBransford.com. Nathan Bransford. 10 Apr. 2008. Web. 8 Mar. 2014. ---. “How to Write a Query Letter.” Blog.NathanBransford.com. Nathan Bransford. 19 Aug. 2010. Web. 8 Mar. 2014. Brown, Joanne, and Nancy St. Clair. Declarations of Independence: Empowered Girls in Young Adult Literature, 1990-2001. Lanham: Scarecrow, 2002. Print. Cart, Michael. “The Value of Young Adult Literature.” ALA.org. Young Adult Library Services Association, Jan. 2008. Web. 1 Mar. 2014. Carter, Regina Sierra. “YA Literature: The Inside and Cover Story.” YALSA.ALA.org. Young Adult Library Services Association. 18 Apr. 2013. Web. 8 Mar. 2014. Cole, Pam B. Young Adult Literature in the 21st Century. Boston: McGraw-Hill Higher Education, 2009. HigherEd-McGrawHill.com. Web. 23 Feb. 2014. Copyright in General (FAQ). US Copyright Office, 12 Jul. 2006. Web. 22 Apr. 2014. Diorio, Geri. “I Am Not a Robot: Teens and Technology.” Voice of Youth Advocates. Feb. 2014: 2223. Print. Elmify. “@maureenjohnson: One big castle, one Danish invasion, and one ZANY boyfriend—what’s a girl to do? HAMLET #isitchicklit.” 5 Feb. 2013, 6:06 a.m. Tweet. Falls, Kat, Bob Sirott, and Marianne Murciano. Kat Falls and the Young Adult Book Market. WGN Radio – 720 AM, 2014. WGNradio.com. Web. 23 Feb. 2014. Fay, Sarah. “After ‘Fifty Shades of Grey,’ What’s Next for Self-Publishing?”. TheAtlantic.com. The Atlantic. 2 Apr. 2012. Web. 5 Mar. 2014. Fisher, Sarah. “The Rise and Rise of Young Adult Literature.” Magazine.Pepperdine.edu. Pepperdine Magazine, Spring 2013. 23 Feb. 2014. Johnson, Maureen. “The Gender Coverup.” HuffingtonPost.com. The Huffington Post. 7 May 2013. Web. 24 Feb. 2014. ---. “I Am A Very Dangerous Person.” MaureenJohnsonbooks.com. Maureen Johnson, 27 Apr. 2007. Web. 21 Apr. 2014. ---. “News Flash: Victory in Bartlesville (Sorta, Kinda).” MaureenJohnsonbooks.com. Maureen Johnson, 20 Jun. 2007. Web. 21 Apr. 2014. ---. “How NOT To Get Published.” MaureenJohnsonbooks.com. Maureen Johnson, 25 May 2006. Web. 8 Mar. 2014. ---. “Because You Asked: Part Two of MJ’s How to Get Published Series.” MaureenJohnsonbooks.com. Maureen Johnson, 24 Jun. 2006. 8 Mar. 2014. ---. “GATSBY: Sure, the parities [sic] are amazing and the clothes are fabulous, but how will Daisy’s torrid love affair with Gatsby end? #isitchicklit.” 5 Feb. 2013, 5:33 a.m. Tweet. Kaufman, Leslie. “New Publisher Authors Trust: Themselves.” TheNewYorkTimes.com. The New York Times. 16 Apr. 2013. Web. 5 Mar. 2014. Lohmiller, Darcy. “Go Ahead: Judge a Book By Its Cover.” Young Adult Library Services 6.3 (2008): 13-15. Academic Search Complete. Web. 3 Mar. 2014. Matchar, Emily. “Chick Lit Remixed: The Simple Brilliance of Gender-Flipping.” TheAtlantic.com. The Atlantic, 7 Aug. 2013. Web. 16 Feb. 2014. 17 Older, Daniel José. “Diversity Is Not Enough: Race, Power, Publishing.” BuzzFeed.com. BuzzFeed, [publication date]. Web. 21 Apr. 2014. Perkins, Mitali. “Straight Talk on Race.” School Library Journal 55.4 (2009): 28-32. Academic Search Complete. Web. 3 Mar. 2014. Registering a Work (FAQ). U.S. Copyright Office, 28 Oct. 2013. Web. 22 Apr. 2014. Sales, Bethany. “Fifty Shades of Grey: The New Publishing Paradigm.” HuffingtonPost.com. The Huffington Post, 18 Apr. 2013. Web. 9 Mar. 2014. Simmons, Jerry D. “Publishing 101: What You Need to Know.” WritersDigest.com. Writer’s Digest, 22 Sep. 2009. Web. 5 Mar. 2014. Sokoll, Talya. “Representations of Trans* Youth in Young Adult Literature: A Report and a Suggestion.” Young Adult Library Services 11.3 (2013: 23-26. Academic Search Complete. Web. 8 Mar. 2014. Sotirakopulos, Trina. Personal interview. 28 Feb. 2014. Strickland, Ashley. “A Brief History of Young Adult Literature.” CNN.com. Central News Network. 17 Oct. 2013. Web. 16 Mar. 2014. Turner, Beth (turning_Beth). “When the hubby’s away, will Penelope play? She has to learn to rely on herself—and to wait for true love. THE ODYSSEY. #isitchicklit.” 5 Feb. 2013, 6:50 a.m. Tweet. Weidenfeld, Lisa. “Talking YA Fiction with a Teen Librarian: It’s Not Just for Young Adults!” Metro.us. Metro News, 20 Oct. 2013. Web. 23 Feb. 2014. 18 Kill to Conserve: Ethical Implications of Trophy Hunting Conservation Measures By Nicholas Bashqawi (Honors Philosophy 1110) O n the eleventh of January 2014, a permit to kill a black rhinoceros in Namibia was auctioned off for $350,000. The Dallas Safari Club stated that the purpose of the auction was to raise funds for the conservation of the black rhinoceros, as well as other threatened and endangered species.3 Corey Knowlton, the winner of the auction, will be able to hunt a rhino of his choosing and has made it clear that he will target an individual that is harming other rhinos.12 Conservationists across the United States decried the auction from the moment it was announced, claiming that hunting an endangered species in the name of preservation is a ridiculous plan. Despite the various underlying reasons for these protests, the controversial nature of this issue brings about an opportunity to further examine the trophy hunting strategy of species conservation and maintenance. For example, one can examine the ethical dilemma brought about by the simple, albeit critical, question: “What about the individual animal being hunted?” Specifically, can the death and suffering of an individual be justified by the possibility of a healthier future for its species as a whole? Fundamental to any side of this issue is the determination of whether or not trophy hunting is an effective means of conservation. Statements from executives at the Dallas Safari Club, as well as from the media, sound well-informed and cast the hunt in a relatively positive light. At the same time, those conservationists who oppose the hunt also seem to have a valid argument when they talk of the detrimental effects of hunting. For instance, Ben Carter, the Dallas Safari Club Executive Director, states: “These bulls no longer contribute to the growth of the population… In many cases, they will kill younger, non-breeding bulls.”9 When you couple the possible prevention of such damages with the amount of money hunting auctions can raise, it may seem as though there is no downside to this method of conservation. As I will lay out later in this argument, though, there is, indeed, a strong basis for opposing trophy hunting; but first, I will examine if there is any empirical foundation for the claims made above. In their paper, “Trophy Hunting of Black Rhino Diceros bicornis: Proposals to Ensure Its Future Sustainability,” N. Leader-Williams et al. describe the possible benefits of allowing a small number of black rhinos to be hunted in Namibia and South Africa. According to their study, one of the reasons that the white rhinos of South Africa continue to increase in number is “limited and sustainable use, through trophy hunting and live sales.”8 They assert that controlled hunts, sanctioned by governments after rigorous evaluations, will assist in managing the surplus of male rhinoceroses and will aid in raising money for furthering conservation efforts. Furthermore, such events would act as incentives for landowners to manage wildlife on their property. The evidence reveals a positive correlation between trophy hunting and stable population growth. White and black rhino populations have both seen population growth in the past few decades, but white rhinos are no longer classified as a threatened species8, whereas black rhinos are still critically endangered.4 So far I have discussed the claims in favor of trophy hunting, and have examined the empirical foundations of those assertions. Examining the case in this manner is necessary to prevent the use of empirical data as a criticism of my argument. I thus make clear my recognition of the scientific foundation and possible consequences of hunting permit sales, and assert that they cannot justify the death of an individual animal; but why not? Surely, it would seem, if there is a method of conservation that is beneficial to a species, which does not sacrifice anything morally significant, it 19 ought to be allowed in order to prevent the greater damage of species extinction. Even this basic argument, though, leads to another, more fundamental question: why should we, as humans, be concerned about conserving other species? One answer to the question of why species conservation matters can be found in the argument provided by Greg Bognar, in his essay, “Respect for Nature.” Bognar’s analysis of human attitudes towards nature, albeit short, provides valuable insight as to why we should all care about the other living beings of our world. He writes: “We should respect other living beings, not for our own sake or their intrinsic worth but for that sake of our interconnectedness in ecological systems… living beings are part of ecosystems, and they have value… from the relations they have in those ecosystems.”2 All living things, including animals, have a specific place in their ecosystem, contributing to the overall success of other species. Eliminating any one species could have far reaching effects, possibly even affecting humans.13 Although Bognar rightfully points out the importance of all living beings, his essay still revolves around human interests. Additionally, focusing on the interconnectedness of animals and humans places a worth on animals only if they seem useful to their ecosystem. It may seem an unlikely scenario, but a consequence of worth based on productiveness in an ecosystem is that it is not particularly wrong to kill off animals that may, at the time, seem harmful to an ecosystem or area. In order to truly determine why humans should care about animals and nature, one must be able to assign, to every individual of a species, some non-arbitrary characteristic that is universal to humans and animals. I include humans in my definition because if there is some characteristic that is general enough to apply to all of the various animals, as living beings, it follows that it should be able to be applied to humans, as well. Ethicist Peter Singer describes exactly such a principle in his book, “Animal Liberation,” in which he argues against speciesism based on an individual’s ability to suffer: “If a being suffers, there can be no moral justification for refusing to take that suffering into consideration. No matter what the nature of the being, the principle of equality requires that its suffering be counted equally with the like suffering- in so far as rough comparisons can be made- of any other being.”10 Therefore, it is not a being’s intelligence, species, or relative usefulness that make it worth caring about and protecting, as these are arbitrary and subjective qualifiers which could easily be shifted to anyone’s liking; rather, a being’s ability to suffer gives it equal standing, in regards to the interest to live and resist suffering, with humans and other animals. Every individual has an interest in living a good life, in which suffering is reduced, and that interest cannot be ignored by morally responsible beings. We now have the tools to argue for the importance of species conservation, and as a result the issue of trophy hunting can be analyzed at a deeper level. If a being has some interest in resisting suffering, it is the duty of a morally responsible individual, viz. a human, to take that interest into account when making decisions. Furthermore, all beings are interconnected in ecosystems, which themselves are interconnected, meaning that actions have far reaching consequences and that all beings play a significant role in nature. Since human activity, such as hunting, pollution, and deforestation, has had damaging effects on many animal populations, including the African rhinos15, resulting in suffering, it is our moral obligation, as humans, to work to alleviate that suffering. Returning to the earlier assertion that any conservation method, which does not sacrifice anything morally significant, ought to be pursued in order to prevent species extinction, one can see that it is a valid statement. One must recognize, though, the qualifying statement, borrowed in part from Singer’s “Famine, Affluence, and Morality,”11 that nothing of moral significance can be sacrificed in the process of conservation. Singer takes great care to highlight the importance of fulfilling one’s moral obligations with moral actions, so as to prevent others from taking his argument to the absurd conclusion that one may pursue any course of action, as long as it is beneficial to some person or group. Acknowledging that one ought not to use immoral means to achieve a goal builds a solid foundation for conservation arguments. As a result, we can eliminate generally unacceptable 20 situations, such as killing factory owners and workers for polluting. Keeping in mind that nothing of moral significance be sacrificed in the name of conservation, and that all beings have an interest to live and avoid suffering, it becomes clear that the basic argument suggested for supporting trophy hunting, that any conservation method be used if it can prevent species extinction, is actually an argument against hunting. If one wishes to help recover the rhino populations, as they ought to, due to the suffering inflicted by human activities, they by all means should pursue any methods that can help produce sustainable population growth, as long as they do not sacrifice anything morally significant in the process. While trophy hunting has been positively correlated with population recovery, it is not a viable conservation option as it does not take into account the suffering of the animal that is to be hunted. Indeed, if we are to accept that animals feel pain, suffer, and have some interest in living, it would be ridiculous to disregard those items in the name of helping the species. One cannot claim to have the best interest of a species in mind if they cannot recognize the most basic interest in living and not suffering. I foresee two major criticism of my assertions, both of which I will address at present. One criticism that may be posed is that the animal will be killed in such a way that will minimize its suffering, and that the targeted animal may be one which is harming other rhinos. In response to the assertion that the animal will be killed in a humane way, I argue that it can never be guaranteed that a shot will not miss its mark, or that a certain method, which we hold to be humane, will actually minimize the animals suffering. Even if such a claim could be made, and it was found that the animal would not suffer, one still cannot justify ending the life of a being who has some interest in not dying. That the rhino may be harming others is a more interesting criticism because it implies that a greater number of rhinos will suffer if the life of one is not taken. As a primarily utilitarian argument, though, it fails to take into account the many subtleties of the issue. If we take as fact that the selected rhino is a danger to other members of its species, and that it causes suffering beyond itself, we must step back and analyze why this behavior seems to occur in rhino populations. As a start, it obviously cannot be asserted that the rhino is evil or misguided, since animals do not have a moral capacity and follow an instinctual, natural lifestyle. It can therefore be safely said that violent male rhinos act either from instinct, or due to some disease. Although it is nearly impossible, without further scientific investigation, to determine which is to blame, it seems more likely, due to the frequency of violent behavior, that instinct drives certain male rhinos to harm others of their species. One may still wish to say that even if the behavior is instinctual, it is causing multiple individuals to suffer and must be stopped. If that were the case, then one also ought to kill all predatory animals and certain insects, which eat their mates and others of their kind. I think most people would agree that killing off all predatory species and insects would be a highly disturbing course of action, and so it does not make sense to kill on the basis of instinctual violence. It also does not seem wise to challenge the systems of nature which have kept animal populations in proper balance since before humans existed, or at least had any significant impact. The second criticism I will address is the defense of the rights of hunters and cultural values. Brought to the table in the debate over the rhino auction, and indeed in any dispute over hunting animals, are the interests of hunters worldwide. Even the Dallas Safari Club’s mission statement, “to conserve wildlife and wilderness lands, to educate youth and the general public and to promote the rights and interests of hunters worldwide,”14 includes a clause about hunters. While a general assessment of hunter’s rights is beyond the scope of this argument, I will examine the interests of hunters in regards to conservation trophy hunting, specifically of the black rhino. As rhinos are not a primary food source, and their conservation is largely funded by governments and organizations, it does not make sense to say that one can justify killing a rhino for food. There are a large number of food sources other than animals to begin with, as Singer points out10, but there are also many other animals which are commonly used for food. Although I do not agree with the point to which Singer carries his argument, as there are circumstances, which cannot be extirpated here for the sake of 21 space and coherence, that warrant hunting for food, I do find that if one cannot justify killing an animal for sustenance, it would be even less reasonable to hunt it for sport or trophy. The so called right of a hunter to gain material satisfaction does not outweigh the suffering and death of the animal. Furthermore, many hunters, like the Dallas Safari Club, say that they have respect for animals and conservation efforts. If that is the case, which I do not doubt that it is, hunters need to recognize that in order to respect the animal, they ought to respect its interests in living and avoiding pain. Cultural values are, in and of themselves, also an insufficient reason to end an animal’s life. Negotiating the weight of religion and tradition is always difficult, since it carries with it the inability to disprove or discredit the intangibilities of spirituality, something which many modern philosophers choose to ignore in order to simplify their arguments. For this argument, though, it is not necessary to directly contradict religion or culture. The main point of concern comes from Southeast Asia, where it is believed that rhino horns possess medicinal properties.3 Rhino horns also contain ivory, which is a large part of traditional Chinese art and culture.6 The reason I say that it is not necessary to contradict religion or culture is that, first of all, there exists a plethora of medicines and treatment methods other than rhino horn. Even if the horns contain some form of medicine, there are other ways to treat patients. In regards to Chinese culture, the recent expansion of the middle class has led to the creation of ivory factories, which are anything but traditional.6 Materialism, more than traditional values, drives the current market for ivory art. If there is still a demand for the traditional practice of ivory sculpting, there are ways to remove rhino horns and elephant tusks without killing the animals. Now that it has been established that the death of an individual animal cannot be justified, based on the animal’s interest in a life with limited suffering, by conservation efforts, hunter’s rights, or cultural values, I will put at ease any doubts as to the financial survival of conservancies. Conservation organizations and government agencies provide most of the financial support needed for conservation programs. According to one study, “foreign donors such as USAID have played a major role in providing financial support for conservancy development since the early 1990s.”7 While hunting programs, particularly in Namibia, provide a source of additional income for the country and its reserves7, removing them would not have as much of an impact as some may think. Tourist expenditures on park entrance fees, safaris, and conservancy funds also generate large amounts of revenue.1 Barbara King, a professor at the College of William and Mary, suggests that instead of auctioning off a hunting permit, organizations should auction permits for wildlife photography safaris to prevent the hunting of endangered species.5 Although hypothetical, King’s idea is evidence of the existence of creative fundraising solutions that take into consideration the interests of animals. Overall, the recent auctioning of the rhino hunting permit must serve as a point of reflection for all of us as human beings. In order to successfully protect and conserve the species damaged by human activity, we must revise our understanding of animals, hunting, and the purpose of conservancies. If the practice of trophy hunting as a conservation method continues unquestioned, I fear that people will not begin to recognize, at a conscious level, that animals, like humans, have an interest in living and avoiding suffering that we as moral beings cannot ignore. I cannot help but agree with Singer that it is a great challenge to overcome the preconceptions of one’s time10, but it is a challenge that we ought to take on together in order provide a better future for non-human animals. 22 Works Cited 1. Barnes, J. I., C. Schier, and G. Van Rooy. “Tourists’ Willingness to Pay for Wildlife Viewing and Wildlife Conservation in Namibia.” South African Journal of Wildlife Research 29.4 (1999): 101-11. Web. 2. Bognar, Greg. “Respect for Nature.” Ethics, Policy and Environment 14.2 (2011): 147-149. Taylor and Francis. Web. 3. “Black rhino hunt permit auctioned in US.” BBC.com. British Broadcasting Service, 12 Jan. 2014. Web. 23 Feb. 2014. 4. Emslie, R. “Diceros bicornis.” IUCN. IUCN Red List of Threatened Species, 2013. Web. 24 Feb. 2014. <http://www.iucnredlist.org/details/6557/0>. 5. King, Barbara J. “Why We Need Compassionate Conservation.” npr.org. NPR, 13 Jan. 2014. Web. 23 Feb. 2014. 6. Levin, Dan. “From Elephants’ Mouths, an Illicit Trail to China.” The New York Times 1 Mar 2013: A1. Web. 7. Nelson, Fred, and Arun Agrawal. “Patronage or Participation? Community-based Natural Resource Management Reform in Sub-Saharan Africa.” Development and Change 39.4 (2008): 557-85. Web. 8. N. Leader-Williams et al. “Trophy Hunting of Black Rhino Diceros bicornis: Proposals to Ensure Its Future Sustainability.” Journal of International Wildlife Law and Policy 8 (2005): 1-11. Taylor and Francis. Web. 9. “Permit to hunt endangered rhino sells for $350,000 despite protests.” Reuters. Issues & Controversies. Facts On File News Services, 12 Jan. 2014. Web. 23 Feb. 2014. 10. Singer, Peter. Animal Liberation. New York: HarperCollins, 2002. Print. 11. Singer, Peter. “Should We Trust Our Moral Intuitions?” project-syndicate.org. Project Syndicate, n.d. Web. 12. “To Save The Black Rhino, Hunting Club Bids On Killing One.” npr.org. NPR, 29 Dec. 2013. Web. 23 Feb. 2014. 13. U.S. Fish & Wildlife Service. “Why Save Endangered Species?” Endangered Species. U.S. Fish & Wildlife Service, Jul. 2005. Web. 23 Feb. 2014. <http://www.fws.gov/endangered/esa/library/pdf/Why_Save_Endangered_Species_Brochure. pdf>. 14. “Who We Are.” biggame.org. Dallas Safari Club, 2013. Web. 23 Feb. 2014. <http://biggame.org/who-we-are/>. 15. WWF. World Wildlife Fund, 2014. Web. 30 April 2014. <http://worldwildlife.org/species/blackrhino>. 23 The Role of Discontent in Communism's Downfall: An Analysis of How We Survived Communism and Even Laughed by Harry Bodell (History 2235) A fter reading Karl Marx's Communist Manifesto, the perks of communism may appear limitless. After listening to the declarations of Joseph Stalin, Mao Tse Tung, and other communist leaders, a communist society may seem like an ideal society. Indeed, many figures throughout history have attempted to spread a common belief- that communism is the correct path for a society to follow. However, just by reading the title of Slavenka Drakulic's memoir How We Survived Communism and Even Laughed, one can begin to understand the different perspective held by those who lived under the tight grip of communism. Life under communism was not ideal, as Marx, Stalin, Mao, and others had claimed. Life under communism was something that needed to be survived. Drakulic's memoir opens a window into the realities of life for an average member of Soviet-controlled communist society. By considering the failures of communism pertaining to basic human needs, individual freedoms, jealousy toward other lifestyles, and the treatment of women, Drakulic's work paints a very clear picture of the downfall of communism- that the system pushed its people into a state of disapproval and ultimately destroyed their faith in communist society. In order to begin to understand how communism pushed its followers away, one must consider the system's failure to provide for basic human needs. After all, individuals are unlikely to fall in love with a system that cannot provide them with food or economic stability. While supporters of the system claimed that communism provided a better standard of living for society, Drakulic's writing tells a very different story. Drakulic writes that members of communist society lived under "a poverty in which the whole country is deprived, everybody is poor, a poverty when to be poor and deprived is a state of life that hardly ever changes" (189). Economic prosperity was virtually impossible to achieve. Everybody was forced into a state of perpetual economic struggle. This struggle manifested itself in the availability of food. Food was not available in excess like it is in the aisles of an American grocery store today. It was unavailable to the extent that nobody thought in terms of food being good or bad. Drakulic explains that "the word 'like' is not the best way to explain the food situation (or any situation) in Poland" (13), as individuals had to prioritize finding and consuming food over enjoying food. The communist system itself seemed to prioritize the wrong products. Products like milk and water, two very important nutritional products, were not readily available to most people. However, "There may be neither milk nor water, but there is sure to be a bottle of Coke around" (13). Fundamentally, communist society placed its followers in a state where their basic needs were difficult to satisfy. This condition expanded beyond the availability of food, water, and milk. Drakulic points out that "Every mother in Bulgaria can point to where communism failed, from the failures of the planned economy (and the consequent lack of food, milk), to the lack of apartments, child-care facilities, clothes, disposable diapers, or toilet paper" (18). Not only does Drakulic list the other basic needs that were not provided for- the need for clothing, housing, and even bathroom products- but she also reminds the reader that these shortcomings were not hidden under the rug. Everybody knew about them. Everybody could point them out. As communist society failed to take care of its people, faith in the system was left to slowly corrode. Not only does Drakulic's memoir illustrate how people were deprived of what they needed to live well, but it also illustrates how those same people were deprived of the individual freedoms they desired. Fundamentally, individual life choices were subject to the communist government's control. 24 While politics may seem quite distant in capitalist societies, inside communist society "politics never becomes abstract. It remains a palpable, brutal force directing every aspect of our lives, from what we eat to how we live and where we work. Like a disease, a plague, an epidemic, it doesn't spare anybody" (17). Indeed, those living under communism had virtually no say in how they lived their lives. Individualism was essentially forced out of society, as "To be yourself, to cultivate individualism, to perceive yourself as an individual in a mass society is dangerous. You might become living proof that the system is failing" (26). Individuals did not have the freedom to be their own person. After all, communism was based upon the forced equality of people within a society in order to do away with social barriers. This forced equality was factored into everyday life, all the way down to an individual's appearance. As Drakulic explains that "everyone is wearing the same thing, not because they want to, but because there is nothing else to buy" (26), it becomes clear that communist society aimed to limit the ways in which people could be different. That meant limited clothing options. That meant that women would not be able to wear make-up. Indeed, "Without a choice of cosmetics and clothes, with bad food and hard work and no spare time, it wasn't at all hard to create the special kind of uniformity that comes out of an equal distribution of poverty and the neglect of people's real needs. There was no chance for individualism- for women or men" (23), as was seemingly the system's goal. The communist system appeared to be incredibly disheartening to those within its grip. Without providing for individual freedoms, Drakulic explains that "What communism instilled in us was precisely this immobility, this absence of a future, the absence of a dream, of the possibility of imagining our lives differently" (7). Under such conditions, it cannot come as a surprise that people would lose faith in communism. In reading Drakulic's words, it cannot come as a surprise that the communist system pushed people so far away that there would someday be a crisis. Had those under communist control not been the only people suffering in these ways, perhaps the system would not have corroded as it did. If the Americans and Western European people were being deprived of their basic needs and freedoms as well, then at least communist society wouldn't seem uniquely unjust. However, that was simply not the case. Those living under the shadow of the Soviet Union were still able to see the light provided in other nations. Drakulic best demonstrates this in the case of Vogue Magazine. The beauty inside such a foreign magazine was difficult to understand for women like Drakulic. She writes that "Living under such conditions and holding Vogue magazine in your hands is a very particular experience- it's almost like holding a pebble from Mars" (27). The mere reality that such beauty was attainable elsewhere was heartbreaking. Drakulic goes on by writing, "I guess that the average Western woman... still feels a slight mixture of envy, frustration, jealousy, and desire while watching this world of images... But tomorrow she can at least go buy what she saw... Here, you can't. Here, the images make you hate the reality you live in" (28). Jealousy rendered life under communism even more unbearable. To know that what they were being deprived of was readily available and embraced in other nations- often nearby nations- was maddening. To know that seemingly the only factor standing in the way of living a similar lifestyle was the communist system made it even easier to turn against the claims and directions of the government. While these troublesome conditions affected men and women alike, women in particular were pushed into disapproval of the communist system. Communist leadership and even Drakulic herself would be quick to point out that "women were equal under the law" (23). After all, communism's goal of universal equality would not be possible otherwise. However, this attempt to prove both genders equal was not always in women's interest. Drakulic points out that women "had to work like men, proving that they were equal even physically" (23), and many would even say that women were made to do most of the work in communist society. Furthermore, women were deprived of their natural character. Most women want to embrace at least some aspects of femininity. However, "the communist ideal was a robust woman who didn't look much different from a man. A 25 nicely dressed woman was subject to suspicion, sometimes even investigation" (23). Women were so masculinized that Drakulic remembers she heard "repeatedly from women in Warsaw, Budapest, Prague, Sofia, East Berlin: 'Look at us- we don't even look like women" (31). It became clear that women were not looked after in the communist system despite any claims of equality. After all, women lived under "a vision in which their needs- what with Ideology, Politics, and Economicswere nowhere near the top" (47). While women may not have been able to set the Soviet Union and its communist system on a path toward collapse all alone, the way that they were treated under communism certainly pushed public morale to newfound lows. Ultimately, the Soviet Union would be destined to collapse- and with it, communism's grip on Eastern Europe. How We Survived Communism and Even Laughed shines a spotlight on perhaps the most significant factor in that collapse. The communist system pushed away those it needed in order to function. By failing to provide for the people's basic needs, failing to recognize the common desire for individual freedom, failing to consider the people's jealousy toward other societies, and failing to change the treatment of women, the communist system forced society to lose faith and turn its back on the words of Marx, Stalin, Mao, and others who had declared communism to be the ideal system of living. In Drakulic's words, the communist system was "invented to make you believe that you- not the government or the party- are to blame... set up to internalize guilt, blame, failure, or fear, to teach you how you yourself should censor your thoughts and deeds" (6). Unfortunately for communist leadership, individuals do not enjoy living under such standards, and the collective discontent of the people would contribute greatly to the collapse of what was once a major world power. Work Cited Drakulic, Slavenka. How We Survived Communism & Even Laughed. New York: Perennial, 1993. 26 Today’s Registered Nurse by Kellie Brennan (English 1102) T he days of female nurses in white uniforms, simply present to be an assistant to doctors, are long over. Nursing is now a competitive, versatile career that is attracting both female and male students across the country. With enticing education opportunities, salaries, job outlooks, duties, settings, specialties and more, a career as a registered nurse is looking better than ever. Students with “[t]he desire ‘to make a difference, to do more than just have a job’” are attracted to nursing as well (Graduate qtd. in Vallano 10). Whether a career as a registered nurse interests you, or you simply want to learn more, continue reading this chapter to find out what it takes to become a successful registered nurse. Nursing is a practice as old as time; caring for others is a part of human nature. However, the true founder of the modern profession of nursing is Florence Nightingale. The American Diabetes Association newsletter reports that Nightingale, an upper class British woman of the 1800s, “dedicate[d] her life to nursing the sick” (Mazur). While her career began by receiving “some medical training” and “running a small London hospital,” her true success began during the 1854 Crimean War. During this brutal war between Britain, Turkey, and Russia, there were many deaths and “thousands [of soldiers] were wounded and dumped in a British military hospital in nearby Scutari, Turkey.” The hospitals of this time were filthy and infested with rats, spreading diseases easily. Florence Nightingale was contacted to help this British military hospital in Turkey. She brought forty women with her, and created a new kind of hospital: “She looked into the smallest details, seeing that men, floors, walls, and bedding were cleaned, rats and bugs gone, and sanitation introduced. She found a real cook to prepare hot healthy meals and bought fresh vegetables with her own money.” By reforming this hospital, she gained respect. Later on, Nightingale “opened the first nursing school” where nurses “wore clean uniforms, attended classes, read medical books, bandaged wounds, assisted in surgery, and even delivered babies.” Florence Nightingale truly “made nursing a trained, respected, and valued profession” by reforming hospitals and opening nursing schools. In the United States, the Civil War era in the late 1800s is what really changed the view of nurses. Hospitals became more prevalent during this time, employed by passionate, hardworking, women nurses. Following Florence Nightingale’s lead, U.S. nurses worked to make hospitals more organized, sanitary, and respected. Doris Weatherford, writing for the National Women’s History Museum newsletter, explains, “The war thus led to greater respect for nurses, something that Congress acknowledged in 1892, when it belatedly passed a bill providing pensions to Civil War nurses. More important, the war served as the beginning of moving the profession from the home to the hospital and clinic. The result was an explosion of nursing schools in the late nineteenth century.” By starting nursing schools and providing pensions for nurses working in hospitals, nursing became an attractive professional career in the United States. What exactly does a registered nurse do? A simple description of a nurse’s job: “RNs provide and coordinate patient care, educate patents and the public about various health conditions, and provide advice and emotional support to patients and their family members” (United States… “Registered Nurses”). RNs work in many different settings and specialties, but most typically complete the following duties: “Record patients’ medical histories and symptoms, [a]dminister patients’ medicines and treatments, [s]ets up plans for patients’ medicines and treatments, [s]et up plans for patients’ care or contribute to existing plans, [c]onsult with doctors and other healthcare 27 professionals, [o]perate and monitor medical equipment, [h]elp perform diagnostic tests and analyze results, [t]each patients and their families how to manage illnesses or injuries, [and] [e]xplain what to do at home after treatment.” Of course, this is just a basic description of duties; nurses can specialize to have more specific responsibilities. In addition to duties, an interesting fact that not everyone considers is the important qualities a registered nurse should have: “Critical-thinking skills, [c]ompassion, [d]etail oriented, [e]motional stability, [o]rganizational skills, [p]hysical stamina, [s]peaking skills.” These qualities are important because nursing is a difficult, demanding career. If these qualities describe you, nursing might be the right career for you. One of the advantages of becoming a registered nurse is the versatility of the career. RNs can work in many settings: “Registered nurses work in hospitals, physicians' offices, home healthcare services, and nursing care facilities. Others work in correctional facilities, schools, clinics, or serve in the military” (United States…“Registered Nurses”). As well, there are many specialties including “critical care nurse, emergency nurse, hospice/palliative care nurse, labor and delivery staff nurse, neonatal nurse, nephrology nurse, nurse educator, nurse executive, oncology nurse, orthopedic nurse, perioperative [operating room] (O.R.) nurse, psychiatric/mental health nurse, [and] school nurse” (“Registered Nurse (RN)”). The following specialties of nurses typically work in fast-paced, emergency environments in hospitals: critical care nurse, emergency nurse, and preoperative O.R. nurse. A critical care nurse is a “specialty within nursing that deals specifically with human responses to life-threatening problems” (Nurses for a Healthier Tomorrow). These nurses “practice in settings where patients require complex assessment, high intensity therapies and interventions, and continuous nursing vigilance.” In a hospital setting, you may find a critical nurse in the Intensive Care Unit (ICU) floor. Specializing as an emergency nurse involves “rapid assessment and treatment when every second counts, particularly during the initial phase of acute illness and trauma.” Emergency nurses are found in the Emergency Room (ER) of a hospital. Perioperative operating room (O.R.) nurses “provide surgical patient care by assessing, planning, and implanting the nursing care patients receive before, during and after surgery.” These nurses will work in the operating room. Of course, this is just a brief overview of some of the specialties possible for registered nurses in the emergency departments of a hospital. As a nursing student, specializing as an emergency nurse is what interests me. The combination of a fast-paced environment and never knowing what will come through the door seems like an exciting career. Another important department that interests many nursing students is working with babies. Labor and delivery staff nurses “care for women who are laboring, having complications of pregnancy or have recently delivered” (Nurses for a Healthier Tomorrow). These nurses would be found in floors of hospitals that are dedicated to labor and delivery. Similarly, a neonatal nurse “works in a nursery” with newborns. These newborns are usually “small for their age, premature, or sick term infants who require high technology care, such as ventilators, special equipment or incubators, or surgery.” Labor and delivery staff nurse and neonatal nurse specialties are also fastpaced environments in hospitals. Not all hospital nurses work in fast-paced, emergency environment. The following are specialties that are in hospitals, but are not emergency departments: nephrology nurse, nurse executive, oncology nurse, and orthopedic nurse. To begin, “[n]ephrology nurses use the nursing process to care for patients of all ages who are experiencing, or at risk for, kidney disease” (Nurses for a Healthier Tomorrow). In a hospital, nephrology nurses would work on a floor with only kidney patients, or be assigned to the rooms with these patients. Nurse executives are the bosses of other nurses, requiring skills such as “collaboration, coaching, mentoring, diversity, co-creating, communicating and coordinating outcomes management, and enabling the spirit of the community.” Nurse executives may be in charge of an entire floor of nurses, or nurses within a certain specialty in 28 a hospital. An oncology nurse “encompasses the roles of direct caregiver, educator, consultant, administrator, and researcher” of cancer patients. An oncology nurse would work with oncologists, doctors specializing in cancer, and work on cancer floors of a hospital. Nurses specializing in orthopedics study “musculosketeletal health care by promoting excellence in orthopedic research, education, and nursing practice.” Orthopedic nurses may work on various floors of the hospital including ER, surgery, pediatrics, adult, and more. While these specialties are not the emergency departments in hospitals, they are still very important. In addition to hospitals, specialized nurses can be found in other facilities. A specialty in hospice/palliative care focuses “on comprehensive physical, psychosocial, emotional, and spiritual care to terminally ill persons and their families” (Nurses for a Healthier Tomorrow). Hospices, or at home hospice care, are caring facilities for elderly people who are dying, with no chance of recovery. As a nurse educator, nurses “combine expertise and a passion for teaching into rich and rewarding careers.” Nurse educators work for universities and colleges to teach nursing students. Also in the educational system are school nurses who work to “promote student and staff health and safety.” They work with students at elementary and high schools. Personally, a specialty that interests me is psychiatric-mental health. Treating patients with abnormal psychology disorders such as anorexia (an eating disorder) and addiction seems both challenging and rewarding. Registered nurses specializing in psychiatric-mental health “work with individuals, families, groups, and communities to assess mental health needs, develop diagnoses, and plan, implement, and evaluate nursing care” (Nurses for a Healthier Tomorrow). Psychiatric-mental health nurses may work in rehabilitation centers for mental disorders, psychiatrists’ offices, psychiatric floor of a hospital, and more. A good education is an important requirement of today’s nurses: “Registered nurses usually take one of three education paths: a bachelor’s of science degree in nursing (BSN), an associate’s degree in nursing (ADN), or a diploma from an approved nursing program” (United States…“Registered Nurses”). Many four-year universities offer nursing as a major to complete a BSN. Some community colleges, such as College of DuPage, offer an associate’s degree in nursing. If you are thinking about becoming a nurse, make sure you are prepared for the course load: “In all nursing education programs, students take courses in anatomy, physiology, microbiology, chemistry, nutrition, psychology, and other social and behavioral sciences, as well as liberal arts” and participate in “supervised clinical experience.” A BSN will take “4 years to complete; ADN and diploma programs usually take 2 to 3 years to complete.” So, which education pathway is best? It all depends on your future goals as a nurse, and your time frame. Personally, I am a nursing student working towards my BSN at Elmhurst College. I decided to pursue a BSN because I did not decide on nursing until my sophomore year of college, I want to earn a bachelor’s degree, I plan to continue my education after my bachelor’s, and because I believe that most employers would prefer a bachelor’s degree to an associate’s degree. After graduating from your chosen program, it is time to complete your state’s license, certification, and registration requirements: “To become licensed, nurses must graduate from an approved nursing program and pass the National Council Licensure Examination, or NCLEX-RN.” Each state may have additional requirements regarding license and certification for RNs: be prepared. After becoming a RN, nurses can continue their education. Once RNs have completed their BSN, they can continue to their master’s of science in nursing (MSN) and even a doctor of philosophy (PhD) in nursing. An interesting fact is that with an advanced education, nurses become advanced practice registered nurses (APRNs) which “is an umbrella term given to a registered nurse who has at least a Master’s educational and clinical practice requirements beyond the basic nursing education and licensing required of all RNs and who provides at least some level of direct care to patient populations” (NursingWorld). The following is a list from the American Nurses Association 29 of typical specialties as an APRN: Nurse practitioner (NP) – Working in clinics, nursing homes, hospitals, or private offices, nurse practitioners provide a wide range of primary and preventive health care services, prescribe medication, and diagnose and treat common minor illnesses and injuries. Certified nurse-midwife (CNM) – CNMs provide well-woman gynecological and low-risk obstetrical care in hospitals, birth centers, and homes. Clinical nurse specialist (CNS) – Working in hospitals, clinics, nursing homes, private offices, and community-based settings, CNSs handle a wide range of physical and mental health problems. They also work in consultation, research, education, and administration. Certified registered nurse anesthetists (CRNA) – The oldest of the advanced nursing specialties, CRNAs administer more than 65 percent of anesthetics given to patients each year. Registered Nurses and other advanced specialties are not the only types of nurses. Starting at the bottom of the nursing totem pole are Certified Nursing Assistants (CNAs). CNAs work under the supervision of Registered Nurses (RNs), and typically perform the following duties: “Clean and bathe patients or residents, [h]elp patients use the toilet and dress, [t]urn, reposition, and transfer patients between beds and wheelchairs, [m]easure patients’ vital signs, such as blood pressure and temperature, [s]erve meals and help patients eat” (United States…“Nursing Assistants”). To become a CNA, students “must complete a state-approved education program in which they learn the basic principles of nursing and complete supervised clinical work. These programs are found in high schools, community colleges, vocational and technical schools, hospitals, and nursing homes.” The length of these programs vary based on the state’s requirements, but can usually be completed in a few months. Once you have completed a program, “nursing assistants [CNAs] take a competency exam.” After completing the program and exam, CNAs may begin working. If becoming a RN is not the path for you, perhaps becoming a CNA is a better fit. With the bad economy, college students across the U.S. are worried about salary and job outlook for their future careers. The Occupational Outlook Handbook, published by the U.S. Bureau of Labor and Statistics (BLS), reports that “[t]he median annual wage for a registered nurse was $65,470 in May 2012” (United States…“Registered Nurses”). “The best-paid 10 percent of RNs made more than $94,720, while the bottom 10 percent earned less than $45,040,” reports Jada Graves of U.S. News & World Report magazine. Even more exciting is that “employment of registered nurses is projected to grow 19 percent from 2012 to 2022, faster than average for all occupations” (United States…“Registered Nurses”). This growth rate anticipates “526,800 new jobs” (Graves). With these positive statistics, students studying nursing can be hopeful for their futures. Before deciding to become a registered nurse, students should be aware of the controversies in the field. A common controversy in nursing is the stereotype that all nurses are women, not men. Annette T. Vallano, MS, RN, APRN, BC, reports an interesting fact in her book, Your Career in Nursing: “According to the most recent National Sample Survey of Registered Nurses in 2008, 6.6 percent of the American RN work force was male” (105). Vallano explores possibilities of why there are so few men in nursing. She identifies the “public’s image of nurses” as a “’her’ in a white uniform and cap” (109). This image of a female nurse in a white uniform and cap is no longer prevalent in today’s nursing world, but is still associated with the career because of the lack of nurses in media. The reality is that nurses are not portrayed in the news today. Doctors are cited for articles, journals, newspapers and news broadcastings instead of nurses. The evidence of this controversy is shown by the research project conducted in 1997 by Sigma Theta Tau International, the international 30 honors society of nursing: 1. On average, nurses were cited only 3 percent of the time in hundreds of health related articles culled from 16 major news publications. 2. In seven newspaper surveyed, nurses and nursing were referenced by only 4 percent of the healthcare articles examined. The few references to nurses or nursing were mostly in passing. 3. Articles examined during the study referred to both physicians and healthcare academics as doctors. No example was found where a nurse with a doctorate was referred to as doctor. (110) Until the public’s image of nursing as females in white uniforms and caps changes, men in nursing will continue to feel excluded, and female nurses will feel diminished. As a demanding career, nursing has its downside. A career in nursing may leave you burnt out from your daily work, which involves long hours, difficult patients, draining paperwork, and more. In the article “Six Signs It May Be Time for a Nursing Career Change,” Debra Wood introduces the possibility that nursing may not be the right career for everyone. Wood provides us with the “six signs that it might be time to look at some new career options” instead of nursing. She includes the following six signs: “1. Feeling burned out or exhausted…, 2. Hating the idea of going to work…, 3. Craving a change…, 4. Wanting to know you made a difference…, 5. Worrying about patient or personal safety…, 6. Harboring a burning desire to move up.” These signs reveal that nursing does not always work out for everyone. Working as a registered nurse can be stressful. Patti DeNichols, RN, BSN, M.Ed., works as a school nurse at Schiesher Elementary and Lisle Junior High School in Lisle, IL. When asked if she would describe her job as stressful, she replied: At certain times it is especially stressful like the beginning of the school year. My school year starts about one week earlier than the teachers’ does because I need to have the appropriate health information ready for the teachers when their students arrive so they are prepared. This could include allergies, emergency plans, 504 plans [plans developed to ensure that a child who has a disability identified under the law and is attending an elementary or secondary educational institution receives accommodations that will ensure their academic success and access to the learning environment], and pertinent medical information. Various times throughout the year, many reports are due but the big day is Oct. 15th, which is mandatory state required exclusion day for any student not in compliance with the required physical exam and immunizations. As you can tell, nurses have responsibilities that can make their jobs time consuming and stressful. DeNichols shows that the most stress comes from preparation: making sure teachers are aware of student allergies, creating emergency plans for students with health problems such as seizures or heart defects, making 504 plans to ensure disabled students are receiving accommodations, and other medical information. Carol Schmidtke, RN, also works as a school nurse at Schiesher Elementary and Tate Woods Elementary in Lisle, IL. After being asked if her job was stressful, she mentioned, “It is not generally [stressful]. It was at first when I realized that I was often on my own with health decisions. After years of experience, it is better. It can be stressful with student injuries and parent conversations/meetings.” Schmidtke shows that with time, a career in nursing will become second nature. While all careers can be stressful, be prepared to watch for the signs that nursing is not the 31 right career for you. If you have made it this far in the chapter, nursing must be a career that interests you. Where can you find more information? The best place to start is by checking out the American Nurses Association, which is a professional association for American nurses. Founded in 1892, the ANA is one of the largest and most respected nursing associations in the United States. Membership costs $191 per year, which includes monthly publications of The American Nurse, American Nurse Today, and The Online Journal of Issues in Nursing. However, students can still use the site and access online information without a membership. Other benefits of joining this association include membership discounts from various companies, professional conferences in areas such as Orlando, Florida, and educational resources. Nurses, and students studying nursing, can learn more about this association by accessing this link: www.nursingworld.org. After reading all of this information, are you still interested in becoming a nurse? DeNichols encourages students to “go for it! It [nursing] is a wonderful profession you can do the rest of your life in many different settings and in many different ways.” Opportunities for today’s nurses are limitless. With the aging baby boomer generation, jobs are opening up for nurses across the country. The future of nursing is looking bright: the predicted “U.S. nursing shortage will grow to 260,000 registered nurses by 2025” (Vallano 23). While this does mean that jobs are available for future nurses, it also may lead to an increased workload for all nurses. Nevertheless, nursing is still an interesting and exciting career. There are many specialties and opportunities for higher education that nurses of the past could only dream about. Nurses are not merely medical assistants; they are now diagnosing and treating patients. With advances in medicine and technology, we can only imagine what tomorrow’s nurse has in store. Works Cited DeNichols, Patti. Email interview. 24 Mar. 2014. Graves, Jada A. “Registered Nurse.” Money Careers. U.S. News & World Report, 2014. Web. 12 Mar. 2014. Mazur, Marcia Levine. “Did You Know? Lady with the Lamp?.” Diabetes Forecast 56.11 (Nov. 2003): 109. Academic OneFile. Web. 19 Apr. 2014. Nurses for a Healthier Tomorrow. Lippincott Williams & Wilkins, 2013. Web. Apr. 2014. NursingWorld. American Nurses Association, 2014. Web. 12 Mar. 2014. “Registered Nurse (RN).” Explore Health Careers. American Dental Education Association, 18 Apr. 2014. Web. 19 Apr. 2014. Schmidtke, Carol. Email interview. 24 Mar. 2014. United States. Dept. of Labor. Bureau of Labor Statistics. “Nursing Assistants and Orderlies.” Occupational Outlook Handbook. 2014-15 ed. U.S. Bureau of Labor Statistics, 8 Jan. 2014. Web. 19 Apr. 2014. United States. Dept. of Labor. Bureau of Labor Statistics. “Registered Nurses.” Occupational Outlook Handbook. 2014-15 ed. U.S. Bureau of Labor Statistics, 8 Jan. 2014. Web. 2 Mar. 2014. Vallano, Annette T. Your Career In Nursing. New York: Kaplan Publishing, 2011. Print. Weatherford, Doris. “The Evolution of Nursing.” National Women’s History Museum. National Women’s History Museum, 28 Aug. 2013. Web. 19 Apr. 2014. Wood, Debra. “Six Signs It May Be Time for a Nursing Career Change.” NurseZone Newsletter 17 July 2013: 1. CINAHL Complete. Web. 12 Mar. 2014. 32 Sankana Essay by Kevin Brewton (Philosophy 1100) Part I Question #3 Metaphorical Doodling, Where Necessary A common Expo marker darted across the white expanse, leaving behind a short diagonal line of color. The erasable weapon came back, glancing off the glossy surface eight more times, until its red gashes formed a six-sided shape with an odd ‘Y’ in its center. “Now,” said the owner of the hand holding the marker, “What do you see?” His sole student stared at the white board for a few seconds, before nodding to herself. “I see a cube,” she stated calmly. “Really?” the teacher asked, raising an eyebrow, “All I see is a hexagon with the letter ‘Y’ inside it.” The student studied the board for a few moments, until her eyes widened in understanding. She looked back to her teacher with a frown on her face. “But it is supposed to be a three dimensional cube, is it not?” Her teacher shook his head, holding a stern expression. “What it is and what you see it as are two completely separate things. Even now, you cannot tell me you are thinking of this image as what it truly is: some ink on a whiteboard. Instead, you have assigned it shapes, forms, and dimensions that this small bit of color simply does not possess.” The student tilted her head in confusion, “I am simply looking at that drawing as we all do: a form of art. What it is made of does not detract from how I see it.” The teacher simply shook his head. “What you fail to see is that this same error of perception exists all around us, in every minute atom of this world. What we ‘see’ is merely an illusion, just as intangible as the three-dimensional cube you claim exists in this two dimensional board. The only thing on that surface is ink, just as the only thing present in the universe is Brahman.” “But,” the student asked, frowning, “how can there be only Brahman when I can clearly see you standing before me, the ink on this board, and the floor upon which we stand? You would say that all these tangible things do not exist?” Her instructor sighed, pointing again to the shape emblazoned upon the board beside him. “Look at this drawing, is there anything truly real about the ‘cube’? Does the cube have color? Does it have form? Could you feel it?” His student furrowed her brow, “Well, the cube is merely an illusion of the ink. But I still see it, it still exists as a drawing.” “Wrong,” the master stated, “You are merely confusing the ink’s qualities with the cube’s. The ink is colorful, tangible, and has existence. The cube is only an illusion, and you know this, yet still, you defend it.” “Whether or not the cube is made of ink,” the student replied, “it is still a cube. It is a concept that I can clearly see in my mind’s eye.” The teacher scowled, “That is nescience. Your ignorance goes so far as to acknowledge itself! Your ‘mind’s eye’ is nothing but a further extension of maya, of the great illusion we all 33 experience. The qualities which you assign the cube are in fact those of the ink. And no matter how much you believe differently, the cube possesses none of them.” He paused, taking a breath as the student digested his wisdom. He chose to ignore the sour expression on her face as he began to speak once more. “It is this same problem of superimposition, of giving the qualities which are Brahman’s alone to the things we ‘see’ or otherwise ‘experience’ through both our mind and our senses in this transmigratory existence, that plagues each and every soul caught in this cycle. Nothing exists, except for Brahman. And therefore when we say something has existence, we are superimposing a trait which is solely Brahman’s upon an unworthy thing.” The student maintained her irritation as she spoke, “If what you say is true, why do we see the world as we do? How does Brahman become everything?” “I’ve already revealed this to you,” the instructor responded, pointing once more to the drawing beside him, “You clearly see something other than the simple bit of ink that is present on this whiteboard. It is in that way that Brahman manifests itself as the universe. It begins as Unevolved Name and Form, before coalescing into the first of the five elements, Ether, and from there moving on to become Air, Fire, Water, and Earth. Just as the lines produced by this marker form together to create complex tapestries of images which grow off of each other with each successive addition.” He swiftly added five more lines to the board, giving the illusion of a second cube attached to the right of the first. The learner looked at the drawing thoughtfully for a few more moments, before once more speaking. “If that is so, why do you say the world is an illusion? Even if it is entirely Brahman, it is still ‘evolved name and form’ and thus still existent.” “Do you not see?” her teacher said, “There are no cubes on this board, no drawings. There is only ink, and until you comprehend that you will forever see a lie. The world around us is just as false as this image is. There is no ‘Earth,’ no ‘you’ or ‘me.’ There is simply Brahman. When you at last come to this understanding you will be set free of the Transmigratory existence, free from this illusion of individuality and nescience.” The clock on the wall chimed thrice, causing the student and her teacher to let out sighs. “Thank you, instructor,” she said, as she rose from her desk, “But it seems I will have to stay trapped in this beautiful world a while yet.” Part I Question #4 Questions and Ethics “...And in his works, Sankara encourages his students to participate in certain ‘ethical and moral actions.’ You can find the list on page eighty-eight of Mayeda’s commentary,” the lecturer droned, standing at his podium before a small group of learners. “The first is the observation of abstention, which means abstaining from injury, falsehood, theft, incontinence, possession of things, and—” The elder glanced up from his notes as the frantic waving of a hand caught his attention. “Yes?” he asked tiredly. “Sir!” the young student exclaimed, “How can Sankara advise his followers to perform actions when he clearly says, on page eighty-five, that all ‘action should be renounced by a seeker after final release’? When we attain enlightenment, are we not becoming one with Brahman and thus no longer participating in the action and reaction of the Transmigratory existence?” The professor blinked, adjusting his spectacles as he processed the rapid-fire speech. “Very true, young one,” he said after a moment, “This is, in fact, a topic on which the Philosopher contradicts himself multiple times throughout his writings.” 34 The same child who had questioned him spoke up once more, “Why would he do that? Didn’t he realize he was asking us to do two completely opposite things?” The teacher smiled, quickly rearranging the structure of his lecture as he responded. “Yes, it would have been difficult for him to not have realized that. Sankara’s reasons for this contradiction could be any number of things, one of which is the fact that anyone studying under him during his time period would have had to be of the Brahmin, or ‘priest’ caste, of society, and thus already possess an abundant amount of ‘good’ karma and most likely a fairly strict code of ethics as well.” He paused, checking to make sure the ten or so pupils present were paying attention. “However, the reason that Sankara touched upon ethics at all in his writings was most likely to help those reading his works get to a point in their brief life where they could more easily attain final release from the cycle. As Mayeda says, on page ninety-four, ‘Sankara’s view of ethics may be vague or self-contradictory, but this is because its real aim is the highest possible effectiveness in leading his pupils to the final goal’. So,” the instructor continued, “the contradiction of ethics is used merely as a means to an end, rather than an end in itself. Sankara only wishes you to follow his advice in order to further your climb to enlightenment, and then abandon it once you no longer require moral, karmic actions to prepare you for the renunciation of the illusory world.” The students pondered that for a few moments as the professor read over his notes. Before long, a different hand was thoughtfully raised in the air. The teacher nodded towards the inquirer, and the student asked, “If that is true, then what kinds of things should we do to get to the final release? Would they not also give us karman if we performed them?” The master nodded sagely, “Very good questions. What Sankara encourages in terms of actions and behaviors that could lead us to the final release are, as I said before, abstention in life from many different and negative actions; austerity, or sternness, in life; improving the concentration of the mind; allowing the emaciation of the body; and performing regular permanent rites. There are also things that, while not truly a part of ethics, might benefit you in your climb to escape the cycle. Parisamkhyana meditation, for instance, is to help you destroy your acquired merit and demerit. Furthermore, if you have not done so already, a good idea would be to read the Sruti and Smriti scriptures, and also learn to distance yourself from your body. All of these would help you to clear your head so that your mind becomes pure until, at long last, you break free of the fallacy of perception that is the transmigratory existence. However, as you asked, these actions do grant you karman, and thus force you to participate in the illusion.” A younger student spoke up as the teacher’s wisdom echoed away in the large hall. “So why should we do these things? Why would we heap more and more karma onto ourselves if all we are trying to do is escape the transmigratory existence?” The teacher frowned at the interruption, but a glance at the clock quickly forced him to answer it before his learners were left to ponder such things for another week. “As I have said before, these actions are a means of preparation for the final release. It does not matter if they are ethical or unethical; after all, many would view life after the final release as ‘unethical’ once the seeker leaves his family and duties behind. You must use these suggestions as you would a ladder, and climb ever upwards with such actions until you at last reach salvation. When you have attained your release, you simply cease all participation in the cycle and knock down the ladder of preparation, as it is no longer necessary. Actions and behavior are indispensable before one achieves the final release, but once it is attained they are cast aside in the light of the knowledge that ‘thou art that’ and thus not a part of any caste or illusion which requires action.” 35 Part II Analyzing Adi Sankara Like the student in the first half of Part I, I was eventually able to grasp the concepts that go into Sankara’s theologically philosophical system of thought, and understand what he thinks and why. However, I cannot say that I agree with some of the leaps in logic he takes, nor do I put stock in the scriptures from which he occasionally cites. I can follow his ideology from the concept of Brahman all the way through to salvation, but there are several spots where I lose my willing suspension of disbelief and thus reject the conclusions Sankara comes to. That is to say I understand his system fairly comprehensively, but I do not agree with some of the parts that make it work. The main points I shall address and attempt to debate are the cause of avidya and the theory—as well as the practice—of moksha, or salvation from the transmigratory existence. But in order to get to those things, we must first take a look at Sankara’s Cosmological theory. Sankara states that Brahman can be described using the theory of ‘neti, neti,’ or, ‘not this, not this,’ which entails pointing out everything that Brahman can not be and working your way through the layers of non-eternal things until you at last find the pure, perfect ‘stuff’ behind the world we live in. Instead of reaching the Buddhist answer of Brahman being nothing, Sankara finds that this uncaused, unknown being is all things, the ‘One without a second,’ which is the reason that there can be only only one dimension of reality. Sankara then goes on to say that Atman, which is Brahman— and also deserves its own definition as ‘That which is You’—is the Pure Consciousness which ties all of our past and future lives in the transmigratory existence together. He also uses the phrase ‘saccidananda’ to describe It, which is broken down into ‘Being’ (Brahman) - ’Consciousness’ (Atman) - ’Bliss’ (Which Sankara does not accept and ignores unless whatever he is commenting upon forces him to mention it; he does not believe Brahman to be ‘Blissful’ no matter what his sacred texts tell him). This is not a ‘division’ of Brahman so much as a way to think about it; similar to Thomas Aquinas’s God having three facets yet still being ‘one’ person—though of course, Brahman is vastly different from that deity. Moving forward, the proposition that Brahman and Atman are both different and non-different, as the sea and its waves are different and non-different, is certainly an apt metaphor for this idea. To me, this theory of Spiritual Monism is fairly logical, and— though to my Westernized mind it seems vastly more complicated than the dualism of mind and body we are familiar with—Brahman and Atman make sense, up to a point. As I outlined in Question 3, Brahman ‘emanates’ and ‘evolves’ into the physical world, and Sankara further believes that the ‘reality’ we experience is in fact an illusion caused by avidya— ignorance, or more specifically the superimposition (or ‘switching of qualities’ which was outlined during my answer to Question 3) of Atman and non-Atman —whose cause he does not go into much depth with during his writings. In my humble opinion, this is not a satisfactory explanation for the world we know. Sankara tries his hardest to preserve the notion of Spiritual Monism—which declares that there is ultimately only one reality—but Mayeda himself states, “[Sankara’s notion of Unevolved Name and Form creates a new issue:] What is avidya? This problem gave birth to very great controversy among the followers of Sankara himself, just as later the problem of the relationship between Brahman and Atman was to play a central role in the division of the whole Vedanta school”(25). The world we live in, or at least experience in our ignorance, is a very imperfect place. So, if Brahman is the only truth and all else is illusion, where exactly does this false perception come from? It seems difficult to contemplate how an utterly perfect entity could somehow give rise to a universe as complex and as wildly shrouded in nescience as our own. Sankara himself deflects the question of avidya’s origin each and every time one of the pupils in his writings ask it, because he “knows what the questioner is really asking, but [...] refrains from engaging in endless and profitless 36 speculation on avidya”(80). To Sankara the real issue is earning final release from the cycle, and this is his real goal to the exclusion of all else—including this rather glaring lack of explanation. Mayeda points this out several times over the course of his commentary, but yet he also agrees that, “Indulgence in profitless speculation [on avidya] is nothing but the result of avidya”(80). Despite the fact that I can see the logic behind this statement—trying to understand where ignorance comes from is something only someone still caught in the cycle of ignorance would worry about—this still does not answer the question of nescience’s existence. And thus I cannot help but question the integrity of Sankara’s Monism which—despite its comprehensiveness—asks us to simply wave away the cause of all imperfection, evil, suffering, and the rest of the transmigratory existence. However, agreeing to disagree with the eighth century thinker on this point and accepting the fact that avidya exists, I will move onto the soteriology of Sankara’s philosophy—his main purpose for writing and teaching. Once again, I find that I can follow him through this chain of logic up to a point. Sankara puts forth the theory that each of us is a part of—or rather, One with—Atman, but we are all trapped in maya—the illusion of the world—and have lost knowledge of Atman due to our nescience. In order to be saved from avidya, we have to step back from our egos, our “I”-notions, realize that we are a part of the Pure Consciousness, not an individual, and let go of our false perceptions of reality to become ‘One with the One’. Now, to me, as a general concept this ‘transmigratory existence’ makes sense. Action gives ‘you’ karma, forces ‘you’ to be reborn as something either better or worse than what ‘you’ were—a cycle that must be escaped from as it will never cease and you cannot maintain ‘good’ karma for long—and the only way out of that vicious circle is knowledge, which is incompatible with action. At least, once I decide to simply accept that the Sruti—scripture ‘heard’ by ancient vedantic sages—says so, and therefore this theory of how life works must be true, I have little logic to actually critique here aside from ‘who gave the scriptures the authority to tell the truth?’. So with that said I shall move onto Sankara’s spin on things. The main problem I have with Sankara’s soteriology, his path to ‘salvation,’ is two-fold. First, before I begin to question his methods, I must say I generally disagree with his motivation for ‘final release.’ Given the general idea of Brahman, Atman, maya, and avidya, I can safely say that— like the girl in my answer to Question 3—I am quite happy to continue enjoying this work of art that is our world and thus participate merrily in the illusion. In fact, the idea of reincarnation and the transmigratory existence isn’t all that unattractive to me—not that I put a whole lot of faith in the theory for, as I stated above, there seems to be little evidence of it aside from the scriptures Sankara follows. Now, moving past the motivation behind achieving moksha, I would like to focus on Sankara’s method for final release. In order to escape the transmigratory existence, the philosopher believes one must realize tat tvam asi or ‘Thou art That’ and thus recognize the self—the “I”notion—is an illusion based in ignorance. In my opinion, there seems to be a few problems with this practice. First of all, the concept of ‘renouncing’ the self—one’s identity in this mortal realm—seems contradictory to me. Sankara writes, “He to whom both ‘I,’ the notion of ‘oneself,’ and ‘my,’ the notion of ‘one’s own,’ have become meaningless, becomes a knower of Atman” (138). And my response to this declaration is: who, exactly, is ‘He’? The individual who steps back and declares his identity and “I”-notion to be false is still an individual, for, if Atman is changeless, how can it reject—as an action—anything? Moving on, should this rejection of the individual be possible as Sankara claims, even the very idea of ‘final release’ seems strange. When a seeker reaches enlightenment and realizes that he or she is Brahman, they are supposed to simply cease all action in the illusory world. This leaves me questioning the reason they retain their bodies at all, if they truly are Atman and have completely given up their worldly “I”-notion. In fact to me it might make more sense if their physical forms simply evaporated, or were otherwise given up at the same time as their mental identities. 37 But, regardless of that oddity, another point I would like to bring up is the notion that ‘you’ have escaped. Firstly, Sankara states, “...[all] action should be renounced by a seeker after final release”(85). And I, personally, would think that if one truly does attain moksha, they would no longer exist as something capable of action. They become One with Atman, and therefore should not be able to perform any actions at all in the physical realm as they are changeless and eternal. How can they renounce, which is itself a form of action, anything? Of course, I also question the reverse of that point. If one escapes the cycle and renounces or rids themselves of avidya—again, the individual seems to be the one rejecting avidya—how can they be certain that ‘they’ will not simply enter the transmigratory existence once again as another false “I”-notion? In theory, they have become One with Atman and ceased to exist as an individual, so what part of them is ‘saved’ exactly? The Atman behind their ‘self’ should never have been in need of ‘saving’ as it is perfect, and their “I”-notion was just terminated. ‘Their’—or rather ‘that’—wave of Atman on the sea of Brahman has returned to its source, yet I see no pressing need for this to happen. In fact, since they are now a nameless part of Atman once more—or at least have lost whatever separated them from Atman, as Atman itself cannot change—what is to stop ‘them’ from simply rising again as some other wave, some other “I”-notion trapped in the transmigratory existence? This question, and all the others I have voiced here, prevent me from simply accepting Sankara’s soteriology. I understand that Sankara does not dwell on any possible flaws in his writings because he is simply trying to help those suffering in transmigratory existence to escape, but this does not persuade me to turn the same blind eye towards them. Indeed, perhaps I am simply so entrenched in avidya myself that I cannot see the purity of Sankara’s truths, but if that is the case I cannot say that I am eager to give it up. The ideas of Immanuel Kant—that reason and experience are purely what we should use to look at the world—certainly hold far greater appeal to me, for one does not have to rely on scriptures, or the belief that we should not question things—such as avidya—that we do not understand, to find meaning in life. Overall, I would not consider myself a follower of Advaita Vedanta philosophy, as it seems to me both unattractive and bearing some not-negligible issues. Truly, I am more than content to simply live this illusory life to its fullest, regardless of how ignorant that they may be. Written, Revised, and Retroactively Revolutionized by: Kevin Brewton Post Scriptum: I would like to make a short comparison between Thomas Aquinas and Sankara that doesn’t really seem to have a place in the main essay. It seems to me like these two brilliant men wrote for similar and yet very different reasons. Aquinas scribbled away on thousands of pages for most of his life, trying to logically debate the existence of God, among other things, with the people of his time. And yet when he supposedly had his vision from God during a church service, the great thinker never picked up a pen again. Sankara, on the other hand, writes supposedly after he has already escaped the cycle and achieved moksha. He wants to lead people out of transmigratory existence and so ‘goes back across the river’ of nescience in order to actively write and teach for his followers; a regression which I believe he himself states is preposterous. So both philosophers write in order to help others follow their theological logic, but while both ‘see the light’ of their Ultimate Beings, they react to it in very different ways. This isn’t really a critique of either philosophers, merely an intriguing observation that we did not go over in class. 38 Reference Mayeda, Sengaku (trans & ed). A Thousand Teachings: The Upandesasahasri of Sankara. Albany: State University of New York Press, 1992. 39 The Effects of Tetracycline and Ibuprofen on Common Duckweed, Lemna minor L. by Elizabeth Cole (Honors Biology 1152) ABSTRACT tudies are showing that waste water is contaminated with various types of chemicals, such as pesticides, flame retardants, and pharmaceuticals. The contamination of water with pharmaceuticals is of growing concern, as water treatment plants are not able to fully remove these chemicals from drinking water sources, and their effects on aquatic organisms and humans are unknown. This study looks at how water contaminated with tetracycline and ibuprofen affect the growth of common duckweed (Lemna minor L). Plants were placed in test tubes, treated with solutions of the medications, and growth was noted by the number of plants in each test tube after a five week period. Tetracycline was found to have no effect on plant growth, while ibuprofen had negative effects. Further testing may need to be conducted to find why this occurred. S INTRODUCTION An increasing number of anthropogenically generated chemicals are being found in waste water and sludge (Herberer 2002). Studies have tested the quality of treated waste water released into streams, and have found that they still contain small amounts of chemicals. These chemicals cannot be fully removed during a normal filtration process, as most water treatment plants are not designed to remove persistent chemicals (Vergili 2013). There is concern over the growing amount of pharmaceuticals found in treated waste water (Jones et al. 2005). Lipid regulators, anti-epileptic, animal growth hormone, and psychotic drugs have all tested positive in samples of treated waste water. Usually the treated waste water is either released into nearby surface waters or used for human consumption (Houtman et al. 2013). However, the effects that pharmaceuticals in drinking water have on humans and aquatic organisms are unknown (Stackelberg et al. 2004). This experiment used common duckweed (Lemna minor L., Lemnaceae), an aquatic plant that is distributed worldwide, as an indicator species to explore how tetracycline and ibuprofen affect aquatic species. The objective of this experiment was to examine how levels of tetracycline and ibuprofen similar to those found in drinking water affect duckweed growth. METHODS Tetracycline (Mars Fishcare North America, Inc., USA) and ibuprofen (Pfizer Inc., Kings Mountain, NC0) were the pharmaceuticals used in the treatments for this experiment, and the experiment was conducted over a period of five weeks. On the first day, 48 test tubes were filled with 55ml of water from a fish tank and three randomly selected duckweed plants were placed into each test tube. Water from a fish tank was chosen to replicate pond water. Test tubes were divided evenly into a control, tetracycline, and ibuprofen group. Concentrations of 50µg/ml of tetracycline and ibuprofen were created to simulate amounts found in drinking water in their respective category. The plants were then placed on a window sill for the duration of the experiment. Plants were treated twice with the chemical solutions and were recounted per replicate after five weeks. Mann-Whitney rank tests were used to compare counts of the chemical treatments to the control as assumptions of parametric testing could not be met (Zar 1984). 40 RESULTS AND DISCUSSION Counts of duckweed did not vary significantly between the control and tetracycline treatment, but did between the control and ibuprofen treatment (Table 1). Significance was determined at P<0.05. Overall, plants treated with tetracycline had growth similar to the control, with no decay in plant numbers. Ibuprofen treated plants exhibited either a decay or showed no growth in all but two test tubes. Therefore, neutral effects could not be safely rejected. There are several reasons why these results could have occurred. First, ibuprofen is composed of various chemicals, including cyclooxygenase inhibitors (Rainsford 2012). Cyclooxygenase is an enzyme that speeds up the production of chemical messengers called prostaglandins. Thus, an inhibitor of cyclooxygenase should slow the production of prostaglandins (Groenewald et al. 1997). Studies have found prostaglandins to have roles in flowering, photosynthesis, and regulating of cell membrane permeability. The cyclooxygenase inhibitors present in ibuprofen could have affected the duckweed’s ability for sunlight absorption or nutrient absorption, resulting in the overall decline in plant numbers. The negative effect of ibuprofen on duckweed does imply that if waters that supply to irrigation fields get contaminated, it may decrease crop yields. Also, ibuprofen is an enantiomer, which means that its isomers differ in shape as a result of an asymmetric carbon. This causes only one of the two isomers to be biologically active because only molecules of that form can bind to molecules of that organism. Ibuprofen is normally sold as a mixture of its two enantiomers, both the effective and less effective (Reece et al. 2011). This could have also influenced the duckweed growth, as some test tubes may have had more ineffective enantiomer than effective, which could account for replicate test tubes that showed neither growth nor decay. However, the exact nature of ibuprofen’s negative effect on duckweed remained unknown and needs to be further explored through continued studies. Finally, ibuprofen has been shown to cause gastritis in children (Young 1995), impairment of color vision and visual acuity in adults (Grant 1986), and premature closure of ductus arteriosus if taken after 34-35 weeks of pregnancy (Young 1995). Because this study indicates that ibuprofen in drinking water has negative effects on organisms, further studies investigating its effects on humans is necessary. LITERATURE CITED Grant, W.M. 1986. Toxicology of the Eye, 3rd ed. Charles C. Thomas, Springfield, IL, USA. Groenewald, E.G. and A.J. Van Der Westhuizen. 1997. Prostaglandins and related substances in plants. The Botanical Review 63: 199-220. Herberer, T. 2002. Tracking persistent pharmaceutical residues from municipal sewage to drinking water. Journal of Hydrology 266: 175-189. Houtman, A., S. Karr, and J. Interlandi. 2013. Environmental Science. W.H. Freeman and Company, New York, NY, USA. Jones, O.A., J. Lester, and N. Voulvoulis. 2005. Pharmaceuticals: a threat to drinking water. Trends in Biotechnology 23: 163-167. Rainsford, K.D. 2013. Ibuprofen: from invention to an OTC therapeutic mainstay. International Journal of Clinical Practice 67: 9-20. Reece, J., L. Urry, M. Cain, S. Wasserman, P. Minorsky, and R. Jackson. 2011. Campbell Biology, 9th ed. Pearson Benjamin Cummings, San Francisco, CA, USA. 41 Stackelberg, P., E. Furlong, M. Meyer, S. Zaugg, A. Henderson, and D. Reissman. 2004. Persistence of pharmaceutical compounds and other organic wastewater contaminants in a conventional drinking water treatment plant. Science of Total Environment 329: 199-113. Vergili, I. 2013. Application of nanofiltration for the removal of carbamazepine, diclofenac, and ibuprofen from drinking water sources. Journal of Environmental Management 127: 177-187. Young, L.Y. and M.A. Koda-Kimble. 1995. Applied Therapeutics: the Clinical Use of Drugs, 6th ed. Applied Therapeutics, Inc., Vancouver, WA, USA. Zar, J.H. 1984. Biostatistical Analysis, 2nd ed. Prentice-Hall, Inc., Englewood Cliffs, NJ, USA. 42 Table 1. Summary (mean + standard error; all n=16) of average plant counts per treatment. Also provided are Mann-Whitney statistics and probability values from comparisons of the control to the chemical treatment. Significance was determined at P<0.05. ______________________________________________________________ Treatment Mean + standard error U P ______________________________________________________________ Control 3.875 + 0.256 Tetracycline 4.063 + 0.266 119 0.734 Ibuprofen 2.500 + 0.398 46 <0.002 ______________________________________________________________ 43 Role of Women in Service by Neil Davis (History 1120) I n the novel, Not So Quiet…Stepdaughters of War, by Helen Zenna Smith, a theme that is introduced is that of women’s role in service. During the times leading up to the First World War, the men of England had to leave their farms and factory jobs behind and fight. As the war progressed, women were called into service to help out in any way that they could, mainly to take care of the wounded men. This essay will specifically discuss and analyze the role of women before, during, and after WWI. Prior to the war, women’s role in society included work in the home and taking care of the household and children. Some lower skilled women worked in the textile mills completing very dangerous jobs for long hours. However, other women were turning to higher education to move up the workforce ladder: “[c]ollege women were a new elite; their advent shaped roles on all frontshome, work, and ‘world’” (Woloch, 279). These women coined the term “professional women” because they were college trained and working in professional careers. Some of the women dominated certain careers including teachers, nurses, librarians, and social workers. Women thrive in these fields due to their caring, nurturing mentalities. Teaching has always been a predominantly female occupation, but during this time women with college degrees were replacing the uneducated, younger women teachers in the schools. Several women wanted to study law but that was one of the few careers that were male dominated: “[t]he legal profession was highly organized and the courtroom, like the saloon, a male citadel” (Woloch, 279). For the few that wanted to enter this field, they often had family members in the business and were able to be trained at their practices. At the turn of the century, some women had been accepted into the field and were able to receive their legal licenses. However, they were still unable to enter the courtroom and were given jobs performing office work for males. The medical field was another popular profession that women were interested in, as it was not male dominated. Women “[t]reat[ed] mainly [other] women and children, they charged less than male competitors, and often worked out of offices in their homes” (Woloch, 280). As the First World War began, women’s priorities and jobs severely shifted, as more opportunities arose. As many of the men left for service, women had to leave their own careers and help by taking the male jobs that were left behind. At the start, they took up factory work doing jobs such as building heavy machinery or making bullets: “[i]n England the munitionettes as they were dubbed, numbered nearly a million” (Coffin, 765). As the war continued, women were needed on the war front, serving under the American Expeditionary Force “as nurses, clerical workers, canteen workers and telephone operators” (Woloch, 377). Being a nurse during this time of war was a very gruesome and hazardous occupation. They had to be close to the front lines and enemy fire and always on alert to help the wounded soldiers. The conditions they worked in were often unsanitary, with a very limited amount of medical supplies, which led to many diseases and amputations. On the home front, 12,000 women enlisted in the Navy and Marine Corps and many of them worked in army offices or hospitals (Woloch, 377). Nellie Smith, the main character in Not So Quiet, is a 21 year old volunteer ambulance driver. Throughout the book her troubles working as an ambulance driver as documented to depict the horrible conditions women faced while working on the war front. One night, Nellie draws the Number Thirteen Hospital as her station, which is the most dreaded of them all. This is because is it “the farthest one out of camp…on top of the hill with a rough, detestable, 44 badly-winding road,” that is difficult to drive “with a load of wounded on a pitch black night in a hurricane of wind” (Smith, 36). Also, Nellie is subject to acquiring diseases and infections from transporting wounded soldiers and eating poorly. The food prepared for the workers is unsanitary as one may “find hair-combings in the greasy gravy; bits of plate-leavings from the day before and an odd hairpin” (Smith, 51). Everything about their life and jobs reeked of uncleanliness. For example, the ambulances that the women drive had to often be cleaned out as they became so filthy from vomit, blood and mud: “cleaning an ambulance is the foulest and most disgusting job” (Smith, 59). The war hardened women, especially those who took care of gruesome injuries and witnessed death on a daily basis. After the war, perceptions of women working once again shifted, this time for the worse. As the soldiers returned home from war, women’s roles in the factories were severely cut. Women were no longer needed to work male jobs and the government treated them poorly. For all the work they did throughout the war, including taking care of the men, one would think they would be treated the opposite and praised for their hard work. Unfortunately this was not true, and women were treated with less respect. The government and employers sent them home after the war and gave the jobs to the returning veterans. Women were often widowed after war and had to take care of their family and relatives. However, because women were losing their jobs it was difficult to provide for their family as earnings were very low. Working male jobs provided women with a higher pay and more money to spare for the household. Now they had “to return to their poorly paid traditional sectors of employment: the textile and garment industries and domestic service” (Coffin, 766). World War I provided women the opportunity to work in different jobs usually dominated by males, but ultimately the war took away those same opportunities as it came to an end. Throughout history, women in the workforce have had several ups and downs, and World War I is a prime example of this demonstration. Before the war, women were beginning to broaden their options as their education and experience increase, making them more qualified for different jobs. During the war, women were called to the forefront to become nurses and ambulance drivers to aid in treating wounded men. After the war, men returned to their former jobs, taking them away from the hard-working women. The war solved and created problems of employment for women as they were treated with less respect than their male counterparts. Women in the workforce deserve more recognition for the work they completed while the men fought and should not have been subjected to the poor treatment afterwards that they received. Works Cited Coffin, Judith. Western Civilizations. 17th ed. Vol. 2. New York: W. W. Norton&Company, 2011. Print. Smith, Helen Zenna. Not so Quiet: Stepdaughters of War. New York: Feminist at the City University of New York, 1989. Print. Woloch, Nancy. Women and the American Experience. 5th ed. New York: McGraw Hill, 2011. Print. 45 The Usefulness of Rubrics by Kelly Donlan (Education 1100) Abstract his paper explores five ways in which rubrics work as a positive aspect in schools. Six different articles that report on the increased use of rubrics are used to support the claims. These articles differ in their topics of why rubrics are beneficial, but all the articles conclude that rubrics are a positive contribution to assessment and grading. This paper discusses how rubrics are positive for teachers, students, and the students’ parents. Educational professors and experts, such as Heidi Andrade, provide data and facts on how rubric use has positively affected grades and schoolwork in America. Quotes from students are used to prove that not only teachers, but also students appreciate the addition of rubrics for their assignments. This paper explains that rubrics are the most efficient and fair way to grade for all teachers in every grade level. The reader will be left with many examples on why rubrics have so many positive aspects. T When buying a new product, whether it is a new phone or new appliance, the buyer will most likely have some questions about how the product works and what to do. Luckily, almost every product comes with directions or an instructional book. These manuals are helpful because they emphasize how the product should be used and to explain why the inventor created it in the first place. Similarly, assignments in school are given with a guide that states the teacher’s expectations and intentions. This guide is called a rubric. There are numerous positive aspects of giving and receiving rubrics for assignments in school. Rubrics benefit the students by clarifying expectations, giving the opportunity to self-assess, and the feedback helps students to improve for future assignments. Rubrics also create a common ground for teachers and provide communication to both students and parents. Many people would agree that they prefer to know what is expected of them before getting involved with something. Rubrics define every expectation that the teacher wants for the assignment. This, in return, makes the students feel more confident and helps them start, continue, and finish their assignment. Any confusion the student may have about the content or points of the assignment are dismissed by the rubric. The Journal of Teaching in Social Work stated, “Rubrics communicated the instructor’s expectations, clarified how to write course assignments, explained grad and point deductions and, in general, made course expectations clearer” (Genzie, Khaja, Chang, Adamek, & Johnsen, 2012, p. 421). Along with clarification, this demonstrates another expectation that rubrics inform students of: points. A rubric will specifically say what aspects of the assignment the teacher is looking for and how many points each section is worth. A few examples of sections that could be on a rubric include the following: grammar, comprehension, organization, content, and creativity. With this, students are able to determine what to include and determine what sections to focus most on based on point value. At the University of Albany, two undergraduate students created a project to determine student perspectives on rubrics. Their findings concluded, “Students’ comments regarding rubric use were consistently positive. They liked the fact that rubrics let them know ‘what’s expected; and contrasted it with the ‘guessing game’ they felt they had to play when teachers did not provide a rubric or some sort of guidelines for an assignment” (Andrade & Du, 2005, p. 3). Rubrics help teachers express what they are expecting the students to submit. As a result, the students are more likely to produce better work because they knew the expectations without any guesswork from 46 the start. Self-assessment is another positive aspect of rubrics. Self-assessment is an extremely effective way for students to identify and correct their own errors prior to submission. Since rubrics map out exactly how the teacher is going to grade the assignment, students can double-check their work before turning it in for formal grading. In an article titled, “Self-Assessment”, Education Professor Heidi Andrade explained, “During self-assessment, students reflect on the quality of their work, judge the degree to which it reflects explicitly stated goals or criteria, and revise” (Andrade, 2008, para. 2). After finishing a project or paper, a student can read over the rubric section-by-section to evaluate, correct, and add specifics to meet the standards that the rubric outlines. This routine creates a long-term habit of reviewing and correcting work before submitting. Beyond that, reading the rubrics and assessing their own work helps students to understand the task they were assigned. This idea is explained further in an educational magazine that said, “Montgomery stated that asking students to complete the same assessment that the teacher will use might give valuable clues to the students’ understanding of the task and readiness for self-reflection skills” (Jackson & Larkin, 2002, para. 2). The more a student self-assesses, the more likely it is that the student will understand the tasks they are given. Self-assessment also gives students a chance to approximate what grade they will receive. Being able to predict the grade the student will receive helps to eliminate the suspense of waiting to see the actual grade. In the end, if the teacher’s final grade is different from the grade the students predicted for themselves, this can lead to additional student-teacher discussion and student performance improvement. Heidi Andrade said, “If students produce it, they can assess it; and if they can assess it, they can improve it.” One of the greatest benefits of rubrics is the opportunity to improve when the graded assignment and rubric is handed back to students. Rubrics explain to the students how they earned a certain grade, and this is very helpful because many students would admit that they have no idea how they earned the grade that they did. A magazine called Teaching Exceptional Children noted, “… Most students do not understand why they received a particular grade on an assignment. The students may reason that their being good or bad or the teacher’s perception of them is why they were given a particular grade” (Jackson & Larkin, 2002, para. 2). Letter grades communicate to students that they performed well or performed below expectations; however, rubrics explain to them why they performed well or below expectations. Reviewing the rubric makes students aware of what they need to improve on in the current assignment or in future assignments. The undergraduates at University at Albany also recorded one of the students who explained, “I think rubrics basically justify your grade because even if they do have like a grade and what you did wrong, this tells you exactly what you need or what you were lacking in” (Andrade & Du, 2005, Better, fairer grades section). Rubrics show what criteria the student met or exceeded, and what criteria require improvement. Rubrics guide the teacher to give clear and constructive feedback for the students. This is more effective than assigning an A, B, C, D, or F on the assignment. If needed, the teacher can write comments on the rubric that will emphasize the areas for improvement. An interesting thing that students can do is keeping all of their rubrics throughout the year to compare the scores on each assignment to see how they have improved over time. Rubric feedback will lead to an increased improvement of student performance as time goes on. Rubrics not only benefit students because they also provide many positive aspects for teachers. In creating rubrics, teachers define the criteria they are looking for before assigning a project or paper. Defining the criteria beforehand saves time for the teacher while grading and creates consistency among teachers in grading expectations for students of similar grade and ability level. A doctor from Minia University explained, “When a common set of performance criteria and scoring guides are used throughout department and grade-level team, school, or district, this benefit is extended, increasing the consistency of judgement across teachers” (Wael, 2012, p. 214). A common ground is created among teachers with the use of a rubric. All the teachers are expecting the same 47 criteria from their students and that is what creates the consistency needed for education in general. Rubrics help defend the teacher’s fairness in grading. Rubrics show that the grade a student is given is based on content and proves that there was neither favor nor disfavor for an individual student. Another positive feature for teachers is that rubrics can help them categorize student work. They can identify the skills in which students excel and what the teacher needs to work on with the students for further practice. Communication is key in education. Fortunately, rubrics trigger communication between the teacher and students, parents, and among teachers. An article in Teaching Exceptional Children informed readers, “The rubric also provides a guide for clear communication among teachers, parents, and students as expectations for academic success are clarified and refined toward the final product” (Jackson & Larkin, 2002, para. 5). The communication that a rubric creates is about the teacher’s expectations and the reasoning for a certain grade after evaluation of the student’s product. The communication through rubrics is beneficial for both ends (giver and receiver of the rubric) to eliminate any questions and produce the best possible grade. Overall, there are many positives of creating rubrics for teachers. A huge influence on children’s lives and schooling are the children’s parents. Undoubtedly, every parent wants to make sure his or her child is doing well in school and will do whatever it takes for that success to happen. Therefore, the use of rubrics is a positive aspect for parents too. Rubrics act as a helpful tool that parents can look off of to help their child with schoolwork at any grade level. The article “Using Rubrics to Promote Thinking and Learning,” said, “Rubrics make sense to people at a glance; they’re concise and digestible. For these reasons, teachers like to use them to assess student work, parents appreciate them when helping their children with homework, and students often request them when given a new assignment” (Andrade, 2000, p. 13). If the child is stuck or wants his or her parents to look over an assignment, the parent can look at the rubric to get an idea on how to help them. Parents then can assess the work according to the rubric before their child turns in the assignment for grading. Another reason why rubrics are helpful to parents is because they can understand and visualize exactly how their child earned a specific grade. Andrade also provided a quote from an interviewee parent who reported this following scenario, “ ‘Look, you’re a smart child, you’ve always done well in school. Two weeks ago I asked how you were doing in school and you said ‘Fine, Dad.’ How can you bring home this report card?’ Sobbing, the child told him, ‘Dad, I don’t know what the grades count on.’” (Andrade, 2000, p. 13). Using rubrics resolve both problems in this situation. The child will know what criteria the grades are based upon and the parents will understand the child’s grade and why they earned that grade. Rubrics will get rid of any speculation a parent might have of the teacher grading unfairly because the grade and points will be mapped out specifically on the rubric. Finally, after the child receives the assignment and graded rubric, the parent can go over it with the child to clarify strengths and weaknesses. A plan to improve any problem areas can then be determined between the parent and child with helpful teacher feedback. With the never-ending positive outcomes rubrics have on all people involved, it is no wonder why rubric use in schools has increased greatly over time. Rubrics are beneficial to the teachers who make and give them and to the students who receive them to guide their learning. Rubrics help students by clarifying the teacher’s expectations, letting them learn and excel in self-assessment, and helping the process of improvement over the course of time. Rubrics help teachers by making the grading process faster and consistent. Also, it creates a common ground among teachers between and within grade levels. Additionally, rubrics benefit parents by including them in their awareness of a grade and to show what to work on with their child. Rubrics help the quality of assignments tremendously and should be used worldwide in classrooms to produce the best possible student work. 48 References Andrade, H. (2000). Using Rubrics to Promote Thinking and Learning. Educational Leadership, 57(5), 13. Andrade, H. (2008). Self-Assessment Through Rubrics. Educational Leadership, 65(4), 60. Andrade, H., & Du, Y. (2005). Student perspectives on rubric-referenced assessment. Practical Assessment, Research & Evaluation, 10(3), 1-11. Gezie, A., Khaja, K., Chang, V.N., Adamek, M.E., & Johnsen, M.S. (2012). Rubrics as a Tool for Learning and Assessment: What Do Baccalaureate Students Think?. Journal of Teaching In Social Work, 32(4), 421. Jackson, C. W., & Larkin, M. J. (2002). Rubric. Teaching Exceptional Children, 35(1), 40-45. Wael Salah Mohamed Seyed, E. (2012). The Effect of Using Integrated Instruction with Evaluation Through the Use of Performance Tasks and Rubrics on Arabic Writing Skills for Fifth-Grade Students. International Journal of Instructional Media, 39(3), 213-224. 49 Civil Engineering: A Surprisingly Noble Career by Randy Dziak (English 1102) C ivil engineering is one of the most important professions in today’s society. Everything civil engineers do has a direct impact on the people and environment around them. However, if a civil engineers do not do their job properly, the impacts of their actions could be extremely negative. According to the Minnesota Department of Transportation, “On Aug. 1, 2007, Minnesota suffered a tragedy of historic proportions when the I-35W bridge collapsed” (Minnesota…). In this tragic event, 13 people were killed and 145 injured. Obviously, bridges are necessary structures in today’s society and it is up to civil engineers to design and build structures like this which function properly and do not threaten people’s safety. Due to a grievous mistake made by Sverdrup & Parcel and Associates, Inc. in the construction of the I-35W bridge, innocent lives were lost which shows the importance of a civil engineer’s job. When civil engineers make mistakes, the consequences can be catastrophic and, in the situation of the loss of a human life, irreversible. Thus engineering is a career that does not afford mistakes. A major concern in everything engineers do is safety. They must be certain that nothing they build or inspect will be a danger to the people using them and must ensure that the structures must be strong enough to endure any forces they might be subjected to such as human use and environmental fatigue. The first fundamental rule in the NSPE’s (National Society of Professional Engineers) Code of Ethics for Engineers is, “Engineers, in the fulfillment of their professional duties, shall: 1. Hold paramount the safety, health and welfare of the public” (National Society of Professional Engineers). Along with their concern for human safety, structural engineers must also take into account the environmental and economic outcomes of their work. While it may not always be possible to produce buildings with absolutely no harm on the environment, there are standards and codes that must be satisfied when designing new structures. Although safety is the main priority for structural engineers, they often have to be able to produce aesthetically pleasing buildings as well. Obviously structural engineers have a lot of pressure on the quality of their work since lives depend on their expertise. Civil engineers are some of the world’s most creative men and women who use their skills and knowledge to improve the quality of living for all society. Not only do civil engineers fix many problems but they also enhance the quality of living by developing more efficient, safe and comfortable structures and systems. Due to the vast expanse of jobs civil engineers are contracted to do, they are indispensable contributors to society. While civil engineering is not seen as a glamorous or exciting job, much of what civil engineers do help to better society so the job is fulfilling in that sense. In addition, if one is capable of completing the required school and training, civil engineering is one of the most secure and rewarding jobs in today’s society. “Civil engineering” is an extremely broad term which is inclusive of many areas of work. Richard and Susan McDavid state in their book, Career Opportunities in: Engineering, that ‘[c]ivil engineering is divided into several subdisciplines in which many Civil Engineers choose to specialize” (2). These subdivisions can be roughly separated into the following groups: Structural Engineer, Geotechnical Engineer, Construction Engineer, Water Resources Engineer, Transportation Engineer, Traffic Engineer, Urban Planner and Environmental Engineer. Depending on which subdivision a civil engineer specializes in, the average work day can widely vary. On the University of Illinois at Urbana-Champaign’s website in the engineering information section it states, “Although each area has its own special body of knowledge and engineering tools, they all rely on the same 50 fundamental core principles” (Programs of Study). Certain civil engineering subdivisions may require almost all work to be done in an office while others must be done entirely in the field or even a mixture of both. Therefore, it is important for one to know his or her work environment preferences when choosing a field in the realm of civil engineering. Structural engineering is perhaps one of the most widely known subdivisions of civil engineering. According to the Bureau of Labor Statistics, “Structural engineers design and assess major projects, such as buildings, bridges, or dams, to ensure their strength and durability” (United States… “Civil Engineering”). Common items that structural engineers design and build include: buildings, bridges, sports stadiums, arenas and towers. Although, oil rigs, spacecraft, satellites and ships could be built by them as well. However, many of the previously stated items require the expertise of many engineers of many different fields of engineering. One of the most recent feats of structural engineering is the Burj Khalifa sky scraper which become the tallest building in the world on January 17, by 2009 standing, “over 828 meters (2,716.5 feet) and more than 160 stories” tall (Burj Khalifa). Geotechnical engineers mostly work with earth materials such as soil and rock, hence the name “geotechnical.” Geo is the Latin root for the English word “earth.” As defined by whatisgeotech.org (a website dedicated to promoting innovation and excellence in the practice of geotechnical engineering), Geotechnical engineering is the science that explains mechanics of soil and rock and its applications to the development of human kind. It includes, without being limited to, the analysis, design and construction of foundations, slopes, retaining structures, embankments, roadways, tunnels, levees, wharves, landfills and other systems that are made of or are supported by soil or rock. A geotechnical engineer’s job mostly consists of soil investigation and fault distribution. The main purpose of these actions is to determine the subsurface conditions of a site on which something is going to be built. Geotechnical engineers predict how stable the ground is in a certain area in order to prevent buildings from being destroyed by earthquakes, landslides and sinkholes. Another major job of a geotechnical engineer is ground improvement. This is a method where the permeability, stiffness and shear strength of soil is modified in order to save costs down the road. This will keep foundations from eroding as quickly which will prevent other complications in the building as well. Construction engineers are very similar to structural engineers. According to Iowa State University’s College of Engineering’s website, “Construction engineers usually focus on a certain type of construction project” (Iowa State University College of Engineering). The website continues on to divide construction engineering into four areas of expertise: Building, Heavy/Highway, Mechanical and Electrical. Each of these areas specialize in different areas of construction, but the necessary degree and education is same for all. One of the major differences between construction engineers and structural engineers is construction engineers must complete classes concerning construction site management. This is to help save time and money on the construction process. Usually there is at least one structural and construction engineer working together on the same project. Structural engineers are more focused on the designing of structures whereas construction engineers specialize more in the area of managing and consulting. A water resources engineer, also called a hydrologist, specializes in knowledge concerning the movement and distribution of water. According to the Occupational Outlook Handbook, “Hydrologists study how water moves across and through the Earth’s crust. They can use their expertise to solve problems in the areas of water quality or availability” (United States… “Hydrologists”). Most hydrologist positions are government jobs such as working for a city or town regarding waste control and clean water distribution. Other jobs a hydrologist might do may be 51 environmentally related such as flood and landslide risk and warning. Often times, hydrologist will work with structural engineers during the construction of bridges by calculation the force exerted on the bridge by the moving water. It is very rare that hydrologist ever work on a project alone, usually their work is done to provide information for the construction of a new building or the safety of an existing structure such as a bridge or canal. Transportation engineering is exactly what it sounds like. It is the engineering of transportation systems and facilities to produce maximum cost effectiveness, efficiency, safety and comfort for the people using the systems. According to CNN, a transportation Engineer, “[A]nalyzes, plans, designs, draws, checks calculations and oversees highway, rail and related transportation civil construction projects” (CNN Money). Most of a transportation engineer’s job consists of predictions and planning. They must be able to estimate how many trips will be made to specific areas in order to provide adequate transportation systems to and from each area. Efficiency is a major concern in a transportation engineer’s job. They must be able to determine the fastest way to transport passengers and goods from one place to another by more than one method. Also they must design facilities such as airports and train stations which can efficiently, effectively and safely get passengers to their destination. Another critical part of transportation engineering involves the development and implementation of systems to ensure smooth travel such as road signs and light signals. Transportation engineers are extremely important in today’s society since without them our transportation systems would be chaotic and uncontrolled. Also as the population grows there will be a greater need for transportation of food and goods so transportation engineers will always be needed. Traffic Engineering is very similar to both urban planning and transportation engineering. However, traffic engineers have a more specific focus than urban planners or transportation engineers. According to the Institute of Transportation Engineers, traffic engineers “facilitate the application of technology and scientific principles to research, planning, functional design, implementation, operation, policy development and management for any mode of ground transportation” (Institute of Transportation Engineering). Traffic engineers mainly specialize in the area of traffic flow, traffic signs, road markings and traffic lights. Unlike transportation engineers, traffic engineers are not involved in the building or design of any transportation related structures. They are concerned principally with the flow of traffic on existing roadways or designing systems for new roadways. Traffic engineers will also look at areas of high crash rates and develop explanations and solutions for the problems causing the crashes. The job of an urban planner, also known as a city planner, is to optimize the use, effectiveness and functionality of cities, towns and villages. According to the Bureau of Labor Statistics’ Occupational Outlook Handbook, “Urban and regional planners develop plans and programs for the use of land. Their plans help create communities, accommodate population growth, and revitalize physical facilities in towns, cities, counties, and metropolitan areas” (United States… “Urban and Regional Planners”). Things that an urban planner must take into consideration when planning an area are things such as fire and police stations, hospitals and schools. For example, it would be more effective to have a fire station in the middle of a village rather than on the outskirts in order to minimize distance from the station to any point in the village. Other issues such as crime, land value, air pollution and traffic congestion must also be taken into account when designing and planning an urban area. For instance, if one area of a town has a higher crime rate than any other area, it might be most practical to put the police station near that area. Environmental engineering is one of the broadest subdivisions in civil engineering and is a combination of science, engineering principles. The Bureau of Labor Statistics states, “Environmental engineers use the principles of engineering, soil science, biology, and chemistry to develop solutions to environmental problems. They are involved in efforts to improve recycling, waste disposal, public health, and water and air pollution control” (United States… “Environmental 52 Engineers”). In order to enhance the quality of human life, environmental engineers pursue to provide society with a healthy environment to live in. They do this by providing healthy water and air for human use to the best of their ability. Environmental engineers are concerned with how technological advances are effecting the environment and if the effects are harmful they attempt to provide a solution. They also strive to provide optimum sanitation to urban environments. Along with these tasks, environmental engineers regulate industrial hygiene, environmental sustainability and public health issues. Environmental engineering is perhaps one of the most important careers concerning society’s health and well-being. Since there are numerous jobs a civil engineer could do due to the many subdivisions, the salary of civil engineers varies greatly. According to the Bureau of Labor Statistics, “The median annual wage for civil engineers was $79,340 in May 2012” (“Occupational…”). However in a personal interview, Professor Scott Banjavic, an engineering professor at The College of DuPage, stated that the starting salary for a civil engineer is “anywhere between $45,000 and $65,000.” The Bureau of Labor Statistics also states that the top paid ten percent of civil engineers earned an average annual wage of $126,190. This is because the salary of civil engineers tend to increase with their years of experience. The area in which a civil engineer works has a major influence on their salary depending on the cost of living, taxes and many other factors. Banjavic confirms this by saying that although the average starting salary for civil engineers is around $45,000 to $65,000, for civil engineers, “in Chicago it’s closer to $60,000.” Obviously, this is on the higher end of the starting average starting salary scale. As job experience increases over one’s years of work, his or her pay tends to increase as well. The same trend can be seen as education increases. In the Field of engineering, “if you have your master’s degree you tend to earn about $10,000 more” than the average starting salary (Banjavic). In order to become a civil engineer there are specific training and educational requirements that are usually expected to be met and “at the very least you have to have a bachelor’s degree” (Banjavic). Banjavic stated that 4 years of schooling is the usual amount of time it takes to obtain a bachelor’s degree in civil engineering. This degree must be obtained from an ABET (Accreditation Board for Engineering and Technology) accredited institution. Although one can obtain their civil engineering degree from any ABET accredited institution, some schools are superior and expedite the job search. Currently, the civil engineering program at the University of Illinois at UrbanaChampaign is rated the best civil engineering program in the nation. This is due to their ability to give students the most and up-to-date education regarding civil engineering. Their curriculum is constantly changing in order to stay current and fit the civil engineering needs of society. Along with their Bachelor’s degree, “Most civil engineers go on to study and qualify for a professional engineering (PE) license. It is required before one can work on projects affecting property, health, or life” (“Civil Engineers” 38). Since “most companies will want you to become a professional engineer,” right out of school a test must be taken called the fundamentals of engineering exam (Banjavic). Once this test has been taken, four years of field work is required before one can take the final exam called The Professional Engineering exam. However, if one obtains their Master’s degree that can count for one of the four required years. There are five different civil engineering PE exams. Each exam focuses on a different area of civil engineering due to the many subdivisions. Engineers usually start studying for this test at the beginning of their junior year due to the extreme difficulty of the test. According to NCEES (National Council of Examiners for Engineering) over 29% of test takers need to take the test twice in order to pass. Along with the education requirements of civil engineering, most companies expect potential employees to have obtained certain on-the-job training and experience. The most common way for one to acquire this training and experience is through internships. Without internships it is very hard to procure a satisfactory position in the field of engineering. Therefore, it is extremely beneficial to complete at least one internship between ones freshmen and senior year of undergraduate studies. In 53 a personal interview, Professor Banjavic stated that “because of my internships I not only got valuable training and experience, but I also figured out what I didn’t want to do in engineering.” It is common for civil engineering interns to be a junior engineer at an engineering firm. As described by Chevron, one of the most prominent energy corporations, engineering internships, “Develop your skills under the supervision and guidance of experienced professionals” (Chevron). Junior engineers will usually follow around an experienced engineer and run errands and file paperwork for them while at the same time getting hands-on training. Therefore, it is not only beneficial for one studying civil engineering to complete internships for the sake of the job outlook but also for the sake of one’s personal interests. While there are many associations for the numerous subdivisions of civil engineering, there is one widely recognized association which is inclusive of all civil engineering subdivisions. This association is the American Society of Civil Engineers (ASCE) with over 145,156 members is the largest and most respected society regarding civil engineering. With multiple conferences held each month around the nation and some even around the world, it is not hard to find a conference that fits one’s schedule. Along with conferences, the ASCE puts out a magazine called Civil Engineering which is digitally published monthly. Membership to the ASCE is free for students and they offer many benefits for their young members along with information about student organizations in one’s area. However, if one is not a student, there are membership dues. The cost of membership ranges from $33 to $260 depending on the number of years after one has earned their baccalaureate degree. Even though it may seem a bit expensive, the benefits one receives from being a member of the ASCE are worth much more than the dues required. The ASCE’s website offers its members tools which help in the understanding of the education requirements in the field of civil engineering. It provides information on the codes and standards used in the field and can connect the reader to helpful books and journals. Another aspect of the ASCE is it offers information and advice on getting certified and also continuing education. ASCE also provides information on leadership and management. This area of information concerns awards in many different aspects of the field given to civil engineers who demonstrate an impressive level of skill, leadership and many more valuable traits. It is also very important to stay well-informed on the issues of one’s field. Aware of this, the ASCE provides information about pertinent and current issues in the field of civil engineering. Also since new technologies and methods are evolving at such a fast rate, it is necessary to stay informed as to not waste time and money using outdated tools and techniques. If there is a controversy or major issue in the civil engineering profession ASCE’s gives essential information and details concerning the problem. Sustainability is currently a main interest in engineering. ASCE’s website ”offers a host of resources to help civil engineers plan and execute with sustainability in mind, and supports other professional groups pursuing the same objectives” (ASCE). The website also offers a journal in which many intelligent and experienced engineers share ideas and methods for more efficient methods and safer materials. In the field of civil engineering, environmental sustainability is a major controversy. With the population of earth growing at an increasingly fast pace, it is important to recognize that our resources are limited and valuable. On their website, ASCE defines sustainability as “[a] set of environmental, economic and social conditions in which all of society has the capacity and opportunity to maintain and improve its quality of life indefinitely without degrading the quantity, quality or availability of natural, economic and social resources” (American Society of Civil Engineers). In order to prolong the health and existence of our population, there are certain requirements that must be met by civil engineers when building new structures or inspecting old ones. One major advancement being made in the field of structural engineering pertains to lighting. Jeff Siegel states in his book, Investing in Renewable Energy, “Commercial and residential buildings in the United States used 40 quadrillion Btus (quads) of energy in 2005, at a cost of $300 Billion” 54 (Siegel 121). However Siegel also states, “196 LEDs draw only W, compared to the 150 W that an incandescent light draws” (124). Therefore, with the implementation of LED requirements in new structures, millions of dollars will be saved in energy each year. Along with conserving energy, civil engineers must also be aware of the environment and the effects their job has on it. With the building of new structures, often times the construction will destroy plants and push out wildlife. Therefore, civil engineers have to meet certain standards when designing and building structures. According to ASCE, these “[s]tandards provide technical guidelines for promoting safety, reliability, productivity and efficiency across all areas of civil engineering. These voluntary standards are written by experts with knowledge and expertise in a particular field” (ASCE). A relatively new concept is the idea of green roofs. Not only do green roofs benefit the environment around them but they also allow for wildlife such as birds and insects to build homes in an otherwise inhabitable environment. As long as humans have existed there has always been a demand for shelter and safety. When civilizations arose, shelters were not the only structures needed. As stated in Ancient Technology: Ancient Construction, “In addition to homes, ancient people built monuments, palaces, roads, dams, bridges, canals, tunnels, harbors, light houses, sewage systems, and many other structures” (Woods 7). Therefore, civil engineers were needed to design and build these buildings, aqueducts and other systems. Ancient Rome is renowned for their complex and efficient aqueducts. Some were over 60 miles in length and retained a steady downward sloping grade to carry water from mountain springs to the city center (Aicher). Even today, the Roman aqueduct system is regarded as one of the most incredible feats of civil engineering. The term “civil engineering” was coined in the 18th century to distinguish between civil and military engineering. In the ancient times, civil engineers were not only seen as builders and designers but also as inventors. Although civil engineers still invent new buildings and methods to this day, nowadays civil engineers mainly attempt to perfect or enhance existing ideas. Benjamin Wright is one of the most important civil engineers in America’s history of civil engineering. According to the National Park Service, Wright was the chief engineer on the Erie Canal project in 1817 and led numerous other projects as well (Engineer Benjamin Wright).The Erie Canal project entailed leading and guiding thousands of unskilled workers who built the canal with wheelbarrows, shovels, horses and mules. After completing the Erie Canal, Wright was appointed chief engineer for the Chesapeake and Ohio Canal due to his unprecedented reputation. Along with Wright’s engineering work, he was also appointed a New York county judge in 1794. According to Connecticut History.org, “in 1969 the American Society of Civil Engineers (ASCE) declared him “’the Father of American Civil Engineering’” (Connecticut History.org). Civil engineers will always be needed. As society grows, the demand for structures will grow as well. Along with the building of new structures, the inspection of existing structures will always be needed. In his article “Mega-Engineering” in Popular Mechanics, Davin Coburn states, “The economy may be struggling, but there is still plenty of work to be done in America. Aging Bridges need to be replaced, ships need building, energy needs to be produced” (61). Also, in the past fifty years the world has seen incredible advances in the area of science and technology. With these progressions, new systems and structures are needed to be designed and built. An article published by Tech News Daily, a poplar and respected online scientific news source, states, “Thanks to new technologies and new testing methods, some of today's largest buildings can stay safe in a quake without compromising their size or aesthetic design. In fact, to attain their great height, many of the world's largest buildings utilize architectural elements that also protect them in case of disaster” (“The World’s 7…”). These earthquake resistant buildings could save thousands of lives, which shows the value and usefulness of civil engineering. While civil engineering may not be a career that one dreams of entering as a child, it truly is a noble career. Thousands of people’s lives depend on civil engineers doing their job properly. As seen in the example of the bridge collapsing in Minnesota, the consequences of even the smallest mistake 55 in this career can be devastating. Civil engineers work each and every day to enhance the quality of life for society and ensure their safety. Therefore, it is clear that civil engineering is one of the most important careers existing today. Works Cited Aicher, Peter. “Watering Ancient Rome.” PBS.org. WGBH, 2014. Web. 22 Apr. 2014. American Society of Civil Engineers. American Society of Civil Engineers, 2014. Web. 10 Mar. 2014. Banjavic, Scott. Personal interview. 26 Apr. 2014. Burj Khalifa. EMAAR Properties, 2013. Web. 22 Apr. 2014. Chevron. Chevron Corporation, 2014. Web. 23 Apr. 2014. “Civil Engineers,” Careers in Focus: Engineering 3rd ed. New York: Ferguson, 2007. 35-43.Print. CNN Money. Cable News Network, 2014. Web. 9 May 2014. Coburn, Davin. “Mega Engineering.” Popular Mechanics Sept 2012: 61-67. Print. Connecticut History.org. Connecticut Humanities, 2014. Web. 9 May 2014. Engineer Benjamin Wright. National Park Service, 22 Apr. 2014. Web. 22 Apr. 2014. Iowa State University College of Engineering. Iowa State University of Science and Technology, 2014. Web. 23 Apr. 2014. Institute of Transportation Engineers. Institute of Transportation Engineers, 2014. Web. 9 May 2014. McDavid, Richard, and Susan Echaore-McDavid. “Civil Engineering.” Career Opportunities in: Civil Engineering. New York: Checkmark Books, 2007. 2-26. Print. Minnesota Department of Transportation. Minnesota Department of Transportation, 2012. Web. 22 Apr. 2014. NCEES. National Council of Examiners for Engineering and Surveying, 2014. Web. 22 Apr.2014. National Society of Professional Engineers. National Society of Professional Engineers, 2014. Web. 9 May 2014. “Occupational Employment and Wages, May 2012 17-2051 Civil Engineers.” Bureau of Labor Statistics. U.S. Bureau of Labor Statistics, 29 March 2013. Web. 12 March 2014. Programs of Study: University of Illinois at Urbana Champaign. University of Illinois at Urbana Champaign, 2010. Web. 2 Apr. 2014. “The World’s 7 Biggest Earthquake-Proof Buildings.” Tech News Daily. Tech Media Network, 2012. Web. 22 Apr. 2014. United States Bureau of Labor Statistics. “Civil Engineers.” Occupational Outlook Handbook, 201213 ed. U.S. Bureau of Labor Statistics, 8 Jan. 2014. Web. 9 May 2014. United States Bureau of Labor Statistics. “Environmental Engineers.” Occupational Outlook Handbook, 2012-13 ed. U.S. Bureau of Labor Statistics, 8 Jan. 2014. Web. 9 May 2014. United States Bureau of Labor Statistics. “Hydrologists.” Occupational Outlook Handbook, 2012-13 ed. U.S. Bureau of Labor Statistics, 8 Jan. 2014. Web. 9 May 2014. United States Bureau of Labor Statistics. “Urban and Regional Planners.” Occupational Outlook Handbook, 2012-13 ed. U.S. Bureau of Labor Statistics, 8 Jan. 2014. Web. 9 May 2014. What is Geotechnical Engineering. International Society for Soil Mechanics and Geotechnical Engineering, 2014. Web. 22 Apr. 2014. Woods, Michael and Mary B. Woods. Ancient Technology: Ancient Construction. Minneapolis: Runestone Press, 2000. Print. 56 The Legend of Starved Rock by Katrina Favis (English 1102) I magine one is standing on a cliff 250 feet above the Illinois River, overlooking the breathtaking scene of woodland, waterfalls, and canyons that nature gave rise to. The colorful trees sway to the beat of the wind and the birds are chirping as the scenario feels that it could be a part of a movie. This famous cliff, known as Starved Rock, is one of Illinois’ most prominent, historical, and scenic landmarks. What many people standing upon this rock do not know though, is the origin of the name. Legend says that the name became “Starved Rock” from a war between two Native American tribes, the Ottawa and the Illini. The Starved Rock legend provides a symbol of the past by adding lore to the well-known area and by giving respect to the once known Native American tribe, the Illinois. The Formation of Starved Rock is one of life’s natural wonders. According to geologists, the creation of Starved Rock started between 15,000 and 17,000 years ago (Finney 240). Geologists claim that the melt waters from the Wisconsonian glaciations became too great to be restrained by the moraines to the north and east. The water broke through and flowed to the headquarters of the Vermillion River, a tributary of the Illinois Valley, and concentrated in the Illinois Valley. This flood is what geologists refer to as the “Kankakee Torrent” (Finney 241). It is believed that the flood created the historic features and the prominent summit of Starved Rock. Starved Rock is located in the northern part of Illinois along the Illinois River, right next to the city of Ottawa. Specifically, Starved Rock is an elevated cliff on the left bank of the Illinois River that consists of parallel layers of white sandstone. The summit is not less than 250 feet high and there are three perpendicular sides as it is washed at its base by the river (Finney 242). This famous cliff is covered with a soil of several feet in depth, bearing a growth of young trees. To this day, Starved Rock is now a historic state park. "It's a beautiful park, Illinois is filled with prairie and farms and all of a sudden right here, you have Starved Rock," Pam Grivetti, president of the Starved Rock Foundation, said about the prominent landmark (Walcyznski 216). Starved Rock is a natural and historic place that Illinois is so proud to have. It is recognized as a unique archeological site because of its focal point in Illinois and in human history of the Illinois Valley. Furthermore, archeologists suggest that human habitation around Starved Rock dates as far back as 8000 B.C. More recently, the earliest group of inhabitants recorded at the region was the historical Kaskaskia, whose large settlement on the north side of the Illinois River was known as the ‘Grand Village of the Illinois’ (Janvien 84). The Kaskaskia were members of the Illinois Confederation, or the Illini, who inhabited the region in the 16th through the 18th centuries. Then, in 1673, Louis Jolliet and Father Jacques Marquette were the first Europeans to mention Starved Rock, or as they called it “Le Rocher” meaning “The Rock” in French (Finney 246). They founded the place to build an area for trade and shelter. Soon after, other Frenchman in pursuit of souls and furs soon followed Jolliet and Marquette’s lead. Chief among these were Sieur de La Salle and his lieutenant, Henri de Tonti. In 1862, the two constructed Fort St. Louis atop Le Rocher or what is known today as Starved Rock. Fort St. Louis was a part of the chain of western forts to prevent the westward expansion of the British (Finney 246). Eight years later, Fort St. Louis was abandoned at Le Rocher as the Frenchmen relocated to a more convenient location. The relocating of the Frenchmen led Native American tribes, like the Illini, to seek refuge atop of the rock during one of their most treacherous battles. 57 While Starved Rock is clearly of historical importance from both a Native and Western perspective, it is best known because of a legend from which it derives its name. On April 20, 1769, Pontiac, chief of the Ottawa, was brutally murdered by an Illinois Confederation Warrior due to a disagreement about land (Osmun). The alleged murder caused uproar with the Ottawa and their allies, the Pottawatomie. Pontiac was very respected by his fellow tribes, so once the news came out about his death, vengeance on the tribe that murdered their beloved master was beginning to rise. This was the beginning of the almost extermination of the Illinois tribe. According to legend, the Ottawa and Pottawatomie then attacked the Illini along the Illinois River. They terribly did revenge themselves upon their enemies because the Pottawatomies and Ottawas deemed it as a duty and glory to slaughter not only the strong Illini men, but also the women and helpless children of the Illinois tribe. About 2,000 escaping Illini climbed the top of Le Rocher as a place of refuge (Osmun). They thought that it was a safe shelter from all the chaos happening around them. What they did not realize though, is that the Pottawatomies and Ottawas then sieged each side of the rock so no one could escape. As one can tell what happens next, most of the Illini died of starvation, thus having the prominent feature be named “Starved Rock”. Of course, there is a heroic twist to the legend. On a dark and stormy night, it is believed that eleven of the most athletic warriors broke through the besieging lines (Osmun). They had marked from below the isolated rock a nook where their enemies kept their canoes. Afterwards, the eleven warriors rushed with headlong speed, unnoticed by their foes, and hurried down the rapids of the Illinois River. Legend says that those escaping Native Americans knew well of every channel of the river so they were able to navigate throughout the darkness to seek safety at Fort St. Louis, a nearby French fortress. When the Illini reached St. Louis, they told their tale and received protection from the commanders. After escaping one of the most treacherous battles, the Illinois never claimed that name again. They ended up seeking new friends of a different tribe in the southern part of the state and that is the last of what people know of the Illinois Tribe (Osmun). The Pottawatomies and Ottawas were bent upon the utter extermination of their hated enemies and according to legend; they received what they wished for. This horrible tale of the extermination of the Illinois due to starvation is a legend that may or may not be true. The legend comes from things that are said to have happened. None are supported by evidence other than frequent retelling. One can only wonder though, that perhaps this tale was based on something that did happen somewhere at some time. What is most important about the Starved Rock tale is whether a person is willing to believe that the incident actually occurred. It is true that there are not many hard records of the total extermination of the Illinois, but there are a few accounts with chiefs that told their take on the alleged battle. One of the most authentic accounts of the great tragedy is obtainable from Meachelle, an old Pottawatomie chief. He told his side of the story to Judge J.D. Caton in 1833 (Walcynski 217). According to Caton, “Meachelle was present at the siege and the final catastrophe, and although a boy at the time, the terrible event made such an impression on his young mind that it ever remained fresh and vivid” (Walcynski 217). Ultimately, Meachelle claimed that the Illinois Indians ceased to exist. This is intriguing to read about and can be reliable if one believes in the treacherous tale of the Illini. Judge J.D. Caton believed that the legend was true also. Other accounts that support the truth of the legend include Henry Schoolcraft, who explored the area near Starved Rock in 1821. He was said to believe the legend and found fragments of antique pottery and stones that must have belonged to the Illini (Schoolcraft). There are also other secondary sources that reveal the proof of the massacre. These include fur traders, settlers, and visitors of the rock (Walcynski 219). Most of them claimed to have seen bones of the defeated Illinois. Dr. J.H Goodell states in his article that an early settler, Simon Crosiar, told him that the ground of Starved Rock was literally covered with human bones (Walcynski 220). These accounts help indicate proof of the Battle of Starved Rock to those that wish to believe. 58 The myth itself gives respect to the Illini. Edgar Lee Masters, a famous poet, wrote a book named after the alleged remnant of the battle between the Ottawas and the Illini. A line from his poem “Starved Rock” reads: “We are the children and the epigone of the Illini, the vanished nation. And this Starved scarp of stone is now the emblem of our tribulation” (Masters 3). This line specifically refers to the suffering that the Illini tribe had undergone. Masters poem has a way of providing empathy to anyone that reads about the challenges that the Illini faced. It provides insight of a mythical and historic tale. His poem also gives respect to the Illinois tribe because he writes about how the rock is theirs, the place where the last of the Illini had died. The legend gives value to the rock where the Battle of Starved Rock occurred. Places like Starved Rock play important roles in the formation of regional identities and cultures with its legend. It is important for people to learn about the history of the famous Starved Rock because residents and newcomers alike can tie themselves deeply to a land far richer and more ancient than themselves. The Starved Rock legend has a very special place in Illinois and it gives respect to the Native Americans whose ancestors supposedly underwent the treacherous war. The legend is a teaching of the past so people could learn and even put themselves into what the Illinois Native Americans experienced. Their name is perpetrated in the great state of Illinois as the legend is apart of their history. Also, whether legends are true or not, they do play important roles in providing intriguing tales of the past that people of the future can reflect on. The legend of Starved Rock adds excitement to a seemingly beautiful landscape. This is the gift legends have for people; they keep people interested in the events of the past and help explain the present. They provide intrigue and stimulate one’s curiosity to discover history. Because of these historical roots, the Starved Rock myth is a tradition passed down from generation that gives respect to the Illinois Native Americans and provides an intriguing look of the past with its mysterious story. Sometimes heroic tales lead people to believe history as entertaining and even beautiful. The legend itself is a time machine to the past that people can keep hold of once they climb the rock, look over the vast Illinois River, and envision that they stood on the very spot where the last of the Illinois may have perished. Works Cited Finney, Fred A. “The Starved Rock Fort.” Illinois Archaeology: Journal of the Illinois Archaeology Survey 22.1 (2010): 240-255. Academic Search Complete. Web. 21 Apr. 2014. Janvein, Mary W. “The Legend of Starved Rock.” Journal of the Illinois State Historical Society 2.4 (1910): 82-87. Print. Masters, Edgar Lee. Starved Rock. New York: The Macmillan Company, 1919. Print. Osmun, Eaton G. Starved Rock: A Historical Sketch. Chicago: The Free Trader Printing House. 1895. Print. Schoolcraft, Henry R. “Travels in the Central Portions of the Mississippi Valley.” Journal of the Illinois State Historical Society (1825): 310-319. Print. Walcyznski, Mark. “The Starved Rock Massacre of 1769: Fact or Fiction.” Journal of the Illinois State Historical Society 100.3 (2007): 215-236. Print. 59 Miasmatic Theory by Emily Garvanovic (Chemistry 1152) D ating as far back as ancient Greece and surviving well into the late nineteenth-century, many people believed in the idea that bad or corrupt air was the causation of illness and disease. Such an idea was called the miasmatic theory. Those who regarded the miasmatic theory to be true believed that typhoid, cholera, malaria, the bubonic plague and other various diseases were not transmitted from person to person, but rather were caused by foul odors from decomposing matter, such as sewage, rotting vegetation, and decaying corpses, and were transmitted by air (Halliday 1469 and Moffett 540). Even though the belief that the causation of diseases through miasma dates back to ancient Greece, the Oxford English Dictionary shows that the word “miasma” was first used in 1665. Throughout history, there have been many influential supporters of the miasmatic theory: Hippocrates (an ancient Greek physician and the father of medicine), David Barnes (a scientific historian and author of The Great Stink of Paris and the Nineteenth-Century Struggle Against Filth and Germs), and even one of the world’s most famous nurses, Florence Nightingale (Hueppe 384 and Pennington 1740), just to name a few. In Notes on Nursing: What It Is, and What It Is Not, Nightingale wrote that “the very first canon of nursing...the first essential to the patient, without which all the rest you can do for him is as nothing...is this: to keep the air he breathes as pure as the external air” (Scovil 355). Nightingale was a firm believer that foul air was the most important and prominent cause of infection. She also attributed smallpox, measles, and scarlet fever to “the practice of building houses with drains beneath them from which odors could escape and infect the inhabitants” (Halliday 1469). Not only did the theory have the backing of many famous individuals within the scientific community but more importantly, most, if not all, civilians believed this theory to be true. Today, however, we know the miasmatic theory to be untrue, due to the advances made in biology and medicine through John Snow, Louis Pasteur, and Robert Koch with the discoveries of germs and germ theory. I think it is safe to say that one of the most terrifying periods of history was, no doubt, when the bubonic plague ravaged the human population; Europe, the Middle East, Asia, and Africa were all devastated by the so-called Black Death. Through years of influence under the miasmatic theory, many civilians and those within the scientific community believed that foul air and rats were not only causing the disease, but that they were also spreading it. Now, how would rats play any sort of role in the causation and spreading of disease if, during this historical period the miasmatic theory explained the causation of diseases? Well, for a long time there had existed scientific belief “that rats – as well as snakes and similarly loathsome vermin – were born of the miasmas produced by putrefying refuse, stagnant water, or swampy earth, all considered local causes of pestilence” (Barker 664). If bad air was the cause of this horrifying disease, then what did doctors and physicians recommend the people of this time to do? During some of the major waves of the plague, Italian physician Marcantonio Ciappi, in his 1630 disquisition of the disease, urged and recommended that people combat the miasma with floral perfumes by filling their homes with heaps of roses, violets, and myrtle (Barker 674). An example of an early theory on the etiology or cause of the disease was brought by Wu Xuanchon, a Chinese physician from Wuchan. Wu believed that the Black Death was caused by qi, foul air vapors that rose up from the ground. He believed that the miasma could become trapped inside poorly ventilated buildings; therefore, dark and sunless houses, like those in cities, were more likely to be infected than were the houses in the open spaces of the country side. 60 According to Wu, rural areas were more immune than the cities because they had more fresh air. He also concluded that because the poisonous air had to pass through the rats’ burrows on its way to the surface, that is why the rats were affected first and for the same reason, the ground floors of buildings were more dangerous than the upper stories. Wu believed in preventative measures and therefore warned people to be on guard for the rats that were believed to help cause the disease and even urged civilians to go and bury any dead rats they saw, but only after plugging their nostrils and keeping their faces turned away from the rats to avoid breathing in the miasma (Benedict 137, 138). One of the earliest contributions to the idea that something other than foul air as the causation of diseases is attributed to John Snow. Snow was unconvinced that miasma was the cause of the 1854 Broad Street cholera outbreak in London, England. As the General Board of Health favored the miasmatic idea of disease transmission, they tested the air in the district to see if they could filter anything that could be associated with cholera. All that was found was fungus, dust, and mold spores, yet nothing that could be associated with cholera. Unconvinced by this miasmatic theory, Snow believed that the fecal matter of those infected with the disease had somehow contaminated the water systems and had infected the drinking water. Snow used statistics to show that the outbreak correlated with a particular water pump on Broad Street, which was located at the center of the neighborhood where the outbreak originated, and after even more investigation, it was revealed that there was broken lining and the Broad Street pump had, in fact, been dispensing “sewage-tainted” water that had caused the outbreak (Moffett 540, 541). Though Snow’s efforts and data had some imperfections, he can still be attributed with some of the earliest work of developing the germ theory and improved hospital and public sanitation. Professor H. Booth once wrote that “from inhaling the odor of beef the butcher’s wife obtains her obesity” (Halliday 1469). Today we would think that statement made by Booth to be totally absurd and completely ridiculous, but during the time when miasmatic theory was dominant, it would have been a normal thought for you to become obese from to the smell of meat in the air. Thanks to the hard work of many individuals throughout the nineteenth century, we now have the germ theory. It was through this theory that the miasmatic theory was ultimately thrown to the wayside by the scientific community. The development of this theory was a crucial moment in the advancement of biology and medicine. As Paul E. M. Fine wrote, the germ theory’s “impact was enormous, leading all of the biomedical sciences away from previously held notions of miasmas and humours, towards more rigorous concepts of specific aetiologies and appropriate preventive and therapeutic measures” (347). The germ theory proved that it was not through foul smelling air that diseases came about, but through understanding of “human-microorganism discovery that transmissible agents are responsible for diseases that were never suspected of being infectious in origin” (Gibbons 627). Because of the germ theory, we have come to know that the bubonic plague was not caused or created by foul air, but by Yersinia pestis, a type of bacterium that is transmitted by fleas that were dwelling on the rats (Sebbane et al. 5526). Even though the Chinese physician Wu gave great advice to be wary and on guard for rats, he basically signed the death certificates of those who listened to him and made an attempt to go and bury any dead rats they saw. The fleas that carried Yersinia pestis actually lived on rats and by trying to bury them individual put themselves at risk of being bitten by the fleas on the rats and then contracting the plague. Cities were uprooted by the plague, not because they had less fresh air, but because of the high populations of humans and rats and the unbelievably low (basically nonexistent) standards of public sanitation and personal hygiene. It was through the work of very dedicated and hard working individuals that medicine and science took a giant leap from the primitive thought of contracting disease by inhaling foul smelling air, to a more modern medicine and the causation of illness from microbes, germs, and viruses through the germ theory. 61 Works Cited Barker, Sheila. “Poussin, Plague, and Early Modern Medicine.” The Art Bulletin 86.4 (2004): 659689 Benedict, Carol. “Bubonic Plague in Nineteenth-Century China.” Modern China 14.2 (1988): 107155 Fine, Paul E. M. “John Brownlee and the Measurement of Infectiousness: An Historical Study in Epidemic Theory.” Journal of the Royal Statistical Society. Series A (General) 142.3 (1979): 347-362 Gibbons, Robert V. “Germs, Dr. Billings, and the Theory of Focal Infection.” Clinical Infectious Diseases 27.3 (1998): 627–633 Halliday, Stephen. “Death And Miasma In Victorian London: An Obstinate Belief.” BMJ: British Medical Journal 323.7327: (2001), 1469-1471 Hueppe, Ferdinand. “The Causes of Infectious Disease.” The Monist 8.3 (1898): 384-414 "miasma, n.". OED Online. September 2013. Oxford University Press. 4 November 2013 <http://0www.oed.com.lrc.cod.edu/view/Entry/117825>. Moffett, John R. "Miasmas, Germs, Homeopathy And Hormesis: Commentary On The Relationship Between Homeopathy And Hormesis." Human & Experimental Toxicology 29.7 (2010): 539-543 Pennington, Hugh. “From Miasmas to Microbes.” Science 313 (2006): 1740–1741 Scovil, Elisabeth Robinson. “Florence Nightingale's Notes on Nursing.” The American Journal of Nursing 27.5 (1927): 355–357 Sebbane, Florent, C. O. Jarrett, D. Gardner, D. Long and B. J. Hinnebusch. “Role of the Yersinia pestis Plasminogen Activator in the Incidence of Distinct Septicemic and Bubonic Forms of Flea-Borne Plague.” Proceedings of the National Academy of Sciences of the United States of America 103.14 (2006): 5526-5530 62 The Composition, Structure, Sources, and Applications of Eugenol by Frank R. Giuliani (Honors Chemistry 1552) W hat do cloves, cinnamon, nutmeg, and basil have in common? Maybe you recognize each of them as common culinary herbs, which they are, but there is actually a characteristic molecule that is found to some extent in each, referred to as eugenol. Eugenol is an organic molecule consisting of a six carbon ring of alternating double-bonds; two oxygen atoms are singlybonded to carbons in the ring, one of the oxygen atoms is single-bonded to another carbon with three hydrogen atoms and the other is bonded only to hydrogen; another carbon atom is singly-bonded to a remaining carbon in the ring, which is bonded to two other carbons each of which are also bonded to hydrogen atoms. Whew, that was a lot of words. The best way to understand how the atoms fit together is to see its skeletal formula and a ball and stick model, each of which are in the public domain and can be found on Wikipedia [1], but are shown in the margin for convenience. It is considered a phenylpropene molecule [1], the “phenyl” part of the word meaning that it contains a modified benzene ring (6 carbon atoms with hydrogen normally attached to each) with an allyl group attached (a side group H2C=CH-CH2R, where R is the rest of the molecule [2]). The terminology given and how all this fits together will be interesting to learn about in organic chemistry. Its molecular formula is C10H12O2. essential oil, which is said to have the “essence” of the plant that gives it its taste and smell. For instance, the essence of cloves can be generally extracted using an organic solvent such as olive oil which will dissolve some of the eugenol in the plant matter [3]. This would work given the principle of like molecules dissolving like molecules. The smaller the plant can be broken up the better, which increases the surface area of the particles being dissolved. The olive oil will have the essence of cloves, but this is not pure eugenol. Cloves, cinnamon, nutmeg, and basil are some of the well known natural (plant) sources of eugenol. It varies in concentration in each type of plant from which it can be extracted as an Fig. 1 Pictured are the spices cloves, cinnamon, and nutmeg with a basil plant. 63 The distinctive and familiar flavor of cinnamon is not due to eugenol, which has a very different taste than cloves. This may be due to the part of the plant we eat, which is from the inner bark of a tree (see figure 6), as eugenol is said to be found primarily in the leaves of a cinnamon tree [9]. There is also a lot of variation in the genus Cinnamomum that seems to have a large affect on the concentration of eugenol found in its leaves, from trace levels to about 80% in its essential oil [10]. Given the general association that eugenol tastes like cloves, when compared to cinnamon or to even nutmeg and basil which generally contain smaller amounts of eugenol, the vastly different flavors signify that the differences in taste are due to other chemicals present within the parts of the plants we eat. In light of the information we have such as the relatively consistently high concentration of eugenol in cloves, and using taste as a guide, it seems that the most efficient source of eugenol is cloves. Fig. 2 Skeletal formula of eugenol. Fig. 3 Ball and stick model of eugenol. To further refine the plant towards pure eugenol, distillation and selective precipitation processes would need to be employed. If purification was needed, distillation would take the place of the olive oil. Eugenol has a boiling point of 254°C and can be steam distilled from freshly ground clove, which sounds like a fun organic chemistry experiment (there looks to be several example experiments available online) [4]. Cloves’ distinctive taste is said to carry the flavor of eugenol [5], which makes sense given the consistently high concentration of eugenol in cloves, which is on the order of 72-89% in the extract [6][7]. The cloves that we eat come from the flower part of an evergreen tree, Syzygium aromaticum, native to Southeast Asia [8]. The flower is dried (see figure 5) and then often ground to a powder to be paired with compatible foods, like coupled with apple cider and cinnamon! Fig. 4 Clove essential oil in clear glass vial. From Wikipedia article. 64 in marine life research is that using pure eugenol to immobilize fish underwater can decrease growth and can kill off coral reefs near the study area [12]. Eugenol has shown to be an effective antibacterial agent, such as against Salmonella typhi. Results of studies suggest that eugenol can break down a bacterial cell membrane [13]. Its antibacterial properties are being developed in nanoemulsion delivery studies to use eugenol as a food preservative. For example, various oils and surfactants mixed with eugenol were tested against Staphylococcus aureus in fruit juices and shown to be effective against microbial spoilage [14]. Fig. 5 Cloves. From Wikipedia article. Nature has also found some applications for eugenol. Like its distinctive taste, it also has a distinctive smell. For example, a Southeast Asian orchid uses eugenol derivatives, among other chemicals, to attract its pollinator, the male fruit fly. The fly in turn metabolizes methyl eugenol into a female-attracting pheromone [15]. Interestingly, these spices containing eugenol were so prized by humans that even regions and treasures were traded for them. It is said that Magellan sent back a ship of cloves from the Spice Islands (Indonesia) which was the most valuable spice in his time. People believed that “clove essence could improve vision, its powders could relieve fevers, and mixed with milk it could enhance intercourse. [16] ” To settle the Second Anglo-Dutch War (1665-1667), the Dutch agreed not to press their claims on New Netherland. In return, they were granted an island rich in nutmeg. New Netherland was renamed by the English to New York. So it seems that if it wasn’t for a derivative of eugenol (a component of nutmeg), it may be that Americans would speak Dutch and not English [17]. Fig. 6 Cinnamon. From Wikipedia article. If you’ve ever tasted cloves, you may find that it has a bitter taste with a slight tongue numbing affect, due to eugenol. This same affect has found its way into anesthetic applications such as for relieving pain in dentistry or even to immobilize organisms during ecological research of marine life. As in most situations where there are benefits, there are also concerns. In dentistry, there have been cases of severe allergic reactions even to the point of death. Even though it is a natural substance, in its purified form it can cause tissue hypersensitivity and should not be applied undiluted [11]. Patients that are concerned about whether or not they are allergic can take a patch test. A major concern 65 References 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. 15. 16. 17. http://en.wikipedia.org/wiki/Eugenol http://en.wikipedia.org/wiki/Allyl http://www.wikihow.com/Make-Clove-Oil http://www.sciencemadness.org/talk/files.php?pid=79288&aid=1941 http://www.seriouseats.com/2011/02/spice-hunting-clove-sweet-savory-spice.html Nassar, M. I., 2007. Chemical Constituents of Clove (Syzygium aromaticum, Fam. Myrtaceae) and Their Antioxidant Activity. Kamel Chaieb, Hafedh Hajlaoui, Tarek Zmantar, Amel Ben Kahla-Nakbi, Mahmoud Rouabhia, Kacem Mahdouani and Amina Bakhrouf, 2007. The Chemical Composition and Biological Activity of Clove Essential Oil, Eugenia caryophyllata (Syzigium aromaticum L. Myrtaceae): A Short Review. InterScience, John Wiley & Sons, Ltd. Phytother. Res. 21, 501–506. http://en.wikipedia.org/wiki/Clove http://en.wikipedia.org/wiki/Cinnamon Rui Wang, Ruijiang Wang, Bao Yang, 2008. Extraction of essential oils from five cinnamon leaves and identification of their volatile compound compositions. Innovative Food Science and Emerging Technologies 10 (2009) 289–292. N. Sarrami, M. N. Pemberton, M. H. Thornhill, E. D. Theaker, 2002. Adverse reactions associated with the use of eugenol in dentistry. British Dental Journal 193, 5. S.E. Boyer, J.S. White, A.C. Stier, C.W. Osenberg, 2009. Effects of the fish anesthetic, clove oil (eugenol), on coral health and growth. Journal of Experimental Marine Biology and Ecology 369, 53–57. Devi, K.P.; Nisha, S.A.; Sakthive, R.; Pandian, S.K., 2010. Eugenol (an essential oil of clove) acts as an antibacterial agent against Salmonella typhi by disrupting the cellular membrane. J Ethnopharmacol. 130(1):107–15. Vijayalakshmi Ghosh, Amitava Mukherjee, Natarajan Chandrasekaran, 2014. Eugenol-loaded antimicrobial nanoemulsion preserves fruit juice against, microbial spoilage. Colloids and Surfaces B: Biointerfaces 114, 392–397. Ritsuo Nishida, Keng-Hong Tan, Suk-Ling Wee, Alvin Kah-Wei Hee, Yock-Chai Toong, 2004. Biochemical Systematics and Ecology 32, 245–252. http://www.history.com/topics/exploration/ferdinand-magellan http://en.wikipedia.org/wiki/New_Amsterdam 66 Saving the Tallgrass Prairie by Hayley Hickson (English 1102) B elow the cement of the many strip malls covering Illinois lay the remains of the tallgrass prairie, an ecosystem with a history spanning thousands of years in creation. At one point in American history, the landscape enthralled visitors with its fields of color and grasses that stretched farther than the eye could see. Eliza Steele, a young traveler entering Illinois for the first time, writes, “A world of grass and flowers stretched around me, rising and falling in gentle undulations, as if an enchanter had struck the ocean swell, and it was at rest forever,” (Johnson). Unfortunately, while the beauty of the Tallgrass prairie still exists in small, fragmented patches around the state of Illinois, most of the land has been converted into farmland and residential or commercial areas. The purpose of this essay is to prove the need for increased tallgrass prairie restoration across the state by providing historical examples, scientific reasoning, and viable methods of prairie restoration. Historically, the Tallgrass prairie has distinguished itself from the many other ecosystems of North America for its rich biodiversity. The development of such rich land was not an overnight phenomenon, however, as the beginnings of the prairie come from the period following the Pleistocene glaciation. Approximately 18,000 years before the current time, most of the state of Illinois was covered in glaciers and, when melted, converted the vegetation to a tundra-like state, and then into a spruce forest. Between 14,000 to 10,000 years before the current time, the climate changed into a warmer and drier atmosphere. Mesic forests, characterized by their cooler temperatures and abundant wood life, dominated for a few thousand years until the climate became increasingly warm and dry. At this point, approximately 8,300 years ago, most forest growth disappeared, save for the trees dotting stream banks and rivers (Robertson). Prairies develop from a combination of conditions too wet to be a desert and too dry to be a forest. In Illinois specifically, the last 1,000 years has become wetter, allowing for the influx of trees, but without human contamination, prairies would still be the dominant vegetation. According to Ken Robertson of the Illinois Natural History Survey, prairies are controlled by three non-biological stresses: climate, fire, and grazing. Prairies in general consist of extreme temperature ranges, so those living in Illinois typically experience hot summers and frigid winters. Another characteristic of prairies is the amount of rainfall per year, which can also vary. Average rainfall may surprise travelers, as, “the annual rainfall in Champaign, Illinois is seven inches more per year than London, England,” (Robertson) but Londoners will rarely experience the droughts found in Illinois. Severe droughts and extended dry periods are often characteristics of prairie life. As stated previously, their varying climate, grazing, and needed fires characterize prairies. The extreme temperatures once deterred settlers, but nothing compared to the fear generated by fires. Although it is undetermined, fires were most likely started by lightning or Native American tribes (Savage 81). Nevertheless, fire is an essential aspect of maintaining tallgrass prairie health. Periodic fires cut down woody growth and provide the soil with nutrients. When too many trees sprout in the prairie, the shade blocks the sunlight and starts to choke the grasses. Fire stops the woody plants from dominating, cleans out the dead grasses and nonnative plant life, and then returns the nutrients to the soil (Robertson). However, the necessary burn periods for prairies unsettle residents to this day, understandably, and this presents yet another problem when advocating for increased restoration projects within the state. 67 Likewise, grazing the land is an aspect of the tallgrass prairie historically accomplished by bison but, after settlement bison were largely exterminated and replaced by cattle (Robertson). Bison restoration is slower in the eastern states of the prairie, as opposed to the expanse of land found in the west, but certain sites have adopted the need for increased bison presence. Nachusa Grasslands in Illinois will begin to add bison to the land in the fall of 2014, after incredible preparation provided by the help of volunteers and donations. Before heavy settlement, over 30 million bison roamed the country, but the species was nearly exterminated due to the mistreatment from early settlers and other land profiteers. Today, the U.S. Department of Agriculture estimates that 198,000 individuals live on private properties and 20,000 survive on public land (Delgado). These numbers are startlingly low for such an iconic North American land mammal, but the conservation of natural prairie habitat sustains the remaining populations. Bison remain contenders for prairie restoration projects over their competition, cattle, because they are a natural part of the ecosystem. Cattle were introduced after European settlers arrived, and are prevalent today because of the accustomed American diet to beef. However, cattle accomplish the same tasks as Bison, but they accomplish it less effectively. Bison are wonderful contributors to the land because they wallow in the soil, “creat[ing] openings for other kinds of plants,” (Savage 112), clear out unnecessary woody vegetation and forbs, and eat grass, which promotes healthier growth. Grazing is an essential aspect of healthy prairie, but also of healthy land in general. As stated in the beginning of this essay, increased prairie restoration will be explained using scientific and historical evidence – the existence of scientific research as the basis for bison at sites like Nachusa, FermiLab, or Midewin National Tallgrass Prairie (Fall 2014 possibly) prove that increased restoration is approved by a wide range of professionals. Returning to the other two attributes of the prairie, fire and climate, both stalled the development of the tallgrass prairie for a hundred years or so but, after the development of the John Deere plow, settlers attempted to weather the country. Tallgrass prairies can grow higher than most people, and their height is the landscape’s trademark. The types of tall grasses that grows such as Indian grass, Big and Little Blue Stem, Switchgrass, Prairie Dropseed, and Side-oats Grama among many others, achieve this height. Because the landscape appeared so difficult to navigate, early explorers thought the land beautiful, but infertile (Robertson). The tremendous height deterred farming, but after the creation of the plow, settlers soon realized that they had stumbled upon one of the most fertile, if not the most fertile, areas of the world. “Although the majority of original prairies in Illinois have been destroyed, there are still a number of areas where people can see prairies,” writes Ken Robertson in his essay entitled, The Tallgrass Prairie in Illinois. After the discovery of fertile lands, pioneers rapidly converted the grasslands into farmland, and the Industrial Age changed the land into residential and commercial use. Today, the tallgrass prairie, which previously spanned thousands of acres, occupies less than 0.02% of its former domain. Locally, sites like the Schulenburg Prairie at the Morton Arboretum, Gooselake Prairie, Fermilab, SpringBrook prairie, and the Russell Kirk site at the College of DuPage are available for visitation (Robertson). The remainder of the Tallgrass prairie exists in remnants and fragments scattered across the state, including a national restoration project at Midewin National Tallgrass Prairie. Midewin National Tallgrass Prairie was created through the Illinois Land Conservation Act of 1995 on February 10, 1996, and was signed by President Clinton. It is the first federally designated sanction for Tallgrass prairie and continues to be of importance today (Robertson). The prairie spans over 19,000 acres, and remains the, “largest piece of contiguous open space in northeastern Illinois,” (Midewin). Using volunteer support, the site nurtures prairie restoration while maintaining areas for human recreation, such as hiking, bird watching, biking, hunting, and horseback riding. 68 One of the interesting features of the prairie is the renovation of the land from its previous arsenal space in Joliet, to an area that both benefits the community and the health of the land. When discussing prairie restoration and opposition to developing new prairies, it is important to remember that restoration can begin in places misused and neglected, and change a destitute area into aesthetically pleasing land. Prairies are necessary to restore because they return the land to its natural state – this means that instead of farmland, tall grasses and wildlife can return. Farmland can be beneficial for human resources but also incredibly damaging, as land farmed annually begins to deteriorate the soil and rob it of its nutrients, thus creating very poor land. When the land becomes poor, it will cease to sustain life for an extended period, and cause occurrences such as drought and subsequent hunger. Many people oppose new prairie restoration sites because they replace areas for commercial use, but as with the Midewin Nation Tallgrass Prairie, lands that are no longer life sustaining, or abandoned, can become prairie. Therefore, both commercial and natural needs would be met. This is the situation regarding Springbrook Prairie in Will County, where citizens agreed to create a prairie with recreation instead of further construction. Regardless, prairie restoration is a difficult task, consisting of many volunteers and often hard labor. Even at Midewin the site is in continual construction, as reconstruction is time consuming and labor intensive. According to Daryl Smith in The Tallgrass Prairie Center Guide to Prairie Restoration in the Upper Midwest, “a reconstruction consisting of a diverse, species-rich seed mixture will undergo successional stages and exhibit changes in species composition and species domination for decades,” (Smith 9) meaning that reconstruction does not end with planting, but is the beginning of a beautiful journey. After the decision to restore land has been made, site preparation, seed selection, planting methodology, and prescribed burns follow. When deciding to allow for bison, setting a prescribed burn team, or deciding which native plants to grow, setting goals that can be accomplished allow for an easier transition. As soon discussed in this paper, conflict often follows the restoration process, but, “the prairie [is] a region of expectant watchfulness, and life a perpetual contrast of work and idleness, hope and misgiving,” (Grierson 2). The discussions surrounding the restoration process may be long and arduous, but the prairie has the unrivaled ability to persevere against human folly, rivaling the dedication of the few who advocate for its very existence. Conversely, the increase in restoration projects does not come without complications. Even amongst those in favor of environmentalism, there is a disagreement between those who advocate for the prairies and those who advocate for the trees. Restoration is a point of agreement between the two opposing forces, but the restoration of which ecosystem began an argument that would lead to a moratorium. In May 1996, the two divided groups of environmentalists disagreed on whether to destroy forests in favor of restoring natural grasslands, the tallgrass prairie. In Char Miller’s review of Restoring Nature: Perspectives from the Social Sciences and Humanities, edited by Paul Gobster and Bruce Hull, the controversy allowed for the assessment of, “Americans’ perceptions of the tangled relationship between nature and culture,” (Gobster 169). The majority of views indicate a need to preserve nature based on cultural inundation, but a lack of full scientific reasoning among the public, and sometimes the professional, spheres inhibits the ability to conserve intelligently. This is the reason behind moratoriums and political disagreements – neither party fully understands what lies in the best interest for both people and nature. Acclaimed journalist and nature writer Richard Manning believes that, “industrialism is rational and progressive; it has an agenda based on our assumption that we understand how things work,” (Manning 262). In truth, the conflict between industrialists and environmentalists and the inner conflict between environmentalists stems from overconfidence in understanding something that is so complex. This lack of understanding is not an excuse for lowering restoration sites, but should stand as a proponent of increased conservation. While disagreements will still occur, without actual scientific experimentation, the general public will not have the chance to finally grasp why 69 restoration is important. Without the courage of trial and error, the science will remain undiscovered. One of the senior editors of the National Geographic, Robert Paul Jordan, writes in 1967, “the men who explored this land for France would marvel at the expanding industrial network that is filling in the open spaces along the Illinois River Valley from the Chicago area to St. Louis,” (Jordan 581). A statement made in the last few decades still rings true today. Expansion is the motto of Illinois, and conservation is continually labeled as unimportant, despite the overwhelming scientific evidence that proves otherwise. Unfortunately, the destruction of the tallgrass prairie has, “seriously depleted the diverse and irreplaceable genetic resources found in nature” (Madson 258). The consequences of these actions will remain high unless further action is taken. Harmony must strike between the needs of the people and the conservation of the land, because the effects of pollution, mismanagement of the land, extinction of species, and the destruction of necessary nutrients will cause irreparable damage to future generations. Naturalist and poet John Madson writes in Where the Sky Began: Land of the Tallgrass Prairie, that the, “tallgrass prairie is of historical and cultural interest and of high biological worth” (Madson 257), but the conservation of such areas relies on the dedication of those willing to save it. Saving the Tallgrass prairie is greater than advocating for a single species because the land itself is endangered. Along with the disappearance of the land, species that have existed solely on the land for generations will cease to exist. Throughout this paper, the scientific evidence of the importance of bison and the preservation of soil nutrients proves the necessity of conservation. The historical evidence from Nachusa and Midewin National Tallgrass Prairie elaborate on these statistics, and the acknowledgement of the struggles with restoration allows for the irrefutable decision to increase conservation. Regardless of the difficulties, without the attempt to save something still misunderstood, this country will lose countless natural artifacts. Saving the tallgrass prairie is more than saving aesthetically pleasing lands; it is saving a forgotten way of life. Works Cited Delgado, Jennifer. “Preserve Set for Return of Bison.” Chicago Tribune: 6. Apr 22 2014. ProQuest. Web. 30 Apr. 2014. Gobster, Paul H, and R Bruce Hull. “Restoring Nature: Perspectives from The Social Sciences and Humanities.” Restoration Ecology 10.1 (2002): 169. Academic Search Complete. Web. 30 April 2014. Grierson, Francis. “Proem.” Introduction. The Valley of Shadows. New York: Harper & Row, 1966. 1-3. Print. Johnson, Christopher. “Restoring the prairie: the Tallgrass waves once more in the Midwest.” E JulyAug. 2007: 14+. Environmental Studies and Policy. Web. 21 Apr. 2014. Jordan, Robert P. “Illinois.” Prairie State: Impressions of Illinois, 1673-1967, by Travelers and Other Observers. Comp. Paul M. Angle. Chicago: U of Chicago, 1968. 578-94. Print. Madson, John. Where the Sky Began: Land of the Tallgrass Prairie. Boston: Houghton Mifflin, 1982. Print. Manning, Richard. Grassland: The History, Biology, Politics, and Promise of the American Prairie. New York: Viking, 1995. Print. “Midewin National Tallgrass Prairie.” Midewin National Tallgrass Prairie - Home. United States Department of Agriculture, n.d. Web. 24 Apr. 2014. <http://www.fs.usda.gov/midewin>. 70 Robertson, Ken. “The Tallgrass Prairie in Illinois.” The Tallgrass Prairie in Illinois. Illinois Natural History Survey, n.d. Web. 10 Mar. 2014. Savage, Candace. Prairie: A Natural History. Vancouver, B.C.: David Suzuki Foundation, 2004. Print. Smith, Daryl. The Tallgrass Prairie Center Guide to Prairie Restoration in the Upper Midwest. Iowa City: Published for the Tallgrass Prairie Center by the U of Iowa, 2010. Print. 71 The Four Humours Theory by Hope Javier (Honors Chemistry 1552) A t the time when disease was believed to have been the doing of the devil, ancient Greek philosophers provided an observable, yet primitive, approach. Led by Hippocrates in 400 B.C.E, this theory remained uncontested for nearly two thousand years influencing both Western and Eastern medicine, proposing that the human body consisted of four major fluids or humours that must be maintained in equilibrium in order to promote a good well-being. Known as the Four Humours Theory, it sought to explain the cause of diseases. Although the theory remained dominant throughout much of the medieval era, the conception of the 16th century, followed by the development of modern medicine, marked its end. One must ask, how did such theory come about? The Four Humours Theory was a compilation of work based upon the ideas of many philosophers, including of ancient Egypt and Mesopotamia, even before the birth of Hippocrates. However, the structural foundations for the theory were comprised of the extended work of the pre-Socratic Greek philosophers such as Anaximander, Pythagoras, and Alcmaron. From these philosophers, Hippocrates assembled their ideas, such as the four underlying elements of nature, the importance of hot, cold, wet, and dry to one's health, and the significance of maintaining balance in the body, which established the foundation for the Four Humours Theory. Such ideas were well accepted during the period when mythologies, religion, and superstition guided much of the people's understanding regards to disease and medicine. However, it was not until a Roman physician, Claudius Galen, started to spread and popularize the idea that the theory generated much influence. Derived from the work of Hippocrates, the Four Humours Theory finally came into formal existence in an effort to provide explanation for the cause of diseases. Through written works, Galen influenced much of the entire Western and Eastern world, in which Islamic medical practitioners grasped strongly during the 11th century. The fundamental concept of the Four Humours Theory relied heavily on the four fluids or humours that were thought to comprise the human body. These humours were: phlegm, blood, yellow bile, and black bile. According to a publication of the Harvard University Press, The Classical Tradition, the four humours were believed to exist in cycles according to the seasons and constellations, revealing the strong influence of astrology on the theory. For instance, an illness during the winter was considered to be affected by an imbalance in phlegm, which caused respiratory and chest problems. On the contrary, an illness that occurred during the summer was associated with yellow bile, which resulted in fever, bad temper, liver disorder, and so forth. The other two humours, black bile and blood were associated with autumn and spring, respectively. As a result, in order to balance the humours, the constituent treatments were based on opposites. For example, to counter the over saturation of phlegm, which was associated the coldness of winter, it must be treated with hot substances. Likewise, warm illness that occurs during the summer was thought to affect the yellow bile humour, in this case cold treatments were used. Furthermore, in an article by N.S Gill, “The Hippocratic Method and The Four Humours of Medicine”, it states that the four humours were associated with the supposed natural elements: water, earth, fire, and air. The humour phlegm was attributed with water, blood with air, yellow bile with fire, and black bile with earth. Ultimately, the relationship of the humours with the seasons and the natural elements provided an understanding that 72 imbalance within the body was the underlying cause of diseases. Moreover, observation played a vital role in providing support for the Four Humours Theory. According to Chris Rohmann's, A World of Ideas: A Dictionary of Important Theories, Concepts, Beliefs, and Thinkers, the Four Humours Theory followed the idea that an individual's well-being was defined by their characteristic complexion or temperament, which correlated with the qualities of the four elements and their relationships with the four humours. For example, one who had too much phlegm, blood, yellow bile, and black bile, were identified to have the personality traits that were phlegmatic, sanguine, choleric, and melancholic, respectively. The physicians that followed the Four Humours Theory diagnosed their patients based on the personality observations. Furthermore, the physicians that followed the Four Humours Theory observed the patient's physical properties that correlate with the associated nature elements and seasons of the humour. For example, people that were hot to the sense of touch were believed to have imbalanced yellow bile, the humour that was associated with the summer season. Evidently, the Four Humours Theory included scientific premise of observation and was a major advancement compared to the preceding ideas that relied heavily on mere superstitions and mystical powers to explain the cause of diseases. However, the treatments that were formulated based on the Hippocratic idea revealed the blunders of the Four Humours Theory. According to David C. Linberg's, The Beginnings of Western Science, bloodletting was a common treatment in maintaining the humours of the body. Linberg writes that prior to bloodletting, the physician would observe the patient for days. Although the Four Humours Theory established a scientific approach in diagnosis, the positive outcomes of the bloodletting treatment were mere the result of the placebo effect. For example, the hierarchy of society placed the physicians in a respected, higher degree. As a result, patients often believed that physicians had immense knowledge on the treatment, which in turn caused them to feel better, even though bloodletting was far from beneficial. The blunders of such treatment outweighed the benefits. For example, the 17th century king of England, Charles II, was treated with bloodletting in order to combat a seizure. Unsurprisingly, this implicated the king's death. In addition, losing blood reduces the oxygen level dramatically, causing severe complications. Unfortunately, many more treatments based on the Four Humours Theory were practiced without much great success, including forced puking, blood leeching, and even breathing in the odor of pig manure to treat nose bleed were, unsurprisingly, ineffective. Clearly, the Four Humours Theory formulated treatments that would otherwise be deemed absurd today. Interestingly, the Four Humours Theory remained unchallenged in Europe for many centuries primarily due to the primitive established scientific knowledge, or lack thereof, during the years of its existence. Especially with the digression of human civilization following the bubonic plague, much of the understanding of human disease leaned toward spiritual and superstition venue, which made the Four Humours Theory highly attractive. According to the Encyclopedia of Britannica, it was not until a 16th century German-Swiss physician named Philippus von Hohenheim, commonly known as Paracelsus, introduced the foundation for modern medicine. Unlike Hippocrates, Paracelsus emphasized that diseases were caused by the body being harmed by outside forces and chemicals found in nature provided cures for the diseases. Furthermore, Paracelsus rejected the four humours of the body. Rather, Paracelsus believed that outside forces, not humours, were the culprit of diseases. The further rejection of the Four Humours Theory was accompanied by the advancements through the scientific revolution during the Renaissance. Ultimately, the changes in both the principles of the people as well as the development of modern science ended the long rule of the Four Humours Theory. 73 Undoubtedly, the Four Humours Theory has widely been rejected to this day. The idea of having fluids that govern the whole well-being of an individual, both controlled by the elements of nature and the seasons, surely seem highly absurd in modern essence with the establishment of modern medicine. In retrospect, the principles and treatments that were formulated as a result of the Four Humours Theory prove highly inaccurate in describing the cause of diseases. Unfortunately, the strong effects of mythologies, religion, and superstition in the ancient societies allowed domination of the Four Humours Theory for nearly two thousand years. Works Cited The Classical Tradition. Eds. Anthony Grafton, Glen W. Most, and Salvatore Settis. Cambridge: Harvard University Press, 2010. Print. Gill, N.S. “Four Humors.” About.com Ancient / Classical History. N.p., 2013. Web. 04 Nov. 2013. Lindberg, David C. The Beginnings of Western Science. Chicago: The University of Chicago Press, 1992. Print. Parker, Richard G. “The Hippocratic Humoural Theory: A Proper Philosophical and Historical Context.” The Rational Argumentator 14 (2003). Web. 04 Nov. 2013. “Paracelsus.” Encyclopaedia Britannica Online Academic Edition. Encyclopædia Britannica Inc., 2013. Web. 04 Nov. 2013. Rohmann, Chris. A World of Ideas: A Dictionary of Important Theories, Concepts, Beliefs, and Thinkers. New York: The Ballantine Publishing Group. 1999. Print. 74 Communicative Action via Internet Technologies by Jonathan Kaye (Philosophy 1800) A bstract: The purpose of this study was to analyze the work of German philosopher Jürgen Habermas, focusing on his theory of communicative action. In his work, Habermas specifically probes the epistemological question; how human beings can know anything at all? Through an interconnected scheme of: survival challenges, knowledge and action; there is found grounding for the physical sciences, the social sciences and philosophy itself. Communicative action becomes the “glue” that holds society together in reference to those sciences, for Habermas, communicative action is how shared meaning about the world is established and it is through communicative action and based on communicative action that humanity can be gauged on a progressive path to freedom from social domination. Habermas used his theory to critique mass communication such as news and entertainment within that framework. This paper expands that critique asking the pros and cons of the use of social media for mass communication; suggesting that in many aspects it is a positive force for the furthering of knowledge through communication, however there may be limits and distortions to communication that inhibit the growth of further knowledge due to the use of technology that leads to a lack of face to face connection. Jürgen Habermas was born in Dusseldorf, Germany in 1929. As the Nazi party came to power in 1933 his family moved to Gummersbach, where he would grow up. In 1939 he was enlisted in the Hitler youth, in which he served and later was drafted as a military reservist for the homeland defense until in 1944. After the war, Habermas became interested in the philosophical existentialist movement. His major influences at that time were Jean Paul Sartre (1905-1980) and Martin Heidegger (1889-1976). Habermas would break with much of existentialist study eventually, due to Heidegger’s refusal to abandon his pro-Nazi stance even a decade after the regime had fallen. Beginning his academic career at the Frankfurt School where he would be educated in the thought of Critical Theory, Habermas came under the tutelage of such thinkers as Theodor Adorno (1903-1969), Max Horkheimer (1895-1973) and Herbert Marcuse (1898-1979). The Frankfurt School had long followed the Marxist method of social critique. Habermas’ mentors would apply this Marxist critique to the capitalist state, to freedom and equality. They would argue these ideals of freedom and equality, were for the wealthy and not all people. They contended that, what they called Humanistic Marxism, did not conflict with liberal ideals, but just the reality in which those ideals existed, claiming that liberalism both legitimated and contradicted modern capitalism. (Ingram 5) In his preliminary academic career, Habermas would pursue studies in sociology and economics, earning a master of social science degree. He spent his time developing a theory of psychoanalytic social critique. He argued that ideology was a mass social delusion and that along with unthinking conformity to authority are social neuroses, grounded in societal repression of basic instincts aimed at individual and social fulfillment. His conclusion was that further emancipation of society came through unlimited and undistorted communication. By 1964, Habermas replaced his mentors and became the chair of philosophy and Sociology at Frankfurt. In 1971, in pursuing linguistics, Habermas developed a philosophy of language and investigated the relation between moral development and a theory of rational communicative competence. Habermas’ Theory of Communicative Action was published in 1984. In this work he argued that communicative action was a basic necessity for the evolution of society. Furthermore, he added that economic and administrative tendencies serve in undermining democratic ideals implicit 75 in interaction and communication. The pursuit of this paper will be to investigate and explicate Habermas’ theory of communicative action with practical applications to today’s society. I will discuss the implications of Habermas’ theory in view of the current form of global social communication. I will take on Habermas’ communication theory to assess the state of communication in society with the advancement of technologies that instantaneously, and potentially without limit or distortion, connect people for the purpose of communication. I shall frame Habermas’ arguments in reference to the arguments that brought his stance about; questions of knowledge and action as well as those on the relationships between science and philosophy. I will layout his major theses, defining major terms of importance, for the purpose of representing his argument in a critique of mass communication. Communication is one of those terms that becomes difficult to define without using the term itself in the definition. As a form of communication, language uses a system of symbols that represent ideas; therefore, it is not communication itself. Communication is a system, an anthology of separate but related methods to convey ideas and thoughts. Methods of communication will not be discussed, because investigations into kinesics and non-verbal communication are not necessary for this paper, I will consult the use of verbal and written language communication only and the mediation therein. What is important is the conveyance of thoughts and ideas through language. When I say something about something, I am also saying that I know something about something. For instance in this paper I am saying something about Habermas, therefore I imply that I know something about Habermas. Before an attempt to discuss communication can be made; we must first discuss knowledge. Historically there have been basically two ways that philosophers claim that humans answer the question of how we know what we know. The two views of knowledge have, through time, been thought to be separate and opposed, which has led many philosophers to look at the dilemma in the terms, of reason vs. experience. On the one side there have been philosophers, who follow René Descartes, (1596-1650) who in his Meditations doubts the reliability of experience of physical objects. The view holds on to a subjective world of thought as the means of knowing, through reasoning, of one’s own existence. On the other side of the debate, are those who say that all knowledge comes from experience, such as David Hume (1711-1776) in Concerning Human Understanding. One comes to expect something to occur only because it has repetitively and reliably on past occasions. The debate casts scientific empiricism vs. philosophic rationality. Immanuel Kant (1724–1804) in his, Critique of Pure Reason, argues that pure reason is unreliable, Kant however applies reason to knowledge inquiries of natural phenomena in experience which constitutes both knowledge of the self and the world. Kant attempts to bridge the gap of dissent between reason and experience by creating a scientific use for reason concerning natural phenomena in which a synthesis of reason and experience occurs to connect thoughts and organize the experience. Habermas addresses two major movements that come from these developments: scientism and nihilism. Habermas explains positivism is a form of scientism initiated by Auguste Comte (17891859) and then later expanded upon by Karl Popper (1902-1994), arguing that interpretations of events are meaningless as they cannot be observed, that theories are only useful as much as they provide falsifiable hypotheses and that philosophic reflection was insufficient compared to the scientific method, therefore positivism is morally uncertain. Popper argued that social moral norms could not be universally verified and must be accepted on faith. In his work Knowledge and Human Interest, Habermas calls Popper uncritical and reactionary. (46) He argues that norms can be evaluated for consistency and consequences, in order to further ends that can be tested against ideals and universal interests. Habermas realizes here that he has a problem of finding the ideals and interests; that humans have developed false desires in the modern cycle of work and consumption. He answers this problem with the idea that language is holistic and contextual. Rule governed speech acts that coordinate around a way of life become meaningful when they are structured, constituted 76 and identified by the intentions of an actor. Therefore, this intentional action can be understood and interpreted, within the context of a way of life. Language and action then express norms and interest that are publically demonstrable and capable of regulating private moral decisions. Habermas criticizes positivism as a contradiction. Since the idea of science as the only reliable source of knowledge is not a science itself; it fails its own criteria. Therefore, positivism does not justify science. Habermas wants to develop a science that explains the meaning of human actions; that can find a new way to gain an understanding of the action and intention. Nihilism is founded in opposition to the philosophies of Georg Hegel (1770-1831) and Karl Marx (1818-1883); they both argued reason to be emancipatory. Hegel argued reason to be the human spirit. Marx argued that reason with technology would lead to universal freedom. Friederich Nietzsche (1844-1900) would argue however that reason is but a tool of a will to power, to gain control of nature. He also claimed that by freeing agents from nature, reason binds agents to social laws; reason dominates! Nihilism implies that the will creates meaning and value, therefore nothing has intrinsic value and there is no meaning in experience. To gain freedom from the domination of reason, one must abandon reason’s control over action through morality and norm boundaries thereby shedding morality, one becomes elite. What occurs, in nihilism, is that instead of a reasonable science aimed at understanding the world of material and ideal, for the purpose of advancing the human species toward freedom, there is a scientific rational aimed at controlling humans and nature through natural behavioral science. Habermas argues that natural sciences cannot aid any understanding of morality and social actions on the part of the actor. While his most recent predecessors, such as Heidegger, had advised a morally relativistic new science approach that explains action based on each individual society’s use of science and technology, Habermas wants to employ a science that is more universal. His argument is that a moral cultural relativity leads to a lack of true critique. Standards of critique become subject to each circumstance, therefore each circumstance is understood by its own standard. Therefore, if a society were to choose a nihilistic mode of thought, and were to experiment on humans, against their will, then that society could claim justification based on their own moral standard, giving insight to cultural justification of actions but failing to understand any true general human moral standard. In order to have a science that can claim knowledge to aid the pursuit of finding grounding for a human moral standard, Habermas shows how it is possible to use critical theory with practical intent in the general interest of human beings. His first thesis in developing this science is, “the achievements of the transcendental subject have their basis in the natural history of the human species”. (Knowledge and Human Interests 312) This transcendental subject is the entire shared subjective experience; it is the subjective experience of the entire human species. In looking to human evolution and the emergence of culture, by combining the objective and subjective experience, there can be an understanding of the human nature and the human drive through reasoning to be free from nature. There are three categories of knowledge each of which is oriented by non-subjective values or interests. This knowledge is necessary for survival and possesses unique methods for refining basic kinds of survival based action. The human being is met with three basic survival challenges that correspond to the three knowledge categories: natural scarcity, absence of social coordination and social domination. These three challenges each in turn bring about certain interests: to gain technical control over the environment, to unite with others around common values, and freedom. These interests are then realized with certain kinds of action. Instrumental action which is experimental, is based on behavioral feedback, this action constitutes knowledge of a sense world of material objects. Communicative action finds mutual understanding of a social world of persons, norms, and meanings connected by a reflective narrative. Finally, critical reflection examines the causes of beliefs, to discover and remove impediments to knowledge. Therefore, from instrumental action actors receive material knowledge, from communicative action they form social knowledge, and from reflective 77 thought actors gain knowledge of what hinders knowledge. (KHI II) Reason serves to preserve the species and functions as an evolutionary adaptation. Here we can think of reason as philosophical reflection, gaining knowledge of knowledge, but, this reflection is also more than knowledge it is also practice. “In the power of self-reflection, knowledge and interest are one.” (KHI 314) Unconscious human desire coupled with conscious reflection shows an option. Instead of being compelled toward the unconscious desire, reflection frees the agent to choose an action based on values beyond instinctive survival. Therefore, Habermas also states that, “knowledge serves as an instrument of and transcends mere self-preservation”. (KHI 313) Instrumental actions serve an experimental role. For instance, C then E is a statement of knowledge that cannot be verified unless you remove C in order to observe whether you still have E. Observing in itself is not enough to gain knowledge that is reliable. Habermas is showing the validity of applied reason in scientific knowledge. “Meanings of properties are defined in terms of instrumental action.” (KHI 130) Instrumental action establishes powers and properties that distinguish causal knowledge from random association through habit, by employing the experimental method over time actors come to have genuine knowledge constituting a transcendent material world. Through instrumental action, knowledge is also obtained as to how the world of material objects responds to what the actor does; it is dialectic of material well-being between subject and object. Knowledge that is gained is then applied to unburden agents of functions; imposing them unto technological instruments. The invention of the wheel is an example of a result of this kind of action. Instead of having to carry my load on my back, I can load a cart that is wheeled and push or pull it. This kind of technological progress necessitates a community of investigators which is held together by communication that tests hypothesis for results to reach accepted conclusions. (Ingram) Communicative action is how mutual agreement and understandings of meanings, values and norms constitute identification, in which the self requires others in recognition to form a biography. These biographies apply to both the entire human condition as well as the context of the intentions of the individual agents involved. These intentions need to be understood to understand the meaning of action. Understanding action creates familiarity of the new by proxy to what is already known through communication between an internal self and self-story as well as, an external communication between the self and others. Communicative action forms social worlds of social organization. Within this social organization Habermas says that “knowledge constitutive interest take form in the medium of work, language and power.” (KHI 313) Work corresponds to instrumental action, language to communicative action, and power to reflection. “The structure of communication anticipates truth, freedom, justice and happiness.” (KHI 314) Language is communicative action involving agents who are free of distortion with the use of reason in the aim of reaching understanding through argumentation that raises validity claims. When an actor speaks they are speaking to be understood, they must say something that can be understood, therefore they must attempt to be understood and in turn reach a mutual agreement with another. Communication can continue only if validity claims that are mutually argued are understood to be justified. (What is Universal Pragmatics? 2) If an actor thinks validity claims to be justified, they engage in discourse or, argumentative speech in which, they give reasons for the claim. Then the hearing actors in this situation can then either accept the reasons or give his/her own reasons against them for understanding. This argumentative discourse is used “to bring about an agreement that terminates in the intersubjective mutuality of reciprocal understanding, shared knowledge, mutual trust and accord with one another.” (WUP 3) Through argumentation the actors realize one material world as well as the transcendence or interchangeability of the personal experience of phenomena. From argumentative speech there is explicit then, a communicative rationality that is the use of reason in having knowledge of what is sound in an argument. This rationality is employed to judge claims, as well as defend them against rational criticism. Therefore, arguments can be made beyond the material into moral claims based on communicative action; these claims can be judged rationally, 78 can satisfy reason, or be argued about until mutually reasonable and agreeable to all parties engaged and then be defended by reason leading to a consensus of the masses. Habermas thinks he has grounded his new science that looks to reflect philosophically on communicative and instrumental actions in moral terms. (Ingram) Applying his social critique to society in its current form, there is an argument that the communicative consensus is constrained, because the state has become an agent of a ruling class that through ideology and violence exercises power over the ruled. Traditions have come to conceal the deviation from ideals implicit in communication such as truth, freedom and justice. Communication implies equality of all parties involved since all are engaged in the process of coming to understanding. Traditions and ideologies serve to undermine equality. Religion for example creates a tradition that validates hierarchy and domination such as in medieval kingdoms. Capitalism then abolishes that validation, with the free market that supposes the ideal of universal equality, in which wages become a communicative action of equal exchange for mutual benefit. Ideals of freedom in communication are still lost in capitalism however, as we see, freedom, equality and happiness are the general interest of the social agent, but through consumer freedom and fulfillment, private property and wage labor have created relations between people that are exploitive, undemocratic, unjust and detrimental. Competition breeds inequality because there has to be a loser. In order to win certain values such as honesty and fairness may have to be shed. Hiring the cheapest labor and selling products at the highest value possible are neither fair nor honest. A visit to any of the websites of organizations such as: the United Nations, the International Monetary Fund, the World Bank and Human Rights Watch will provide evidence of vast disparity that exists globally; starvation and disease from extreme poverty and grave social injustices including violence and subjugation are found to be imposed upon many peoples. Habermas argues three major points in his article, Political Communication in Media Society. His first issue is a lack of face-to-face interaction between present participants in a shared practice of collective decision making. He follows that there is a lack of reciprocity between the roles of speakers and addressees in an egalitarian exchange of claims and opinions. Finally, arguing that the dynamics of mass communication are driven by the power of the media to select, and shape the presentation of, messages and by the strategic use of political and social power to influence the agendas as well as the triggering and framing of public issues. (PCMS 14-15) In the globalized public sphere, political and social powers are exercised with the backing of economic power at a near universal level, taking place in the realms of campaign finance and commercial advertising especially in the more economically and technologically developed nations. Economically powerful interests, through their own capital, influence what the audience will perceive their choices to be. Therefore, the economic power structure of the public sphere may well distort the dynamics of mass communications. (PCMS 418) Furthermore, social deprivation and cultural exclusion of citizens explain the selective access to and uneven participation in mass communication, whereas the colonization of the public sphere by market imperatives leads to a peculiar paralysis of civil society. (PCMS 421-422) What then is necessary to promote further removal of distortions and limitations to communication for the purpose of removing roadblocks to freedom and happiness? Habermas argues that, “we should not seek the explanation in the paralyzed state of civil society but in the content and formats of a degenerating kind of political communication itself,” pointing out that, “issues of political discourse become assimilated into and absorbed by the modes and contents of entertainment and advertising; besides personalization, the dramatization of events, the simplification of complex matters, and the vivid polarization of conflicts promote civic privatism and a mood of anti-politics”. (PCMS 422) In expanding Habermas’ argument to include the use of internet technologies I argue that while Habermas is correct in pointing to the connection between entertainment and politics, the argument should be that forms of entertainment and social networks can be formats for political 79 debate; the issue is still that economic power is the cause of any inequality in those formats. Who gets heard is whoever can afford to pay to be heard. While it is true that there is a large amount of anonymity online, allowing agents to be free from fear, it is also recognized then, that this anonymity can be used to distort information. Instead of pointing to the anonymous aspect as problematic the focus should be on the reasons for the possibility of distortions, namely power relations relating to economics and politics. Mass communication has proven itself to be a social power in the face of economic and political power. For example, social media networks aided the organization of people’s risings in: Egypt, Libya, Tunisia and Syria during the recent Arab Spring, although the outcomes of these situation are quite unsure. Political issues are debated between every-day people on a daily basis, from war to homosexual marriage, nearly the world over. The obstacles to consensus are not in the format of communication, the disconnection is between the people and their policy makers. Looking back to the invasion of Iraq by the US in 2003, there was a breakdown in the discourse concerning the debate over the invasion. Almost unanimously there was no news media network that spoke against the war, in circumstances where people did show dissent they were silenced and ignored. An example is Phil Donahue from MSNBC, a news commentator who was fired for expressing his objection to the beginning of hostilities against Iraq and Afghanistan following the 9/11 terrorist attacks. There was a state sanctioned control over communication to sway perception of the circumstances. Therefore, while mass communication is a power in the face of power, it is still one that can be managed by a controlling class working in their own interest and not those of the human species, gaining personal profit at the expense of the freedom of others. The removal of powerful economic interest from social and political communication networks is necessary to alleviate the skewed power relations that allow this kind of control to occur. Communication is an exchange of ideas that implies equality, when the very issues debated are controlled this equality is lost. The internet does provide an avenue to expand upon debated issues with hopes of restoring equality. The use of the internet in some respect allows a level of equality of opportunity of exchange, yet, fails to grant a total equality. I say there is the opportunity for equality because everyone who can gain access to the internet can make their mark in cyberspace; this does not guarantee an audience however. True face to face contact would. Finally, communication has taken one wound for sure in the deterioration of language in itself that is beginning to occur with the growth of net-speak such as, lol, hmu, smh, r u thr, so on... I would argue that this “on-the-go” lifestyle that has taken over portions of the developed world has caused the issue where our culture of technology allows people to be so connected to everything around the clock; from work to the news, from family to personal relations that many begin to feel so pressed for time that even the language used suffers. The human species survives based upon its ability to gain information, judge that information and act upon it in a continuous and cyclical system that enhances techniques to finding information, deciding what information is valuable and how best to act upon that information. The systems of enhancement are held together by their deliberative fashion utilizing communication to build consensus. Since communication is universal among humans, and it is utilized to build mutual understanding and trust in its conception, it is a basis for establishing a universal code of ethics. A universal code of ethics can be set forth through communication, reason and evidence. Gauging the level of communication and exchange of real information gives insight to the health of a society. In this sense communicative action, as theory, provides a comprehensive tool that can bring insight to issues including: economic imbalances, power struggles, and distortions to what is knowledge in itself. 80 Works Cited Descartes, Rene. “Meditations on First Philosophy.” n.d. Wright.edu. web. 28 12 2005. Habermas, Jurgen. Knowledge and Human Interest. Boston: Beacon Press, 1971. print. —. “Political Communication in Media Society: Does Democracy Still Enjoy an Epistemic Dimension? The Impact of Normative Theory on Empirical Research.” Communication Theory vol. 16 from the International Communication Association (2006): 411-426. web. —. “What is Universal Pragmatics from” Communication and the Evolution of Society. Boston: Beacon Press, 1979. print. Hegel, Georg Wilhelm Friedrick. “The Philosophy of Right.” McNeil, William. Continental Philosophy. Oxford: Blackwell, 2005. 205-214. print. Hume, David. An enquiry Concerning Human Understanding. Oxford: Oxford Press, 1777. web. 2/16/2012. Ingram, David. Habermas: Introduction and Analysis. New York: Cornell Press, 2010. Print. Kant, Immanuel. Critique of Pure Reason. Koenigsberg, 1787. web. Key of abbreviated terms: KHI - Knowledge and Human Interest PCMS - Political Communication in Mass Society 81 Pediatrics: Because We Care About the ‘Little World’ by Syeda Khalid (English 1102) T he young boy cries profusely in pain, while the parents stand frightened at the sight. Handing in a lollipop with a smile of compassion, the pediatrician calms the young boy to find out where truly the problem lies. The pediatrician knows exactly what she is doing. Sure about bacterial infection, she prescribes some medicines and says farewell to the boy with a hope to see him with a big happy smile the next time they meet. This is what a pediatrician is here for, to care for the little troubles of the little ones. Pediatrics is a branch of medicine, a specialty, which deals with the health care of infants, children and adolescents. It is a “young specialty, scarcely 200 years old” (Luecke 56). The first separate pediatrics hospital was founded in 1855 in Philadelphia. However, the father of pediatrics in the United States is considered a German pediatrician, Dr. Abraham Jacobi. In 1861, Dr. Jacobi formed pediatrics chair at New York Medical College as well as organized numerous pediatric societies, started a number of pediatric journals, and established children's departments in various New York hospitals. According to the Illinois Career Information System (CIS), a pediatrician’s main focus is on “preventing problems before they begin” ("Pediatricians" 1). They see the youngest patients for frequently scheduled visits—such as well-baby checks where the child’s growth, weight and development are examined. Working at a hospital or a private clinic, pediatricians educate and advise about immunization, exercise, hygiene and diet. They may “order lab tests….explain test results and review treatment options.” In case of serious health conditions, a pediatrician may refer the pediatric patient to a health care specialist for further treatment. Other than seeing their patients, pediatricians may “assign tasks to nurses and other health care workers, keep detailed records about each patient, and write reports for insurance companies and government agencies.” Some pediatricians also choose to teach at medical schools. As specified by the Illinois CIS, the “median wage for pediatrician is $167, 640 per year;” however, “self-employed pediatricians generally earn more than those who are not self-employed” (“Pediatricians” 1). Along with wages, pediatricians also receive job benefits like “sick leave, paid vacation, health insurance, and a retirement plan.” While in the United States, the expected growth rate for pediatricians is 24.4%; in Illinois alone, “faster than average employment growth is expected through the year 2020 for pediatricians” (3). The outlook for pediatricians is promising as the demand for them is increasing with the growing population and expansion in healthcare industries. Pediatrics has arduous education and training requirements. To work as a pediatrician, a person needs to complete four years of medical school to get a Doctor of Medicine (MD) degree or a Doctor of Osteopathic Medicine (DOM) degree (“Pediatricians” 6). After medical school, a state licensing exam needs to be passed. In Illinois, pediatricians must be licensed by the Illinois Department of Financial and Professional Regulation to practice (8). Once licensed, a residency program in pediatrics needs to be completed. This usually lasts three to four years when residents work in a hospital. In his book, Careers in Medicine, talking about pediatrics in general, Terence Sacks declares that there are “two hundred [and] two training programs in pediatrics involving a three-year residency” (140). During and after residency, several exams have to be taken in order to become board certified. After the completion of residency, pediatricians can pursue fellowships in various 82 subspecialties that take another three years. Subspecialties include adolescent medicine, sports medicine, pediatric cardiology, pediatric critical care, pediatric endocrinology etc. For example, sports medicine only focuses on the “care of children with medical problems involving exercise and recreational competitive sports” (Sacks 141). Just like a residency, there are various training programs offered for each pediatric subspecialty. For example adolescent medicine has twenty-five training programs offered, pediatric cardiology and pediatric critical care has forty-eight, pediatric emergency medicine offers forty-three, pediatric endocrinology has sixty-two, pediatric gastroenterology has fifty-one, while neonatal and prenatal medicine offers as many as ninety-six training programs. Many fields have been ignored by doctors as a possible subspecialty due to unawareness. More pediatricians tend to focus on the common areas of specialty and tend not to explore the rare. In the article, “The Ambulatory Pediatric Association Fellowship In Pediatric Environmental Health: A 5-Year Assessment,” the authors mention that because environmental pollution plays a big role in pediatric disorders and other health problems, there is a need for more pediatricians to subspecialize in environmental health. Despite of the growing diseases with environmental origin among children like asthma, Sudden Infant Death Syndrome (SIDS) and neurodevelopmental disorder, only a few pediatricians are trained in this area. Many pediatricians even report that they “frequently encounter diseases that appear to be initiated by environmental factors, but most report discomfort and lack of information in dealing with these conditions” (Landrigan, et al. 1384). To counter this problem, the Ambulatory Pediatric Association (APA) launched a three year fellowship in Pediatric Environmental Health (PEH) in various medical schools across the country. The APA fellowship program in PEH is proving to be successful in “preparing pediatricians for leadership careers in PEH.” Just like Environmental Health, there are many such areas that might need more doctors due to possibly more health problems associated with them, and therefore awareness about such areas need to be created. For example, areas like pediatric urology are also experiencing shortages and it is highly recommended that doctors consider these areas to further their training. Not only is education and training a must but certain skills and personal characteristics may also play a big role in making someone a successful pediatrician. According to Illinois CIS, one of the most important skills required by a pediatrician is communication. It is not only essential for them to “express ideas clearly when speaking or writing,” but it is even more crucial that they are able to “listen to [the patient], understand, and ask questions” (“Pediatricians” 4). Claiming that listening skills are very important, Dr. Ruth Kim, a pediatrician, adds: “Lots of times the parents give you hints and clues, but they won’t come out and say what’s really bothering them” (qtd. in “Pediatricians” 12). For example, they might say, “I don't understand. My child is always tired and complains of stomach aches, and won’t eat dinner.” It is the duty of the pediatrician to also look at the young patient’s emotional health as the problem might actually be depression as “children suffer from depression more often than you might think.” It is a task for a pediatrician to, through conversations, spot where the problem lies. Additional important skills that help a pediatrician are reason and problem solving. Dr. Kim believes that in this profession “[y]ou’re like a detective” (qtd. in “Pediatricians” 11). “You have to be thorough and look over the child, because they rarely possess the vocabulary to tell you what’s going on.” Sometimes the cause of a problem is hidden and being a doctor you need to be vigilant enough to find it and cure it. After a possible problem is diagnosed, it is then important for them to “judge the costs and benefits of a possible action” through reason (“Pediatricians” 4). Apart from skills, an ultimate personal trait for a successful career as a pediatrician is having love and patience for children. Because this field deals with children, it is important that a pediatrician is comfortable working with them. It is important to keep in mind that children can be hyper active and stubborn, so as a pediatrician you should be ready to deal with them (Sacks 140). Dr. Kim describes how she decided to specialize in pediatrics while completing her rotation in 83 hospital's pediatric unit: “[I discovered] that I'm pretty good with kids, personality-wise. I'm able to joke and play with them even when they're feeling pretty miserable. That's my favorite part of the job, talking and interacting with the kids, and helping them feel better” (qtd. In “Pediatricians” 11). She further adds that sometimes it may be frustrating as “you can get pooped or peed on. It can escalate to an ugly situation if you're unable to stay level-headed. It's an endurance test” (12). At the end, a pediatrician needs to demonstrate his/her patience towards the innocent children. Though it seems promising, choosing pediatrics as a career has some negative consequences. As this field requires education and training requirements in school and beyond, it may seem exhausting. Dr. Nazjabeen Qureshi, an Attending Physician at Women & Children's Hospital of Buffalo in New York state, mentions that this field requires lifelong learning and many people are not ready for that. Dr. Ruth Kim, a pediatrician, also agrees that to become a doctor “you have to buckle down and study” (qtd. in “Pediatricians”). Not only does a person striving to become a pediatrician has to study to pass exams and enter the field but with advances in medicine, and new diseases and treatments being introduced; even a practicing pediatrician may find himself/herself in a never ending cycle of learning. Not only is lifelong learning overwhelming but the work load is also considered a negative side of this occupation. In his book, Sacks mentions that since children are more disposed to becoming ill at any time of the day or night, a pediatrician’s working schedule is tough and irregular. They generally have to work long hours, “averaging a 58.6-hour work-week, with about 65 percent of that that time spent in the office, 19 percent in the hospital, and 16 percent in surgical and other procedures” (Sacks 140). This is further testified to by Dr. Qureshi who agrees and adds that one has to especially expect exhaustion to be an unceasing element throughout residency due to the long working hours and extreme work load. Dr. Kim also adds, “Don’t believe everything you see on shows like ER. Know what you are in for. There are aspects of this job that money doesn’t compensate for. The hours are long, and there are heartaches,” she adds (qtd. in “Pediatricians”12). The nature of the work is such that pediatricians even have to “work on weekends and nights when on call,” leaving them with minimal leisure time (4). Extreme workload imposed on pediatricians is also a reason why doctors experience a high rate of burnout cases. A physician, Dr. Diane Shannon, talks about physician burnout in her article “Why I Left Medicine: A Burnt-Out Doctor’s Decision to Quit.” In this article Dr. Shannon shares her experience of exhaustion while practicing that finally led her to quit from the career. She narrates: As an internist, working in adult outpatient clinics around Boston, I had trouble leaving my work at work. I’d go for a run and spend the entire 30 minutes wondering if I’d ordered the right diagnostic test. I suffered from chronic early morning wakening, even on my weekends off. I startled easily. I found it impossible to relax. I worried constantly that I’d make a mistake, like ordering the wrong dosage of a medication, or that a system flaw, like an abnormal lab report getting overlooked, would harm a patient. I no longer remembered the joy I’d felt when I first began medical school, and I couldn’t imagine surviving life as a doctor. Long working hours, time pressure, a fast work pace and the high level of chaos have all been linked to high rates of burnouts experienced by physicians. Another aspect of this occupation that some people might find disturbing is cases of child abuse. Such cases are very sensitive and dealing with the children is a challenge. Dr. Kim claims that there are around five child abuse cases that she has to handle in a year. These cases are required to be reported to authorities immediately, and the young victim may have to be hospitalized (“Pediatricians”12). The children of the child abuse cases are very sensitive and a pediatrician needs 84 to be extra careful in trying to help them recover from what they have faced. This may be emotionally disturbing for pediatricians themselves. Nevertheless, the “joyous part is helping kids recuperate and seeing them get better” (Kim qtd. in “Pediatricians”12). Legal decision-making has also evolved into a morally controversial topic in the field of medicine. With the “emergence of pediatric palliative care as an interdisciplinary subspecialty” comes the question of who decides for the treatment of seriously ill, chronic or end-of-life pediatric patients (Berlinger, Barfield and Fleischman 789). In the article, “Facing Persistent Challenges in Pediatric Decision-Making: New Hastings Center Guidelines” by Nancy Berlinger, Raymond Barfield and Alan R. Fleischman, published in the official journal of American Academy of Pediatrics, Pediatrics, the authors talk about the right of the patients and their parents to be involved while making decisions about a treatment. Since an infant or a young child might not be able to make a decision for himself/herself, it is important that parents are fully aware of the consequences of a treatment. An older child or an adolescent, on the other hand, might have preferences, fears or concerns that are significant when deciding on a treatment. Sometimes doctors “save the life of a child without sufficient regard for the profound consequences on a child’s subsequent quality of life” (Berlinger, Barfield and Fleischman 789). For example, sometimes death might be preferable than a long life being paralyzed. This is a decision where the pediatric patient and the caregivers should have a say. In order to make sure that patients make a decision in their best-interest, it is important that they should be fully informed. In his article, “When Doctors Need to Lie,” featured in The New York Times, Dr. Sandeep Jauhar talks about yet another moral dilemma often faced by doctors: whether to inform a patient about their health condition or to lie, knowing the shock it can cause to the respective patient. Dr. Jauhar claims that “the moral basis for withholding information from such a patient is clear: Above all, physicians must do no harm” (SR.9). But is it not a harm to a patient that his right to make an informed decision for himself is being taken away? While Berlinger, Barfield and Fleischman talk in favor of patients making a decision for themselves, like Dr. Jauhar, they also acknowledge the fact that a patient “might not be able to grasp all consequences of a choice” (790). At this point a doctor might have to make the ultimate decision. Dr. Jauhar talks about his own personal experience with a patient who received a stent to open up a blocked coronary artery. After a few days, he started to bleed into his lungs. “He needed to be intubated with a breathing tube or he was going to die. However, I was informed that he had told doctors that he never wanted to be intubated,” explains Dr. Jauhar (SR.9). “I didn’t know the quality of the discussion he’d had with the other doctors, and I couldn’t talk to him because he was nearly unconscious from lack of oxygen. So with a troubled heart I intubated him.” Within few days, the patient’s condition improved and the breathing tube was removed. The life of the patient was saved, but was a decision ethically and morally right? Should doctors go against a patient’s will? Should they lie to the patient? How different is it if the patient was a child and not an adult? These are debatable questions. The case of a child is different and the parent's say matters. There have been many cases of parents rejecting medical treatment due to religious beliefs. But can the parents’ decision be repealed? Courts have been frequently seen ordering “life-saving blood transfusions for the children of Jehovah’s Witnesses, or cancer treatment against parents’ wishes” (Hall). Yet a majority of “30 states still have religious shield laws, and every state but Mississippi and West Virginia allows religious and/or philosophical exemptions for school vaccination requirements.” Even The Affordable Care Act (Obamacare) demands “insurance companies to cover ‘nonmedical’ health care such as prayers by Christian Science practitioners.” The related laws have been changing, but it all comes down to the interpretation of morals, ethics and priorities. For those in the field, The American Academy of Pediatrics (AAP) is an American professional association for pediatricians, founded in 1930. According to its website, AAP is “an organization of 62,000 pediatricians committed to the optimal physical, mental, and social health and 85 well-being for all infants, children, adolescents, and young adults” (American Academy of Pediatrics). The annual membership fee ranges from $100 to $485, according to the pediatrician’s designation and services provided. Conveniently for medical students, the annual membership fee is only $20. AAP has two journals, Pediatrics and Hospital Pediatrics. Pediatrics is AAP’s official journal, which is “the most cited journal in the field, it is among the top 100 most-cited journals in all of science and medicine.” This association also conducts a conference and exhibition annually that can be attended by non-members. AAP helps its members find a job; become part of committees, councils and sections dedicated to specifics issues; and advocate for patient’s health. It also recognizes professionals who make a difference in pediatrics through AAP awards. An interesting fact about the occupation of pediatricians is that they are not just doctors, but advocates striving for the welfare of children. In his article, “Pediatric Advocacy: Yesterday, Today, and Tomorrow,” Charles Oberg explains that because “children have little political voice of their own [they have to] rely on the proxy voice of others including pediatricians to speak out on their behalf” (406). Some examples of pediatricians serving as advocates include building safe playgrounds with schools; starting reading programs with libraries; influencing law-making agencies to ban lead-based paint in homes, requiring smoke detectors in apartment buildings and also suggesting teen driving rules (American Academy of Pediatrics). AAP has also been working with “government, communities and other national organizations to shape many child health and safety issues.” Pediatricians are not only there to cure diseases but also to make sure that the little things that may affect the children are resolved. When the pediatrician looks at the young little boy; back with a bright big smile, recovered from the bacterial infection that brought tears into his small, beautiful eyes; the world seems so much better. That is the beauty of this field: bringing even little differences in the lives of young children brings tranquility to the heart of a pediatrician. With a great outlook and demand, the field of pediatrics may be the right choice for the compassionate and determined individuals out there considering Pre-Medicine as a major. Works Cited American Academy of Pediatrics. American Academy of Pediatrics, 2014. Web. 3 Apr. 2014. Berlinger, Nancy, Raymond Barfield, and Alan R. Fleischman. “Facing Persistent Challenges in Pediatric Decision - Making: New Hastings Center Guidelines.” Pediatrics 132.5 (2013): 789-791. Academic Search Complete. Web. 4 Apr. 2014. Hall, Harriet. “Faith Healing: Religious Freedom vs. Child Protection.” Science-Based Medicine. Science-Based Medicine, 2013. Web. 20 Apr. 2014. Jauhar, Sandeep. “When Doctors Need to Lie.” New York Times, Late Edition (East Coast) ed. 23 Feb. 2014: SR.9. ProQuest. Web. 4 Apr. 2014. Landrigan, Philip J., et al. “The Ambulatory Pediatric Association Fellowship in Pediatric Environmental Health: A 5-Year Assessment.” Environmental Health Perspectives 115.10 (2007): 1383-1387. Academic Search Complete. Web. 15 Mar. 2014. Luecke, Percy. “The history of pediatrics at Baylor University Medical Center.” Baylor University Medical Center Proceedings 17.1 (2004): 56-60. Pub Med--National Library of Medicine. Web. 3 Apr. 2014. Oberg, Charles. “Pediatric Advocacy: Yesterday, Today, and Tomorrow.” Pediatrics 112.2 (2003): 406-409. Academic Search Complete. Web. 4 Apr. 2014. 86 “Pediatricians.” Illinois Career Information System. U of Oregon, 2014. Web. 3 Feb. 2014. Qureshi, Nazjabeen. Personal interview. 2 Apr. 2014. Sacks, Terence. Careers in Medicine. New York: McGraw-Hill, 2006. Print. Shannon, Diane. “Why I Left Medicine: A Burnt-Out Doctor’s Decision to Quit.” Common Health. Boston University, n.d. Web. 3 Apr. 2014. 87 Physical Therapy: A Profession with a Promising Future by Marie Labbe (English 1102) P hysical Therapy is a growing medical profession, which I was drawn to especially because it consists of working closely with and helping a variety of people while earning an above average pay. The more I researched the profession, the more I became interested. I found that the profession fit my personality and my qualities well, and most enticing to me is the Early Intervention specialization with babies and children. Physical Therapy, a sophisticated science and a highly desired field that requires extended education and training, is an important and necessary medical profession that has evolved from various cultural practices from many generations and that concentrates on enhancing patients’ lives by strengthening their physical capacities and overall wellness. Physical Therapy recently became its own medical field wherein the practices come from different cultures and times; the practitioners are highly qualified and often possess specific personality traits. The education for Physical Therapists (PTs) has also evolved and is now very comprehensive and involved; the specializations are varied and many studies show that the effect of physical therapy can be greater than just improvement in physical wellness. Physical Therapy developed over time into a more science-based practice. Bernice Krumhansl, in her book called Opportunities in Physical Therapy Careers, claims that, in England, toward the end of the nineteenth century, orthopedists started training young women who were physical education (PE) graduates, to give muscular reeducation treatments to patients (10). Mark Drnach, in his book titled The Clinical Practice of Pediatric Physical Therapy states that in 1894, in the U.S. the first physical therapists were also women who were PE teachers but they were trained to help with the poliomyelitis epidemic, which affects mainly babies and young children (3). Mary McMillan, an American trained in England is known as the first American physical therapist; her training and muscle reeducation continued to be very important during World War I, when many soldiers were injured and impaired (Krumhansl 4). Since this type of care for the wounded did not belong in any other medical department, the Surgeon General of the US Army Medical Corp formed, at the beginning of World War I, a new Division of Special Hospitals and Physical Reconstruction (4). Marguerite Sanderson started the first physical therapy course at Walter Reed Hospital, and then 14 other hospitals opened physical therapy programs all over the country (11). During World War I these 14 schools trained 800 physical education teachers in these muscle-reeducation techniques and were called reconstruction aides (11). At the end of World War I these Reconstruction Aides returned to their former work because their new work was not yet established as a specific field. On January 15, 1921, a few of the Reconstruction Aides met in New York City and inaugurated the American Women’s Physical Therapeutic Association and elected Mary McMillan as their president. There were 245 members. In 1921, the association started publishing a journal quarterly regarding developments in physical therapy (Krumhansl 12). Trisha Hawkins states in her book Careers in Physical Therapy that by the end of the 1930s there were a thousand members and men were joining the association, so they changed its name to American Physiotherapy Association (3). In 1925, physical therapy was recognized by the American Medical Association as a specific medical field. The Council on Physical Therapy was established and it outlined a nine month course to be offered in medical schools for the graduate nurses and PE graduates. Before long, “sixteen schools were graduating 135 students a year.…The field of physical therapy was gaining momentum” 88 (Krumhansl 12). There were about 1,000 physical therapists in the U.S. In the 1940’s and 1950’s the casualties of the war and the many victims of the polio epidemic again required the development of many more physical therapists and proved how strongly needed they were. Army hospitals and private hospitals started offering courses in physical therapy and most civilian schools accelerated their programs. In 1946, at the end of the war, most accelerated programs were stopped; however, 21 schools could educate 480 PT students. The polio epidemic of the 1940’s and 1950’s was the worst in history; there were approximately 14,500 people affected. Physical therapists were so important in treating polio patients that “[i]n 1945 the National Foundation for Infantile Paralysis gave the American Physiotherapy Association $1,267,000 for two thousand scholarships to train physical therapists and other necessary personnel” (13). In 1946 the APA became known as the American Physical Therapy Association (Drnach 4). The 1940’s and 1950’s were important years for the development of the physical therapy field. Trisha Hawkins explains in Careers in Physical Therapy that practitioners were highly needed during the war and the epidemic and continued to be even after (3). Physical Therapy was finally an independent medical field and the association kept getting stronger. Today this medical field is still growing and developing, and the professions in physical therapy are highly sought after. Physical therapy is relatively new as a profession but is based on various practices used over thousands of years in different cultures. As Hawkins describes it, “Physical therapists (PTs) are the heirs to the ancient tradition of healing by physical methods. At the same time, they are modern health professionals—highly skilled, educated, and trained to evaluate and treat health and movement problems that result from injury and disease, as well as disorders present at birth” (5). Therese Harasymiw, in her book A Career as a Physical Therapist, explains that heat may be one of the oldest modalities used by man to relieve pain. Heat helps with relaxation of muscles and blood circulation (23). Warm towels or heat packs have been commonly used. Another heat method used by the Greeks was fever treatment as therapy. They would cure diseases by raising the body temperature of the patients. The Romans discovered that the application of hot wax could relieve pain (Krumhansl 3). Ultrasound is a modern use of heat therapy. In the beginning of World War II, German scientists were trying to find a way to use sound waves for therapy. By the mid-1950s ultrasound was developed and used as a physical therapy treatment. The high frequency sound waves are used to reach through the skin to the internal body tissues to promote healing (3). Even though it would seem that cold therapy was used as early as heat, it was not because it could not be made without ice or snow. Russians, therefore, obviously were able to use it more frequently. They used ice therapy to “reduce high fevers, to control infections, to treat wounds and to treat meningitis. It was also used to treat diseases of the central nervous system” (4). Ice packs are now most often used to reduce pain and swelling in topical areas. Cold therapy can also be referred to as cryotherapy (Harasymiw 23). Water therapy is similar to that of heat and cold therapy; it has also been used for generations to relax muscles through temperature. However, hydrotherapy is uniquely used for its power to weaken the force of gravity, which can relieve the stress on weight bearing joints and muscles. Water therapy is used to “stimulate the immune system, slow down the production of stress hormones, invigorate blood circulation and digestion and lessen inflammation” (Krumhansl 5-6). Electrotherapy was also used in ancient times. In Rome, electric eels were used for arthritis treatment (Harasymiw 23). There are now many uses for electric currents including pain control reducing muscle atrophy and accelerating healing by drawing blood to the tissue (25). Massage and exercise are the most commonly used treatments in physical therapy and have been used to improve health in many ways. Massage and exercise are an innate human response to pain. People naturally try to bend their bodies or rub and massage muscles to relieve pain. Massage and exercise can also be used to treat illnesses. The first to help with the recovery of fractures, 89 dislocation or other injuries were ancient Greek gymnasts. The realization of the benefits of exercise came in the 17th century in England and the understanding of the relation between massage, blood circulation and general health came in the 18th century in France and Germany (Krumhansl 8). Most therapeutic exercises used today to promote health come from Pehr Henry Ling of Sweden, who showed that properly done exercise could heal diseases and body dysfunctions. He was the founder of the first institute for medical and orthopedic gymnastics. The Swedish system introduced two types of exercise, resistive exercises and isometric exercises. They are muscle-building exercises that consist of squeezing against resistance without movement (9). The effectiveness of massage, acupuncture and acupressure are subject to controversies because they are based on the theory that “[t]he central nervous system and the circulatory system connect internal organs to the skin and all points in between. Pressure applied to certain points in the body will result in electrical messages or bodily fluids carried to other part” (10). The work of a physical therapist often consists of introducing the patient to exercises and teaching them how to perform them properly. There are different types of exercises depending on the ability of the patient to move. Some patients might not be able to move at all, while some have very little strength and others just need to strengthen a little bit. Some exercises are meant for patients to do on their own and others are meant for the therapists to perform on patients with little to no help from them. If patients are able to do exercises on their own, physical therapists can request that they perform the exercises on their own at home and teach them how to properly do so (Harasymiw 19). Physical therapists can work in many different settings; they can work in hospitals, in rehabilitation centers, in outpatient clinics, in private practices, in home health agencies, in schools, in nursing homes, and even in emergency rooms. Physical therapists work in general about 40 hours a week but the hours can widely vary from one therapist to another. Katie Guimon, a physical therapist at Cadence Health Hospital in Winfield, Illinois, works about ten hours a day and four days a week, which is the average 40 hours a week. Michele Neidlinger, a physical therapist in early intervention whom I interviewed, works about ten to twenty hours a week as an independent contractor and two 10-hour days a month at the hospital. The pay for a physical therapist varies from the setting they work in, the state, and seniority. In the United States a physical therapist earns on average $38.39 an hour and $79,860 a year (United States. Dept. of Labor Bureau of Labor Statistics). In Illinois the average pay of a physical therapist is slightly lower, $36.87 an hour and $76,700 a year (“Summary Report”). Benefits vary depending on the setting that physical therapists work in. Neidlinger, for example, receives health insurance, money for continuing education, and a 401K from working at the hospital, but doesn’t receive any benefits as an independent contractor. Physical therapists are highly educated. To be a qualified physical therapist, one must earn a Master’s or a Doctoral degree in Physical Therapy and successfully pass the national exam called the National Physical Therapy Exam (NPTE). The programs vary depending on the school but the classes required before entering the program often include “psychology, biology, physics, chemistry, statistics, English, professional writing, and humanities. After you are accepted in the program the courses will become more focused on the physical therapy practice” (“Become a PhysicalTherapist”). Neidlinger became a physical therapist with only a Master’s degree but she has been practicing for fourteen years. Now it is getting harder to become a physical therapist with only a Master’s Degree and by 2017 that won’t be possible anymore. All physical therapy students will then be required to get a Doctoral Degree in order to become physical therapists (American Physical Therapy Association). The American Physical Therapy Association (APTA) is the national association of physical therapy; it includes 88,000 members including practicing and non-practicing physical therapists, physical therapy assistants (PTAs), and physical therapy students. The association holds three national conferences a year, publishes a journal every month called the American Physical Therapy 90 Journal and regularly updates its website www.apta.org. According to Neidlinger, she is no longer an APTA member. She explained that “[t]he fees were high; about 200 dollars a year, and although it was really helpful for [her] research at school, it wasn’t that helpful after and [she] didn’t have time to read the journals.” Physical therapy has a bright outlook; it’s a growing field. As most professions in medical fields, physical therapy ensures a job security because even during recession medical bills are one of the last ones that can be eliminated. The baby boom generation is getting older and is more and more often in need physical therapy treatments (United States. Dept. of Labor Bureau of Labor Statistics). Camden Flath, in Therapy Jobs in Educational Settings: Speech, Physical, Occupational, & Audiology, claims that the need for physical therapists is growing in schools too (54). However, the fact that the field is growing and new advancements are regularly made means that physical therapists have to keep learning as they practice. Neidlinger explained that PTs have to take 40 hours of class and pass a test every two years. An article by Jeffrey W. Hathaway on PhysicalTherapist.com cautions about the fact that even though the future looks good the issue of reimbursements can be a problem. Health insurance doesn’t always cover all the recommended treatments and patients aren’t always capable of paying themselves. Some patients don’t receive all the necessary treatments because of that. Like other medical fields, there are specializations in physical therapy. Many physical therapists work in a special area of the field, but not many are certified in their specialization. Neidlinger states that “[she is] specialized in pediatrics but [is] not certified because it takes a lot of time and more education. The board exam is supposedly very hard and very specific. You really have to be an expert in your specialization to pass that exam.” PTs may specialize in orthopedic medicine, sports medicine, neurology, cardiopulmonary medicine, electrophysiology, pediatric medicine, geriatric medicine, or women’s health (Harasymiw 11-17). “Physical therapists who choose to specialize are prepared to treat patients within that area of medicine,” notes Harasymiw (11). “They study more about certain parts and functions of the body as well as specific disorders, disabilities, and diseases” (11). The job of physical therapists is to evaluate their patients, find the most appropriate treatment therapy, conduct the therapy and evaluate the progress of the patients. Hawkins explains that to evaluate a patient, a physical therapist has to take the medical and personal history of the patient, evaluate his strength, evaluate his range of motion (ROM), evaluate the patient’s pain, evaluate the patient’s functions and evaluate other factors, such as posture, skin condition, respiratory system, heart-beat, brain and nervous system function and senses. The process of evaluation is very important because there isn’t always a previous diagnosis and the patient might have many layered problems causing chronic pain and loss of function. The Physical Therapist’s evaluation skills are very important to understanding the source of the problem and consequently for the appropriate treatment therapy (27-29). After the evaluation, a PT needs to choose the appropriate treatment. The most commonly used method of treatment is the therapeutic exercise. PTs, however, also use manual and massage therapies to manage their patient’s pain, increase their range of motion, reduce their inflammation, or induce relaxation. Physical therapists can also use other therapies such as pulmonary therapy, hydrotherapy, cryotherapy, heat therapy, electrotherapy, as well as other complementary therapies (41-43). PT’s can help patients choose appropriate assistive devices and teach them how to use them. The PTs also teach patients with prostheses and orthotics to walk again or use their limbs properly in their new condition (43). “Each patient is evaluated as an individual and each treatment plan is unique,” explains Hawkins (44). Physical therapists apply treatment therapies, adapt them to their patients and modify them to their individual needs. Physical therapists also continually monitor their patients’ progress and adjust treatment plans as necessary until the desired goal is reached. Physical therapists are service providers. Krumhansl believes that they must be open to a 91 wide variety of people: “They need to interact with people from a range of cultures and with varying levels of education and socioeconomic statuses” (24). Two important qualities for a PT are empathy and compassion. Physical therapists need to gain the trust of their patient to get them to cooperate with the treatment. By showing that they genuinely understand and care about their pain and that they want to help, the patients are more likely to react positively to the treatment therapy. Hawkins illustrates this point with the words of a physical therapist: “‘You have to be able to make therapy fun, and be very much a caring person. You’ve got to be a person who really wants to know people and learn what is most important to them… it’s very much a relationship’” (Margaret Plack qtd. in Hawkins 25). Krumhansl observes that physical therapists have to be thoughtful and sensitive: Tact is an important personality trait for physical therapists to possess. People who are ill and in pain must be dealt with carefully and sensitively. Patients are living individuals who, when you encounter them, will be in a physical and emotional crisis. This will mean they may be resistant to treatment and may have difficulty understanding the need for particularly painful forms of treatment. A PT must be able to inspire confidence in them and build a solid working framework with them and their families to provide them with and have them accept the highest-quality care. (26) Emotional stability is also crucial for a PT since there is so much going on simultaneously and quick clear-headed decisions are often necessary. There are also often interruptions from other medical personnel and multiple patients to work with as well, so there is a need to be able to multi-task (Krumhansl 25). Physical therapy can be a strenuous work; PTs need to be in good physical shape themselves. Harasymiw points out: “[T]hey bend, crouch, lift, stand, swim, pull, push, walk, run, and support” (27). I have always been very interested in working with children and was delighted to find out that pediatrics was an important section of physical therapy. Neidlinger works in early intervention and reported that she didn’t need the certification of specialization to do so. She works with 0 to 3 year old children at their homes, which sounds very enticing to me, and is able to find enough patients within a 5-mile radius. My interest in early intervention brought me to an article written by Michele A. Lobo et al. titled “Grounding Early Intervention: Physical Therapy Cannot Just Be About Motor Skills Anymore.” This article discussed the fact that physical therapy impacts more than the physical wellness of children; it also has a big impact on their future cognitive, language and social development (95). The idea of the article is that muscle and brain are almost always used together and that strengthening the connection between the two can lead to a long term improvement in capacities in all domains. The article gave the example of school work, which can seem to be more of a mental activity but the article explains that physical activities are very often involved in it. Writing, for example, uses the brain and the muscles of the hand; talking uses the brain and all the mouth muscles. The habit of using the brain and the muscles simultaneously becomes easier and more natural over time, but physical therapy at an early age can help make the connection stronger and, therefore, improve the chances of having good learning capacities and lessen the chances of having learning disabilities. “[T]he research strongly suggests that early object interaction, sitting, and locomotor behaviors are important vehicles for promoting future cognitive, perceptual-motor, language, and social-emotional development,” explains the article (98). Understanding the deep improvement that can be done through physical therapy reinforces my desire to become a physical therapist. Furthermore, because I am very interested in working with babies and young children, I was very interested in the important impact of physical therapy on the development of young children. Physical therapy is a profession with many opportunities. The job offers great variety and 92 involves working with many different people, patients and other medical and professional staff. Physical therapists can experience many feelings every day: “demanding, rewarding, draining, inspiring….Perhaps it is that variety that draws many physical therapists to their work” (Harasymiw 51). Many surveys have shown that physical therapists are some of the most satisfied with their job. For example, Harasymiw reports that, in a survey for U.S News and World Report, “[M]ore than three-quarters said they were ‘very satisfied’” with their profession (51). Neidlinger confirmed being very satisfied herself with her career as a PT. Physical therapy is a medical profession that is growing much faster than average. According to the Bureau of Labor Statistics, “[E]mployment of physical therapists is projected to grow 36% from 2012 to 2022.” This career is projected to be one of the most recession-proof fields because health is less often affected by recession. The educational requirements are extensive but lead to excellent job security and satisfaction. Most of all, the help one can bring to others as a PT can be very rewarding. Works Cited American Physical Therapy Association. American Physical Therapy Association, 2014. Web. 3 Mar. 2014. “Become A Physical Therapist.” PhysicalTherapist.com. PhysicalTherapist.com, 2014. Web. 10 Apr. 2014. Drnach, Mark. The Clinical Practice of Pediatrics Physical Therapy. Baltimore: Lippincott Williams & Wilkins, 2008. Print. Flath, Camden. Therapy Jobs in Educational Settings: Speech, Physical, Occupational, & Audiology. Broomall: Mason Crest Publishers Inc., 2011. Print. Guimon, Katie. Personal interview. 1 Mar. 2014. Hathaway, Jeffrey W. “The Future of PT – Will it be all it should be?” PhysicalTherapist.com. PhysicalTherapist.com, 2014. Web. 17 Apr. 2014. Harasymiw, Therese. A Career as a Physical Therapist. New York: Rosen, 2011. Print. Hawkins, Trisha. Careers in Physical Therapy. New York: Rosen, 2001. Print. Krumhansl, Bernice. Opportunities in Physical Therapy Careers. New York: McGraw-Hill, 2006. Print. Lobo, Michele A. et al. “Grounding Early Intervention: Physical Therapy Cannot Just Be About Motor Skills Anymore.” Physical Therapy 93.1 (2013): 94-103. Academic Search Complete. Web. 17 Mar. 2014. Neidlinger, Michele. Personal interview. 7 Apr. 2014. “Summary Report for: 29-1123.00-Physical Therapists.” O*NET OnLine. National Center for O*NET Development, 2013. Web. 8 Mar. 2014. United States. Dept. of Labor Bureau of Labor Statistics. “Physical Therapists.” Occupational Outlook Handbook. 2014-2015 Edition, U.S. Bureau of Labor Statistics, 8 Jan 2014. Web. 2 Mar. 2014. 93 Term Limits in the United States by Amanda Manton (Honors English 1102) I n recent years, term limits for the state and federal government, such as the United States Congress and state legislatures, have been an issue discussed and contemplated about in the United States. Mostly, the Republican Party pushes for term limits to correct issues involving the processes of government while the Democratic Party opposes term limits because they view them as ineffective to solve problems in the government. The only way that there will ever be term limits for the U.S. Congress is if an amendment is passed, such as the 22nd amendment that limits the president to two terms; for an amendment to be passed, both parts of Congress need to a have two-thirds majority vote and three-fourths of the states need to ratify the amendment (Maddock). Although amendments are possible to pass, it takes a great deal of effort and cooperation among political officials to pass. According to Preston Maddock from The Huffington Post, in April 2013, Arizona Republican Representative Matt Salmon proposed a bill named Joint Resolution 41, which would limit House of Representatives to three terms and Senators to two terms. This would ensure that the politicians are more aware of their constituents’ environment. This shows that term limits are still relevant today and is not just a new idea. On the other hand, there have been Congress members that have been adamant against term limits for politicians, from local levels all the way up to the President. For instance, ever since 1997, New York Democratic Representative Jose Serrano has tried to repeal the 22nd amendment, but all of his attempts have died before reaching a vote in the House of Representatives. Regardless, both sides of the debate want to improve the government but, use term limits for different justifications. Reasons Supporting Term Limits The proponents of term limits often try to look back at history to defend their push for term limits. For instance, many, such as Jed Babbin from the political journal, American Spectator, state that the Founding Fathers supported term limits and shaped the foundation of our government to favor term limits because none of them were politicians for their whole careers (30). The Founding Fathers also specifically stated that the government needed a constant “‘rotation of office,’” so that the politicians would not be a permanent fixture in one area of government and could move to other areas of government if they wanted to continue in politics (“Arguments”). The cycling of politicians in a certain office benefits that position by getting a different point of view from the next official. This helps other areas of government too because they will gain the experience and knowledge of the former officials if they win a different position in one of the other areas of government. Proponents of term limits use the Founding Fathers as an example because they believe that the country should value the tradition and legacy they set for the government. Supporters of term limits also state that term limits will limit the amount of corruption in politics. According to Restart Congress, a political movement that advocates for term limits, they will weaken the influence of lobbyists and special interest groups because Congress members “will have less time to develop financially beneficial commitments” (“Arguments”). This will limit corruption because the politicians will be forced to make decisions based on what is best for their constituents instead of trying to line their pockets with money to finance their next election. Additionally, term limits will cause Congress committee members to be chosen based on merit rather 94 than be based on “political favors” and a greed for power in Congress and seniority (“Arguments”). The reason that the members will be based on merit is because there will be less career politicians, so each member of Congress has more equal seniority and weaker political alliances. Similarly, much of the corruption in Congress is due to the career politicians. According to Babbin, the more time a person stays in office, the more likely they will become corrupted because their experience brings a sense of power and seniority (30). Also, with that, the more likely the members will try to use their position to gain favors and bribes from lobbyists and groups that want to speak with them (Babbin 28). The members believe that they can use their position because those groups want to have some influence on the lawmaking process by meeting with them. The members figure that the groups will be more willing to take unethical measures, such as bribes, to promote their agenda. The corruptive power of staying in office can be seen in how the Republican Party’s viewpoint on term limits shifts in 1997. Previously, the Democratic Party had control of Congress, so the Republicans supported term limits more strongly, but when they won control in 1997, their support for term limits diminished because they wanted to stay in office as long as they could (Babbin 30). Their change in opinion symbolizes that the power associated with a position in Congress corrupts the politicians because they become greedy and used to their new power and do not want to give it up. Overall, the supporters of term limits view it as a limit of corruption by being an automatic barrier to the corrupting forces political power has on people. Although elections are part of our government to prevent low quality politicians from keeping power, it is extremely difficult to defeat incumbents. For instance, on average, incumbents win about 90% of their re-elections (“Arguments”). That rate is high when a person considers that many people seem fed up with the government and the performance of Congress. One reason for the high re-election rate is that some voters are not well-educated about the candidates before voting, so they just vote for whoever’s name they recognize on the ballot or whoever is part of the voters’ preferred political party (“Arguments”). The incumbents also have an advantage over challengers because they are able to raise more funds for campaigns. They develop financial relationships with special interest groups that want to have influence in politics (Blumel qtd. in Keck). In the 2010 Congressional elections, the average campaign funding for the House members was about $1.4 million and for Senators it was about $9.4 million while challengers had on average $166,000 and $519,000 respectively (“Arguments”). This high discrepancy shows that the politicians need to be limited, so they have less time to gain money from special interest groups. Elections also will be more effective because both candidates will have similar finances to spend on advertising. The more advertising a candidate has, the more easily they will be elected because the public will see their name and faces more. The need for term limits is great, so elections will be competitive and force candidates to focus on the issues instead of trying to buy their way into office. Reasons Opposing Term Limits Although in polls, Congressional term limits are popular among voters, they still are not the solution to fixing our political system. According to College of DuPage American Politics professor David Goldberg, term limits are popular because they are a “catchy idea …a quick fix, and something to shake [the government] up” to correct the problems in Congress. Term limits, in fact, will make the current situation worse. They will weaken the power of the legislative branch and shift power to the executive branch, according to studies collected by the National Conference of State Legislatures (Bowser qtd. in Keck). The shift in power is due to the president retaining office while Congress gains new members, which will change the voting dynamic in Congress and give more power to the president, who will not have to deal with any changes in his branch of government. Although, due to elections, some Congress member would be replaced, term limits will cause more members to lose their seats than normal. Additionally, term limits would limit the quality of Congress’ decisions because it takes time for the new members to learn how to do their job well, 95 especially when it comes to building relationships with other members (Goldberg). Those relationships allow Congress to work together and form bills that the majority of them can agree on, so they can pass laws. Additionally, the more time a person has in Congress, the more knowledge they will gain in certain areas, so they can be experts on issues and have useful information for projects and committees (Keck). Therefore, term limits would hinder the quality of Congressional decisions because the members with expertise in areas will be lost when their time is over. Without expertise, Congress memebers will make uneducated decisions or will waste more time training the new members about the information the old ones learned before voting. Instead of improving Congress, term limits would worsen Congress’ ability to serve their voters well. In fact, representatives, who have taken pledges to self-limit their terms have regretted their pledge, such as Illinois Representative Timothy Johnson, because he “‘underestimated the value of seniority’” (Bloomer qtd in Stone 5A). Like any other job, the more time a person has in it, the better they are able to do his/her job. The officials also gain power that help them benefit their constituents when they hold office for longer periods of times; the added time allows them to gain more money for their districts and be part of committees that can benefit their district (Stone 5A). In short, the power associated with a position in Congress can be also positive to do good for their voters, rather than just leading to corruption. Although 67% of the voters supported term limits, there really has not been a negative consequence for those officials that went back on their pledges. For example, Representative George Nethercutt promised to only have three terms, but he ran for a fourth term in 2000 and won (5A). This symbolizes that even though voters support term limits, when they have the choice, they will not look at how many term limits a candidate has served to determine if they should be re-elected or not. Figure 1 Proportion of Congressman Who Are Former State Legislators in the 15 TermLimited States (National Conference of State Legislatures). People in opposition of term limits see that the public support them because they think there is a problem with the system, so politicians must fix the underlying issues to truly correct the problems. Supporters of term limits say they will limit career politicians. However, there will still be career politicians because they will have the chance to run for other offices still. Figure one shows the high rate of state legislators with term limits switching to Congress, so term limits do not really resolve the issue of career politicians (Kurtz 9). Additionally, the U.S. already has elections to get rid of bad officials, so there is no need for term limits; to help curb corruption, voters need to be better 96 educated and attentive (Goldberg). Term limits are ineffective at solving the issues they are supposed to solve because they do not deal with the root of the problems. Recent Candidate Opinion In the upcoming election, Illinois governor primary candidate Bruce Rauner openly supports term limits to help change the workings of government in Illinois; however, he does not offer any indepth information about how he will do it or how they will help. According to Professor Goldberg, Rauner is popular because he has a message that is popular among voters. He is consistent about his message but does not go into depth in the media about his term limit policy. If he wins the primary, he will probably give more details of the logistics of his term limit policy. Overview Congressional term limits are not a new idea in the United States, but they still carry a lot of support among voters. Supporters focus on the Founding Fathers as an example of what the government should be. Also, term limits will limit the corruption and career politicians, who are disconnected, in Congress. On the other side, opponents cite the regret of the representatives that took a term-limit pledge to show that term limits are not practical. Congress members need to build experience and knowledge about the lawmaking process in order to build relationships with other members and groups to enact new laws. The debate of term limits will probably continue for decades to come, but right now it seems unlikely that they will be enacted in the near future. Works Cited “Arguments for Term Limits.” Restart Congress. Restart Congress, 2012. Web. 1 Feb. 2014. Babbin, Jed. “Ill Winds Over Washington.” American Spectator 39.2 (2006): 26-30. Academic Search Complete. Web. 1 Feb. 2014. Goldberg, David. Personal interview. 28 Feb. 2014. Kreck, Kirsti. “Anti-Incumbent Mood Fuels Term Limit Debate.” CNN. Cable News Network, 19 July 2010. Web. 25 Jan. 2014. Kurtz, Karl. “An Unexpected Term Limit Effect.” State Legislatures 39.3 (2013): 9. Academic Search Complete. Web. 1 Feb. 2014. Maddock, Preston. “Congressional Term Limit Bill Introduced by House Lawmakers.” The Huffington Post. The Huffington Post.com, 24 Apr. 2013. Web. 25 Jan. 2014. Stone, Andrea. “Term-Limit Pledges Get Left Behind.” USA Today 13 Apr. 2006: 5A. Academic Search Complete. Web. 1 Feb. 2014. 97 Is there an Animal Consciousness? A Phenomenological Approach by Matt Mazur (Philosophy 1800) Abstract he notion of consciousness, animal or otherwise, has been often discussed throughout the history of human thought. More recently it has become a topic for debate in philosophy with many competing ideas emerging. This paper is an attempt to discuss the nature of animal consciousness using a phenomenological approach. Phenomenology is a branch of philosophy with the goal of discovering and describing the true nature of phenomena beyond the human subjective experience of them. Using the French philosopher Maurice Merleau-Ponty’s work on perception, this paper argues for the acceptance of a form of consciousness in many animals while also seeking a better understanding of animal consciousness from a phenomenological approach. The paper concludes by discussing aspects that the phenomenological approach can add to our limited perspective of animal consciousness. T Introduction It’s rather easy to consider oneself conscious, the next step up is to consider other human beings, and finally considering the consciousness of other species entirely. Human consciousness has a rich history in philosophy and with the arrival of the scientific discipline of psychology, science as well. The idea of non-human consciousness however has a smaller place in history. Animal consciousness has had a back seat in science and philosophy for quite some time and until recently it wasn’t really considered. The rise of various animal rights movements has brought the idea of animal consciousness to the forefront of many issues in areas such as ethical concerns, ecology and wildlife management. In this paper, I will argue for an interpretation of the consciousness of animals based off of the philosophy of Maurice Merleau-Ponty (1908-1961) and integrating this theoretical approach with the recent work undertaken by the contemporary philosopher David Morris as well as recent scientific developments. In Part I, I will give a brief history of the philosophical groundwork for animal consciousness and scientific debate surrounding the consciousness of animals. In Part II, I will lay out the philosophy of Merleau-Ponty and Morris involving consciousness and perception before finally making my argument for the consciousness of other animals in Part III. I First I will begin with an overview of a few of the more important voices in the history of philosophy concerning animal consciousness. As is fitting for a discussion concerning a history involving philosophy and science it is best to go back to one of the first great natural philosophers, before the term scientist had even been invented. Aristotle (384-322BC) spoke on a wide range of topics from ethics to politics and was also one of the first to catalog, categorize, and discuss animals in a somewhat scientific way. In On the Soul Aristotle allows for a small portion of animals to have thinking abilities and all animals to be capable of perceiving and “sense imagination” (434). All animals are endowed with perceptive sensations to survive but very few are given mental powers of calculation or “deliberative imagination” (434). In the Nicomachean Ethics, Aristotle argues that anything done under ignorance or compulsion is involuntary with an awareness of the “particular circumstances” of the action (1111). From that argument he goes on to allow animals at least the willingness of their actions arguing “For things done on account of spiritedness or desire are 98 probably not rightly called unwilling acts. In the first place, none of the other animals would any longer do anything willingly” (1111). Aristotle had perhaps a better outlook on animal thought and will than what followed him in the Cartesian Revolution. Rene Descartes (1596-1650) exerted extreme influence on western philosophy and many movements in philosophy can be traced back to him. While he may have made many great strides in terms of understanding consciousness with the personal “I,” his position on animal thought was bleak when seen today under a more comprehensive perspective. In the Cartesian system our physical bodies are essentially machines to house the consciousness. There is a duality between mind and body. Under this system of thought animals are reduced to mindless automata that merely react to stimuli. His argument for this is contained in Part V of the Discourse on Methods, in which he claims that because all animals possess the same organs as humans (brains, vocal chords, etc.) but do not communicate with language they must be lacking in some faculty. They have the organs to communicate but instead do not and therefore must be just automata because even the most stupid of humans can at least communicate basic thoughts. Since the animal has the organs needed but doesn’t communicate thoughts it must be that they have no thoughts to communicate. The argument follows that because the actions and behaviors of animals can be imitated or described in a machine like fashion, they must be analogous to machines and therefore unconscious. This line of thinking has had a drastic impact on the concept of animal consciousness. The Cartesian formulation of the world ushered in a new wave of thought. The world could now be understood by the use of human reason and rationality alone. Everything was open to understanding through the laws of physics and mathematics and rationality. Knowledge could be found in human rational thought alone and so all the world could be explained and understood. The dualism of Descartes creates a void of meaning for the material objects. Material objects could be described with mathematical formulations and so lacked any true meaning. This also applied to the body, which was seen as a mechanical thing with a separate substance of mind driving it around. In relation to animals they were seen as just a mechanical body, without the mind substance and therefore were just unconscious robots. This of course is not the end of the debate as others have argued differently since then. David Hume (1711-1776) presented another account for the consciousness of animals in A Treatise of Human Nature. Hume lays out an argument for animal thought and behavior saying “no truth appears to be more evident, than that beast are endow’d with thought and reason as well as men” (176). Hume’s argument is one of analogy in which he says that animals preform many of the same behaviors as humans and we cannot just assume them unconscious because we want to. According to Hume, the mind is a blank slate upon which sense perceptions are the only data to enhance it. The sense data we receive shapes the mind and experiences and molds consciousness. In this way, animals would also receive sense data and experiences which would shape their thoughts and behaviors much in the same that humans’ experiences would also do. The modern philosophical debate has exploded with various arguments for different levels of consciousness or states, along with much debate. Robert Lurz has summarized the debate in contemporary philosophy as the contrast between various higher order theories of consciousness and basic creature consciousness, or simple perceptual awareness (5). Michael Antony points out that the current state of writing is saturated with what he labels as different “kinds” of consciousness, different “concepts” as well as different “meanings” that often get thrown around indiscriminately (1). This can cause issues when trying to broach the subject if many people are talking about the same thing using different vernacular. Some modern philosophers, such as Peter Carruthers, argue for the higher order theories which deny “phenomenal” consciousness to animals. Lurz, on the other hand, has fiercely critiqued Carruthers and the higher order theories of consciousness (151). Another philosopher Thomas Nagel, coming out of the analytic tradition, argued that we can never move beyond our human subjective point of view to really know the consciousness of another organism 99 because we can only know that “no matter how the form may vary, the fact that an organism has conscious experience at all means, basically, that there is something it is like to be that organism” (303). The modern debate seems to go back and forth with no clear answer either way although one notion that has arisen from the consciousness debate is that consciousness is directed. This means that one cannot be conscious of nothing at all, there is always some “thing” to be conscious of. The scientific side of the debate takes its own variation which is different from the philosophical one. The scientific debate of consciousness arose with the advent of modern psychology. Psychology, the study of the mind, has had a fruitful yield in the modern era. There has been a lot of good work on the study of human consciousness but this paper is concerned with the notion of an animal consciousness and so that is where the focus shall be. Donald Griffin (1915-2003), the former animal zoologist and Harvard graduate, does an excellent summary of the debate in his book Animal Minds, pointing out that the study of animal consciousness has been largely influenced by the behaviorist movement of the early 20th century. Griffin lays out 3 main points of the movement as follows: Claim 1 is that learning and other factors of an animal’s lifetime account for almost all the behavior of said animal beyond its physical adaptations. Claim 2 is that we can only consider outside influences and observed behavior to account for explanations of animal behavior. Claim 3 is that subjective thinking should be ignored for 2 reasons 1- They are subjective experiences that cannot be quantified nor verified 2- They are just byproducts of brain activity and do not influence behavior (2021) Griffin does go on to state that claim 1 has been largely abandoned by most scientists and that claim 2 has been heavily modified due largely to the “cognitive revolution” which argues for various cognitive states in psychology (21). Griffin also claims that it is obvious animals process information within their nervous systems but what is not clear is if there is any conscious awareness of this fact (22). Since many processes within human beings also occur at an unconscious level, the assumption is that none of the animals’ information processing is conscious. The crux of what Griffin points out is that there are many in the scientific community who are hostile to the idea of animal consciousness under the basis that it is hard to identify or measure quantitatively and therefore should not be talked about in a scientific sense. Yet there is more to the story than just the behaviorist perspective. Many studies have been done on the consciousness of animal by scientists. The field of cognitive ethology, which has the goal of understanding animal behavior empirically with detracting from the ability to make predictions with utility and also including the possibility of internal states of cognition , is also an area of interest within the scientific community (Kamil 21-23). In 2012 a document entitled the “Cambridge Declaration of Consciousness” was signed during the Francis Crick Memorial Conference. This document is a statement about animal cognition which goes on to say that numerous animals possess the same neurochemical and physical properties in which to experience states of consciousness. The scientists’ analysis states that there are numerous brain systems that are homologous within humans and some animals that react similarly with regard to microcircuitry of the brain. Based on these findings the group of scientists that signed the document concluded: 100 The absence of a neocortex does not appear to preclude an organism from experiencing affective states. Convergent evidence indicates that non-human animals have the neuroanatomical, neurochemical, and neurophysiological substrates of conscious states along with the capacity to exhibit intentional behaviors. Consequently, the weight of evidence indicates that humans are not unique in possessing the neurological substrates that generate consciousness. Nonhuman animals, including all mammals and birds, and many other creatures, including octopuses, also possess these neurological substrates. (Low) This statement is of course not the end of the debate surrounding animal consciousness in either the scientific field or the philosophical one. II I will be drawing on the philosophy of Maurice Merleau-Ponty as well as the philosopher David Morris for my argument on the consciousness of animals. Merleau-Ponty was a phenomenologist with his major work being The Phenomenology of Perception. Phenomenology is the study of phenomena in themselves. Modern phenomenology is based on the work of Edmund Husserl (1859-1938), and can be described as an attempt to get to the heart of what a phenomenon truly is beyond a subjective experience or description of that thing. Husserl described this as the epoché, the first step in the phenomenological reduction in which we bracket or refrain from judging our thoughts about the existence of the world while also not taking them for granted. In this way the phenomenological method seeks to strip away the subjectivity and allow the phenomena to present themselves in a true manner. Merleau-Ponty argues against Husserl in this regard saying that in order to use the phenomenological method we must comprehend the notion that we cannot be fully separate from the world. There is no way that we may become a passive, objective observer of a phenomenon because “I cannot conceive myself as nothing but a bit of the world, a mere object of biological, psychological or sociological investigation”(ix). In our bodies we are trapped within the world and cannot step completely outside of it to try and investigate it. Furthermore, Merleau-Ponty argues that we cannot resort to only empirical methods in order to describe things because “[a]ll my knowledge of the world, even my scientific knowledge, is gained from my own particular point of view, or from some experience of the world without which the symbols of science would be meaningless” and “we must begin by reawaking the basic experience of the world of which science is the second-order expression” (PP ix). The empirical methods of science can be seen as a “second-order expression” because all the knowledge described by science is a description from a human subjective view point. To further clarify let us consider an example such as the color blue. When I see the color blue, I experience it myself from my subjective viewpoint. Now with a photospectrometer we can describe blue as being a wavelength of light at a certain frequency, acting in accordance with mathematical principles. That description though, is second-order because those attributes or mathematical derivations themselves are expressions from a human standpoint in an attempt to describe how we perceive blue objectively, and yet they cannot truly be objective since there is no way to step outside the human body and verify them from some other position. This subjectivity does not mean that those attempted objective descriptions themselves are not useful or should be disregarded, only that we must be aware that they are still a subjective description from the human perspective. Since we are forced into a subjective view of things by our very nature of being in the world there is of course a conundrum, namely that we are unable to step outside ourselves and get an objective third person perspective of experience, yet neither can we only rely on an empirical method such as science because it too is contained in the sphere of human subjectivity. In order to examine this issue Merleau-Ponty turns to perspective itself, as the primary lens in which we communicate 101 with the world. First and foremost, we are perceivers of things before anything else. At every moment we are surrounded and bombarded by perceptions from all angles via touch, sight, sound, etc. These perceptions are our first contact with the world, and so in order to properly try to understand the world we must first understand perception itself. In order to understand perception it becomes necessary to talk of the human body, the thing through which we perceive in the first place. Merleau-Ponty argues that the body is rooted in the world and as such “the permanence of one’s own body, if only classical psychology had analyzed it, might have led it to the body no longer conceived as an object of the world, but as our means of communication with it” (106). Seen this way the body is not an object of experience, for that would mean I could step outside and examine it, but rather the body is the subject of experience. I am my body and my body is me, I do not experience my body but rather my body experiences the world through my perception. In this way being is rooted in a body in the world. Consequently, consciousness cannot just be some metaphysical thing as Descartes would argue, but instead must be rooted within the body as well. Now if the body is the subject of perception the question arises of what consciousness is. Previously I discussed that consciousness is directional, which is to say that one must always be conscious of something. Seen this way, consciousness is a kind of awareness of things, or more properly an awareness of the perceptions we are having at that time. Merleau-Ponty argues in The Phenomenology of Perception that “[t]o be a consciousness or rather to be an experience is to hold inner communication with the world, the body and other people, to be with them instead of being beside them” (111). I argue that consciousness is an awareness of the perceptions we are having from the world at any moment on the subject that is our body. Using Merleau-Ponty’s definition, being aware of our perceptions allows us to “be and experience” those perceptions. The extent of this awareness may change depending on the situation. For instance I may be so enthralled in a painting that I am not aware of the things around that painting consciously but I am still receiving the perceptions of those things. Those perceptions I am aware of I can say I am conscious of. Once we are aware of the perceptions of the world the question is how to make sense of them and attribute meaning to them. This is how the color blue becomes the color blue, or the distance between me and a chair becomes that distance by the attribution of meaning to the perceptions. It is here that I will rely on the work of David Morris to further clarify my position. In order to talk about meaning I will refer to the description of Morris in which he argues that we should “draw on Merleau-Ponty, who discovers that human meaning is at root a bodily phenomenon that emerges from the way the body fits into and is oriented by a situation beyond and prior to what is meant” (329). To illustrate this concept Morris references the French word sens. Sens deals with direction and sensation, which is to say that we make meaning out of our sensations. Morris argues this is a shift from meaning to sense, in which we expand the concept of meaning beyond just being attributed “in our heads”. If our first communication with the world is through perception it follows that we add meaning to these perceptions with our sensations of them. Morris argues that “Sense entails differences that make a difference” (329). In this way, sense, or meaning, can be constructed as differences in experience that make an actual difference to the perceiving subject. Morris uses the example of glucose and a bacterium. The presence of the glucose makes a difference in the behavior of the bacterium (it swims toward it), and the presence of gold does not. The difference (presence of glucose) therefore makes a difference to the bacterium and so the glucose molecules have meaning to the bacterium and the gold would not. If we “don’t find that difference making a difference, you have a dead bacterium (or one not sensitive to glucose)” (330). Here Morris is making an argument on our understanding of ontology, the study of being, and arguing for a new way to conceive of the world as having meaning already inherent in many things through differences that make a difference. I am going to extend this argument from meaning inherent in reality to the way we consciously construct meaning using the same mechanisms of differences that make a difference. For 102 clarification, I will use an analogy of a human example: a person wearing a wedding ring. If I were single and saw a wedding ring on a person I found attractive I would be less willing to ask them on a date assuming they would say no, being married. Now if I come across someone not wearing a wedding ring I find attractive I would be more willing to ask them on a date. The difference here is the presence of the wedding ring and the meaning (datable or not) is constructed by me as a subject of this experience, as I perceive the difference. So the difference in experience is what allows me to construct meaning to the perceptions of that experience. To summarize, we are born into a human body that is subject to experiences through perceptions and rooted in the world. We cannot step outside this body, or the world, and so must accept that we are grounded in our body as a subject to the perceptions of the world around us. This is our first line of communication with the world and indeed how we build the meaning of things around us. Consciousness as it is must be subject to something, for we cannot be conscious of nothing at all. Consciousness is then seen as directed by the perceptions we receive as a body subject to the world since these are our first lines of communication with the world. Taking these perceptions or experiences into account, we then make meaning out of them consciously with the identification of differences that make a difference to our perceptions. So my body which is embedded in the world is subject to perceptions of the world around me of which I then am conscious of through perception and I then make meaning of those perceptions and therefore subjectively construct the world around me through my experience. III With consciousness embedded in the body-subject in the world, a transition to animal consciousness can be made. First and foremost, there has been some issue of animal consciousness in the early writings of Merleau-Ponty. Ted Toadvine discusses this in Merleau-Ponty’s Philosophy of Nature by saying “While Merleau-Ponty’s approach retains these echoes of the sharp divide between human and animal already sketched out in the work of his predecessors, in his own thinking the human-animal relation, or more generally the relation between life and mind, remains a point of significant tension” (83-84). Toadvine argues that there are issues in which Merleau-Ponty drew a sharp distinction between human and animal consciousness such as the “peculiar ontological role that the human being plays, throughout Merleau-Ponty’s work, as being’s means of self-expression”(84). Toadvine is pointing out that in some earlier works Merleau-Ponty held the condition of being human as ontologically significant in the way being it itself could express itself. While this would seem to present a problem to my argument, Merleau-Ponty’s later thoughts on the subject did change significantly. An example of this change is found in The World of Perception, which is a compilation of written radio lectures by Merleau-Ponty in 1948. In that book Merleau-Ponty argues that “in fact this world is not just open to other human beings but also to animals. . .who dwell in it after their own fashion; they coexist in this world”(70). This change in thought opens up the world of perception to the animal and recognizes animal consciousness. I argue that animals, having a physical body rooted in the world with similar morphological structures to allow for perception of the world, would be constituted as a body-subject of the world and as such have a consciousness of those perceptions in which they create and discover meaning and so construct their subjective view of the world. Many animals have a very similar physical body structure as the human being. While they may take various shapes and forms, they have the same basic structural components that allow them to navigate the world. They have sensory organs such as ears, eyes, and noses, etc. Even Descartes noted that animals have all the same basic features of human beings. From the same basic make up comes the idea that the animal is itself also another body-subject of the world just as is the human being. Being another body-subject of the world means the animal is also subject to perceptions of the world. This would be the first line of communication to the world just as it is in the human being. Animals then construct their world out of those perceptions and discover the meaning within creating 103 their own subjective perspective on the world. As human beings can experience a difference that makes a difference and discover meaning so too should an animal be able to in its construction of the world. The creation of meaning from perception is the recognition of differences and not necessarily a function of some higher order thought only human beings have access to. (This is not to say no meaning is constructed from higher cognitive functions for certainly this is not the case as we can self-reflect on something and create and apply new meaning and indeed do all the time. I am only claiming that it is not the only way to create meaning). This can be reflected in the animal’s behaviors towards its environment, for example my dog may take a liking to a specific ball and exclusively carry around that ball. Then for instance one day I decide the ball is too old and dirty and throw it away. My dog will search endlessly for that specific ball and may even not like any new toys I attempt to substitute it with. As a human being I cannot fully understand the meaning to be found in the ball but from the dog’s subjective experience it has discovered a meaning in it and so the difference of its being found or not has a meaning to the dog. Animals communicate with the world and construct and discover their own meaning within from their subjective viewpoint through their perceptions just as human beings do, and so they would also be conscious of the world and their experiences in it. The argument against the appearance of animals making meaning in the world is that while it may appear so, there is a perfectly good mechanistic explanation of the behavior and so any subjective experience should be discarded. This was outlaid by Griffin earlier in the behavioristic model. The argument is that somehow the human brain is the key difference in that our vastly superior intellect and ability for self-reflection excludes other creatures from conscious experience as they lack ability to inform us of their subjective experiences. Firstly I must point out that the lack of ability to inform another being of a subjective experience does not exclude that experience from existence. For example, people with extreme mental disabilities may lack the ability to inform us of their personal subjective experience but we do not simply conclude that they therefore have no interior conscious experiences. Another problem is that this outlook assumes an anthropocentric notion of consciousness, considering only the human subjective experience. Meaning to an animal would be different than meaning to a human and this is reflected in the behavioral differences. Merleau-Ponty argued that “Thus in spite of what mechanistic biology might suggest, the world we live in is not made up only of things and space: some of these parcels of matter, which we call living beings, proceed to trace in their environment, by the way they act or behave, their very own vision of things” (75). He goes on to say that if only we were to pay enough attention to and “live alongside” the animal world we could see this to be true. Casting an anthropocentric lens on the study of animal behavior can limit our ability to understand it in any way other than by comparing it to human behavior, and so likewise as the saying goes if we judge a fish by its ability to climb trees we will think the fish very unskilled. Another argument against this approach to consciousness deals with situations where a subject may still receive perceptions but not appear to be conscious of them, like victims of severe strokes. In the case of the stroke victim they would still be subject to the same perceptions persay, such as the light in the room, or sounds, or touches, but they may not respond at all to them. So if consciousness arises out of perception then the stroke victim should have a consciousness the argument goes. In response to the stroke victim critique I argue that since perception is a bodily phenomenon, and the brain is wired into the body at every level the consciousness that arises out of bodily perceptions would be altered if there where damage to the body in some way. This naturally includes the brain as the nerve center of the body and the filter through which many perceptions come. In the case of the stroke victim, there has been severe damage to the brain, through which their perceptions are received and likewise it would create a different perceptual world for the victim to have access to. Since consciousness, as previously mentioned, is an awareness of perceptions one must have the proper bodily organ in which to be aware with. So the stroke victim with the damaged 104 brain has a vastly altered perceptual field which accounts for their apparent lack of what we could call conscious activities. Consciousness is still a very physical aspect of our being and as such damage to the physical body would alter the conscious experience. The phenomenon of consciousness is a very complex and difficult thing to discuss, but phenomenology has shown to add new avenues of thought in dealing with it. The fields of Cognitive Ethology and the Cambridge declaration previously mentioned reflect a subtle shift in the acceptance of animal consciousness in the scientific field. The modern philosophical discussions of animal consciousness present a shift toward new perspectives on the issue. If Merleau-Ponty is correct and consciousness is rooted in the body stuck in the world, and Morris is correct that meaning can be constructed from differences that make a difference, then there is no reason to doubt that animals can also have conscious experience of the world. They are rooted in the same body-subject experience set in the same world of perceptions. Animals also experience differences that make a difference to their subjective experience and allow them to discover meaning in the world. Animals are conscious beings in the world the same as humans are, only from a different subjective experience. Taking a phenomenological approach to the issue of animal minds opens up new directions in which we perceive the mind of another being and provides new insights into how we understand the phenomenon of consciousness itself. Works Cited Antony, Michael. “Concepts of Consciousness, Kinds of Consciousness, Meanings of Consciousness.” Philosophical Studies: An International Journal for Philosophy in the Analytic Tradition. 109.1 (2002): 1-16. JSTOR. Web. 11 Oct 2013. Aristotle. On the Soul. Trans. Joe Sachs. Santa Fe: Green Lion Press, 2001. Print ---,Nicomachean Ethics. Trans. Joe Sachs. Newburyport: Focus Publishing, 2002. Print Carruthers, Peter. “Animal Subjectivity.” Psyche 4.3 (1998): n. pag. Web. 12 Oct 2013. Carel, Havi, Darian Meacham, and David Morris. “From The Nature of Meaning To A Phenomenological Refiguring of Nature.” Royal Institute of Philosophy Supplement, 72 (2013): 317-341. Print. Descartes, Rene. Discourse on the Method of Rightly Conducting the Reason and Seeking Truth in the Sciences. (1637) Project Gutenberg. Web. 12 Oct 2013. Griffin, Donald. Animal Minds: Beyond Cognition to Consciousness. Chicago: University of Chicago Press, 2001. Print. Hume, David. A Treatise of Human Nature. 1888. Ed. L. A. Selby-Bigge. London: Oxford University Press, 1965. Print. Kamil, Alan, “On the Proper Definition of Cognitive Ethology” (1998). Papers in Behavior and Biological Sciences. Paper 18. Web 14 Nov 2013. Low, Phillip. “The Cambridge Declaration on Consciousness.” Ed. Jaak Panksepp et al. Francis Crick Memorial Conference. 7 July 2012, Cambridge, UK. Web 19 Nov 2013. Lurz, Robert. “Animal Consciousness.” The Journal Of Philosophical Research 24 (1999): 149-168. Print. Lurz, Robert. “The Philosophy of Animal Minds: an Introduction.” The Philogophy of Animal Minds. Ed. Robert Lurz. New York: Cambridge University Press, 2009. 1-14. Print Merleau-Ponty, Maurice. The Phenomenology of Perception. Trans. Colin Smith. New York: Routledge Classics, 2005. Print. 105 Merleau-Ponty, Maurice. The World of Perception. New York: Routledge. 2004. Print. Nagel, Thomas. “What Is It Like To Be a Bat?” Analytic Philosophy: An Anthology. 2nd ed. Eds. A.P Martinich and David Sosa. Malden, MA: Wiley-Blackwell, 2012. 302-310. Print. Toadvine, Ted. Merleau-Ponty’s Philosophy of Nature. Evanston: Northwestern University Press, 2009. Print. 106 Life, Lottery, for the Pursuit of Organs by Virginia Meglio (Philosophy 1100) C urrently, there are a limited number of organs for transplant due to the fact that it’s based on donation, however, what if organ donation were made mandatory- would such a system be just? Philosopher John Harris argues it would. He believes that ethical societies would endorse and benefit from his mandatory organ donation concept known as the “survival lottery”. In this model, citizens are each given a number and when two or more organs are needed to save patients in need of transplants, a number would be pulled at random. The individual whose number has been selected would then have their organs harvested and distributed to the patients in need. His scheme operates under the assumption that the procedure for organ transplants has been perfected, and that the program is designed to ensure equal age distribution in society. Harris provides several explanations for why this lottery would be beneficial to society and examines the possible objections to his argument. Although Harris attempts to address the flaws in his system, I will argue throughout this paper that his efforts fall short and his argument is unsuccessful. Harris provides several reasons for why the lottery would be beneficial to society and examines the possible objections to his argument by first explaining the different roles that the people involved would play. Specifically, he examines the role of patients “Y” and “Z”: the individuals in need of organs in order to survive. He also examines the role of the doctor who, Harris argues, would be responsible for the deaths of patients Y and Z if he rejects the survival lottery. We recognize that the doctor cannot be blamed for the death of his patients if there are no available organs to perform a transplant, but Harris proposes that there may very well be available organs that a healthy man, “A”, possesses that could save multiple patients. If the doctor is obligated to treat, has the opportunity and fails to do so, he would be responsible for their deaths since there is no distinction between killing and letting die. The responsibility of having to choose whose organs to remove is taken away from the doctor as a result of the implementation of the lottery. Harris recognizes that many philosophers say there is a significant difference between killing a person and letting them die, therefore, the doctor is morally obligated not to kill. Even though the doctor has a moral obligation to his patients to save their lives, it is of higher moral obligation that he not kill. Harris counters this idea, saying that from the perspective of patient Y and Z, A is no more innocent than they are, so it brings into question why one innocent life would seem to have greater importance over two innocent lives. This propels his argument further to say that the doctor should have a stronger moral obligation to ensure that a greater number of people are spared, and therefore a stronger ethical incentive to take one innocent man’s life in order to spare two. There would be an immense psychological impact on the individual selected, the doctors performing the operation, and society at large surrounding the idea that someone must choose who will die, and someone will in fact die. It is for this reason that Harris offers the idea of the lottery, so that the individual would be picked at random and no one person would have the burden of choosing which life to take. He argues that although people would be apprehensive about being called to give their organs, the likelihood of one person having to do so would be just as likely as an individual being struck by lightning, or eaten by a lion. Therefore, the apprehension experienced by people would not be a likely factor for disapproving of such a scheme. On the contrary, Harris argues that people would actually feel more secure by having this system implemented. Harris concludes that there’s more distress without the lottery than there is with it, and less people would suffer as a result 107 of having the lottery in their society (Harris 276). One of the largest objections faced by the survival lottery is the opposition people have toward the notion of “natural” death, dying due to natural causes, and taking life purposely, which is what this lottery would do in order to save those who are dying of natural causes. By man orchestrating who is given life and who has it taken away, it is essentially playing God or changing destiny, which many feel great hesitancy to do. Harris addresses this objection, communicating through his argument that “It’s morally okay to violate any rule”, in this case, killing one innocent life, “if doing so will raise overall well-being”, that is, sparing two innocent lives (Shafer-Landau 128). Harris explains that if we don’t kill one, our failure in doing so is killing multiple, so either way loss is occurring. Therefore, it is not a plausible argument to say that the lottery involves killing and reject it simply for that reason alone because electing to discard the lottery results in the same consequence. Because of what the lottery demands, the idea of saintliness would become obligatory; every individual would be mandated to give their lives when called. However, what about the rights of the individual who is chosen? The system does not require an individual to give their direct consent, which would violate their right to choose what to do with their body. This argument is addressed again through the eyes of patients Y and Z, who would contend that if arguing from a perspective of self-defense, that same right is applied to them. Their well-being outweighs the harm that would be done to A. Since they have just as much right to defend themselves against harm, if they must kill A in order to defend themselves, then they would be blameless for doing so. Otherwise, the individual who survives will be doing so “over their dead bodies” (Harris 277). A problematic area that Harris attempts to deal with is the issue of selection. He brings up an idea that most would think would be suitable for the lottery to work, where Y and Z and the patients alike would be the pool from which to select from, putting no more outside individuals’ lives at risk. Harris again takes the position of patients Y and Z, explaining how their rights would be violated by this so called solution. Y and Z would feel that their lives are of less importance and value than those of the outside parties. This same issue extends to the terminally ill and those who are dying and cannot be spared. We would again be isolating a group of the population by telling them that since they will die anyway, their life, whatever is left of it, is less valuable than anyone else’s. Harris opposes this idea of selection from the terminally ill group by saying that the last few weeks of life for the dying can arguably be the most valuable, even more valuable than the rest of their life, so it can’t be justified to solely select from this group (Harris 278-79). The issue of trying not to discriminate against one group from another through the selection process of the lottery is one of great concern for the survival lottery’s achievement. Harris recognizes that his argument comes under scrutiny, but he doesn’t think any rival philosophical views disprove his argument. If this is the case, it would be implausible to reject the lottery. He begins by saying that any moral theory opposed to the lottery in that it involves the death of innocent lives is mistaken because without the lottery, there could potentially be more loss of innocent lives. According to the perspective that Harris argues from, “we are not allowed to ignore the suffering of others” if we consider ourselves to be “members of the moral community” (ShaferLandau 129). Therefore, simply allowing innocent patients Y and Z to die would be just as bad as killing them directly. Harris acknowledges that the moral game changer in his argument is the intent of patients Y and Z, and whether it is morally permissible. Some would argue since no one intentionally wishes for the deaths of the patients, but the patients wish for the death of A, the patients would be wrong. Y and Z disagree however, they don’t want the death of A, they just want his organs. Here the patients rely upon an idea known as the Doctrine of Double Effect. As defined by Russ Shafer-Landau, the doctrine of double effect, (DDE), is the view that if your goal is worthwhile, you are sometimes permitted to act in ways that foreseeably cause certain harms, though you must never intend to cause 108 those harms (219). For Y and Z the death of A is a foreseeable consequence of harvesting his organs, not the intended consequence. Harris delves deeper into this argument, addressing the belief that the system is violent and vicious, and only a tainted mind would advocate such a program. But Y and Z again would counter argue that to disagree with the lottery system would be immoral because you are not giving equal respect to all human life. If all human lives are not morally equal then we need some criteria for determining the weight of human life. If there are no such criteria available, it is safer to operate as if all lives have moral value rather than none. Harris concludes his paper by reexamining the survival lottery, acknowledging the weak spots in its potential success such as its susceptibility to corruption, difficulties in determining who is a viable candidate to receive a transplant, the psychological damage that may ensue as a result, and other objections which we see him try to counter throughout his paper. Ultimately, Harris argues that we are incorrect to conclude that the survival lottery is wrong strictly on the basis that it results in the intentional death of one innocent life- there must be more to disqualify it for being morally impermissible. John Harris claims that although there may be apprehension and opposition to his idea, the failure to adopt such a system would ultimately result in the loss of more innocent lives. Therefore, Harris believes societies and the people within them would be more successful, morally stable, and ultimately reasonable by implementing the survival lottery into their communities. Although Harris addresses some of the objections made to his argument and attempts to argue for the survival lottery’s moral success, ultimately it fails. There are several weaknesses to his concept that make the idea of the organ lottery unsuccessful. One of these weaknesses is the pool from which participants would be selected, and the effect that would then have on the promotion of well-being in society. By default, the survival lottery can only choose from a pool of healthy individuals who have able organs from which to give. If that’s the case, this would have the opposite effect of promoting healthy human beings. If the pool of participants must be fit enough to provide organs to those in need, people would intentionally try to eliminate themselves from the pool by engaging in activities that would deem their organs unviable, thereby counteracting the promotion of well-being that Harris argued his lottery would provide. Although Harris argues that installing this type of system into society encourages well-being, it can be argued that as a result of its installment, individuals would wish that they not be healthy in order to avoid being selected in the lottery. Another area where Harris’ argument falls short is his attempt to eliminate unfairness. Harris recognizes that it would be unfair to kill A so that patients who are responsible for the decline of their own organs from things like excessive drinking and smoking, can continue these behaviors of overindulgence with their newly gained organs. But Harris fails to give a solution for this major issue. It also may be difficult to determine who actually is responsible and who isn’t. He also argues that it would be unfair to strictly select from the terminally ill and the individuals in need of organs because that would be singling them out from society. The fault with this argument is that by excluding the ill from the lottery, you are still discriminating against another group of individuals: the healthy. Harris recognizes this flaw, but fails to acknowledge that by accepting this as part of the survival lottery, he is ultimately giving higher value to the unhealthy, contradicting his prior claim that the survival lottery makes it so no life has higher value over another’s. According to Nelson, “if Y and Z are members of the community and if they are not immune from the lottery due to the unhealthy lifestyle choices that would also exclude them from the lottery, it follows that they should be liable to the lottery, too” (399). If it truly were the case that all lives were viewed equally under such a scheme, then the people in search of an organ would also be subject to the lottery alongside the other healthy individuals with viable organs to give. One point that Harris uses for the basis of his argument that can be argued against is the concept that the lottery ensures all lives are viewed as equal. It’s difficult to say however that if, say, the president of the United States were to have his organs harvested, there would not be an exception made for him. It could be argued that his role in society is of great instrumental value and importance 109 since it promotes a high level of well-being, and his maintenance of life would promote the most overall good than if he were to participate in the lottery. Although he has no more right to life than Y or Z, it would be reasonable to excuse him from the program. It seems that excuses could plausibly be made for certain individuals in society, rendering the lottery unsuccessful since not everyone is on an equal playing field for being subject to the lottery. An objection could be made to this argument though, that if one can be excused from the lottery based on their immense contribution to society, if the lottery were implemented, we should only select from a group which makes little to no contribution to society to promote overall well-being for the greatest number of people. This line of thinking would then essentially justify the killing of the disabled and other groups alike, which as a society we know to be morally impermissible, but would seem justifiable if the lottery were to be employed. Because it seems we can make excuses for some individuals to be excluded from the lottery, it makes the idea Harris has that all lives are equal and should be taken to spare others, implausible, thereby making the lottery scheme ineffective. Speaking within the context of the survival lottery, the medically needy are favored; therefore, life does not appear to be as equal as Harris claims under such a system, making it inoperable within society. Another idea that can be criticized in John Harris’ argument is whether or not it would truly be morally obligatory for a society to function under such a system. Consider Judith Jarvis Thomson’s analogy of the violinist where you are kidnapped and hooked up to a world famous violinist and told that if you unhook yourself from him, he will die. If you don’t unhook yourself, then you will die from the strain he is putting on your kidneys. No one would say you are morally obligated to stay hooked up to the violinist because you have been put there against your will, and you are dying as a result. Therefore if you unhooked yourself, you certainly wouldn’t be considered a murderer. The cost, which is your own life, is far too great a price to ask to pay for a stranger (35354). As it would be morally praiseworthy for you to stay hooked up to the violinist, it would be morally praiseworthy for an individual to volunteer to give their organs to save another life, but certainly not obligatory. Because in the lottery you are forced to give up your organs and in the analogy of the violinist are kidnapped and forced to be hooked up to him, “it wouldn’t be morally wrong to neglect helping the violinist, if selected, if one never intended to in the first place” just as it wouldn’t be morally wrong to refuse the mandatory organ lottery because no one is electing to be in it in the first place (Nobis and Jarr-Koroma 60). And just as if you were to unhook yourself from the violinist you would not be thought of as a murderer, your opposition to the organ lottery and Y and Z’s resulting death would not be thought of as murder by your hand. Examining this idea further, individuals are not obligated to promote the well-being of strangers, unless voluntarily obligating themselves to such an individual, which we don’t see happening within the context of the survival lottery. When reviewed from a feminist perspective, “…care cannot be parceled out to everyone equally”, therefore “impartial benevolence” is an unrealistic expectation and certainly not a mandatory one for rational individuals to be moral (ShaferLandau 279). According to Judith Jarvis Thomson, “having a right to life does not guarantee having either a right to be given the use of or a right to be allowed continued use of another person’s bodyeven if one needs it for life itself”, thereby disproving the claim Y and Z have for their right to use another’s organs for their survival and ultimately the survival lottery in itself (359). One problematic portion of Harris’s argument is Y and Z’s use of the doctrine of double effect. Y and Z explain that the death of A is not part of their plan, they just want to use his organs, and if he cannot live without them, then that’s too bad (Harris 279). This line of thinking that only the intention of the action can deem it moral or immoral is highly problematic, and can justify what reasonable human beings would deem as horrendous and immoral actions. According to Lisa Day, “Using the PDE to allow the hastening of death as a foreseeable but unintended effect of preserving organs blurs the rule…[that] organ procurement cannot be the cause of death; the donor must have already died of causes unrelated to organ recovery before procurement takes place” (336). Would a 110 reasonable person say that there is a difference between harvesting someone’s organs and having them die versus killing someone by harvesting their organs? Take for example, that my dog is obese and needs to lose weight. My intention is to have him lose that weight as fast as possible, so I choose not to feed him for a week. As a result, he dies from starvation. It’s common knowledge that without food and water, animals will die. Just because my intention was good, it seems unlikely that someone would say my actions were justified and not call ASPCA for animal cruelty. But, according to the DDE, this sort of thinking can be justified because my intentions were good; I foresaw my dog’s death, but my actions were still acceptable. Any reasonable human being would seem to disagree with such a sentiment. Likewise, it would be just as absurd to argue that Y and Z are justified in taking A’s life. It’s common knowledge that harvesting someone’s organs would result in their inevitable death. In this case, there is no way to distinguish between the actions which Y and Z intend and the actions which they foresee, which is a good reason to believe that the DDE fails in this situation. Therefore, reasonable people would not accept the survival lottery as a plausible solution to the issue as Harris suggests. As Russ Schaefer-Landau explains by even using an example similar to that of Harris’ argument, “if we were to abduct a number of healthy people…and cut them up to distribute their vital organs to those who would otherwise die…we could minimize misery…but [this] way of minimizing harm [is] deeply offensive”(219). Our natural repulsion towards the idea of the selection of death, even to save one or more people and minimize misery, is in part, evidence of its own for the moral difference between letting die and killing. There is a clear distinction between allowing someone to die of natural causes and bringing about another individual’s death, and the lottery breaks this moral line. This paper has been an analysis and evaluation of John Harris’ argument for the moral success of the survival lottery. Although Harris exemplifies potential benefits and attempts to address the flaws within his argument, there are still detrimental gaps within the scheme that make the survival lottery unsuitable for reasonable societies to adhere to. One potential argument against my criticism of John Harris’ survival lottery could be that I have not specified the conditions in which there is a relevant moral difference between killing and letting die. It hasn’t been my purpose to say that there always is a distinction between the two, just that in this circumstance, things are not as equal as Harris suggests. Although it may be argued that there’s no difference between letting die and killing, and it is of highest obligation to promote the well-being of the most people, the intentional death of A, and the unintentional hand that Y and Z were dealt, speaks against this objection. Y and Z’s death would be brought on by no one in particular, while A’s death would be brought on by the hand of Y and Z. There seems to be a clear distinction of how death would be brought upon the individuals involved, proving why this potential weakness in my argument does not threaten the validity of my evaluation. Although at face value, saving two lives versus one seems appealing, when implemented through John Harris’ survival lottery, it is deemed unsuccessful, and ultimately immoral. Works Cited Day, Lisa. “Boundaries of Double Effect.” American Journal of Critical Care 14.4 (2005): 334-337. Academic OneFile. Web. 31 Mar. 2014. 111 Harris, John. “The Survival Lottery.” The Ethical Life: Fundamental Readings in Ethics and Moral Problems. Ed. Russ Shafer-Landau. New York: Oxford University Press Inc, 2012. 272-80. Print. Nelson, Mark T. “Y And Z Are Not Off The Hook: The Survival Lottery Made Fairer.” Journal of Medicine & Philosophy 35.4 (2010): 396-401. Academic Search Complete. Web. 29 Mar. 2014. Nobis, Nathan, and Abubakarr Sidique Jarr-Koroma. “Abortion And Moral Arguments From Analogy.” American Journal Of Bioethics 10.12 (2010): 59-61. Academic Search Complete. Web. 29 Mar. 2014. Shafer-Landau, Russ. The Fundamentals of Ethics. New York: Oxford University Press Inc, 2012. Print. Thomson, Judith Jarvis. “A Defense of Abortion.” The Ethical Life: Fundamental Readings in Ethics and Moral Problems. Ed. Russ Shafer-Landau. New York: Oxford University Press Inc, 2012. 351-63. 112 A Fountain Very Exhaustible: The Pollution and Restoration of Lake Michigan by Bryan Mundt (English 1102) O n August 2, 1885, torrential rains fell on the city of Chicago. In just 24 hours, over six inches of rain fell upon the city. Sewers backed up, basements flooded and gravel paved Lake Shore Drive was washed away (Cahan and Williams 12). But the citizens and officials of the early city were not distraught at the flooding or the damage it caused, instead they were elated that the dusty, slimy and filthy city had, in essence, been given a bath. The superintendent of sewers boasted that “public health will be vastly improved” now that the city’s 420 miles of sewers, including the cesspool that was the Chicago river, were “purged of all the filth” and swept out to Lake Michigan (Platt 363). Even the Chicago Daily News reported that Chicago was probably the cleanest city on the continent (Cahan and Williams 12). Yet during all the jubilation that followed the flood, Commissioner of Public Works and the man in charge of Chicago’s Water management, DeWitt Clinton Cregier, assured concerned residents that the mass of sludge, oil and sewage posed no threat to the city’s drinking water supply, even though this mass was drifting its way toward the water intake crib, located two miles offshore (Platt 365). The mentality of city officials after the flood of 1885 has continued to echo since that day in regards to pollution of Lake Michigan, as well as the entire Great Lakes region. The Great Lakes account for approximately 20% of the world’s surface freshwater, and 95% of North America’s. Even though they are located, geographically, in the northeastern part of the United States, one in 10 U.S. residents depend on them for their drinking water (Maeda 70). This essay will argue that there is no acceptable level of pollution to Lake Michigan and that restoration efforts need to be ongoing and sustained in order to reverse over one hundred years of damage caused by careless pollution into one of the largest bodies of fresh water in the world. Even though Native Americans inhabited the area we call Chicago for centuries before, history gives credit for its origins to Louis Jolliet and Jacques Marquette. In 1673, while paddling upstream on the Mississippi River, Jolliet and Marquette came to the mouth of the Illinois River. It was there they discovered the fertile soil of the prairies as well as the abundant wood and wildlife. As they continued their diagonal trek upstream on the Illinois River and then to the Des Plaines River, they had to slide their canoes through a swampy, wet grassland called Mud Lake. This portage was often used by the Native Americans as a shortcut to Lake Michigan and it was here Jolliet first felt the need to alter the land that took nature thousands of years to produce. He wrote “all that needs to be done is to dig a canal through half a league of prairie from the lower end of Lake Michigan to the river of St. Louis (the Illinois River)” (Cahan and Williams 15). However, Jolliet was mistaken when he thought a canal would only need to be less than two miles long, as during low water periods of the year, one would have had to carry a boat five or 10 miles. Instead it was later determined that a canal ninety-six miles in length would need to be constructed, connecting the South Branch of the Chicago River to a prairie outpost on the Illinois River called LaSalle. The I&M canal was completed in 1848 and was an instant commercial success. It was one of the main contributors to the growth of Chicago, causing the city’s population to skyrocket from less than five thousand in 1840 to one million in 1890 (Cahan and Williams 16). It was due to this population growth, however, that led to the first real pollution of Lake Michigan. Early citizens received their drinking water from private wells on their own properties; however, these wells soon became contaminated with dishwater, garbage, privy waste, even underground clay which eventually led to outbreaks of cholera and typhoid (Cahan and Williams 17). 113 Around the 1850’s, the city started drawing its water from the abundant, near pure drinking water of Lake Michigan. At first water was drawn directly from the lake shore via cast iron pipes and stored in huge reservoirs across the city’s three main sections. But soon enough, the lake water started tasting bad due to the run-off from the South Side stockyards and Northside distilleries that were dumping their waste directly into the Chicago River which emptied into the lake (Cahan and Williams 17). The south branch of the river took on its own name, “Bubbly Creek,” due to the physical bubbling and churning of the waters from the waste decomposing in it. Journalist and novelist Upton Sinclair, who was gathering information about the Chicago Stockyards, described the scene, It is constantly in motion as if huge fish were feeding in it or great leviathans disporting themselves in its depths…Here and there the grease and filth have caked solid, and the creek looks like a bed of lava; chickens walk about on it, feeding…The packers used to leave the creek that way, till every now and then the surface would catch on fire and burn furiously” (qtd. in Brett 42). Desperate to remedy the drinking water issue, and faced with a cholera epidemic, the city hired engineer Ellis Sylvester Chesbrough to find a solution. Chesborugh’s first design did nothing to improve the pollution pouring into the lake. While raising the city streets ten feet to accommodate a new network of sewers kept the garbage out of the streets, the new sewers still led to the same place, the Chicago River (Cahan and Williams 17). Chesbrough’s next plan entailed locating the water intake crib two miles off shore, then connecting it to the city’s waterworks via a tunnel that was built under the lake bed. “Two-Mile Crib” was completed on March 25, 1867 to much fanfare. The city, it seemed, had an endless supply of pure drinking water available to them. The editor of the Chicago Times, Wilbur F. Storey even wrote, “We have the purest water in the world…We have tapped the lake at a point where exists only ever-lasting purity” (Platt 136-137). However, the growing concern over the toxicity of the Chicago River and its pollution of the lake reached a head after the flood of 1885 with proof that the water being collected at Two-Mile crib was compromised by the sewage that was swept from the city just two days prior. Chesbrough devised a new solution to the city’s sewage problem; he proposed building a new canal 28 miles in length, connecting the Chicago River with the Des Plaines River in Lockport, Illinois (Brett 42). This new canal, named the Chicago Sanitary and Ship Canal was completed on January 2, 1900. The massive new channel was able to permanently reverse the flow of the Chicago River by pulling in massive amounts of lake water, and flushing the filth downstream (Brett 43). However, a hundred years later and partially due to climate change, the river is at risk of reversing its course once again and flowing back into Lake Michigan. Normally the river sits approximately six inches below the lake, allowing the lake water to flow into the river, but in 2013 after 320 days without at least one inch of snow, coupled with a hot and dry summer, water levels in the lake were at the lowest in history and almost equalized with the river (Bailey). If this trend continues, and the river reverses course, the city will have develop a new plan to deal with the, still polluted, Chicago River. Short term options include limiting boat traffic traveling to and from the lake, thereby limiting the amount of time the locks separating the two bodies of water sit open (Bailey). While the Chicago River historically was a major cause of pollution to lake, and still has the potential to be, other pollutants continue to be dumped by several other industries. In 1953, the S.S. Badger ferry entered service, primarily to transport railroad freight cars across Lake Michigan between Manitowoc, Wisconsin, and Ludington, Michigan (Hawthorne, “Badger”). The Badger is the last coal-powered steamship on the Great Lakes, and during a single sailing, season it dumps approximately 509 tons of coal ash directly into the lake. The ash is concentrated with arsenic, lead, mercury and other heavy metals (Hawthorne, “Badger”). To make matters worse, the company that operates the ferry service, Lake Michigan Carferry Service, was granted an exception to the Clean 114 Water Act, which allows the continued dumping of coal ash into the lake through the 2014 sailing season. The company originally sought an exception to the Act by attempting to register the ship as a National Historic Landmark, thereby allowing the ship to continue polluting the lake indefinitely (Hawthorne, “Badger”). The ship’s operator states they are outfitting the engines with equipment to store the coal ash onboard. The upgrades should be completed by the start of the 2015 sailing season (April 2014 Press Release: S.S Badger). The London company, British Petroleum (BP), has also been a longtime polluter of Lake Michigan. BP operates an oil refinery on the lake shore in Whiting, Indiana, and a mere 8 miles southeast of Chicago’s 68th Street water intake crib. In July 2007, the Indiana Department of Environmental Management granted a permit exempting BP from a 1995 federal regulation limiting the amount of mercury discharged into the Great Lakes to 1.3 ounces per year. BP reported that it released approximately three pounds of mercury into Lake Michigan each year between 2002 and 2005 (Carmichael). The company argued that it required the permit in connection to its $3.8 billion expansion to the refinery and Indiana officials backed them up, stating that the amount of mercury BP was releasing was “minor” (Carmichael). Unfortunately, mercury is a neurotoxin that is absorbed by fish and can be harmful if eaten in sufficient quantities. Even more recently, BP spilled over 1,600 gallons of crude oil into the lake from the same Indiana refinery. On March 24, 2014, workers noticed an oil sheen on the water and traced the spill to a distillation unit that processes heavy Canadian oil from the tar sands region of Alberta (Hawthorn, “BP”). Even though the U.S. Environmental Protection Agency said that the spill most likely poses no long-term risks to Lake Michigan, it raises serious questions about the safety of this refinery as well as other industries on the lake. We will now look at past and present efforts that have been put forth in order to not only limit the pollution of the Great Lakes, but to help restore the damage that has already been done. With the Great Lakes being the sheer size they are, the argument is often made that the amount of toxic chemicals being released is minimal. However, many of these chemicals and substances build up over time in the sediment beds of the lakes. This accumulation coupled with the fact that the Great Lakes only have a single outlet, the St. Lawrence River, means that the lakes are “like a giant bathtub with a really, really slow drain and a dripping faucet, so the toxics build up over time,” said Emily Green, director of the Great Lakes program of the Sierra Club (qtd. in Carmichael). In an effort to have oversight and authority over the Great Lakes, the governments of the United States and Britain created the International Joint Commission in 1909 (Sproule-Jones 37-38). Except for some studies on pollution levels and minor recommendations on sewage treatment in the early 1900’s, the first major restoration efforts began with the Great Lakes Water Quality Agreements of 1972 and 1978. The agreements would work to limit the amount of excess nutrient loadings and control of toxic substances, respectively (Sproule-Jones 39-41). The U.S. federal government has initiated its own acts, such as the Safe Drinking Water Act, which regulates 83 potential water contaminants, and more recently, the Obama administration announced the Great Lakes Restoration Initiative in February 2009 (Landers 26). One of the main focuses of the initiative is to reduce the amount of mercury and pharmaceutical products entering the lakes due improper disposal of electronics, medicines, and household hazardous waste. Contaminated sediments in 31 “areas-ofconcerns” are also slated to be cleaned up as part of the restoration efforts (Landers 27). Heavy metals and chemicals in sediments are consumed by micro-organisms, which are consumed by the smallest minnows, which are then in turn consumed by larger and larger fish. These pollutants are not metabolized by the organisms and build up over time. This is why each of the eight Great Lake states warns residents to avoid eating certain types of larger, longer lived fish (Carmichael). If successful, efforts and regulations such as these will continue to make a positive impact on the health and restoration of the Great Lakes. As so often has occurred throughout history, it is human nature to be narrow sighted on the impacts we make to this planet. As a young, growing city, Chicago was naive to believe that the great 115 inland sea in its backyard was “a fountain inexhaustible” (Platt 135), and when they did realize their fallacies, they merely devised new ways to sweep away the filth, instead of developing ways to prevent it in the first place. With each new generation, every new business venture, new exceptions are made to regulations that were developed to restore Great Lakes to the ecological splendor that they once were. Greater accountability and enforcement of current rules and regulations need to continue. I believe we are at a turning point however; more so now than ever, the human race is realizing the fragility of the planet we call home. We are starting to realize the consequences of our actions and we are taking steps to limit the impact of the footprint we leave. Lisa P. Jackson, administrator of the U.S. Environmental Protection Agency, said in 2010, “We’re committed to creating a new standard of care that will leave the Great Lakes better for the next generation” (quoted in Landers 26). I believe this to be a great piece of advice and guidance; if each one of us makes this world just a little better for the next generation, the tides will change and hopefully the environment can start to heal. Works Cited April 2014 Press Release: S.S Badger. April 2014. Web. 2014 1 May. Bailey, Marissa. Continuing Drought Could Lead to Reversal of Chicago River Flow. 9 January 2013. Web. 1 May 2014. Brett, Hansen. “The Reversal of The Chicago River: Flushing the System.” Civil Engineering 79.12 (2009): 40-43. Academic Search Complete. Web. 21 April 2014. Cahan, Richard and Michael Williams. The Lost Panoramas: When Chicago Changed Its River and the Land Beyond. Chicago: CityFiles Press, 2011. Print. Carmichael, Bobby. “BP gets OK to dump mercury into lake.” USA Today 31 July 2007. Academic Search Complete. Web. 21 April 2014. Hawthorn, Michael. “BP Raises Estimate of Lake Michigan Oil Spill.” Chicago Tribune 28 March 2014. Web. 1 May 2014. Hawthorne, Michael. “Badger Ferry Gets 2-year Reprieve from Pollution Rules.” Chicago Tribune 22 March 2013. Web. 1 May 2013. Landers, Jay. “Federal ‘Action Plan’ Sets Ambitious Goals For Great Lakes Restoration.” Civil Engineering 80.4 (2010): 26-29. Academic Search Complete. Web. 21 April 2014. Maeda, Deborah. Troubled Waters: Asian Carp and the Great Lakes. New York: Nova Science Publishers, 2011. Print. Platt, Harold. Shock Cities: The Environmental Transformation and Reform of Manchester and Chicago. Chicago: The University of Chicago Press, 2005. Print. Sproule-Jones, Mark. Restoration of the Great Lakes: Promises, Practices, Performances. East Lansing: Michigan State University Press, 2002. Print. 116 Hospitality by David O’Sullivan (Philosophy 1110) I f a person were chosen, at random, to ponder the question “What is the opposite of cruelty?” it would not be surprising for him or her to quickly answer with “kindness.” Not to say that people do not think, but most may not see the question from a philosophical point of view. It is easy for one to assume the answer is “kindness” because it may, at first, appear to be the “black-and-white” solution. The opposite of bad is good. The opposite of up is down. But the opposite of cruelty has more profundity and significance than the aforementioned examples. The word “cruelty” comes from the Latin crudus, which roughly translates to physical harm and bloodshed. In its simplest form, this is what cruelty is. According to philosopher Philip P. Hallie, however, the worst kind of cruelty is less of a physical attack and more of an emotional one. In his work titled “From Cruelty to Goodness,” he states that “… cruelty (what the anti-cruelty societies usually call ‘substantial cruelty’) involves the maiming of a person’s dignity, the crushing of a person’s self-respect.”1 In other words, Hallie is saying that to be cruel is to damage a person in such a way that he or she feels worthless. After experiencing cruelty, kindness cannot restore one’s dignity or self-respect. Cruelty is an experience with which the human race bears great familiarity. If one were to expand upon Hallie’s definition of cruelty, this experience originates from within human ranks as one that is inflicted upon each other. This human act of cruelty has been witnessed throughout history, and can be illustrated by the actions of world leaders in recent decades. One such example can be seen through the result of Adolf Hitler’s rule of Germany during the Second World War. Elie Wiesel, a holocaust survivor, describes his experience as a prisoner in his work Night. He states that, “We were incapable of thinking of anything at all. Our senses were blunted; everything was blurred as in a fog. It was no longer possible to grasp anything. The instincts of self-preservation, of self-defense, of pride, had all deserted us”2. Elie’s explanation of the prisoners’ feelings, or lack thereof, conveys what it is like to have suffered cruelty. Wiesel recalls that while being forced to run during frigid winter nights, the guards would shout at them to run faster, calling them “swine” and “filthy sons of bitches”3. This type of experience is beyond physical pain, and no amount of kindness could equal an opposing force. Imbalance of power, according to Hallie, is the key factor which preserves cruelty. He believed that when both the victim and the victimizer believe the victim to be a lesser person, cruelty will remain. One who has lost the willpower and confidence to end the cruelty they are enduring may need strength from an outside source. It is imperative that this outside source carries the power necessary to eradicate cruelty. This outside source of power is, in fact, the opposite of cruelty, and it closely relates to the philosophy of Immanuel Kant. In his writing “Respect for Persons,” Kant highlights the importance of treating people as an end. In this he means that there is final value in every person, and one should not use another only as a means to benefit themselves. Things, on the other hand, only have relative worth and can be used as a means only.4 One could argue that if a person were only ever used, and not appreciated for his or her value as a human, they would be experiencing cruelty. To know the opposite of cruelty, one must understand the value of a person and the value of hospitality. “Hospitality” is a word derived from the Latin term hospes, which means “host,” “stranger” or “guest.” Generally speaking, the term “hospitality” is used when referring to a host treating a guest 117 with respect and providing for his or her needs. For an act to be considered hospitable, a “host” and “guest” are not required in the literal sense. Simply treating a stranger with respect and accepting them as a person is demonstrating hospitality. The understanding of this concept of hospitality carries significance because it gives insight as to why Hallie believes it to be the opposite of cruelty. Hallie describes the opposite of cruelty as hospitality. He sees hospitality as a method of eliminating imbalance of power, which is essential in the ending of cruelty. While kindness may play a role in showing hospitality, there is much more to being hospitable. Showing hospitality to a person is to treat them with respect, cordiality, and sincerity. It is the recognition that people exist not only as a means, but as an end. To someone whose self-respect and dignity have been crushed, hospitality from another person can have rehabilitative effects. Hospitality can restore a person’s humanity, dignity and sense of self-worth. It can give them the strength they lacked to fight against cruelty. For these reasons Hallie claims that, in fact, hospitality is the opposite of cruelty. To reiterate, not all opposites may be as concrete as the opposite of left, or the opposite of black. While many may have differing opinions on the opposite of cruelty, I doubtlessly believe that Hallie is veracious in saying it is hospitality. Kindness alone would certainly never harm a person’s self-respect; however it does not have the same effect as hospitality. Cruelty is nothing short of dehumanization, it is the elimination of a person’s self-respect and dignity. Hospitality is the elimination of imbalance of power, and is likewise the restoration of the previously mentioned human traits. References 1 Philip P. Hallie, “From Cruelty to Goodness,” in Keith W. Krasemann, ed., Quest for Goodness (Acton: Copley Publishing, 2011), 28-41. 2 Elie Wiesel and Marion Wiesel, Night (New York, NY: Hill and Wang, a Division of Farrar, Straus and Giroux, 2006), 45. 3 Ibid., p. 89. 4 Immanuel Kant, “Respect for Persons,” in Keith W. Krasemann, ed., Quest for Goodness (Acton: Copley Publishing, 2011), 69-73. 118 Effect of Resource Indulgence on Aquatic Vascular Plants: A Closer Study of Lemna minor by Ashley Padavonia (Biology 1152) Abstract lants make up a vast majority of life on planet Earth and have been a very successful kingdom since their first appearance 450 million years ago. There are many different species that grow rapidly with various adaptations to be successful in their environments. We decided to test the growth rate and life span of Lemna minor, an aquatic vascular plant that produces mostly asexually, and created an experimental group of Lemna plants that received a fertilizer treatment. We also decided to test the effect of CO2 concentration on Lemna due to the recent concerns on how much harm the rising CO2 levels are causing in the biosphere. The Lemna experimental group with fertilizer treatment exhibited the least growth rate as compared to the control group and CO2 seemed to have little significant effect on the Lemna minor. Further study needs to be conducted in order to definitively state that Lemna benefits negatively from fertilizer treatment or has an increased resistance to rising CO2 levels. P Keywords: Lemna minor, population growth, fertilizer, CO2 levels, nutrient availability Introduction Lemna minor, also known as duckweed, is a successful aquatic vascular plant species that is found in fresh water ponds throughout North America, Africa, Asia, and Europe. Like many other plants, Lemna minor requires specific nutrients and while it is successful, it must compete with numerous other plants as well as animals in order to survive in its environment and has evolved into a highly competitive species. Some of the earliest experiments with duckweed examined its competitive nature and in a study in 1969 comparing Lemna minor with algae and phytoplankton it was discovered that, “Observations made of laboratory cultures and field populations of aquatic weeds and filamentous green algae… have confirmed that antagonistic activity exists between such populations when aquatic weeds or filamentous green algae are nitrogen-limited” (Fitzgerald 1969). When the duckweed was in an environment with nitrogen-fixing bacteria and algae, two of its top competitors, the duckweed attempted to absorb as much nitrogen as possible in order to significantly increase reproduction and reduce competition. As a result the population of the bacteria and the filamentous algae suffered severely. Ecologists use population growth models in order to assess a particular plant species and determine how it interacts with the environment. Generally, when resources are limited and the competition is fierce, one species thrives and others suffer negatively and in extreme cases can become extinct from a habitat. The growth curve of the duckweed can be described as exponential because its population nearly doubles over time due to its great ability to beat its competitors. In a recent study of vascular epiphytes, non-parasitic plants that grow on their hosts, their ability to retrieve nutrients in dilute as well as in concentrated solutions, “Vascular epiphytes possess effective mechanisms to take up potassium from both highly diluted and highly concentrated solutions, enabling the plant to incorporate this nutrient element quickly and almost quantitatively” (Winkler et. al 2010). In their conclusion, Winkler and Zotz applied their findings to other nutrients that epiphytes required, such as CO2 and sunlight, as well as other highly adapted plants such as “aquatic weeds”. In an article written by Brian Moss of the University of Liverpool the effects of atmospheric changes on plant growth due to global warming are discussed. Plants take in CO2 required for survival through openings in the leaf called stomata, which close when too much 119 CO2 is present in the atmosphere (Moss 2010). In a sense, the effect of increased CO2 has a similar effect on plants as increased oxygen does to humans. A slight increase may foster productivity in growth but too much of an increase can cause a plant to become malnourished and suffocate. Plenty of research has already been done on how duckweed handles nutrient deprivation but it remains to be seen how this plant handles and excess. In order to test how Lemna minor would fare in an environment with enriched nutrients, we conducted an experiment in which some duckweed plants received excess CO2 and some received excess CO2 and water. Because of Lemna minor’s adaptable nature I hypothesize that an enriched atmosphere of CO2 and fertilizer treatment, will cause a marked increase in the growth of the duckweed exposed to each treatment and that plants that are exposed only to a fertilizer treatment will outperform the plants with no fertilizer and those that are only exposed to increased CO2 levels. Materials and Methods In order to test the two points of our hypothesis we created two experimental groups from the Lemna minor plant fronds (duckweed). One group was exposed only to enriched CO2 levels and the other was exposed to enriched CO2 as well as the fertilizer treatment. A group of lemna that was not given fertilizer was kept at normal atmospheric CO2 levels in order to function as out control group. Each experimental group was prepared by placing approximately three duckweed fronds into test tubes fifteen cm long in length. The tubes were then filled with filtered pond water, from a pond on site where Lemna was known to grow, and each sample was labeled. Next, half of the duckweed frond tubes were given the liquid fertilizer treatment diluted to 140 ppm. We decided to test each group’s progress in an environmental chamber to eliminate the appearance of unwanted variables. Both experimental groups were exposed to an enhanced CO2 atmosphere level of 800ppm. Each was exposed to a light level of 8000 lumens/m2 for duration of twelve hours, along with twelve hours of darkness, for approximately two weeks. The temperature during light exposure was maintained at 25°C, and during darkness at 23°C. At the end of the duration, each test tube’s frond count was counted and the experimental groups were compared by plotting the mean number of fronds for each group and then calculating the average maximal growth rate to determine if there was indeed an increase in the experimental group that had the fertilizer treatment. Results Following the described procedure, we tested three experimental groups, Lemna with fertilizer treatment, Lemna in an enriched CO2 environment, and Lemna with fertilizer treatment and in an enriched CO2 environment, along with a control group with no treatments. In the control group, the greatest number of fronds occurred on day 10 with an average of 6.2 per test tube. The greatest amount of growth was occurred between days 7 and 8 when the mean number of fronds jumped from 4 to 5.3 and death did not occur until day 13 which can be observed in the population growth curves in Figure 1. The intrinsic rate of growth for this population was calculated to be 0.081. When comparing the control with the Lemna fronds that received fertilizer treatment, the control group outperformed significantly which can be observed through Figure 1. In the Lemna with the fertilizer treatment the greatest amount of growth was observed between days 2 and 3 when the mean number of fronds went from 3.0 to 3.3 and by day 6 a decrease in the amount of fronds began to occur. The intrinsic rate of growth was calculated to be 0.048. The pH of the fertilized pond water was taken and there was no significant change from the untreated pond water. The experimental Lemna group that received CO2 performed identically to the control group that received no treatments. The greatest amount of growth was observed between days 7 and 8 when the mean number of fronds also grew from 4 to 5.3. The intrinsic growth rate was calculated to be the same, 0.081, and when comparing Figure 1 and Figure 2, the population curves are also identical. The Lemna group that received the fertilizer and CO2 treatments performed identically to the Lemna that received just fertilizer treatments and also had an intrinsic rate of growth of 0.048. At the end of the experiment, a significant amount of algae was observed in all tubes that contained fertilizer. 120 Discussion The proposed hypothesis was that the Lemna minor group with increased CO2 treatments as well as fertilizer treatment would show the largest growth rate and that the experimental group that received only CO2 and the control group would show the poorest growth rate. The results from the experiment did not support the hypothesis. The control group seemed to perform the best because it had the highest intrinsic growth rate and death was not observed until day 13. The Lemna that received the fertilizer treatments actually showed the least amount of growth and exhibited frond death as early as day 6 which seems inconsistent with fertilization of similar plants. However, the fertilizer treatments seemed to facilitate the unforeseen growth of aquatic algae in the test tubes. As observed in Fitzgerald’s study with algae and aquatic weeds, the algae and Lemna minor grew in a limited space, and even though fertilizer was provided, a competition still arouse between the two organisms, and the Lemna became malnourished (Fitzgerald 1969). A possible change in procedure to improve future results would be the use of distilled water, instead of the use of filtered pond water, to eliminate the possibility of microorganisms competing for nutrients in the experimental Lemna systems. The enriched CO2 environment of 800 ppm seemed to have no effect at all on the growth of the Lemna populations even though it was almost double the control value of 450 ppm. These results also do not support the hypothesis. According to the study conducted by Brian Moss that appeared in Freshwater Biology, a significant increase in CO2 may increase growth initially in aquatic species but after long term increased exposure would affect the plant detrimentally. Based on these previous experiments, it is determined that the increased concentration of CO2 was not significant enough to have a pronounced effect on the population growth of Lemna minor. These results show that Lemna minor must have an increased tolerance to excess CO2 environments. Although this discovery requires further study, future findings could be useful in the field of agronomy. Perhaps these genes can be isolated from the Lemna minor plants and applied to other plant genera in order to increase resistance in plants that are very susceptible to the increased presence of CO2 in our atmosphere today. Literature Cited Fitzgerald, George P. (1969). Some Factors in the Competition or Antagonism Among Bacteria, Algae, and Aquatic Weeds. Journal of Phycology 69: 351-359. Moss, Brian. 2010. Climate change, nutrient pollution and the bargain of Dr. Faustus. Freshwater Biology 55: 175-187. Winkler, Uwe, Zotz, G. 2010. ‘And then there were three’: highly efficient uptake of potassium by foliar trichomes of epiphytic bromeliads. Annals of Botany 106: 421-427. 121 Figure 1 Fi Figure 2 122 A Humorous Lesson by Becky Reichel (Honors Chemistry 1551) T hroughout history, humanity has reached for and fallen short of an explanation for the physical phenomena of life- even if each stretch seems to bring us a little closer to the truth. We have such patterns in most fields of science, including medical practice and pathology. For centuries, spirits and gods were to blame for disease, and the cures were logically sought along the same lines, but as we progressed, we began to look at our actions as having a more direct affect on our own health. It was Hippocrates who gave the medical world one big push forward with his use of observation and his thorough, systematic examinations and diagnoses. But while he is commonly known as the “Father of Medicine,” like all major human advancements, his pathology did not quite encapsulate the truth. One of Hippocrates’ big contributions to the medical sciences was the determination that it was not, in fact, spirits or broken taboos that brought about illness, but something chemical. Where he went wrong, however, lies in the details of his not-quite-theories, and the way others built upon them. During the Age of Pericles, Hippocrates was the one to introduce humoral pathology, or the belief that illness and other ailments were caused by in imbalance of chemicals in the body, according to Fielding Hudson Garrison in his work, An Introduction to the History of Medicine. Some Greek beliefs of the time pointed to four elements (water, fire, earth and air) as making up the world and the humors mimicked this distribution, being categorized as four main humors of which too much or too little of one or the other could cause nearly any ailment not directly attributable to some obvious injury. Though some continued to believe spirits could unbalance the humors, the actual cause of whatever ailment inflicted was accepted to be due to the humors themselves, and the system spread through the western world. Despite regional differences in other beliefs and applications, it was widely accepted that the humors were four fluids, which the body contained, classified as blood, phlegm, black bile and yellow bile. Blood and phlegm (mucus) are just as well known today, though not as humors. Yellow bile was likely the bile vomited by fasting patients, visible when no food was present to obscure it. Black bile was known as the “massive foul-smelling black loose stools” produced in the event of some diseases and the consumption of various herbs, including the toxic hellebore, according to Michele Fornaro, Nicoletta Clementi and Panataleo Fornaro’s “Medicine and Psychiatry in Western Culture.” Yellow and black bile stood as opposites, black being seen as the cause of melancholia and yellow, of mania, as well as, cholera. In the face of this, it is not a surprise that black bile was also known as the melancholic humor, and yellow bile as choleric. As for the other two humors, blood was also called the sanguine humor, while phlegm received the less original moniker, phlegmatic. These four bodily fluids were linked to both psychological and physical ailments or conditions, and were assumed to be influenced by dietary or environmental factors. According to Hippocrates, it was a person’s physical and mental activity, as well as their diet and environment, which imbalanced the humors within a body, not insufficient worship of the sovereign deities. In fact, Helen Christopoulou-Aletra and Niki Papavramidou’s “Methods Used by Hippocratic Physicians for Weight Reduction” reveals that when treating a patient, they took into account the effect of “the profession of the patient, the amount of body exercise and athletic endeavors, walking, sleeping, sexual activity, baths, steam baths, […] and the feelings of the patient” on the humors, seeing the connection between some activities (or lack of activity) and disease, and 123 then chalking it up as a factor of imbalance. Other environmental variables such as the climate were also commonly known to imbalance the humors or to weaken a person’s resistance to illness. One such climate thought of as particularly treacherous was any that had hot, damp summers and cold winters. Hippocrates thought people in such climates were “languid, moist, and generally ‘cowardly’ in their souls” while cold climates fostered “extremely thin persons with sinewy, firm flesh” (Christopoulou-Aletra 514). The cold and dry appeared to lessen the overall amount of fluid in the body, leading to a total humor deficiency that supposedly caused this extreme thinness. Likewise, a surplus of all humors led to obesity, in the eyes of those who followed the humoral pathology. Yet, though the humors seemed to coincide with different combinations of hot, cold, dry, and wet, and thus could be increased or decreased by exposure to their own or their opposite characteristics, as in the example of cold climates discussed above, the categorization of what was hot, dry, cold, or wet could be quite complicated, if not seen for their effects, rather than on their own worth. As seen in Daniel L. Heiple’s “Renaissance Medical Psychology in ‘Don Quijote’” for Tulane University, such things as lack of sleep and excessive studying were thought to dry out the body and mind, increasing the amount of the melancholic humor, black bile. This can be explained with the simple assertion that an increase of black bile was believed to cause a depression and spiral into madness which, in fact, could be prompted or worsened by lack of sleep or the stress which usually causes or is brought about by excessive studying. Much of the interactions of humors with the world and the categorization of foods, behaviors, and regions as hot, cold, moist, or dry was based on after-the-fact correlation, and later used to semi-accurately predict possible ailments without understanding the real middleman, providing what seemed like proof of the humors to the people of the time. There were other reasons to believe in the humors and their influence. Treatments based on the humors ranged from a change in behavior and diet, to the “bleeding, blistering, purging, vomiting, and sweating” the pathology is most known for today and in John Duffy’s From Humors to Medical Science. Yet, nearly every treatment had some noticeable effect, even if not an intended, or beneficial one. It is when a curative has no effect, for good or ill, that a patient truly begins to lose hope in its power, after all, and every time the physicians lucked into the use of something like St. John’s Wort (which has potential antidepressant properties) to combat a surplus of melancholic, black bile, the practice gained more plausibility. Hippocrates’ careful and systematic observations of his clients, and the passing on of this process, certainly didn’t hinder the humoral pathology’s spread, either. Even popular literature towards the end of the humors’ reign still incorporated the humors’ effects, as seen quite obviously when the main character of Don Quijote regained sanity twice through extensive bleeding from battlefield wounds, reinforcing the healing power of humors to the very end (Heiple). Truthfully, the power of the humors lay in their explanation of the link between behavior and disease; they were not correct, but they allowed physicians to encourage behaviors and lifestyles that they saw correlated with good health with a persuasive, non-spiritual explanation of why and how. And that was a step forward in itself. Eventually, the sun would set on humoral pathology’s reign, but not until the late seventeenth and early eighteenth century would the last rays begin to dim. Most educated medical practitioners traded in the reigning humoral pathology in exchange for the iatrophysical and iatrochemical theories, which, respectively saw the body as “a machine motivated by ‘life force’ circulating through the nerves” or as a system of acids and alkalis, of which disease was the result of an abnormal reaction (Duffy 13,14). Yet these theories only dealt the deathblow to humors’ prominence in professional circles, with more attractive explanations for disease. The silent killers of humoral pathology would be the invention of microscopes and the discovery of microorganisms by Robert Hooke and Antoni Van Leeuwenhoek during the medical renaissance, published in Micrographia in 1665. Leeuwenhoek went on to observe and report the discoveries of microscopic protozoa and bacteria, and this quiet assassin of the humors bided its time for nearly two centuries, until in the 1870’s, “pathogenic bacteria were identified as the aetiological agents of infectious diseases” by 124 Louis Pasteur in his famous germ theory of disease, according to Alfred Tauber’s “Metchnikoff and the Phagocytosis Theory.” The discovery that microorganisms could actually cause disease without a surplus of this or a deficiency of that left humors dead in the water, even if there were those who would fight it, such as a “Professor Bastian” who wrote a letter to IN NATURE magazine stating that “such germs when present would be sure to go on increasing until they brought about the death of their host,” since the immune system was not yet known, in the November issue of 1871. Granted, a more detailed knowledge of human anatomy was likely helpful as well; after all, if there was meant to be a surplus or deficiency of black bile at any time in a human’s cycle, it’d be difficult to explain why most bodies simply didn’t have any. Overall, however, the discovery of bacteria and the late arrival of the germ theory disproved the influence of humors on the body for future medical practitioners and researchers, even if their influence on medical history would linger forever. References 1. Clementi, Nicolleta; Michele Fornaro; Pantaleo Fornaro. “Medicine and psychiatry in Western Culture: Ancient Greek Myths and Modern Prejudices.” Annals of General Psychiatry. (2009) Annals of General Psychology archive. Web. 28 October 2013. 2. Garrison, Fielding Hudson. An Introduction to the History of Medicine C. 2. W.B. Saunders Company, 1921. Google Scholar. Web. 28 October 2013. 3. Bastian. “Letters to the Editor.” Nature 5. 109. (1871): 84. Nature archive. Web. 28 October 2013. 4. Tauber, Alfred I. “Netchnikorr and the Phagocytosis Theory.” Nature Review Molecular Cell Biology 4. (November 2003): 897-901. Nature reviews archive. Web. 28 October 2013. 5. Gest H. “The discovery of microorganisms by Robert Hooke and Antoni Van Leeuwenhoek, fellows of the Royal Society.” Notes Rec R Soc Lond. 58. (2004): 187-201. Notes and Records of the Royal Society archive. Web. 28 October 2013. 6. Christopoulou-Aletra, Helen; Niki Papavramidou. “Methods Used by the Hippocratic Physicians for Weight Reduction.” World Journal of Surgery. 28. (2004): 513-517. Google Scholar. Web. 28 October 2013. 7. Heiple, Daniel L. “Renaissance Medical Psychology in ‘Don Quijote.’” Ideologies and Literature 2. 9. (1979): 65-71. 8. Duffy, John. From Humors to Medical Science. The University of Illinois Press, 1993. Google Scholar. Web. 28 October 2013. 125 Video Games as an Art Form by Michael Robles (English 1102) E very form of art originates from entertainment. Paintings were meant to decorate a room, music filled the void with appeasing melodies, literature engaged readers intellectually, and movies turn what lies in our imagination into a tangible reality. Of course, there are a plethora of variations on each of these categories, but all of these mediums have been solidified throughout the ages as forms of artistic expression. However, a relative newcomer to the world of entertainment, the video game, has not yet been recognized as worthy enough to take its place in the world of art. The problem arises when video game developers try to express their artistic vision, but the general public remains almost completely oblivious to their existence. In essence, video games cannot prove themselves because no one is listening. Nevertheless, video games should be promoted as a valid form of artistic expression because of the potential they have to creatively express an idea that no other form of media can claim to accomplish. This uniqueness is found in the video game's direct interaction with the observer. If academically studied, games can be as intellectually beneficial as the analysis of a renowned movie or a legendary piece of literature. A way to promote video games as art is to have increased positive coverage of video games by the public. A possible method of doing so could include introducing games in libraries, or perhaps even having interactive exhibits at museums dedicated to artistry. In essence, the only way to experience a game is to become directly involved with it. The more people that can experience what video games have to offer, the more they can understand their artistic value. Unfortunately, the virtual world of video games is only a reality to a few. In his article titled “Is it Art?”, John Lanchester states that “books, films, TV, dance, theatre, music, painting, photography…are all part of our current cultural discourse. Video games aren’t” (18). Even if a person does not partake in or even enjoy one of these artistic subjects, at least he or she knows that the artistic medium exists. Lanchester explains this situation as the “cultural segregation” of video games which have such a “clean-cut division between the audience and non-audience” (18). As one might expect, it will prove difficult for someone to value something that is unknown to him or her. Additionally, the public’s exposure to video games in the media often only involve a connection between violent video games and various shooting tragedies that occur throughout a nation. If this is a person’s only knowledge of video games, it becomes very easy to dismiss video games as a negative influence and a waste of time. Beyond the separation between those who do and do not play video games, the different types of games available even divide the gaming populace. There are games that are made to capitalize on its “fun factor,” where most of the developer’s effort is concentrated on the gameplay mechanics. Other developers sacrifice gameplay to provide a graphically impressive experience, but at the cost of its classification as a “video game.” To address this problem, it must be made clear that not all video games should be considered as art. Ernest W. Adams, in his article titled “Will Computer Games Ever be a Legitimate Art Form?,” is quick to state that “the vast majority of what the game industry does is not art, but popular culture” (69) and that most games are created for pure entertainment. However, Adams qualifies his statement by explaining that film, which is a recognized art form, is also usually not art. He claims that the video game medium as a whole, just like film, has the ability to be an art form due to the qualities it shares with other forms of art (70). Before video games can be promoted as art, the argument must be made that video games 126 have the attributes to qualify them as an art form. However, the difficulty arises in finding succinct definition of art. To obtain a greater understanding of art itself, Adams offers a description of the evolution of art. He proposes that art was initially meant to “portray a person or scene or object” but in this current generation, art serves “as an expression of the artist’s thoughts” (68). While video games are certainly not a product of one person, much like movies, there are “directors” of games that oversee a game’s development. Relatively famous developers in the gaming world would include Ken Levine (Bioshock), Peter Molyneux (Fable), and Jade Raymond (Assassin’s Creed). In an article published in the journal At the Interface/Probing the Boundaries, Jef Folkerts suggests a distinction between what is truly artistic and what is simply aesthetic. The aesthetic quality of something only involves its outward experience, while an object’s artistic properties involve its content (103). Folkerts’ asserts that while all art has an aesthetic component, “art does not have to be beautiful” (102), but that it is the content of a medium that makes it art. It is the content that distinguishes video games from other games such as football or blackjack, as video games actually have the potential for literary content. While video games of the current-generation boast life-like graphical engines that often surpass any visual effects present in movies, it is not the only aspect of games that define them as an artistic medium. Most, if not all video games have music uniquely orchestrated for them and could be analyzed completely on its own as a form of musical art. What differentiates a game from its closest relation, film, and other forms of art is its completely interactive form of narrative. When we are truly part of the “artistic experience” in film or literature we constantly associate meaning to what is presented in the movie or book. A film or piece of literature ceases to become art when “what we see is what we get” (Folkerts 103) where there is no substance beyond the medium’s outward appearance. By making the player directly involved with the narrative, even to the point where the player has the ability to shape it into his or her own unique experience, the video game effectively evokes this meta-reflexive nature of our minds. This ability of the video game to allow us to constantly interpret in-game events and associate meaning to them is what can define it as art, and the distinguishing factor of video games is that the observer has direct influence over the narrative. This level of interactivity would ultimately lead to a greater understanding of any artistic message that is present in the video game. The ambiguity of art has also caused some to believe that the original purpose of a game undermines it as a literary art form. Acclaimed film critic Roger Ebert claims that game cease to become art because “one obvious difference between art and games is that you can win a game”. Commenting on the state of narrative of video games Ebert continues with the claim that video games currently “exhibit prose on the level of a wordy fortune cookie.” However, when he addresses a counterargument made in a presentation, he clearly states that “the three games she chooses as examples do not raise my hopes for a video game that will deserve my attention long enough to play it.” Unfortunately, Ebert falls into the same logical trap as many critics of video games do. According to Lanchester, “The other way in which games might converge on art is through the beauty and detail of their imagined worlds, combined with the freedom they give the player to wander around in them” (20). In essence, video games must be played to realize the potential of the medium. Games cannot be simply “observed” like a film or a sculpture or painting. An observation Lanchester makes is that what undermines the validity of video games to non-gamers is the difficulty of learning the game mechanics. He proposes that “the whole package of conventions and codes and how-tos which become second nature to video-game players, but which strike non-gamers as arbitrary and confining and a little bit stupid” (19). It is important to point out that artistry is often accompanied by difficulty. The various works of Shakespeare are often offputting to students because of the difficulty of the language. It takes time and patience to understand how Shakespeare is read before one can understand his literary meaning. The same can be said of a video game, once the gamer can figure out how the game is played, then he or she will be able to 127 delve deeper into any meaning present in either the gameplay or narrative. One common theme among all of the articles reviewed for this essay is that the current development of video games makes it difficult for anyone to consider them an art form. The most popular games focus almost purely on the aesthetic qualities, making sacrifices in story and narrative for improvements in graphics and gameplay instead of finding the perfect balance between them. Like almost everything in society, the video games industry is a business first, and the games that sell the most are those based purely on fun. According to sales statistics provided by VG Charts, the top fifty of the best-video games of all time are all multiplayer-centric, including games such as Call of Duty, Grand Theft Auto and Super Smash Bros (Database). That is not to say that these games such as Call of Duty or Grand Theft Auto have no place in society, but if the general public can start recognizing games as an art form, then the developers that do try to focus on the balance between gameplay and narrative will be rewarded. As a result, the games industry may shift towards providing an enriching experience to the player because the balance will prove to be much more satisfying for those looking to find an alternate means of deriving meaning in society. Nevertheless, there are games out there that are worthy enough to earn the distinction of being art, and as technology advances, developers are able to create richly detailed creative worlds that can further immerse the player. However, these games are virtually nonexistent to the average citizen. Consequently, it is imperative that video games be promoted as a form of artistic expression because if the immeasurable benefits it could bring to our society. Even now, there are many connections between games and literature. Spec Ops: the Line, for example, is a third-person shooter which involves a three-man squad infiltrating a conflict somewhere in the Middle-East. However, Spec Ops does not fall into the trap of the glorifying of war as many shooters do. Instead it focuses on the brutality that arises and lets the player choose whether or not to violently engage in certain situations, dismissing any notion of a “black and white” morality system that games so often implement. The game also shares strong connections with the novel Heart of Darkness by Joseph Conrad by its narrative and resounding themes. Another example would be Valve’s Half-Life 2 which is clearly inspired by George Orwell’s 1984 setting of an overwhelmingly oppressive central government. If educators had a chance of recognizing the literary elements of games, they could provide connections to the texts that they are teaching students. The interactivity level will let students develop a greater understanding of difficult themes since they are directly involved in how the story progresses. This interactive experience could be applied to a more general audience, and it could improve society as we have the potential to learn very complex themes in a relatively simple manner. Perhaps most importantly, games with an “artistic experience” will provoke thought, much like the literary and cinematic masterpieces throughout history, but it will do so using its unique quality of interactivity. It will take an undoubted lengthy amount of time to change the mindset of any culture, but a greater amount of exposure is needed before any significant change can occur. A possible method in doing this is to make video games free and accessible. Libraries are a great resource of free knowledge, and introducing certain video games in this system will expose this medium to a greater amount of people who could not financially afford video games. Currently “…15% of all libraries now check out video games to anyone with a library card” and in those libraries even “books are being checked out at a rate that exceeds what it was before gamers arrived” (Navarrette). This could be attributed to the fact that gamers could find literary connections to games present in the library and it piques their interest in a multitude of related subjects. Additionally, there could be interactive exhibits that focus on a video game’s artistic significance along with their technological prowess. The idea of art is an indispensable aspect of our society. It allows the general public to grow intellectually as we strive to make sense of life and all of its facets. If video games were to be promoted as an art form, then the public will only stand to reap its aforementioned artistic benefits. After we can establish that video games are art, we can see unique aspect of interactivity that it 128 brings to the artistic table. Should this happen, all that remains to be done is the positive exposure of the medium. With enough time and exposure to this new medium, there may be a day when Bioware, a video game developer whose games focus on the element of choice, will take its place among Bach, Benigni, Byron, and Bacon. Perhaps the day video games become truly significant in society is when it is acceptable for students to cite them. Works Cited Adams, Ernest W. “Will Computer Games Ever Be a Legitimate Art Form.” Journal of Media Practice 7.1 (2006): 67-77. Communication and Mass Media Complete. Web. 15 Apr. 2014. Ebert, Roger. “Video Games Can Never Be Art.” Roger Ebert’s Journal. Ebert Digital LLC, 16 Apr. 2010. Web. 14 Apr. 2014. Folkerts, Jef. “Playing Games as an Art Experience: How Videogames Produce Meaning through Narrative and Play.” At the Interface/Probing the Boundaries 69 (2010): 99-117. Communication and Mass Media Complete. Web. 15 Apr. 2014. “Game Database.” VG Chartz. VG Chartz Ltd., 2014. Web. 07 May 2014. Lanchester, John. “Is It Art?” London Review of Books 31.1 (2009): 18-20. London Review of Books. LRB Limited, 1 Jan. 2009. Web. 7 May 2014. Navarrette, Ruben. “Do Video Games Belong in Libraries?” CNN. Cable News Network, 15 Aug. 2013. Web. 07 May 2014. 129 The 1968 Riots: Protesting the Democratic Convention by Dina Rudenya (English 1102) T he 1960’s were a decade full of turmoil and restlessness. The feminist movement, Vietnam War, the draft, racial riots, and the civil rights movement had sparked much discontent among Americans, leading to some of the most violent behavior since World War II. Against this unstable background, the Democratic Party’s National Convention in Chicago in August of 1968 had served as a catalyst in the dangerous rioting of the people. The intended purpose of this event was for fellow Democrats to choose a presidential candidate and decide on what positions the party should hold regarding the most controversial issues of the day (McGowen 1). However, due to the pressing matters of the decade, the people of Chicago had hoped to use the Convention to protest the Vietnam war, the military draft, and a government who seemed to pay no attention to the needs of the people. If not for the Democratic Convention and Mayor Richard Daley’s desire for political power, the Chicago riots in August of 1968 could have been either prevented or better controlled by the local government. Before the Democratic Convention of 1968, Chicago already had a fairly recent history of rioting. The years of 1919 and 1966 had both been a difficult time for Chicago’s black districts. These riots were associated with residential and social segregation that had invisibly divided Chicago into various racial districts, and their boundaries were not meant to be crossed. Martin Luther King’s arrival in Chicago to help promote the civil rights movement in 1966, followed by his assassination in 1968, had caused a series of intense riots that had left black Chicago homes burned and raided, while their owners were being harassed and killed (Bailey). These troublesome events had put other cities ahead of Chicago on President Johnson’s list of favorable Convention-hosting cities, and Mayor Richard Daley saw this as a negative reflection on his political status (Farber 115). In order to host the Democratic Convention and convey its symbolism of power to Chicago, Daley traveled to Washington D.C. on October 8th of 1967 to try to gain President Johnson’s approval. Hosting the convention would mean portraying an elevated political status and making millions of dollars from the jobs it would bring to the city. In Washington, Daley attempted to convince Johnson that if the Convention was not hosted in Chicago, he might potentially lose Illinois in the upcoming presidential election. At the time, Houston was favored above Chicago as a convention city, but Daley claimed that hosting it there would be useless since Texas was already supporting Johnson in the election (Farber 116). These reasons were apparently logical to the president, because Chicago was chosen to host the Convention, and Richard Daley was looking forward to showing off “Chicago as the number one convention city” (Farber 116). The event was to be held in August of 1968. However, no one could predict the drastic chain of events it would coax out of the people of Chicago. The Convention riots started out as protests against the many controversial issues that would be decided upon during the Democratic Convention. The major problem was the United States’ involvement in the Vietnam War. Since the early 1960s, government officials had been promising Americans that even though U.S. involvement in Vietnam was necessary, it would be very limited and brief. By the end of 1967, over 500,000 troops had been sent to Vietnam. (McGowen 8). It was becoming evident that U.S. troops would be continuously sent overseas to fight the battle of another country. 130 Another major problem related to Vietnam occurred in January of 1968, which marked the beginning of the first day of Tet on the Vietnamese calendar. The North Vietnamese Army and the Vietcong surrounded Khe Sanh in South Vietnam, and while all the attention was focused there, the Northern Vietnamese Army and Vietcong attacked major cities in South Vietnam, including U.S. air and military bases (McGowen 14). While this was happening, Americans were being assured by their government that the U.S. was “close to achieving its goals” in Vietnam (McGowen 14). Soon, it became clear that the war was not going as planned, and the military involvement was not going to be as limited and brief as expected. American citizens began to understand that their own government was not being 100% truthful with them. With so much unrest within the country, it seemed like United States was warring both at home and overseas (McGowen 15). Due to its vast involvement in the war, United States was experiencing a growing need for troops. In order to fulfill this need, the selective service, or draft, escalated. When a young man turned 18, he was required to register with the government, and if selected, had to serve at least 2 years in the military (McGowen 14). Along with protesting America’s involvement in another country’s affairs, the riots of 1968 were also opposing the mandatory draft that would force Americans to participate in a war they did not support. At this point, the civil rights movement that had been supported by Martin Luther King had evolved from ending racial segregation to ending America’s involvement in Vietnam’s problems (McGowen 20). Due to these issues, the Democratic Convention appeared to be taking place in two different places. Inside the convention center, politicians would meet, make speeches, cast votes, and be interviewed by the media to relay the events to the public. However, outside of the convention, the protesters gathered to oppose the decisions the delegates would make inside the convention: delegates that did not seem to truly represent the needs of the people (McGowen 4). Despite the rioting going on outside, Richard Daley attempted to portray himself as maintaining full control of the city. Armed with 12,000 police officers, Daley assured the convention delegates that everything would go smoothly (Simon). However, that was not the case. The riots escalated, and the rioters were seriously hurt. Ira Jaffe and Vivian Stovall were two protesters who formed a human chain in Grant Park when the police showed up to chase them away. “The police would move in with clubs and tear gas. We were being kicked, pulled apart, and some very racial statements being made” (Simon). That does not represent a city under control. But in a larger context, it also does not represent a country with a government that makes the decisions it does in the best interest of the people. In his book titled “No One Was Killed”, John Schultz stated that Mayor Daley and the police force attempted to block the protest “not only because it represented a mythic nightmare for those who control large material intentions and directions of this country, but because they know the protest-revolt to be a real and future threat to them” (Shultz 3). The protesters may have been handled differently, but eventually they had to be treated as a growing threat to those who hold any political power. It is unfortunate that circumstances had to be so violent for the people to be further encouraged to voice their oppositions to the government’s decisions, but then again, when does anyone in power respond well to being opposed? Being attacked with tear gas, mace, clubs appeared like a dust storm from the convention center where the delegates were holding their meetings (Frank). The convention that Daley had desired to host in order to boost his political status had turned into a reason for the people in Chicago to speak out against the government in the hopes that they would be heard. When Daley decided that the voices of protest and discontent had grown much too loud, he decided to prove to the delegates that he was still the king of his city. After Martin Luther’s assassination, Daley had criticized the city’s police commissioner for being too gentle with the rioters, and claimed that the police should have been ordered to “shoot to kill” the protesters (McGowen 35). The only way Daley wanted to deal with the rioters was to put them in their place 131 and to show them who’s in control, as opposed to hearing them out and coming to some sort of a compromise with his own people. Armed with over 6,000 National Guardsmen and even more army troops, machine guns and rocket launchers were a regular sight to see in the streets of Chicago during the Convention (McGowen 36). Protesters attempted to work with the city to keep their demonstrations within legal limits. They applied for permits to sleep in the city’s parks, but the city denied them and gave the police reason to use clubs, mace, and tear gas at 11 p.m. to force the protesters out of the parks after curfew (McGowen 37). Being denied by their city and local government, protesters were encouraged to keep on with their demonstrations, opposing the Vietnam War, the draft, the Democratic Convention, and the government in general. Despite the circumstances, Daley wanted to keep up appearances. It was more important to him to keep Michigan Avenue free of demonstrators than it was to allow the people in Chicago to have freedom of speech and assembly. He valued the state of his neighborhoods and parks far more than he valued “abstract ideas” about national stability and calm (Farber 256). His political outlook only extended as far as his own city, Chicago. He did not seem to grasp that Chicago was part of a bigger picture, and that the political status of the country would directly impact it. His beloved city was not segregated from the rest of the country, and did not reside in its own bubble of safety that would prevent the decisions of the government from causing turmoil in the streets of Chicago. The Democratic Convention would not ease the tension, neither would it elevate Daley’s political career while he attempted to host a Convention with his own people rioting outside in the streets. So if the Democratic National Convention did not add to Chicago’s political resume due to the extreme dissent among the people, what did it accomplish? It helped to emphasize the viewpoints that had served as the foundation in shaping United States. Without freedom of speech and the right to assemble, all decisions would be made by the government with little or no regard to what the people want to see their leaders do. Of course, the Democratic Convention did not need to be hosted in Chicago in order for riots against the war, draft, and government to break out. The demonstrators simply saw the convention as an opportunity to be heard, since their government had not been listening to them very well until this point. Having so much media coverage at once because of the convention allowed the protesters to exclaim “The whole world is watching!” (McGowen 17). Beginning to see the protesters as a threat to his image of being in control, Richard Daley allowed his police force to use violence against the protesters. His lust for political power prevented him from genuinely caring about the welfare of his citizens, as opposed to simply being able to keep those citizens in line. Without the convention, Daley would have had more time and energy to focus on Chicago and making his people feel represented instead of planning for an event that would selfishly serve him well and put more money in his pocket. Today, riots are still possible, and the extent to which they may escalate is uncertain. There are many of the same and new issues that could be so controversial as to cause an uprising. Wars, government involvement in foreign countries, our national debt, abortion, gay marriage, legalizing certain substances, and other things are capable of causing major discontent. These potential riots would not be limited to Chicago because the issues mentioned have spread throughout United States and continue to be a struggle in every state. The government cannot please both sides of the same issue, but must make compromises to make both sides feel like they have been heard. No one wants to be ruled by a government that ignores the opinions of the people it governs. Looking back on the Democratic Convention of 1968 and Richard Daley’s reaction to the protests, political leaders today can better understand what the general public is capable of when they are not heard. If the convention had not been hosted in Chicago, the riots would have probably taken place in another convention city, or they might not have happened at all. Mayor Daley’s politically lustful leadership had helped to ignite the flames of anger in Chicagoans when the convention was chosen to be hosted there. History may repeat itself unless humankind can learn from its past mistakes and take precautions in 132 their future decisions to prevent disasters as opposed to trying to solve them once they have already gained momentum. Works Cited Bailey, Amy Kate. “Race, Space, and Riots in Chicago, New York, and Los Angeles.” Social Forces 88.3 (2010): 1496+. Academic OneFile. Web. 17 Apr. 2014. Farber, David. Chicago ’68. Chicago: University of Chicago, 1988. Print. Frank, Thomas. “The naked and the daft: Norman Mailer’s chronicle of Chicago ’68 points up the pitiful state of contemporary political commentary.” Artforum International 47.2 (2008): S9+. Academic OneFile. Web. 16 Apr. 2014. McGowen, Tom. The 1968 Democratic Convention. Chicago: Scholastic Inc, 2003. Print. Schultz, John. No One Was Killed. Chicago: Big Table Publishing, 1969. Print. Simon, Scott. “1968 Chicago Riot Left Mark On Political Protests.” Weekend Edition Saturday (NPR). (August 2008). Newspaper Source. Web. 16 Apr. 2014. 133 Changing Baseball Forever by Jake Sumeraj (Honors English 1102) H idden in the back rooms of any modern major league baseball franchise are a select few individuals that are drastically changing the way teams operate. Using numbers and borderline obsessive tracking of each player’s every move, they see things that elude the everyday baseball fan. These are the baseball analysts. Although they do the research that can potentially decide which player becomes the face of the team, these analysts can likely walk the city streets without a single diehard fan knowing who they are. Baseball analysts get almost zero publicity. However, their work is clearly visible at any baseball game. A catcher’s decision to call for a 2-0 curveball to a power hitter, the manager’s choice to continuously play a hitter that’s only batting 0.238, and a defensive shift to the left that leaves the entire right side of the infield open are all moves that are the result of research done by analysts. Some of it may make no sense on the surface, but in a game where slight advantages can give you a huge edge over the opponent, every little bit counts. The work done by baseball analysts stems from Bill James, the pioneer of advanced statistics (“Bill James”). James is the creator of many different ways teams measure player performance, including runs created (RC). RC is a formula that combines a hitter’s output in numerous individual offensive statistics to measure how many runs that player contributes to the team. Of course, it is only theoretical, but the numbers are useful. For example, a team can calculate the Runs Created for two players that are battling for a spot in the lineup and determine which player would theoretically help the team produce more runs. This new-age statistic is an example of sabermetrics. Technically speaking, sabermetrics are complex statistics used to quantify a plethora of baseball situations in order to reveal things that basic statistics keep hidden. They allow us to see beyond the box scores and reveal what exactly is going on. Baseball is a complicated game. A pitcher’s earned run average (ERA) may give an idea of how dominant he is, but it is horrifically skewed. ERA measures a pitcher’s average number of earned runs, i.e., runs scored not through an error, given up in a nine inning game. It is as basic as it gets when it comes to pitching stats, and it’s the most popular. However, ERA does not nearly tell everything. For example, a pitcher plays his home games at a ballpark with a really short fence in left field. If a batter hits a home run over the left field fence that would have been caught in any other ballpark, the pitcher’s ERA still goes up even though the run that scored is a situation of a unique circumstance. In another ballpark, an identical pitcher allows a batter to hit the exact same fly ball to left field, except this time it is caught for an out. That pitcher’s ERA is lowered. If the two pitchers are identical, any variance in ERA is purely a difference of luck and circumstance. In order to level the playing field, ERA+ comes into play. ERA+ adjusts a pitcher’s basic ERA based on the ballpark that he is playing at and the league average ERA (“Adjusted ERA+ Definition”). The average ERA+ is set to be at 100, so any pitchers with an ERA+ over 100 are considered to be above average. Ideally, two identical pitchers playing at different ballparks should have the same ERA+. Sabermetric stats, like ERA+, allow teams to get a better idea of how valuable a player is in comparison to others. It is an analyst’s job to use these statistics to find out things that may be of use to the team. But baseball analysts do much more apart from formula-driven statistics. They work for hours on end replaying video of a hitter’s every at-bat to find their weaknesses and tendencies. After viewing hundreds of plate appearances, trends can be seen. An analyst may notice that a hitter likes 134 to chase curveballs on the first pitch or is hesitant to swing on anything up in the zone when they are ahead in the count. Also, it could be clear that a hitter really knows how to crush a specific pitch in a certain area of the strike zone while they cannot make contact with others. Then, after the hitter makes contact, the analyst maps out where the ball was hit and what kind of contact it was. If a trend shows balls being hit more often on the ground to the right side than to the left, it may be beneficial to shift the infield to the right. After all, there is no use in having a fielder defend a portion of the infield where the ball is not likely to go. A shift could give the defense an advantage over the hitter; balls hit in one spot that would normally result in a hit could instead lead to an out. These advantages add up. Preventing the other team’s offense from producing by using careful analysis could potentially avoid runs from being scored and, thus, lead to more wins. An analyst’s work regarding individual players’ results is evident regarding pitching and defense as well. For pitchers, an analyst could study each of their pitches over time and measure variances in speeds or degrees of movement. This could show trends that could help predict how effective the pitcher will be in the future. Defensive analytics could include how much range a fielder has when fielding their position or how efficient the path an outfielder takes to a fly ball is. There is no way that simple statistics could tell anything about things like this because all of the common statistics only give information about outcomes. Sabermetrics give information about efficiency and ability. Analytics help teams understand more about a player’s skills in each phase of the game. When putting a team together, baseball analysts can help the franchise run efficiently by saving money. Using advanced statistics, teams can find players that are not necessarily valued high by other clubs but theoretically are productive. If a player does not have high value, they will most likely not require a large contract. This allows teams to fill spots on the roster with players that fly under the radar but end up playing key roles. A perfect example of this was the Oakland Athletics 2002 season outlined in Michael Lewis’s book, Moneyball. Going into that year, the A’s needed to replace three of their key players through free agency. The only problem was that they could not spend big money on a player with high market value. To replace the lost offensive production, General Manager Billy Beane used advanced statistics to find players that were undervalued by other teams. Beane focused on signing players that got on base in any way or hit for power. Unbelievably, the Athletics ended up winning 103 games, including a 20-game win streak, and easily made the playoffs. Because they focused on finding bargain players, “the Oakland A’s had paid about half a million dollars per win” (Lewis). This was incredible because, in contrast, “[t]he most profligate rich franchises…paid nearly $3 million for each win, or more than six times what Oakland paid.” Oakland’s success that year was so remarkable that Lewis’s book was even made into a major motion picture in 2011 starring Brad Pitt and Jonah Hill. Among the key acquisitions that year were Scott Hatteberg and David Justice. Hatteberg, an injury-prone catcher, was not receiving many offers from teams because of the risk. The Athletics had an opening at first base and Beane decided that since Hatteberg had a history of getting on base, he would take a chance and let him learn first base. Because he was not wanted by any other team, Beane was able to sign him for less than a million dollars (“Oakland Athletics 2002 Salaries”). Similarly, David Justice was not being pursued by many other teams because he was aging and on the decline. As a former power hitter, Justice also knew how to draw walks and get on base. In 2002, he only hit 0.266 but had a whopping on-base percentage of 0.376. Justice signed for $7 million. Both Hatteberg and Justice signed contracts well under their production value. The creativity that led to the Athletics’ success in 2002 was sparked by baseball analysis and sabermetrics. Beane’s ideas at the time were foreign to the baseball community; every decision they made was scrutinized by the public and media because it seemed so unorthodox. Normally, a team would just pursue the players that produce the most runs and have direct offensive output. Instead, Beane decided to create runs through simply getting on base and relying on the probability that some 135 of those baserunners would end up scoring. In result, this strategy won the Athletics 103 of 162 games in 2002. Since the Moneyball year, teams around the league began to alter how they approached the game. Basically everything changed. Teams began hiring analysts to do intense research in order to come up with new ideas that would help them stay ahead of the opponent. Sabermetrics came to the forefront and began to play a role in how teams made player decisions. Simple scouting reports were not enough anymore, and advanced statistics were finally appreciated. Kevin Towers, the General Manager of the San Diego Padres, put it this way: “You’re crazy if you are blind to the value of sabermetrics as a way to supplement scouting reports. There are trends you can see in numbers that you might not pick up in scouting reports. The organizations that are not open-minded and not willing to think outside the box are going to be in trouble” (qtd. in Coffey). Towers is correct because baseball is always evolving. Staying up to date on the newest ideas is key to being able to run a team efficiently and successfully. The Boston Red Sox took action pretty quickly after the Athletics’ iconic season. On November 15, 2002, the Red Sox hired famed baseball analyst Bill James, himself (St. John). In fact, the Red Sox were so impressed with Billy Beane’s work in Oakland that they tried hiring him just days after the 2002 season ended. Beane backed out, but Bill James was enough. Just two short seasons later, the Red Sox won the World Series in 2004 for the first time since 1918. That championship was not a fluke, too, because the Red Sox won two more World Series championships in 2007 and 2013. Much of their success has been credited to Bill James and his work. To win the World Series in 2013, Boston had to rebound from a last place finish in 2012. The offseason that followed included a restructuring of the roster. Similar to Oakland’s 2002 season, Boston signed a bunch of players that seemed like they just came off the scrap heap. But once again, the front office’s pristine work and research led to great success. The Red Sox are a perfect example that using baseball analysis not only wins games, but it also wins championships. Analysts are important in the draft, as well. When a team considers which player to select in any of the 40 rounds of the MLB First Year Player Draft, it comes down to who has the best chance at being an impact figure in the major leagues at some point. It is difficult to find players that the general manager, analysts, and scouts can agree on because every player has their strengths and weaknesses. The players that are taken in the later rounds are especially difficult because they are always big risks; otherwise, they would not be available at that point in the draft. Scouts go and visibly get an idea of what a player’s strengths/weaknesses are and put together a scouting report. Analysts watch a player and find things that may signify potential or future failure. Although these two sides seem similar, they rarely go hand in hand because the scouts and analysts focus on such different ideas when evaluating a player. A player being taken in the draft is selected because he has clear potential to play in the majors, has an intriguing combination of abilities, or a scout or analyst has a gut feeling about him. In a 40 round draft, almost all of these players never play in the majors. This improbability of individual success is so drastic that “[f]ewer than 10% of drafted players spend even a single inning in the big leagues” (Reiter). Of course, almost all of these players are taken in the first few rounds of the draft. Very few selections in the latest rounds end up ever paying off at all. Being able to compete yearly is dependent on drafting well and having players come to the majors from within. Otherwise, a team will have to spend big money on free agents that are already established as impact players. This is why drafting well is important. In order to have effective drafts, a team needs scouts and analysts that know how to make the correct decisions when choosing between players. One of the greatest drafts of all time was performed by the St. Louis Cardinals in 2009. Current Astros General Manager and former Cardinals scouting director Jeff Luhnow ran the draft. The Cardinals had a history of drafting well, but this one topped them all. In 2013, just four short years after the draft happened, five of St. Louis’s 25 active roster players were acquired in that draft. 136 What’s even more remarkable is that these five players were not bench-fillers. All five were vital to their success that ultimately led to a matchup with the Red Sox in the World Series. This includes: “second baseman Matt Carpenter, their leadoff hitter and WAR leader; first baseman Matt Adams; righthander Joe Kelly, who is likely to start Game 3 against Boston; closer Trevor Rosenthal; and righthander Shelby Miller, a 15-game winner who ranked 10th in the NL in ERA (3.06)…” (Reiter). Luhnow and his team were able to find successful players in the later rounds, which really separates this draft from the rest. They did this by using a sleek combination of scouting, analytics, and following gut feelings that separated individual players from a pool of well over a thousand. Luhnow’s work throughout his time in St. Louis gained him enough recognition to get hired as the General Manager of the Houston Astros in December of 2011. He took over a weak team with an even weaker minor league system and embarked the franchise on a total teardown. Trading almost every major league asset since for prospects and minor league depth, Luhnow has created arguably the deepest and most impressive minor league system in all of baseball. In fact, the Astros placed first overall in ESPN’s annual ranking of all 30 farm systems this year (Law). To improve the major league roster, Luhnow and the Astros have focused on using analysis to find what works. The men that hold the front office positions in Houston are overqualified and have titles such as “Director of Decision Sciences.” The Astros have been so committed to trying to run efficiently during their rebuild that their 2013 team payroll was less than what some individual players in the league make in one season. To this point, there has not been progress in the wins/losses category, but many experts have praised Luhnow for his work and project the Astros to be a powerhouse in the near future. This is due in part to the front office group that includes some of the smartest figures in all of baseball. The Astros front office includes graduates of University of Pennsylvania, Harvard, Yale, and Georgetown, along with a NASA engineer. Apart from being a group of complete baseball nerds, the Astros have a smart group of people running the show. Getting a job in baseball is a difficult task. There are so few baseball analysis jobs out there that the ones that do end up being hired tend to be overqualified. Of course, there is no specific college major that would best suit a career in baseball so there are multiple paths. Jeff Luhnow has bachelor degrees in both engineering and economics along with an MBA. In an interview, Luhnow explained how he was one of the few that were asked to join the industry, so he did not have to break in at the lower ranks. “I also came in at a high level because I had three successful careers prior to baseball,” Luhnow says. Overall, baseball analysis is an extremely difficult career to just break into. There are so many people that aspire to obtain the same few jobs that it is unreliable to depend on being hired. Much of the hiring process in baseball is luck and knowing the right people at the right time. Luhnow believes that the best way to break into baseball is to build connections. “Get to know as many people in the industry as possible and don’t focus on the people on the top, focus on the people on the front lines,” Luhnow explained. The best way to do this is to either attend games and try to befriend scouts or to do some of your own research and try to share it online with people that already work in baseball. After connections have been made, the initial job quality is poor. Luhnow describes that if you are going to break into the industry at the front lines, you usually begin as an assistant to a scout working for free. If you are not scouting, most teams have internship programs in the front office. “[B]ut even after a successful internship, many find there are limited opportunities for full time work,” Luhnow notes. Pursuing a job in baseball requires commitment and passion. It is not always the one that has the flashiest degrees that gets hired; it’s whoever shows the team that they can help them be successful in any way. Luhnow says that “the two key areas are problem solving skills and communication skills.” He also points out that “[r]esumes don’t get people hired, connections and knowledge help more than a resume.” So, preparation for a job in baseball is not dependent on just studying hard and building flashy accolades. College major is not a huge factor. Of course, math and 137 science helps on the analysis side of things, but the skills required to work in baseball can stem from a wide variety of college majors. There truly are many different ways to break into the baseball industry, and most of it comes down to commitment and passion. Once a baseball analyst is successfully hired, it is a really enjoyable career. Life in baseball is hectic and busy, but definitely different than any typical career because the lucky few that hold those jobs are as passionate about baseball as it gets. Luhnow describes baseball life as “long days, long hours, no weekends to yourself, but otherwise a very rewarding career because you are doing what you love.” As teams are more openly using analytics in their everyday operations, there are more jobs being created. Looking to the future, there should be even more openings. A downside to the baseball job market is that it is extremely demanding in terms of producing success. If it is clear that a team is not running efficiently and successfully, employees get replaced. This will never change, as teams will always demand the best. Overall, holding a job in the front office of a major league baseball franchise may be one of the more rewarding careers out there. Today, baseball analysis is vital to the success of any major league team. As a game of advantages, baseball teams need to get ahead of the opponent in any way possible. Since the turn of the millennium, professional baseball teams have gone from disregarding sabermetrics to embracing it in every move they make. Past success shows that these strategies work, too. Looking to the future, the analysis used by teams is bound to only increase as new strategies and ideas unfold. Getting a job in baseball may be difficult, but it is enjoyable and gratifying. The lucky few that sit in the back rooms of Wrigley Field or Minute Maid Park have changed the science of baseball forever and are the reason teams approach the game the way they do today. Works Cited “Adjusted ERA+ Definition - Sporting Charts.” What Is Adjusted ERA+? Definition from SportingCharts.com. Sporting Charts, n.d. Web. 21 Apr. 2014. “Bill James.” SABR.org. SABR, n.d. Web. 20 Apr. 2014. Coffey, Wayne. “New-age Stat Men Taking over Baseball.” NY Daily News. NY Daily News, 8 Dec. 2002. Web. 05 Apr. 2014. Law, Keith. “Ranking All 30 Farm Systems.” ESPN.com. ESPN, 28 Jan. 2014. Web. 22 Apr. 2014. Lewis, Michael. Moneyball: The Art of Winning an Unfair Game. New York: W.W. Norton, 2003. Print. Luhnow, Jeff. E-mail interview. 7 Feb. 2014. “Oakland Athletics 2002 Salaries.” CNNSI.com. Sports Illustrated, 2003. Web. 05 Apr. 2014. Reiter, Ben. “Three Days in June.” Sports Illustrated. 28 Oct. 2013: 32. Academic Search Complete. Web. 5 Mar. 2014. St. John, Allen. “Powered By Bill James And Friends, The Red Sox Win (Another) Moneyball World Series.” Forbes. Forbes Magazine, 31 Oct. 2013. Web. 5 Mar. 2014. 138 Being a woman in the time of Stalin: Becoming the Other by Emily Trca (History 2225) “S he was now afraid of everyone and everything” (Chukovskaya 92). Such was the experience of Sofia Petrovna, the protagonist of Lidia Chukovskaya’s Sofia Petrovna, set in the highest years of Stalin’s regime. Sofia’s son is arrested, and she is caught up in the hellish existence of those who were left behind, those who were forced to remember. “Nonincarcerated family members effectively lived in prisons without walls,” and Sofia’s’s prison, like that of so many other women at that time, became marked by the walls of her own mind (Adler 215). When Sofia finally cries out, saying “I can’t stand it any more. . .I cannot, I cannot I can’t possibly stand it any more” the tragedy is not in her loss and brokenness, though these are valid expressions of her condition (Chukovskaya 98). What is tragic about her confession is that she did not, and could not, identify what “it” represented. And for that reason, one cannot, with any confidence, identify oneself with this woman’s experience of her own human self. To be a woman under Stalin’s regime was to be a placeholder for an idea projected by others, embodying an identity determined by the other, who required and perpetuated her sense of alienation. Strangely enough however, women were expected to, in a way, represent not just the conceptual ideal of femininity, but more importantly, the ideal Russian person. “Women modeled the ideal attitude of ‘love, honor, and obedience’” (Reid 133). The truly worthy Soviet woman “was to place her traditionally ‘feminine’ skills as a caregiver, educator, and homemaker at the service of Soviet society as a whole. . . by overseeing standards of hygiene, decency, and kul’turnost’ in the workplace” (Reid 154). After Sofia Petrovna had secured a stenographic job, she was made a member of the Mestkom, which “took up nearly all her evenings” (Chukovskaya 9). When Sofia’s son Kolya was still living with her, he convinced her of “the necessity for women to do socially useful work,” an idea strongly propagandized to the average Russian housewife, an idea Sofia seemed to have internalized (Chukovskaya 14). Sofia believes her son when he explains the reason for the arrest of, what seemed to Sofia to be, a perfectly harmless individual: “‘She didn’t recognize Mayakovsky as a poet. . .She’s not a real Soviet person...’” (Chukovskaya 32). “Everything that was written in the newspapers now seemed to [Sofia Petrovna] completely obvious” (Chukovskaya 14). Yet despite women’s contribution to the well-ordered state, “envisaged as a disciplined and hierarchical family, with the people as Stalin’s dutiful and loving children” (Reid 157), it was not long before the biological fathers began to disappear, leaving behind “women, women, women, old and young, in kerchiefs and hats, alone or with small children or babies. . .quiet, frightened, laconic women” (Chukovskaya 57). It was these very women who became the oral historians of not only their nuclear families, but also of the Russian people, Stalin’s “little family.” “Oral testimonies, on the whole, are more reliable than literary memoirs,” and as a result, contemporary historians of the Stalinist regime have turned increasingly to oral history as a window on questions of identity (Figes 128, 119). In her introduction to Sofia Petrovna, Chukovskaya explains to the reader the nature of her novel, describing it as not so much a story as “a piece of evidence, which it would be dishonorable to destroy” (Chukovskaya 1). She asks the reader to let her “Petrovna speak today as a voice from the past, the tale of a witness striving conscientiously, against the powerful forces of falsehood, to discern and record the events which occurred before his eyes” (Chukovskaya 2). Chukovskaya does well to literarily transform “herself from victim into important witness” (Clements 278). The issue 139 with her objective, however, is not just her unfortunate use of phallocentric language. More problematic is the class bias of the late nineteenth and early twentieth centuries, where the intelligentsia and middle ranks of Russian society “expressed ‘a telling lack of respect for the preferences of the lower classes’ in their cultural projects and ambitions” (Holmgren 92). One must also take into consideration the fact that memoirs written by intellectual emigres from the Soviet Union and Soviet survivors were often written for a specific audience. In the West, these memoirs were “widely greeted as the ‘authentic voice’ of ‘the silenced’, which told. . .what it had ‘been like’ to live through the Stalin Terror as an ordinary citizen” (Figes 117). While one should not assume that Chukovskaya intended to misrepresent the average Russian person, the fact remains that the story of Sophia Petrovna was not, in fact, her own. Furthermore, Chukovskaya’s purpose in writing the novel was “to help to reveal the causes and the consequences of the great tragedy the people had suffered” (Chukovskaya 1). However, while her words do, with integrity, identify her as a member of “the intelligentsia strongly committed to the ideals of individual liberty,” “they do not speak for the millions of ordinary Soviet citizens. . .who did not share this inner freedom or feeling of dissent” (Figes 118). The ordinary Soviet citizen, “on the contrary, silently accepted and internalized the system’s basic values, conformed to its public rules, and perhaps even collaborated in the perpetration of its crimes” (Figes 118). At one point in Sofia Petrovna, she approaches the wife of her former director as they wait in line for information. The wife tells Sofia that her husband has received ten years at remote camps, and Sofia reacts. “‘Then he was guilty after all. I never would have believed it. Such a nice person,’ thought Sofia Petrovna . . . ‘They won’t send my son away,’ said SP apologetically. ‘You see, he’s not guilty. He was arrested by mistake.’” The director’s wife responds by laughing, retorting that everything around them was “by mistake” (Chukovskaya 74-5). Later, Sofia confronts her son’s friend Alik on his comparison of Kolya to others who had been arrested. “‘After all Kolya was arrested through a misunderstanding, but the others...Don’t you read the newspapers, or what?’. . .She had been quite right to keep aloof from her neighbors in the lines. She was sorry for them, of course, as human beings, sorry especially for the children; but still an honest person had to remember that all these women were the wives and mothers of poisoners, spies, and murderers” (Chukovskaya 60). Sofia persists in her belief that there has been a misunderstanding. She assumes that if the authorities and representatives of the Soviet regime were to understand, they would come to agree with her conclusion. She maintained her “fundamental faith” in socialism’s “inherent superiority” (Adler 220), despite its sustained failure to return to her that which she most desired: her son. Sofia’s belief was sustained by social consensus and validation, which told her that “‘nothing can happen to an honest man in our country’” (Chukovskaya 36-7). Consequently, Sofia concludes that “nothing happens without a reason. . .[Kolya] must have put someone’s back up” (Chukovskaya 58-9). Her fellow children of Stalin consist of the entirety of the Russian population, serving to cement Sofia’s “subjective confidence in the validity of the belief” (Adler 224). The ever-present propaganda motivated the Russian working class towards conformist behaviors, and thus “what might have begun as a functional, pragmatic, or coercive conformity would have a tendency to be propelled by cognitive dissonance from ‘adapting to’ toward ‘adopting of’ a belief system” (Adler 230, 224). Sofia, driven by cognitive dissonance, supported her faith in the rightness of Stalin’s system by refusing to fault the system itself, instead reinterpreting her experience as a misunderstanding, a “perversion of an inherently good ideology” (Adler 219), later even admitting that her son must have done something wrong. Yet this explanation cannot fully capture Sofia’s experience, nor does it give us experiential understanding of her necessary alienation, from herself and from others. A single story, one might argue, may not provide the fullest expression of a Russian, working-class woman, but it certainly can do no damage as “anecdotal evidence” to supplement and corroborate the oral narratives of survivors. However, even if one’s objective is to get at what “really 140 happened,” to ask Russian survivors about their experience will not necessarily yield fully trustworthy accounts. In post-Soviet Russia today, there is “an increasing trend to repress the memory of repression. . .reinforced by nostalgia for a selectively edited history” (Adler 229). Those who have experienced trauma will usually block out portions of their own past and, having no “clear conceptual grasp of their own experience,” they have no way of making sense of their memories” (Figes 128, 127). To fill these mental voids, the oral historians of the country, the women, would do a number of things. Many “made up their own narratives, their own myths of the ‘happy family life’ or the ‘good father’ that was lost” (Figes 123). More victims of Stalinist repression, however, seeking to contextualize their experiences within a meaningful, broader narrative, located their lifestories in the safety of collective narratives (Figes 126). When asked to describe their experiences, women would offer “family chronicles” and “documentary tales” that were virtually identical to the narratives of others (Figes 129). This seems to have developed out of a radical reshaping of one’s memories. It was not uncommon for the victims of repression to substitute others’ coherent and clear memories for their own, which became so fully internalized out of the need for meaning that many oral historians would insist “on their version of events,” even after being confronted with physical evidence that they must be wrong (Figes 124). It is this extraordinarily strong experience of cognitive dissonance that should force one to withdraw any absolute conclusions regarding the nature of a Russian woman’s experience. It is almost as if those who were aware of what was happening could articulate only their autonomous experience of those who were truly experiencing the terrors of Stalin’s regime, who in turn could not articulate these experiences without using and internalizing the accounts of the other. Within the conceptual framework of the nation’s memory-keeper, there was no reason to seek an understanding of her experience, which would have required that she let go of a tightly-held belief. There was nothing to understand. A woman’s sense of self was insignificant, unless it identified so fully with Stalinist ideology that the very idea of being individuated was as absurd as it was immoral. Sofia’s distressed confusion over how Kolya, the “irreproachable Komsomol member,” could have possibly confessed a crime he could not have committed brings her to the point where “she had to decide, she had to think,” but instead sat by the window, thinking of nothing. She seems to go mad. At the end of the novel, Sofia burns Kolya’s letter; the reader is left with the image of a flame being thrown on the floor and stamped out (Chukovskaya 87, 108-9). There is no closure for her––only an act of necessary, ideological obedience to eliminate irreconcilable evidence. She chose to relinquish the right to her own life; to speak would have been to reject and be rejected by society, to disobey and displease the father of the country. After all, Father knew best. Works Cited Adler, Nanci. “Enduring Repression: Narratives of Loyalty to the Party Before, During and After the Gulag.” EuropeAsia Studies 62.2 (March 2010): 211-234. JSTOR. Web. 11 April 2014. Chukovskaya, Lydia. Sofia Petrovna. Trans. Aline Worth. Evanston: Northwestern University Press, 1988. Print. Clements, Barbara Evans. “Women’s Works in Stalin’s Time: On Lidiia Chukovskaia and Nadezhda Mandelstam by Beth Holmgren.” The History Teacher 28.2 (Feb. 1995): 277-278. JSTOR. Web. 11 April 2014. 141 Figes, Orlando. “Private Life in Stalin’s Russia: Family Narratives, Memory and Oral History.” History Workshop Journal 65 (Spring 2008): 117-135. JSTOR. Web. 11 April 2014. Holmgren, Beth. “Why Russian Girls Loved Charskaia.” Russian Review 54.1 (Jan. 1995): 91-106. JSTOR. Web. 21 April 2014. Reid, Susan E. “All Stalin’s Women: Gender and Power in Soviet Art of the 1930s.” Slavic Review 57.1 (Spring 1998): 133-173. Web. 1 April 2014. 142 Mothers, Martyrs, Damsels, and Demons: Women in Western Horror from Romanticism to the Modern Age by Alexandra Wickersham English 1102 T he Western horror genre has undergone many changes over the years, reflecting and adapting to the changes in Western society. While many of the standard plots of horror stories remain largely unchanged, the portrayals of characters of different races, genders, and backgrounds have undergone many alterations, some for the better, some not. In this paper, works from various time periods will be examined, including the early nineteenth century, the 1980s and 1990s, and the modern era. Several sub-genres of horror will be examined, including sci-fi/horror, vampire stories, teen horror, slasher films, and action/horror. Though the general portrayal of women in the Western horror genre appears to have changed since the Romantic era, in truth, the same sexist tropes have been constantly present in Western horror, they have just been expressed in more insidious ways. First, a brief overview of definitions for terms that will be used. “First wave feminism” refers to the portion of the feminist movement that started in the nineteenth century and ended between the late nineteenth and early twentieth centuries, with the exact year depending on the country. One of the primary goals of first wave feminism was getting women the right to vote.1 The term “second wave feminism” refers to the part of the feminist movement that reached its peak during the 1970s and 1980s. The primary issues of the second wave were reproductive rights and employment opportunities.2 The term “third wave feminism” refers to the part of the feminist movement spanning from the mid-1990s to the present day.3 Several works of horror fiction will be examined, with each of them being representative of the time period in which. The nineteenth-century work examined is Mary Shelley’s Frankenstein, published in 1818. Works published during the second wave include Stephen King’s Carrie (1974), Wes Craven’s A Nightmare on Elm Street (1984), Sean S. Cunningham’s Friday the Thirteenth (1980), and William Friedkin’s The Exorcist (1973). The third wave works that will be examined include John Fawcett’s Ginger Snaps (2000), Joss Whedon’s Buffy the Vampire Slayer (1997-2003) and Angel (1999-2004), and Eric Kripke’s Supernatural (2005present). The Cultural Landscape The above named fictional works are all products of the time and places in which they were written. To understand how these works portray women, we must examine the cultural climates that produced them, as well as the backgrounds of the people who created them. When Frankenstein was first published in 1818, the world was full of feminist thought, however, there was not yet a solidified feminist movement. The word “feminist” didn’t even exist at this point, and would not exist until near the end of the century. As early as the eighteenth century, feminist writers such as Mary Astell and Lady Mary Wortley Montagu had gained recognition for their opposition to male dominance in society (Taylor 205). In 1792, Mary Wollstonecraft published her famous work, A Vindication of the Rights of Woman, in which she expressed her outrage that women had been trapped in “a state of ‘ignorance and slavish dependence’” (Gordon 144). The general public’s reaction to these early feminist writers was not good. According to Barbara Taylor, a professor of history at the University of East London, “by 1791 Wollstonecraft already had a reputation as an insurrectionist, the English equivalent of the ‘revolutionary harpies of France, sprung from night and hell’” (Taylor 202). Over 150 years later, in 1974, another famous work of horror fiction was published. Carrie was the first published novel of the now well-known writer Stephen King. When King was born in 1947, World 143 War II had recently ended. During this time, women had filled many jobs that had been left by men serving in the military. However, once the war was over, women were expected to leave their jobs and return to their previous roles as homemakers. The often-romanticized era of the 1950s was an era of emptiness and depression for many American women (Friedan 19). In her revolutionary 1963 book, The Feminine Mystique, feminist writer Betty Friedan termed this “the problem that has no name” (Friedan 15). This name was fitting, as mental health professionals at the time had no term for this phenomenon (Friedan 19). This book is largely credited with sparking the second wave of America’s feminist movement. This movement was largely concerned with actually getting women equal treatment under the law in practice, as opposed to just in theory. Given the feminist ideals emerging at this time, the way King conceived of the idea of Carrie deserves special attention. In a 1981 interview for Twilight Zone Magazine, King stated that he wrote Carrie because “Some woman said, ‘You write all these macho things, but you can’t wrote about women.’ I said, ‘I’m not scared of women. I could write about them if I wanted to.’ So I got an idea for a story about this incident in a girls’ shower room, and the girl would be telekinetic. The other girls would pelt her with sanitary napkins when she got her period. The period would release the right hormones and she would rain down destruction on them…I did the shower scene, but I hated it and threw it away” (Grant). King’s decision to focus on a female character came not from a belief in equal representation in media, but from what he viewed as a challenge and an accusation of fear. It’s also worth noting that, despite this story being written at the height of the second wave of feminism, when women were making every effort to erase the seemingly-inherent connection between womanhood and motherhood, King’s only idea for a story focusing on a female character was centered around the character’s reproductive abilities. Another notable thing about the formation of the idea that would become Carrie is that King hated the story. In his essay “On Becoming a Brand Name,” he wrote that “I persisted because I was dry and had no better ideas…my considered opinion was that I had written the world’s all-time loser” (King). The 1980s saw the advent of what would become one of the most well-known subgenres of horror: the slasher film. By the 1980s, many of the goals of the second wave of feminism had been met or were well on their way to being met. The Roe v. Wade case had been won, giving women more reproductive freedom, women had more opportunities for employment, and several antidiscrimination laws had been passed. However, many of the horror films of the time still contained sexist narratives that foreshadow the later backlash against women’s progress. “Possessed girl” stories and “black and white” morality permeated the horror genre at this time. The mid-1990s saw the beginning of the third wave of feminism. One of the most famous works of horror fiction from this era is Joss Whedon’s television series Buffy the Vampire Slayer, which ran from 1997 to 2003. Its popularity and cultural impact created the term “The Buffy Era” as a nickname for this time period. This era marks the beginning of what has been termed “backlash broadcasting.” On the subject of the backlash against women’s rights, journalist Susan Faludi writes “it’s a recurring phenomenon: it returns every time women begin to make some headway toward equality” (Faludi 46). Though backlash is not unique to the modern era, the prominence of media in everyday life makes the backlash extremely visible. Motherhood in Horror Maternal death is a major theme in Frankenstein. In her critical essay “Frankenstein’s Circumvention of the Maternal,” Margaret Homans writes “There are many mothers in the Frankenstein circle, and all die horrible deaths” (Homans 134). In addition to Frankenstein’s mother, Caroline, mother-figures Elizabeth and Justine also suffer early deaths. Elizabeth’s illness causes Caroline’s death, after which Elizabeth becomes a replacement for her, acting as a surrogate mother to Frankenstein’s younger brothers. Justine Moritz also becomes a substitute mother to William Frankenstein, and later is accused of and executed for his murder (Homans 134). The death of mothers and mother-figures is a common theme is Joss Whedon’s TV series’ 144 Buffy the Vampire Slayer and Angel, which will both be discussed in greater depth later. On Buffy, in the episode “The Body,” the main character, Buffy Summers, comes home to find her mother dead. Joyce Summers’s death forces Buffy into the role of being the “mother” to her sister, Dawn. Her new role as Dawn’s primary protector directly leads to Buffy sacrificing herself, jumping to her death in order to close a portal that can only be closed with the death of one of the sisters (“The Gift”). Buffy is even told that her death is her “gift” to her sister and to the world (“Intervention”). Like the deaths of the mother-figures in Frankenstein, Buffy’s motherhood is painted as inherently linked to sacrifice. Despite the fact that Buffy did have other options (killing the person who planned to open the portal, or stalling them long enough that it would be past the time frame in which they could open it), she, like Caroline Frankenstein, chose the route that she knew could result in her death. After Buffy’s death, her friends Willow and Tara take over as mother-figures to Dawn. Their maternal roles are highlighted and cemented in place after Willow and Tara break up, and Tara gives Dawn the typical “divorced parents talk,” in which she assures Dawn that their break-up was not her fault and they can still see each other at any time (“Smashed”). Tara is killed immediately after getting back together with Willow (“Seeing Red”). Though, unlike Buffy Summers and Caroline Frankenstein, her death is not directly linked to her being a mother, it is linked to her involvement with the Summers family, much like Justine Moritz’s death is linked to her involvement with the Frankensteins. If Tara hadn’t chosen to be part of the family and to reassume her role as Dawn’s substitute mother, she wouldn’t have been in the Summers house, and wouldn’t have been shot by the stray bullet that was meant for Buffy. Angel, Joss Whedon’s spin-off of Buffy, which ran from 1999 to 2004, is also rife with the theme of maternal death. In the show’s second season, a vampire named Darla, Angel’s former lover who was killed in the early seasons of Buffy, is resurrected. After failing to kill Angel, she simply disappears. In the third season, she returns, nine months pregnant with Angel’s child (“Offspring”). She later kills herself by driving a stake through her own heard for the sole purpose of giving birth to a child she didn’t want (“Lullaby”). Later in the season, Cordelia, who became a surrogate mother to Darla and Angel’s son, ascends to a higher plane of existence, being forced to abandon her family, and is assumed to be gone forever (“Ground State”). She later returns to Earth and gets pregnant, which ends in her going into a coma and eventually dying. No discussion of motherhood in horror would be complete without mention of Supernatural, a horror series that first aired in 2005 and is still airing today. The first death in the series occurs in the pilot episode, when Mary Winchester, the mother of the two main characters, is murdered by a demon. Because of her death, the boy’ father raises them to hunt monsters, robbing them of any chance they had at a normal childhood (“Pilot”). Mary is put on a pedestal, with her role as a wife and mother becoming a symbol of the normal life her family lost. Twenty-two years after her death, Sam, the younger of the two brothers, has left the family and is living a normal, stable life with his girlfriend, Jess. At the end of the pilot episode, Sam returns home to find Jess dead, murdered in the same manner as Mary (“Pilot”). Her death causes Sam to be drawn back into the life he left behind, giving up his chance at law school, marriage, and normality in general. Like Mary, Jess wasn’t a character in her own right, but a symbol of a normal, domestic life the main characters couldn’t have. In the third season, the character of Lisa Braeden is introduced. Lisa is an old girlfriend of Dean, the older brother, and the mother of a child who is suspected to be his (“The Kids Are Alright”). In the sixth season, Dean moves in with Lisa and begins living a relatively normal life (“Exile On Main St.”). Eventually, this falls apart when Lisa and her son, Ben, are kidnapped by demons, and Lisa is possessed. Dean attempts an exorcism, even though he knows Lisa will probably die if he goes through with it (“Let It Bleed”). Even though Lisa survives, Dean has symbolically killed her and, by extension, destroyed his normal life. In the seventh season, Linda Tran, the mother of a teenage prophet, is introduced. Despite 145 the revelation that all things supernatural are real, she tries to protect her son and keep his life as normal as possible, even as he’s being hunted by demons (“What’s Up, Tiger Mommy?”). When she goes missing, it’s assumed that she’s been killed by the main antagonist, Crowley. When Crowley reveals that Kevin’s mother is alive, Dean forbids Kevin from trying to rescue her (“Devil May Care”). Though Linda Tran is her own character and, as such, her connection to the idea of a normal life isn’t as implicit as Marry, Jess, or Lisa’s, she is the last part of Kevin’s old life that remains after he learns he’s a prophet. Once he’s forced to give up his life as a normal teenager, Dean also forces him to give up on saving his mother. In Stephen King’s Carrie, Margaret White’s motherhood is written as being directly connected to her mental instability. After her first pregnancy resulted in a miscarriage, sex, pregnancy, and childbirth became linked in with sin in her mind, causing her to consider Carrie her “punishment.” This led to her mental break and her abuse of her daughter (King 153-154). In Friday the Thirteenth, the murders committed by Pamela Voorheese are inherently tied to her role as a mother. Her anger which drove her to kill the counselors at Camp Crystal Lake was caused by the death of her son, Jason, who drowned while two camp counselors weren’t paying attention. It’s worth noting that this film was released in 1980, during the precursor to today’s epidemic of backlash broadcasting, and that Mrs. Voorheese was a working mother who the audience is supposed to see as a villain. When viewed through this lens, Friday the Thirteenth can be taken as an early backlash film, attempting to moralize about the perceived “dangers” or working women. Women’s Agency in Horror As Anne K. Mellor points out in her essay “Possessing Nature,” a key part of the narrative of Frankenstein revolves around a man removing or restricting women’s agency. “He [Frankenstein] is afraid of an independent female will, afraid that his female creature will have desires and opinions that cannot be controlled by his male creature” (Mellor 360). Mellor also notes that in attempting to create life on his own, Frankenstein is “stealing the female’s control over reproduction” (Mellor 355). In addition to his attempts to remove women’s agency through his scientific endeavors, Frankenstein does the same through his secrecy. By not telling anyone about the Creature, he removes Justine’s ability to defend herself against the accusation of murder. He also denies Elizabeth the ability to remove herself from the danger of being associated with him. In both these cases, his secrecy and disregard for these women’s lives and agency results in their deaths. In the modern era, women’s agency is often restricted, both in real life and in fiction. Supernatural has become infamous for this. However, in contrast to Frankenstein, Supernatural portrays men’s restriction of women’s agency as a good, necessary thing. In the second season episode “No Exit,” Sam and Dean team up with Jo Harvelle, the daughter of an old friend of their father’s, to hunt and destroy the ghost of H. H. Holmes, who has been killing young women. Despite the fact that Jo is an adult who was also raised as a hunter (“Everybody Loves A Clown”), Dean attempts to stop her, not just from joining their hunt, but from hunting in general. He even goes so far as to threaten to call her mother, as though Jo were a child, rather than a woman in her 20s. Though Jo helps them defeat Holmes, she is sent back home to her mother, and is only considered a member of the Winchesters’ “team” when they have no other options (“Abandon All Hope”). In the current season, Carlie Bradbury, Sam and Dean’s ally from season seven, returns and reveals that she has been hunting ghosts and vampires on her own (“Slumber Party”). Dean’s reaction is very much the same as his reaction to Jo, despite the fact that Charlie is also a grown woman. Dean’s paternalistic views are proven “right” when Charlie is killed. Even though she is resurrected, her death is still part of the show’s pattern of infantilizing women and treating them as if they can’t “handle” their own agency. One of the most well-known examples of Supernatural’s removal of women’s agency is the fate of Lisa Braeden. After she is possessed and nearly dies, Dean has her memory of him wiped (“Let It Bleed”). Because she remembers nothing about him, she also knows nothing about demons 146 or monsters, and therefore loses her ability to protect herself and her son from future danger. Often in modern media, women who control and retain their agency are portrayed as villains. On Buffy the Vampire Slayer, female vampires are written as inherently sexual (and predatory) creatures, with their sexuality directly tied to their dangerous nature. The first non-vampiric adult woman with sexual agency on the show turns out to be a praying-mantis-like monster who poses as a substitute teacher in order to feed on male students (“Teacher’s Pet”). In the Supernatural episode “Bloody Mary,” the boys encounter a true version of the Bloody Mary urban legend.4 The ghost of a woman named Mary is revealed to be responsible for a series of deaths in a small town. The deaths are Mary’s attempts to get justice for her own murder, as well as the murders committed by her victims. The audience is not intended to have any sympathy for Mary, instead viewing her as another ghost to destroy. Metaphors for Female Adolescence Most well-known contemporary mainstream fiction is created by men, the majority of whom have no way of understanding what it means to be a woman, let alone an adolescent girl. Stereotypes of women as mysterious and unknowable creatures, practically a separate species from men, persist to this day. Nowhere is this more evident than in second-wave horror’s consistent portrayal of female biology as a metaphor for the monstrous. Tammy Oler writes that “Carrie and her cohorts entered puberty at a time when the horror genre was obsessed with the female curse” (Oler 32). She also writes that “As much as images like those in the likes of Carrie and The Exorcist offer the possibility of embracing a distinctly feminine source of power, they threaten to reduce girls to mere expressions of their biological essence” (Oler 36). Indeed, in Stephen King’s Carrie, a future study of Carrie’s telekinetic abilities states that “Carrie White’s exceptionally late and traumatic commencement of the menstrual cycle might well have provided the trigger for her latent talent” (King 8). In his interview for Twilight Zone Magazine, King stated that in his original concept for the book, it was specifically the hormones related to menstruation that caused Carrie’s telekinetic abilities. In her article “Bloodletting: Female Adolescence in Modern Horror Films,” Oler pays special attention to the 2000 film Ginger Snaps. The plot of this film revolves around a teenage girl named Ginger who is bitten by a werewolf shortly after having her first period. Older notes that “When her younger sister Brigitte begins to suspect that Ginger is undergoing more than just ‘the most normal thing in the world,’ she observes, ‘Something’s wrong – like more than you being just female’” (Oler 35). The conflation of being female with being “wrong” speaks to the idea of a young woman going through puberty as something inherently wrong or scary. Even today, in a supposedly enlightened and sex-positive era, this positioning of female puberty as equivalent to the monsters of horror is still common. In the Buffy the Vampire Slayer episode “Innocence,” Buffy’s vampire boyfriend, Angel, loses his soul and, by extension, his humanity, as a direct result of having sex with Buffy. This not only causes Angel’s eventual death when Buffy is forced to kill him (“Becoming (Part 2)”), but also results in the deaths of some of her classmates (“Phases”) and her teacher and mentor, Jenny (“Passion”). In the first season of Supernatural, the Winchesters discover a true version of another urban legend: the hook man.5 The ghost of a preacher who had been executed for murdering prostitutes has attached itself to Lori, an 18-year-old girl who is the daughter of the town’s current preacher. He feeds off Lori’s repressed emotions and kills people both she and ghost view as “immoral,” including Lori’s over-affectionate boyfriend, her sexually active roommate, and her father, who is having an affair with a married woman (“Hook Man”). The deaths of each victim are directly tied to Lori’s confusion about her repressed sexuality, and her religious beliefs about how sexuality should be expressed. In Supernatural’s seventh season, the Amazons join the show’s mythology. Dean has a onenight stand with an Amazon warrior, which results in the conception and birth of a daughter. It’s 147 quickly revealed that the Amazons are responsible for the recent deaths in the town. The Winchesters discover that once an Amazon girl turns 16, she has to kill her father as part of her coming-of-age ritual (“The Slice Girls”). In this case, the connection could not be more obvious: the process of a girl becoming a woman is contingent on someone else’s death. These examples show that despite the massive amounts of medical knowledge on the subject, female adolescence is still viewed as a dark unknown, as something to be feared. Male puberty isn’t portrayed this way in the modern era. The MTV series Teen Wolf, like the film Ginger Snaps, uses lycanthropy as a metaphor for puberty, but the protagonist is a boy rather than a girl. Teen Wolf portrays being a werewolf as being an amazing experience, while Ginger Snaps paints it as a tragedy. Male puberty turns boys into superheroes, while female puberty turns girls into monsters. The Possessed Girl Another common narrative involving teenage girls is the “possessed girl.” Oler describes the possessed girl as a narrative in which “women’s very bodies become the Pandora’s box that unleashes evil into the worldly domain” (Oler 33). Oler cites Carrie, Ginger Snaps, and The Exorcist as examples of this narrative. Oler writes that “while these girls [Carrie, Ginger, and Regan] are ostensibly the films’ subjects, the narrative action inevitable reduces them to being merely bodies themselves, with their actual experiences rarely investigated or explored” (Oler 33). In Carrie, most of the narrative is from the perspective of other people, whether it’s one of Carrie’s classmates, one of her teachers, a journalist reporting on prom night, or a scientist studying telekinesis. Little of the overall story is focused on Carrie herself, rather, the focus is on her power and the destruction she causes with it. In Ginger Snaps, the focus shifts away from Ginger after her transformation, with the story becoming more about her still-human (and still-prepubescent) sister. On the subject of The Exorcist, Oler writes that “Regan transforms from girl to female portal so thoroughly that her character’s only cry for help is literally written on her body (“help me” spelled out in the raised skin of her stomach). At the end of the film, when an enraged Father Karras, the titular exorcist, physically assaults Regan, the audience barely registers any shock. In no other film context would the act of a grown man punching a teenage girl be acceptable, or even understandable. Yet the action that immediately follows – Karras is himself possessed and subsequently hurled out the window to his death – makes it clear that this is really his story and not that of the young girl left crying in the corner of her room. Regan spends the few remaining moments of the film gaunt and silent, hardly even a witness to her own terrifying trials. No longer “open” (at least not until the sequel) thanks to Karras’s sacrifice, she becomes useless as the object of horror – and as the subject of the film” (Oler 34). The possession of adolescent girls by outside forces, most often brought on by or directly coinciding with their maturation into women, serves to dehumanize them to the point where they are able to be abused for the sake of the plot, without the audience wondering what they are going through. Despite being supernatural in focus, this narrative is mirrored in the real world, with violence against women being at epidemic levels in Western society. While young girls on screen are being painted as inhuman monsters, and therefore acceptable targets for abuse, real-life women and girls are faced with a reality in which one in four women experience actual or attempted rape (“Sexual Assault Statistics”), and a woman’s murder will get a man an average sentence of two to six years in jail (“The Michigan Women’s Justice & Clemency Project”). The “possessed girl” narratives, when paired with these statistics, show a vivid picture of how Western society devalues women. Subverted Tropes TV Tropes, an online database that defines, categorizes, and tracks tropes in media throughout history, defines a “subversion” as “playing bait and switch with a trope. A work makes you think a trope is going to happen, but it doesn’t” (“Subverted Trope”). Frankenstein subverts one of the oldest tropes in literature, dating back to the book of Genesis: the idea that women are the source of trouble, and the source of humanity’s general woes. In Frankenstein, the protagonist’s 148 troubles are caused not by a woman, but by the fact that he tries to create life on his own, cutting women out of the process. Wes Craven’s 1984 film A Nightmare On Elm Street subverts another well-known trope: the damsel in distress. While Nancy seems like the typical female horror protagonist, she’s the one who saves herself in the end and defeats Freddy Krueger. Her mother defies this trope as well, being one of the mob that originally killed Krueger to save their children. Whedon’s World Writer/director Joss Whedon is often hailed as a beacon of feminism in an otherwise incredibly sexist television and film industry. However, in recent years, women have come out of the woodwork criticizing Whedon for employing just as many sexist tropes as his cohorts, and masking it in rhetoric about “strong female characters.” When Buffy the Vampire Slayer first aired in 1997, it became notable for supposedly subverting two very common tropes: the “dumb blonde” stereotype, and the damsel in distress. However, Buffy’s liberation from these tropes is contingent upon other women being victims of them. In the first episode, the character of Cordelia Chase takes Buffy’s place as the by then archetypal “stupid girl who always gets killed in horror movies.” Not only does this play into the idea that there must always be a “dumb” woman, rather than having all female characters possess equal amounts of intelligence, it also plays off the stereotypes that women are inherently competitive with each other. As the “dumb” girl, Cordelia is set up as the most popular girl in school, the airheaded cheerleader who the boys fall head over heels for. When Buffy moves to Sunnydale, Cordelia’s position as the prettiest and the most popular is threatened, ‘causing her to bully Buffy and spread rumors about her. Whedon also uses the concept of the “good girl” vs the “bad girl” to set up the contrast between Buffy, and the “dark Slayer” Faith, introduced in season three. Faith is set up as the antiBuffy. She grew up poor, it’s implied she was abused as a child, she’s been in trouble with the law, and she is fiercely independent, not wanting to rely on family or friends for anything. But the key difference that is brought up in any discussion of Faith is that she is not the sexually “pure” girl that Buffy is, but rather a sexually aware and mature woman. One of the first clues the audience gets that there is something “wrong” about Faith is that she is open about the fact that she’s an adult who has sex (“Faith, Hope & Trick”). In her book Sex and the Slayer: A Gender Studies Primer for the Buffy Fan, Lorna Jowett writes that “Faith is ‘bad’ not because she participates in sex and violence but because she enjoys them” (Jowett 86). Part of the Western cultural narrative about what defines a “good girl” is that women are meant to have sex for specific purposes (most often procreation), but aren’t supposed to enjoy sex (Valenti 196). Women who enjoy sex, like Faith, are viewed as being “deviant” or “transgressive” (Jowett 86). As was discussed earlier, both Buffy the Vampire Slayer and Angel carry the theme of maternal death. Buffy is often praised as a “strong female character,” but the deaths of the maternal figures in her life are used by the narrative to punish her for her strength and to drive Buffy into a dark place where her situation seems hopeless. Jenny’s death in the first season was used to prevent Buffy from being able to return Angel’s soul (“Passion”). Joyce’s death was used to destroy the happy and relatively normal life Buffy had been leading during the fifth season (“The Body”). Buffy’s own death becomes her punishment for attempting to find a way to save her sister, as well as being Dawn’s punishment for her inherent magical power. If Dawn wasn’t this powerful, there would have been no need for Buffy to sacrifice herself (“The Gift”). The final sexist trope that will be discussed in the context of Whedon’s universe is the “mystical pregnancy.” The mystical pregnancy occurs when a female character’s pregnancy is caused or affected by supernatural forces. This happens multiple times on the series Angel, most notable to Angel’s two love interests, Darla and Cordelia. In both cases, the pregnancy isn’t part of the female character’s story arc, but rather a way to throw a new hurdle in the path of the male lead. 149 When Darla appears in the third season, pregnant with Angel’s child, she kills herself to save it, despite the fact that she didn’t want to have the baby in the first place (“Lullaby”). The main point of Darla’s pregnancy story line isn’t about her. The focus is placed on Angel having to take responsibility for a child he didn’t know he was going to have. When Cordelia becomes pregnant, the focus of the narrative is once again placed on Angel. Cordelia’s pregnancy becomes a way to give Angel a new villain to defeat (“Inside Out”). Conclusion Over the years, women have made a lot of progress toward equality. However, each time progress is made, a new wave of anti-feminism arises to attempt to undo what little has been done. This is extremely evident in the media produced by Western society, whether that media is offering commentary on sexist ideas, such as in Frankenstein, or is directly endorsing them, like Supernatural or Buffy the Vampire Slayer. Women appear to be in a much better place in society than we were in the nineteenth century, but digging beneath the surface reveals that systemic misogyny hasn’t gotten that much better, and in some ways has gotten worse. This is reflected in our media, which appears to be more egalitarian until the material is examined more closely, after which the sexist nature of some of our most popular books, TV shows, and films can be more clearly seen. Notes 1 In most Western countries, first wave feminism was mostly focused on white women. In the United States, slavery was still legal for much of the first wave of feminism, and black women were rarely included in women’s struggle for civil rights until the late nineteenth century. 2 Like the first wave, the second wave was largely focused on white women and middle-class women, especially when it came to employment opportunities. Poor women and women of color had already been working outside the home due to necessity. White middle-class women were the ones who primarily benefitted from the second wave’s focus on employment. The early second wave movement also distanced itself from the struggles of disabled and lesbian, bisexual, and transgender women. 3 Mainstream third-wave feminism started off also with the focus on white women. However, due to the advent of the Internet, intersectional feminism became more well-known, with more women of color, disabled women, and queer women gaining a platform through which their voices could be heard and they could connect with other women who faced the same struggles. 4 The Bloody Mary legend states that if you say the words “Bloody Mary” three times into a mirror, her ghost will appear. The actual original Bloody Mary was Queen Mary I of England, who had hundreds of Protestants killed during her brief reign (“Mary I, Queen of England (1516-1558)”). The Mary who appears in this episode of Supernatural is unrelated to the original Bloody Mary. 5 The Hook Man is an urban legend about two teenagers who sneak off into the woods together, and hear a radio broadcast about an escaped convict who is known to kill people with a hook. When they hear a noise in the woods, the boyfriend leaves the car to check it out, only to be killed by the Hook Man. The girl escapes, but when she arrives at home, she finds a hook dangling from the handle of her car door. 150 Works Cited A Nightmare on Elm Street. Dir. Wes Craven. Perf. Robert Englund, Heather Langenkamp. New Line Cinema, 1984. DVD. “Abandon All Hope.” Supernatural: The Complete Fifth Season. Writ. Ben Edlund. Dir. Philip Sgriccia. Warner Brothers, 2009. DVD. “Becoming (Part 2).” Buffy the Vampire Slayer: The Complete Second Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 1998. DVD. “Bloody Mary.” Supernatural: The Complete First Season. Writ. Ron Milbauer and Terri Hughes Burton. Dir. Peter Ellis. Warner Brothers, 2005. DVD. “The Body.” Buffy the Vampire Slayer: The Complete Fifth Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 2001. DVD. “Devil May Care.” Supernatural. Writ. Andrew Dabb. Dir. Guy Bee. The CW, 2013. Television. “Everybody Loves A Clown.” Supernatural: The Complete Second Season. Writ. John Shiban. Dir. Philip Sgriccia. Warner Brothers, 2006. DVD. “Exile On Main St.” Supernatural: The Complete Sixth Season. Writ. Sera Gamble. Dir. Philip Sgriccia. Warner Brothers, 2010. DVD. The Exorcist. Dir. William Friedkin. Perf. Ellen Burstyn, Max von Sydow, Jason Miller, and Linda Blair. Warner Brothers, 1973. DVD. “Faith, Hope & Trick.” Buffy the Vampire Slayer: The Complete Third Season. Writ. David Greenwalt. Dir. James A. Contner. Warner Brothers, 1998. DVD. Faludi, Susan. Backlash: The Undeclared War Against American Women. New York: Crown Publishers, 1991. Print. Friday the Thirteenth. Dir. Sean S. Cunningham. Perf. Betsy Palmer and Adrienne King. Paramount Pictures, 1980. DVD. Friedan, Betty. The Feminine Mystique. New York: W. W. Norton, 1963. Print. “The Gift.” Buffy the Vampire Slayer: The Complete Fifth Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 2001. DVD. Gordon, Lyndall. Vindication: A Life of Mary Wollstonecraft. New York: HarperCollins, 2005. Print. Grant, Charles L. “Stephen King: ‘I Like to go for the Jugular.’” Twilight Zone Magazine April 1981. Print. “Ground State.” Angel: The Complete Fourth Season. Writ. Mere Smith. Dir. Michael Grossman. Warner Brothers, 2002. DVD. “Hook Man.” Supernatural: The Complete First Season. Writ. John Shiban. Dir. David Jackson. Warner Brothers, 2005. DVD. “Innocence.” Buffy The Vampire Slayer: The Complete Second Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 1998. DVD. “Inside Out.” Angel: The Complete Fourth Season. Writ. Steven S. DeKnight. Dir. Steven S. DeKnight. Warner Brothers, 2003. DVD “Intervention.” Buffy the Vampire Slayer: The Complete Fifth Season. Writ. Jane Espenson. Dir. Michael Gershman. Warner Brothers, 2001. DVD. Jowett, Lorna. Sex and the Slayer: A Gender Studies Primer for the Buffy Fan. Middletown, CT: Wesleyan University Press, 2005. Print. “The Kids Are Alright.” Supernatural: The Complete Third Season. Writ. Sera Gamble. Dir. Philip Sgriccia. Warner Brothers, 2007. DVD. 151 King, Stephen. Carrie. New York: Doubleday, 1974. Print. King, Stephen. “On Becoming A Brand Name.” Adelina Magazine February 1980. Print. “Let It Bleed.” Supernatural: The Complete Sixth Season. Writ. Sera Gamble. Dir. John Showalter. Warner Brothers, 2011. DVD. “Lullaby.” Angel: The Complete Third Season. Writ. Tim Minear. Dir. Tim Minear. Warner Brothers, 2001. “Mary I, Queen of England (1516-1558).” Luminarium Enyclopedia Project. N.p., n.d. Web. 19 Nov. 2013. <http://www.luminarium.org/encyclopedia/queenmary.htm>. Mellor, Anne K. “Possessing Nature: The Female In Frankenstein.” Frankenstein. Ed. J. Paul Hunter. New York: W.W. Northn, 2012. 355 – 368. Print. “The Michigan Women's Justice & Clemency Project.” The Michigan Women's Justice & Clemency Project. N.p., n.d. Web. 19 Nov. 2013. <http://www.umich.edu/~clemency/clemency_mnl/ch1.html>. “No Exit.” Supernatural: The Complete Second Season. Writ. Matt Witten. Dir. Kim Manners. Warner Brothers, 2006. DVD. “Offspring.” Angel: The Complete Third Season. Writ. David Greenwalt. Dir. Turi Meyer. Warner Brothers, 2001. DVD. Oler, Tammy. “Bloodletting: Female Adolescence in Modern Horror Films.” Bitchfest. Ed. Lisa Jervis and Andi Zeisler. New York: Farrar, Straus and Giroux, 2006. 31-37. Print. “Passion.” Buffy the Vampire Slayer: The Complete Second Season. Writ. Ty King. Dir. Michael Gershman. Warner Brothers, 1998. DVD. “Phases.” Buffy the Vampire Slayer: The Complete Second Season. Writ. Rob Des Hotel & Dean Batali. Dir. Bruce Seth Green. Warner Brothers, 1998. DVD. “Pilot.” Supernatural: The Complete First Season. Writ. Eric Kripke. Dir. David Nutter. Warner Brothers, 2005. DVD. “Seeing Red.” Buffy the Vampire Slayer: The Complete Sixth Season. Writ. Steven DeKnight. Dir. Michael Gershman. Warner Brothers, 2002. DVD. “Sexual Assault Statistics.” One In Four USA. N.p., n.d. Web. 19 Nov. 2013. <http://www.oneinfourusa.org/statistics.php>. Shelley, Mary. Frankenstein. Ed. J. Paul Hunter. New York: W.W. Norton, 2012. Print. “The Slice Girls.” Supernatural: The Complete Seventh Season. Writ. Brad Buckner & Eugenie Ross-Leming. Dir. Jerry Wanek. Warner Brothers, 2012. DVD. “Slumber Party.” Supernatural. Writ. Robbie Thompson. Dir. Robert Singer. The CW, 2013. Television. “Smashed.” Buffy the Vampire Slayer: The Complete Sixth Season. Writ. Drew Z. Greenburg. Dir. Turi Meyer. Warner Brothers, 2001. DVD. “Subverted Trope.” TVTropes.org. N.p., n.d. Web. 19 Nov. 2013. <http://tvtropes.org/pmwiki/pmwiki.php/Main/SubvertedTrope>. Taylor, Barbara. “Mary Wollstonecraft and the Wild Wish of Early Feminism.” History Workshop Journal. No. 33 (1992): 197-219. JSTOR. Web. 24 November 2013. “Teacher’s Pet.” Buffy The Vampire Slayer: The Complete First Season. Writ. David Greenwalt. Dir. Bruce Seth Green. Warner Brothers, 1997. DVD. Valenti, Jessica. The Purity Myth: How America’s Obsession with Virginity is Hurting Young Women. Berkeley, CA: Seal Press, 2010. Print. “What’s Up, Tiger Mommy?” Supernatural: The Complete Eighth Season. Writ. Andrew Dabb & Daniel Loflin. Dir. John Showalter. Warner Brothers, 2012. DVD. 152 Bibliography A Nightmare on Elm Street. Dir. Wes Craven. Perf. Robert Englund, Heather Langenkamp. New Line Cinema, 1984. DVD. “Abandon All Hope.” Supernatural: The Complete Fifth Season. Writ. Ben Edlund. Dir. Philip Sgriccia. Warner Brothers, 2009. DVD. “Becoming (Part 2).” Buffy the Vampire Slayer: The Complete Second Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 1998. DVD. “Bloody Mary.” Supernatural: The Complete First Season. Writ. Ron Milbauer and Terri Hughes Burton. Dir. Peter Ellis. Warner Brothers, 2005. DVD. “The Body.” Buffy the Vampire Slayer: The Complete Fifth Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 2001. DVD. DeLamotte, Eugenia C. Perils of the Night: A Feminist Study of Nineteenth-Century Gothic. Oxford: Oxford University Press, 1990. Print. “Devil May Care.” Supernatural. Writ. Andrew Dabb. Dir. Guy Bee. The CW, 2013. Television. “Everybody Loves A Clown.” Supernatural: The Complete Second Season. Writ. John Shiban. Dir. Philip Sgriccia. Warner Brothers, 2006. DVD. “Exile On Main St.” Supernatural: The Complete Sixth Season. Writ. Sera Gamble. Dir. Philip Sgriccia. Warner Brothers, 2010. DVD. The Exorcist. Dir. William Friedkin. Perf. Ellen Burstyn, Max von Sydow, Jason Miller, and Linda Blair. Warner Brothers, 1973. DVD. “Faith, Hope & Trick.” Buffy the Vampire Slayer: The Complete Third Season. Writ. David Greenwalt. Dir. James A. Contner. Warner Brothers, 1998. DVD. Faludi, Susan. Backlash: The Undeclared War Against American Women. New York: Crown Publishers, 1991. Print. Friday the Thirteenth. Dir. Sean S. Cunningham. Perf. Betsy Palmer and Adrienne King. Paramount Pictures, 1980. DVD. Friedan, Betty. The Feminine Mystique. New York: W. W. Norton, 1963. Print. Fritzsche, Bettina. “Spicy Strategies: Pop Feminist and Other Empowerments in Girl Culture.” All About The Girl. Ed. Anita Harris. New York: Routledge, 2004. 155-162. Print. “The Gift.” Buffy the Vampire Slayer: The Complete Fifth Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 2001. DVD. Gordon, Lyndall. Vindication: A Life of Mary Wollstonecraft. New York: HarperCollins, 2005. Print. Grant, Charles L. “Stephen King: ‘I Like to go for the Jugular.’” Twilight Zone Magazine April 1981. Print. “Ground State.” Angel: The Complete Fourth Season. Writ. Mere Smith. Dir. Michael Grossman. Warner Brothers, 2002. DVD. “Hook Man.” Supernatural: The Complete First Season. Writ. John Shiban. Dir. David Jackson. Warner Brothers, 2005. DVD. “Innocence.” Buffy The Vampire Slayer: The Complete Second Season. Writ. Joss Whedon. Dir. Joss Whedon. Warner Brothers, 1998. DVD. “Inside Out.” Angel: The Complete Fourth Season. Writ. Steven S. DeKnight. Dir. Steven S. DeKnight. Warner Brothers, 2003. DVD 153 “Intervention.” Buffy the Vampire Slayer: The Complete Fifth Season. Writ. Jane Espenson. Dir. Michael Gershman. Warner Brothers, 2001. DVD. Jowett, Lorna. Sex and the Slayer: A Gender Studies Primer for the Buffy Fan. Middletown, CT: Wesleyan University Press, 2005. Print. “The Kids Are Alright.” Supernatural: The Complete Third Season. Writ. Sera Gamble. Dir. Philip Sgriccia. Warner Brothers, 2007. DVD. King, Stephen. Carrie. New York: Doubleday, 1974. Print. King, Stephen. “On Becoming A Brand Name.” Adelina Magazine February 1980. Print. “Let It Bleed.” Supernatural: The Complete Sixth Season. Writ. Sera Gamble. Dir. John Showalter. Warner Brothers, 2011. DVD. “Lullaby.” Angel: The Complete Third Season. Writ. Tim Minear. Dir. Tim Minear. Warner Brothers, 2001. “Mary I, Queen of England (1516-1558).” Luminarium Enyclopedia Project. N.p., n.d. Web. 19 Nov. 2013. <http://www.luminarium.org/encyclopedia/queenmary.htm>. Mellor, Anne K. “Possessing Nature: The Female In Frankenstein.” Frankenstein. Ed. J. Paul Hunter. New York: W.W. Northn, 2012. 355-368. Print. “The Michigan Women's Justice & Clemency Project.” The Michigan Women's Justice & Clemency Project. N.p., n.d. Web. 19 Nov. 2013. <http://www.umich.edu/~clemency/clemency_mnl/ch1.html>. Moss, Gabrielle. “Teen Mean Fighting Machine: Why Does the Media Love Mean Girls?” Bitchfest. Ed. Lisa Jervis and Andi Zeisler. New York: Farrar, Strauss and Giroux, 2006. 43-48. Print. “No Exit.” Supernatural: The Complete Second Season. Writ. Matt Witten. Dir. Kim Manners. Warner Brothers, 2006. DVD. “Offspring.” Angel: The Complete Third Season. Writ. David Greenwalt. Dir. Turi Meyer. Warner Brothers, 2001. DVD. Oler, Tammy. “Bloodletting: Female Adolescence in Modern Horror Films.” Bitchfest. Ed. Lisa Jervis and Andi Zeisler. New York: Farrar, Straus and Giroux, 2006. 31-37. Print. “Passion.” Buffy the Vampire Slayer: The Complete Second Season. Writ. Ty King. Dir. Michael Gershman. Warner Brothers, 1998. DVD. “Phases.” Buffy the Vampire Slayer: The Complete Second Season. Writ. Rob Des Hotel & Dean Batali. Dir. Bruce Seth Green. Warner Brothers, 1998. DVD. “Pilot.” Supernatural: The Complete First Season. Writ. Eric Kripke. Dir. David Nutter. Warner Brothers, 2005. DVD. “Seeing Red.” Buffy the Vampire Slayer: The Complete Sixth Season. Writ. Steven DeKnight. Dir. Michael Gershman. Warner Brothers, 2002. DVD. “Sexual Assault Statistics.” One In Four USA. N.p., n.d. Web. 19 Nov. 2013. <http://www.oneinfourusa.org/statistics.php>. Shelley, Mary. Frankenstein. Ed. J. Paul Hunter. New York: W.W. Norton, 2012. Print. “The Slice Girls.” Supernatural: The Complete Seventh Season. Writ. Brad Buckner & Eugenie Ross-Leming. Dir. Jerry Wanek. Warner Brothers, 2012. DVD. “Slumber Party.” Supernatural. Writ. Robbie Thompson. Dir. Robert Singer. The CW, 2013. Television. “Smashed.” Buffy the Vampire Slayer: The Complete Sixth Season. Writ. Drew Z. Greenburg. Dir. Turi Meyer. Warner Brothers, 2001. DVD. “Subverted Trope.” TVTropes.org. N.p., n.d. Web. 19 Nov. 2013. <http://tvtropes.org/pmwiki/pmwiki.php/Main/SubvertedTrope>. 154 Taft, Jessica K. “Girl Power Politics: Pop Culture Barriers Organizational Resistance.” All About The Girl. Ed. Anita Harris. New York: Routledge, 2004. 69-78. Print. Taylor, Barbara. “Mary Wollstonecraft and the Wild Wish of Early Feminism.” History Workshop Journal. No. 33 (1992): 197-219. JSTOR. Web. 24 November 2013. “Teacher’s Pet.” Buffy The Vampire Slayer: The Complete First Season. Writ. David Greenwalt. Dir. Bruce Seth Green. Warner Brothers, 1997. DVD. Valenti, Jessica. The Purity Myth: How America’s Obsession with Virginity is Hurting Young Women. Berkeley, CA: Seal Press, 2010. Print. “What’s Up, Tiger Mommy?” Supernatural: The Complete Eighth Season. Writ. Andrew Dabb & Daniel Loflin. Dir. John Showalter. Warner Brothers, 2012. DVD. 155 Naturally Occurring Persistent Organic Pollutant by Ryan Woitas (Chemistry 1552) N ative across forests of Central and Northern Europe lives a species so vulgar, it would be hard pressed to think that such a beast could possess a certain quality that gives significant meaning to its existence. With its compact body, stubby legs, and wide head, the bristly haired wild boar may provide insight to chemists about an organic compound called drosophilin A (p-methoxytetrachlorophenol) that closely resembles persistent organic pollutants. Industrial persistent organic pollutants are halogenated molecules that are intrinsic with environmental toxicity.1 First studied by chemists in 1952, drosophilin A has become part of the modern scientific community due to an alarming discovery of the potentially fatal organic compound residing in wild boar meat sold in German food markets. Persistent organic pollutants (POPs) are not uncommon; there are theoretically reported 209 polychlorinated biphenyls (POPs) that differ from each other only by their constituents quantities, primarily the chlorine, and substitution positions. POPs are known for their hydrophobic and lipophilic desires as well as displaying attractive tendencies towards organic solids when in aqueous environments in an effort to avoid the liquid phase. What makes POPs especially harmful to mammalian body systems is the attractive nature of the toxic molecules to lipid storages, which consequently causes the chemical to remain in mammalian adipose tissue storages, indefinitely. POPs also have an affinity for traveling; POPs may enter the gas phase in certain environmental conditions, as temperatures warm, POPs are expelled into the lower atmosphere where they can travel long distances due to their resilient nature to breaking down or forming more stable compounds in the air. Such a characteristic of POPs is a cause of concern since it is possible for the toxic compounds to travel far from their origin, depositing into river beds contaminating water and aquatic life.2 Drosophilin A was the first natural chlorinated compound studied and analyzed to contain a halogenated benzene ring that was first isolated in 1952 from a basidiomycete, a fungus whose spores develop in basidia.3 Basidiomycetes are the parent group of most familiar mushrooms and toadstools. Table 1 displays the minimum various species that are known to contain drosophilin A. Significance of the aforementioned discovery answers questions that arose to a group of scientists in Germany, who made an astonishing discovery while performing a routine safety inspection of a food market, specifically in the markets supply of wild boar meat. Routinely analyzed samples of wild boar meat tested positive for drosophilin A, however, since these animals are usually hunted in heavily wooded areas away from industrialization (the primary source of POPs) it became apparent that the source of drosophilin A contained within the wild boar meat was from something benign: a subspecies of basidiomycete known to be a part of the wild beasts diet. 1 de Jong, Ed et al. “Significant biogenesis of chlorinated aromatics by fungi in natural environments.” Applied and environmental microbiology 60.1 (1994): 264-270. 2 Jones, Kevin C, and P De Voogt. "Persistent organic pollutants (POPs): state of the science." Environmental Pollution 100.1 (1999): 209-221. 3 Anchel M. 1952. Identification of drosophilin A as p-methoxytetrachlorophenol. J. Am. Chem. Soc. 74:2943. 156 TABLE 1 DROSOPHILIN PRODUCTION BY BASIDIOMYCETES Organism Drosophilin A (mg/L) Agaricus arvensis 0.5004 Bjerkandera adusta 0.5004 Coprinus plicatilis 20.05 Peniophora pseudeopini 1.004 Phellinus fastuosus 15.74& 6 Psathyrella subatrata 3.207 The significance of the German scientists’ discovery centers on the fact that most knowledge of POPs’ environmental impacts deal primarily with amphibious habitats; scientists hope to shed light on the continental behavior of POPs through further analysation of the samples of drosophilin A collected from the market boar meat. The team of German Scientists proliferating the investigation of wild boar meat contamination is being led by Walter Vetter and his technical assistant Katja Lehnert at the University of Hohenheim, one of Germany’s top-tier universities. Analytical techniques used by Vetter and his team are known for producing incredibly accurate results: gas chromatography and mass spectrometry, which helped the scientists confirm that the potentially toxic compound collected from the German markets boar meat was indeed drosophilin A. Although Vetter and his colleagues cannot pinpoint the exact species of basidiomycete ingested by the wild boar, Vetter is certain that the boar acquired drosophilin A from consuming mushrooms as long term data suggest, based on the fact that there are numerous mushroom species that can house the chemical organic compound that are known to be consumed by wild boar. Drosophilin A closely resembles a POP, known toxic industrial chemicals. What Walter Vetter hopes to achieve is an understanding of the effects of drosophilin A on humans, because an apparent risk will initiate recourse to prevent harm to individuals that consume wild boar meat even if it’s (consumed) sparingly, since the intrinsic ability of POPs to remain in adipose storages. However, Vetter and his team are hoping to discover that the wild boar has developed a defense against the compounds undesirable effects along with coming up with an answer to the question, why do species of basidiomycetes contain the toxic compound in the first place? Chemist Gordon W. Gribble of Dartmouth College stated, “If the compound (drosophilin A) turns out to be nontoxic to 4 Teunissen PJM, Swarts HJ, Field JA. 1997. Screening of ligninolytic basidiomycetes for the production of Drosophilin A (tetrachloro-4-methoxyphenol) and Drosophilin A methyl ether (tetrachloro-1,4-dimethoxybenzene). Appl. Microbiol. Biotechnol. In press. 5 Bastian W. 1985. Vergleichende Untersuchungen zum Sekunda ̈rstoffwechsel an Coprophilen und Erdoder Holzbewohnende Basidiomyceten. PhD thesis. Univ. Kaiserslautern, Germany (In German). 6 Singh P, Rangaswami S. 1966. Occurrence of O-methyl-drosophilin A in Fomes fastuosus Lev. Tetrahedron Lett. 11:1229–31. 7 Kavanagh F, Hervey A, Robbins WJ. 1952. Antibiotic substances from basidiomycetes. IX. Drosophila subarata (Batsch:Fr) Quel. Proc. Natl. Acad. Sci. USA 38:555–60. 157 the boars, it could provide useful lessons about how to construct halogenated compounds so they don’t end up becoming a health or environmental threat.” July 15, 1952, proceedings of the National Academy of Sciences published research on four variations of drosophilin A pertaining to the possible antibiotic properties of the aforementioned substances found in twelve species of basidiomycetes from nine genera; the following information gives modern chemists an advantage in terms of analyzing the POP analogue. Antibacterial activity was measured against seven bacteria and data showed that drosophilin A was active primarily against only staphylophage, particularly against strains of staphylococci and B subtilis. However, the fact remains that the inevitable production of naturally occurring drosophilin A remains as an environmental significance.8 The National Academy of Sciences were also able to deduce information during their experimentation in 1952 about the chemical structure and properties of drosophilin A. Experimentation was performed by analyzing a crystalline variant of the organic compound after firstly purified with a benzene solution and treatment with decolorizing carbon. Drosophilin A was determined to have: a melting point of 118 °C; solubility in hexane, benzene, acetone, alcohol, chloroform, ether and slightly soluble in water; and drosophilin A was unable to be extracted by concentrated hydrochloric acid, which suggests that the organic compound lacks alkaline properties, however the chemical is only slightly acidic since it was unaffected by attempts to extract appreciable amounts with a chloroform solution containing bicarbonate. Chemists of the National Academy of Sciences were able to determine the molar mass of drosophilin A cryoscopically in exaltone. Drosophilin A was determined to have the empirical formula of C13H9O4Cl7.9 The fact remains that continued research surrounding drosophilin A and other drosophilins is currently underway due to findings of the potentially fatal organic compound in food consumed by human beings. Rigorous standards set by Germany surrounding testing for food contaminants has proven to be effective, since it is still not known pertaining to the lethality of drosophilin A to humans and mammals alike. Close resemblance of drosophilin A to POPs is enough motivation for chemists to continue researching the less than well-known compound, as to prevent possible future catastrophic events from occurring regarding potential toxicity of drosophilin A in human populations. 8 de Jong, Ed et al. “Significant biogenesis of chlorinated aromatics by fungi in natural environments.” Applied and environmental microbiology 60.1 (1994): 264-270. 9 The quantitative analyses and molecular weight determinations were made by the Huffman Microanalytical Laboratories, Denver, Colo.. 158 Tribute to Faculty We warmly pay tribute to all submitting instructors, past and present. Their dedicated and inspired teaching has contributed to the successful publication of their students’ work in ESSAI Volume I through Volume XII. Deborah Adelman Scott Albert Jim Allen Kathleen Altman E. Bates Patty Benson Brian Blevins Scott Boyd Franz Burnier Allan Carter Tom Carter Nancy Conradt Sue Dreghorn Barbara Eaton Linda Elaine Helen Feng Bob Georgalas Julie Gibbs Gloria Golec Ida Hagman Naheed Hasan Lisa Higgins Connie C. Howard Bill Hussong Richard Jarman Ellen C. K. Johnson Daniel Kies Dennis Korneff Keith Krasemann Chikako D. Kumamoto Deb Lantermo David Leary Freyda Libman Lynn Mackenzie James M. Magrini Heather Maher Anne C. Malone Nicole Matos Jackie McGrath Christine Monnier Katie Nagel Nick Obradovich Karen Persky Chris Petersen Kamelia Popova Loretta Pyrdek Eva Maria Raepple Lynda Randa Kent Richter Carol Riphenburg Shamili Sandiford John Santiago Steve Schroeder Hema Shende Dale Simpson Jason Snart Alice Snelgrove Susan Stamler Lois Stanciak Jodi Stapleton Carol Strelczyk Kathleen Szeszol Beth Vlad Diane Wawrejko Ben Whisenhunt Liz Whiteacre James Wies 159 Cumulative Index of Student Authors Volume I – (2002-2003) Kristyn Bales Virginia Bateman Barbara Bergschicker Hillary J. Brown Tess Cole Dean Costello Joi Cuartero Ryan Fowler Susanne K. Frens Alexa Greer Thomas Hemmingsen Karen Hess Jennifer Hixon Lauren Hovis Natalie A. Howes Beth Bednarz Roger Darringrand Lynn Dehlinger Natallia Ervin Brittany Gross Brandi Halle Allison Hamad Chris Harvey Anna Hendrey Jennifer Hixon Marsella Jorgolli Michael Kundert Sarah Magin Daniel McCullum Mary Meir Andy Morris Sarah Huber Dara Huff Stephanie McMahan Rachel Meek Andrew Morris Stacy Nalley Mary E. Norton Lindsay Olsen Ginny Pace Megan Posch Jason Schramm Dan Swindle Abe Whiting Chin Yang Volume II – (2003-2004) Noorjabeen Naseer My Nguyen Kristopher O’Dell Aaron Olson Chris Orlock Richard Payton Nadia Qazi Cara Raimon Frank Redmond Jonathan Scherer Maya Strahl Jasmine Thorne Bonnie White Michelle Ziehn Volume III – (2004-2005) Tanya Basu Mary Bodine Wendy Carpenter Meredith Carey Thomas Ferguson Michael Graczyk Victoria Graham Sara Heppner Benjamin Hooper Brian Kinsman Valentine Kopjo Kathy Kozak Sean Labbé Rebecca Lunev Karen L. Arneson Lisa Bonomo Steve Braun Brian Casario Patrick Caulfield Joseph Cholewa Joe Cinquegrani Janet Crawford Andy Dameron Dodi Dolendi Laura Dubberke Heather A. Duncan Christine Dyslin Rebecca Eineke Shannon Grilli Kristina Grudem Emily N. Hansen Jason Hernandez Julie McKee Matt Naglewski Stacy Nalley Allan Norgaard Jason Nosek Richard Payton Megan Posch Nadia Qazi David Rojas Maggie Smith Ruth Valentine Anthony Wilder Roberta Williams Sharon Wussow Volume IV – (2005-2006) William Hodges Benjamin Hooper Janelle A. Jenkins Hannah Kim Ruslan Kochemirovskiy Emily Krauspe Maren Ann McKee Tabitha Metreger Jason Morrow Keiko Narisako Jeremy Shier Jason Smith Joan Tett Jennifer Vail Alicia Voss Weien Wang Justin Wedekind Volume V – (2006-2007) Karen L. Arneson David Beck Natalie Berent Leslie A. Bertram Angela Cibich Peter Cohen Joseph A. Deardurff Lauren Del Carlo Camille Diebold Kennerly Diebold Melissa Donahue Christine Dyslin Travis Farrington Jessica Gibbs Victoria Graham Lucy A. Hawke Matt Horan Natalie Hustead Shaina Hutson Jenna L. Keefe Anna Kinsella Alison Lederer Keelin Lord Tracy Noeske Zachary Peschke 160 Carey Raupp Daniel Smith Nicholas Treff Ashley Urban Alice Wieduwilt Rutger Wolfgang Stache Fariah Zainuddin Ava Zandieh Volume VI – (2007-2008) Kristin Arita Krystal N. Arrate Cristina Baica William Berkhout Julie Bertacchi Kavita Bhavsar Michael Butler Leah Cameron Victoria Canaday Howard Joanna Ciesielski Katherine Ciesla Michael Davis Christine Dyslin Jessica Eller Carrie Fenwick Joel Griswell Shan Haq Jennifer Horst Mo Ismail Lyndon Kacick Nick Kogos Matthew Kos Jenny Letourneau Heather Lydon Melissa Maida Ali Malik Margaret McNulty Nathan Meno Katherine Morales Sylvia Morales Lydia Newby Agnes Mary Reband Jennifer Rodriguez Matt Sauter Robert Schletz Christina Scott Tom Serb Samantha Sferas Judy Staiano Kara Stevenson Adrienne Tam Christina Villa Cynthia L. Witt Volume VII – (2008-2009) Jacob Albaugh Ashley Aller Kyle Berthel Courtney Bobko Ellen Briggs Dave Broche Mary Beth Broda Patrick Buranicz Brittany Burkes Katherine Ciesla Ryan Corrigan Jonathon Drews Bethany Ester Aly Gentile Tiffany Gray Scott Hixson Edward Hren Grant L. Iler Graham Johnson Casey Jones Susan Kareska Sunya Kashan Nick Kontos Laurel Kunath Chris Lemke Emily Lombardi Melissa Mastrogiovanni Sean McGowan Mateja Mircic Phuong Nguyen Connie Padera Harry Podschwit Zach Porlier Nick Rascia Olivia Schultz Brandon Slate Richards Walker Michael Wendt Anum Zafar Volume VIII – (2009-2010) Kelsey Barron Geneva Beles Christine Berta Lauren Blanford Chelsey Boutan Sarah Burgess Stephanie Christian Eileen Corley Sally Jo Detloff Justin Eberhart Nathan Samuel Ellingsen Danny Hankes Max Hermann Christopher Hunton Nicole Anderson Sofia Bermudez Matt Brousil Najah Burks-Pittman Alicia M. Chapman Robert de Lara Dodi Dolendi Tyler Eickelmann Jennifer Elsman Michael Halla Chad Hochstatter Kevin Kienitz Michael Kintz Theresa Kolodziej Charise Joy Javonillo Tim Kamin Breanne Kanak Julie Kelsey Sarah Lichtenwalter Daniel Lowder Mandie McGlynn Matthew Murray Bekah Myroup Lily Nicholson Sean O'Leary Tina Osornio Sarah Oxley M. Lisa Parenti Zach Porlier James T. Ricker Shannon Rohn Susan Schumacher Alana Shuma Emily Stephen Lauren Stull Jennifer Wendt Amanda Woolum Michael K. O'Malley Jessica Otto-Rosario Kimberly Overby Michele Roberts Annette Sanchez Brittany Schweik Alana Shuma David Skorusa Kasie-Marie Smith Jeff Tosch Allison Watts Beth Wegner Frederick Wennlund Annie Werly Erik Wright Kari Zapletal Patrick Noonan Kevin Prchal Vicki Price Laura Romaine Sana Shafiuddin Sophie Shukin John Skarha Jessica Steslow Marcus Tolson Ethan Woodward Sarah Yockey Volume IX – (2010-2011) Sasha Nicole Kruger Jessica LaPlante Kristin Liewald Daniel Lowder Matt MacDougall Erin Massoni Zuzanna Mierzejewska Bekah Myroup Volume X – (2011-2012) Anna Albrecht Lynn Altman Karen Arneson Dan Barclay Matthew Brousil Chris Burkett Jordan Colbert Christine Dahlson-Rutherford Samantha De Carlo Alec Dorner Christina Duski John Eed Charles Hall Patrick Hanes Ian Haseltine Ernie Koh Nina Kokayeff Jeff Kort 161 Volume XI – (2012-2013) Jim Alop Virginia Anderson Angelo Andriopoulos Heidi Braunschneider Brody Challinor Lauren Challinor Steven Franklin Tony Frontuto Erna Gevondyan Tooba Ghafoor Ruth Groza Ian Haseltine Julia Andersen Nicholas Bashqawi Harry Bodell Kellie Brennan Kevin Brewton Elizabeth Cole Neil Davis Kelly Donlan Randy Dziak Katrina Favis Emily Garvanovic Frank Giuliani Martin Hayford Grace Horn Caleb Jenkins Roberta Johnson Akbar Khan Larry Lambert Sara Liesman James Maguire Danielle Meyer Kevin Midlash Erin Miller William Jasen Mummert Carlos S. OrtegaGuerrero Agnieszka Poznanska Hayley Reagin Sarah Rochford Rebecca M. Samson Vesal Stoakley Thomas Struett Matt Mazur Virginia Meglio Bryan Mundt David O’Sullivan Ashley Padavonia Becky Reichel Michael Robles Dina Rudenya Jake Sumeraj Emily Trca Alexandra Wickersham Ryan Woitas Volume XII – (2013-2014) Hayley Hickson Hope Javier Jonathan Kaye Syeda Khalid Marie Labbe Amanda Manton 162 ESSAI Created in 2002 by a cross-disciplinary faculty, ESSAI, The College of DuPage Anthology of Academic Writing Across the Curriculum is an annual academic journal that represents a unique but inclusive discourse community of student writers; it fosters student success through writing across the curriculum and in all levels of learning at College of DuPage. The resulting published work is their expanded cultural and epistemological self-endowed with talent, sophistication, and fresh scholarship, all expressed with clarity, eloquence, and grace. Join with us in honoring the exceptional talents of our student authors. Volume XII It rests on our belief that writing empowers and prepares our students for meaningful careers and advanced academic pursuits and that, to realize such goals, the entire college community is responsible for incorporating writing into their courses as a vital part of the academic experience at our college. That we practice the philosophy of student success through good writing will become even more apparent in the journal’s name. It harkens back and pays homage to Michel de Montaigne, whose seminal style of writing “essais” connoted one of trials and attempts. The College of DuPage Anthology of Academic Writing Across the Curriculum What is Essai? ESSAI 2013-2014 The College of DuPage Anthology of Academic Writing Across the Curriculum Volume XII 2013-2014