The Effects of User Feedback/Ranking to Enhance VMT Brian Smith, Drexel University, 3141 Chestnut Street, Philadelphia PA 19104, bps23@drexel.edu Abstract: Throughout the published research of the Virtual Math Team (VMT) environment, there is significant attention given to classifying users within the system. The classification can be as explicit as the Moderator/Participant relationship but more commonly, the research analysis also explores the how particular users interact; specifically their familiarity and involvement with the topic at-hand. Within this context, the researchers have been able to draw categorical conclusions about the users. These conclusions however can provide valuable information back to the users in the system to enhance the VMT environment and other users’ experience. By examining the results of these systems, and comparing to the interactions of the VMT environment, it is possible to conclude how implementing a similar type of rating or ranking system associated to each user account would affect the usefulness of the tool. Introduction The Virtual Math Team program is one that fosters and relies on peer interaction between groups of users to solve a mathematical problem. As a result, the explicit human interaction is a major contributing factor in the success of the system. Part of how that interaction occurs is defined by how any one individual user participates relative to the VMT tool and the rest of the users in the tool. At the same time, there is growing precedent for online tools’ use of user feedback systems to allow users to assume some level of expectation. These social indicators have become seen as essential components various tools and can be implemented into the VMT tool, provided that the feedback system introduces similar types of advantages and does not detract from the benefits already provided by the VMT tool. Online User Feedback and Rating in e-commerce Two popular online environments that take advantage of this type of user rating feedback system are eBay.com (eBay) and Amazon.com (Amazon). Both are e-commerce websites that rely on user feedback systems to foster interaction between a buyer and a seller; but at a more simple level, both are still electronic systems with internal users interacting. Within this respect, it shares a number of environmental similarities with the VMT tool. Analysis of the pros and cons of the feedback systems of these e-commerce environments can be extrapolated into the VMT environment, with consideration given to how the VMT tool does differ. The eBay user feedback system is one that transpires after an auction listed by the user known as the seller is won by the user known as the buyer, money is exchanged, and the item in question is received. Both users are then afforded the ability to rate the transaction based on a five star rating system for each of four categories: How accurate was the item description? How satisfied were you with the seller's communication? How quickly did the seller ship the item? How reasonable were the shipping and handling charges? (eBay). The composite of these scores is compiled into a single five-star rating that is attributed to the user’s account. A “composite of composites” over the course of all transactions in which the user was a participant is then displayed as the visible ‘user rating’ within the system. In addition, the number of ratings (which corresponds to the number of transactions) is displayed. This adds to the user confidence: as the number of ratings increases, the more credence is assumed with the starred ratings system. Figure 1 - View of an eBay User Feedback Rating Figure 1 shows the user feedback that is seen wherever user information is displayed in the system (an example of this is expanding in Figure 2 as well). The convention in displaying the star rating along with the number of reviews (as well as the percentage information) provides consistency that all users have come to understand; it now “means something” in defining expectations for user interactivity. Figure 2 - eBay Auction Screen One of the most notable considerations of the feedback system is how quickly it establishes a relationship amongst the users: “eBay’s feedback system is, arguably, their most valuable asset. It provides the grease necessary to make complete strangers comfortable enough to buy and sell from each other.” (TechCrunch) This is a key factor of eBay and Amazon, but a similar sentiment concerning the users of VMT is given early in the VMT text: “Since the students did not know each other from before the chat and could not observe each other except through the behavior that took place in the chat room, they could only understand each other’s messages and actions based on what took place inside the chat room.” (Stahl, 2006, p.50) Therefore, similar to the e-commerce tools, the users are relying solely on stimuli from the system as the basis for their opinions: “But it’s a closed system – only eBay transactions can affect a user’s feedback score.” (TechCrunch). The Amazon.com feedback works in much the same way Figure 3 - View of an Amazon User Feedback Rating Figure 4 - Amazon Item Purchase Screen Online User Content Feedback and Rating There are also non-e-commerce sites that employ a user feedback ratings, however news reporting sites such as Yahoo! News direct the user feedback to be indicative of the actual content of the post. Yahoo! News’ feedback allows users to leave feedback on the news story being reported. The comments left by the users are subject to a “thumbs up/thumbs down” rating that gives the users a chance to provide feedback based on specific items; however since a comment must be left by a user, the feedback can be correlated to a specific user. Over time, an average of comment ratings can be turned into a “user” score. Figure 5 - Yahoo! News User Feedback screen Figure 6 - Yahoo! News content rating option Similarly, the Drexel Blackboard site has instituted a post-by-post feedback system that allows users to rate the quality of the posts made, and again, each post has a specific user correlated to it. This would allow for an aggregate system to be employed to show a user rating. Figure 7 - Blackboard's post rating Blackboard’s system has a star ranking system, tucked away in the real estate in the standard interface. It is present in every post, and with the system requiring a login to make any post, there is traceability back to the user. The interface is conventional, and has the same affordances as the Yahoo! News system. Figure 8 - Blackboard post's "star" rating Issues With Existing Systems The feedback systems are not perfect representations of the users’ activity. The fact that it is based off of human intervention explicitly introduces human error into the ranking system. This has also been noted by the ecommerce tools, both in terms of ongoing revisions the tools make to the feedback process as well as fundamental flaws that user feedback systems inherently have – that may not be resolvable. The eBay site is making some major changes to eBay feedback and one of the biggest is always that sellers will no longer be able to leave comments for buyers. The contention here is that sellers were using bad information as being a way to prompt slow payers to pay up and also utilizing the reviews as being a forum to air personal grievances. (eZine) If someone is selling something I really want then their eBay suggestions may not be as large of a concern unless it was one hundred percent unfavorable. (eZine) People are afraid of what is referred to in the eBay community as "retaliatory feedback." That's when a seller (or a buyer--it can happen on both sides) becomes annoyed or enraged at a negative comment posted by someone else about them, and in retaliation posts a negative comment in return. The system obviously has no way of determining for sure the motivations of the various people posting. It therefore happens--and happens quite a lot, according to some estimates--that an innocent person's reputation suffers. (Forbes) Weber said most buyers don't understand the purpose of Amazon's feedback mechanism and don't understand the consequences of their venting that can damage sellers' reputation and ability to earn money. (Steiner) Tying Into VMT Thus far, the discussion has focused on a feedback system with respect to e-commerce sites. The VMT tool is not offering products or purchases that are subject to buyer/seller confidence. So, the consideration must then be given to how users of an educational knowledge sharing system differ from e-commerce users; specifically, how would user feedback differ. As a corollary to how the user feedback differs; analysis of how feedback would affect how users natively interact within the VMT tool must be considered before implementing such a feature. In the Stahl text, the authors dedicate extensive time to analyzing entire specific interactions within the tool. Involved in this is the classification of users into relevant categories. However, in the text the analysis is done for the benefit of the Stahl text researchers and their understanding of user interaction and refinement of the VMT tool. It is the contention of this paper’s research that these same categorizations can A) benefit the users if rolled back into the system, and B) be determined by the user base itself. To represent how the user base could conceivably reach the conclusions as determined by the Stahl text researchers, three user determinations from the “Agentic Movement” analysis section are in focus: “When we began this analysis we tended to see Bwang as the “math student” because Bwang was very good at taking a given problem and expressing it in an equation. He had a certain math orientation and was often the first to create mathematical objects that the group later worked with”. (Stahl, 2006, p.218) As the system is primarily focused on solving math problems, someone with Bwang’s mathematical background would presumably be beneficial to any group within the VMT system. His frequency in interaction and dedication to solving the problem and adding to the group knowledge indicates that he could also be an asset to a team independent of his mathematical skills, though his responses are fewer while more deliberate. “Aznx was very skilled at being creative in thinking about new problems and facilitating interaction, caring about the group as we saw above”. (Stahl, 2006, p.218) The user, Aznx, is noted as having excellent interpersonal skills within the group, and bringing an additional perspective to the conversation, without explicitly having a particularly strong mathematical background. That said, the content of the chats contained in the Stahl text (p. 216-222) show that this user does have a solid understand of the math concepts. “Quicksilver was harder to get a sense of”. (Stahl, 2006, p.218) The third user, Quicksilver, doesn’t appear to have an y particular categorization associated to him. However, this is actually a categorization unto itself. It would be the same classification that could be granted to any new user to the system, or any user that only takes a minimal involvement in the tool. Within the confines of a problem solving session of three users, it is evident that a simple scoring system could convey similar thoughts in defining each user. A numeric score for math knowledge and a numeric score for user interaction would quantify each of the above. Using a 1-5 scoring system (1 indicating a low ranking and 5 indicating a high ranking) for both math and interaction respectively, Bwang could rate as a 5-4 while Aznx could rank as a 3-5. With Quicksilver, the Stahl text researchers indicate that they are unable to draw any conclusions; a lack of a score within a feedback system would indicate the same To round out examples from the Stahl study to further show the quantification of the user categorization a quick example may be referenced: … it appeared that one student (Mic) who seemed particularly weak in math was clowning around a lot and that another (Cosi) managed to solve the problem herself despite this distraction in the chat room. (Stahl, 2006, p.50) The user Mic could conceivably be conveyed as a 1-1 for not having particularly strong math skills and for distracting from the group learning effort. Similarly a user like Cosi was able to handle the mathematical component, but did so independent of the group. With no indication from the Stahl text researchers that Cosi made any exceptional attempt to communicate with the group, even despite the cost of Mic’s “clowning around”, Cosi may be conveyed as a 4-2. Scoring Systems for VMT While the details of the scoring system can be refined by the total number of available ratings or the visible format (eg. stars, numbers, slide chart), with relatively minimal effort, the Stahl text researchers’ user analysis could be turned into quantifiable data that now affords those viewing to gain an expectation of each users’ general capabilities without having to read through extensive textual analysis and without having to analyze the individual case studies from where the determinations were made. That said, the expert researcher analysis is thus far the only thing responsible for that conclusion. It is the conjecture of this paper that, over time, the user-inputted feedback will resolve to similar conclusions. The input from any one single user may skew the data due to a lack of understanding of the feedback process, personal experience, ulterior motives, or any other non-expert review; however the average of all users input should reveal what the expert analysis is able to conclude. The e-commerce tools as described in this paper share comparable online environmental user stimuli and thus have shown that there is a precedent for a user ranking system. The Stahl paper researchers have successfully quantified the user interactions. As such, the remaining consideration concerns how introducing a feature like the user feedback function, and specifically how users would react to other users of varying ratings, would affect the user experience with VMT. Figure 5 shows a simple example of an existing Stahl text screenshot, that has includes a ranking system indicator for the users involved. The scoring system does not affect on, or intrude upon any of the existing VMT tool. Figure 6 shows the rating in more detail, consistent with the potential average scorings as described previously in this document. Figure 9 - VMT Mock-up of User ratings Figure 10 - Close-up of VMT Mock-up Upon completion of the VMT session, as defined from first login to conclusion of the work (when the problem is solved and/or when the Group session is closed), each user would have the ability to offer this feedback to their peer users in much the same way as the post-transaction feedback is available in the e-commerce tools. Citing certain tenants of the Stahl book under the Preconditions for Cognitive Processes by Groups (p.525), there are /particular points to which implementing the user feedback tool and user status would apply. “Shared intentionality” and “Shared Background Culture” address the commonality between the users of the system. The user ranking speaks to both of those consistencies as the user ranking assumes a scoring system based on accepted standards of the system (all users having the same interfaces and functionality availability to them) and general societal understandings. “Intersubjectivity”, or human co-presence, as a precondition is explicitly met; by ranking experiences with another user, attention is given on a peer-to-peer basis and thus every user has the ability to recognize any other user in the system that they have worked on a resolution with. The most important two preconditions in supporting the integration of a user feedback system though, are “Member methods for social order” and “Motivation and Engagement”. The ranking system of users, by users, speaks directly to establishing a hierarchy within the system. Whereas this would usually get flushed out over the course of solving a problem within a team, using a universal feedback system across the entire tool, can accelerate this natural evolution. Finally, the motivation and engagement is also explicit within a feedback tool in the exact same way that it is for the e-commerce tools: with a public display of how valuable a user is considered to be, the impetus to improves ones’ ranking to achieve a higher status can augment the motivational points that Stahl, et al, elaborate upon. Resolving Issues with User Feedback Within a user feedback system, inherent issues arise. As stated previously, a ‘lack of understanding of the feedback process, personal experience, ulterior motives, or any other non-expert review’ may contribute to the feedback system not accurately portraying the contribution of the users involved. As such, this section will identify potential alternatives to a pure user-on-user feedback system. Analysis of the pros and cons of replacing or supplementing the feedback with the following would yield appropriate results for inclusion consideration. First, and as a notable difference between the e-commerce systems versus a math learning tool, is the expected average user age, and the emotional maturity that follows. Whereas adult users dealing with goods have an inherent understanding of the relationship at large, younger students using a math tool may consistently skew the feedback, in such situations by constantly ranking friends high regardless of contribution to the VMT. As such, a potential resolution is to restrict the feedback only to the moderator(s) in each session. While this would detract from allowing user to provide their own impact, it ensures the integrity of the ratings. This, in turn, yields functionality that is more beneficial to the user community at large. Second is the motivation factor. Without any specific provocation to input feedback, the success of this functionality depends solely on the willingness of the user community to contribute back into the tool. Again, unlike the e-commerce sites in which feedback motivation stems from the fact that user attraction is based off of user ratings, incorporating it into the VMT environment carries no requisite activity. The only users that appear to have vested interest in the tool are the moderators, who return to different sessions and have expectations of the activities contained within. This noted, the moderators again prove to be the best users for feedback integrity. Both of the above limitations place the impetus on the moderators in place of the entire user community, which is an available option, but belies the focus of this proposal. A possible solution to this is to observe a similar system to a number of online news reporting sites, or Drexel’s Blackboard system, employ: a post-by-post rating system that allows all users to provide input to the system, but only on particular posts. This resolves a couple of items problematic with the above: users will be less likely to tag each and every posting which should limit the gratuitous ratings that younger users may afford each other, and with the feedback being immediate (at the time of the post) instead of cumulative at the end of the session, users would have more impetus to contribute to the ranking system. An aggregate rating of a user’s posts would manifest as the user’s overall score. A moderator’s input would still be valuable, and a moderator could still contribute to the rating, but the weight shifts back to the general users. Future Resolution The remaining issue that would still need research attention (in the form of case studies akin to those presented in the VMT text) is the determination of whether or not the user feedback display would change the way in which users interacted with the system. By conducting a broad scope of analyses of sessions, some with the user feedback notification and some without, the Stahl text researchers could enumerate how other users reacted to those same case study users with the varying ratings. In the example in this paper, the user “Bwang” is seen as having an expertise in math and being reasonable helpful to other users without any rating attached to him. Over the course of a number of subsequent problem sessions, if Bwang is tagged with any ranking, rating, or notification of his/her skills or general nature, would users be more or less inclined to interact with Bwang? Would the qualitative data shift to any show any significant alteration in how users use VMT? Conclusion Throughout this paper, it has been proposed that a user feedback and ranking system would enhance the VMT system. The enhancement would come in the form of strengthening the users interactivity as an underlying support to the tool. By allowing for users to expedite general project team hierarchies and expectations, the VMT tool would benefit in facilitating learning. Electronic systems that use a user feedback and ranking system have grown to rely on them in an e-commerce environment that survives on how cohesive users peer relationships are. Systems that do not rely on user feedback specifically still have content feedback systems that reflect on the contributing user. With additional research to verify that the integrity of the VMT system is maintained, integrating a user feedback and ranking system into the VMT tool can provide incremental benefits to the environment and group collaborative learning at large. References Arrington, M. (2006). Rapleaf to Challenge Ebay Feedback. TechCrunch.com. Retrieved on November 1, 2010 from http://techcrunch.com/2006/04/23/rapleaf-to-challenge-ebay-feedback/ Blackboard Academic Suite. (2010). Retrieved on November 27, 2010 from https://drexel.blackboard.com eBay.com. (2010). About Detailed Seller Ratings (DSR). Retrieved on November 4, 2010 from http://pages.ebay.com/help/feedback/detailed-seller-ratings.html LaPlante, A. (2007). EBay Feedback: Fatally Flawed? Forbes.com. Retrieved on November 3, 2010 from http://www.forbes.com/2007/01/02/ebay-smallbusiness-feedback-ent-salescx_al_0102smallbizresource.html Stahl, G. (2009). Studying Virtual Math Teams. New York: Springer. Steiner, I. (2006). Online Booksellers React to Amazon.com Feedback Changes. AuctionBytes.com. Retrieved on November 6 from http://www.auctionbytes.com/cab/abn/y06/m07/i10/s04 Tanady, M. (2010). The Aggressive World of eBay Feedback. EzineArticles.com. Retrieved on November 3, 2010 from http://ezinearticles.com/?The-Aggressive-World-Of-eBay-Feedback&id=4809988 Yahoo! News. (2010). The Top News Headlines on Current Events from Yahoo! News. Retrieved on November 26, 2010 from http://news.yahoo.com Acknowledgments I certify that: To the best of my knowledge, this assignment is entirely work produced by me. Any identification of my individual work is accurate. I have not quoted the words of any other person from a printed source or a website without indicating what has been quoted and providing an appropriate citation. I have not submitted any of the material in this document to satisfy the requirements of any other course.