Questions?

advertisement
Why Pay?: Exploring How Financial Incentives
are Used for Question & Answer
Gary Hsieh, Robert E. Kraut, Scott E.
Hudson, CHI 2010
Research Methods
• Data collection
– Mahalo Answers
• Rating questions in Mahalo
– Randomly selected 800 questions
• 400 were user-paid, 400 were not.
– Rating (1) question type (fact, opinion, advice,
non-question; non-mutual exclusive), and (2)
question value (sincerity, urgency, difficulty)
– Amazon M-Turk: 401 raters; each rated 19
questions on average
Results
• Question asking
– Factual vs. is paid?
– Difficulty level vs. reward levels
• Question answering
– Rewards/Opinion/Sincerity vs. answer length
– Rewards/Opinion/Difficulty vs. answer count
– Opinion/Difficulty vs. answer quality
• Archival use
– Higher value; higher archival value
Review
• Askers value longer answers, quick responses, and reliable
sources
• Answerers learn over time (as such, receiving higher
ratings, slightly earning more)
• Answerers tend to specialize into several topics (which may
lower actual hourly wage)
• Financial incentives improves quantity but not quality of
answers (and reputation of repliers is related to quality)
• Tangible (monetary) and intangible incentives have positive
impact on the number of answers per answerer
• Question type matters (factual vs. conversational)
• How much a question is paid can be used as an indicator of
how valuable certain Q&A exchanges are
Q&A over Social Networks
KSE 652
Uichin Lee
Aardvark: The Anatomy of a LargeScale Social Search Engine
Damon Horowitz, Aardvark
Sepandar D. Kamvar, Stanford University
WWW '10
Original slides by Hailong Sun, April 13, 2010
Web Search vs. Social Q&A
• Google is about traditional Web search
– Give me keywords, I will provide contents
– Search for the most suitable contents
• While Aardvark is about social search
– Users can ask questions in natural language, not keywords
– Content is generated “on-demand”, tapping the huge
amount of information in peoples’ heads (i.e., everyone
knows something)
– The system is fueled by the goodwill of its users
Problem in Aardvark
Search
Engine
• How to find the user who can best answer a
given question?
Aardvark Architecture
2
1
Question?
3
Edit question?
Determine the appropriate
topics for the question
3
(Routing Suggestion Request)
-- find a list of candidate
answerers (and rank them)
4
Ask one by one
Social graph indexing
User’s topic parsing
VarkRank: Relevance and Connectedness
Question
User
User
User
Topic
User
User
User
P(ui|t)
Used for measuring relevance score of
a user’s question (query dependent)
User
User
User
P(ui|uj)
Used for measuring connectedness
score between users (query
independent)
VarkRank: Relevance and Connectedness
• Given a question q, the probability that a user ui can
answer it (relevance score)
• Score that user ui can answer a question from uj: (query
dependent user’s query relevance score * query
independent user connectedness score)
– Query independent user quality score = p(ui|uj): i.e., user i
delivers a satisfying answer to user j (simply measured using
connectedness)
Relevance Scores (Expertise)
• How to find experts on a given topic?
• Expertise score: p(ui|t), and w/ Bayes’ law, we have:
• For each user, we need to profile a user’s interest in a
given topic by using the following information sources:
–
–
–
–
–
3+ topics provided by a user
Topics provided by friends of a user
Online profiles
Online unstructured data
Status message updates
13
Connectedness Scores
• Connection strengths between people i.e., p(ui|uj) are computed
using a weighted cosine similarity over this feature set (normalized)
• Utilize existing social networks
– Facebook, Twitter, LinkedIn…
• Feature set measures similarities in demographics/behavior
–
–
–
–
–
–
–
–
Social connection (common friends and affiliations)
Demographic similarity
Profile similarity (e.g., common favorite movies)
Vocabulary match (e.g., IM shortcuts)
Chattiness match (frequency of follow-up messages)
Verbosity match (the average length of messages)
Politeness match (e.g., use of “Thanks!”)
Speed match (responsiveness to other users)
Analyzing Questions: TopicMapper
• Question classification:
–
–
–
–
Question or not?
Inappropriate question?
Trivial question?
Location sensitive question?
• Map a question to a topic (weighted linear sum of the
following features)
– Keyword matches w/ a user’s profile topics?
– Classifies the question text into a taxonomy of roughly 3000
popular question topics (using an SVM trained on an annotated
corpus of several million questions)
– Extracting salient phrases from questions and find semantically
similar user topics
Ranking Algorithm
• Topic and connectedness matching  availability
• Routing Engine prioritizes candidate answerers
– Optimize the chances that the present question will be answered
– Yet, preserving the available set of answerers (i.e., the quantity of
“answering resource” in the system) as much as possible by spreading
out the answering load across the user base
• Considering factors: currently online users (e.g., via IM presence
data, iPhone usage, etc.), user’s daily activity history, and user’s
response history (lowering scores of non-responsive users)
• Conversation Manager serially inquiring whether candidates would
like to answer the present question; and iterating until an answer is
provided and returned to the asker.
User Interface
IM
Email
Twitter
Aardvark
Deployment Status
• Aardvark is actively used
– Users: from 2,272 to 90,361
– 55.9% active users, 73.8% passive
users
– 3,167.2 questions/day
– 3.1 queries/month
• Mobile users are particularly active
–
–
–
–
Average 3.6322 sessions/month
Comparison: Google
Desktop v.s. mobile users: 3
Mobile users: 5.68 sessions/month
Categories of Questions in Aardvark
• Questions are highly contextualized
– Average query length: 18.6 words
• 2.2~2.9 for Web search
– 45.3% are about context
• Questions often have a subjective element
– What are the things/crafts/toys your children have made that made them
really proud of themselves?
Answers
• Answers
– 57.2% received answers in less than 10 minutes
– A question receives 2 answers averagely
– The quality of answers are good
• 70.4% are “good”; 14.1% are “OK”; 15.5% are “bad”
Distribution of questions and answering times
Distribution of questions and number of answers received
Topic Distribution
• People are indexable
– 97.7% have 3+ topics
21
Comparative Evaluation with Google
• Experimental setup
– Randomly select a group of questions
– Insert a tip “do you want to help Aardvark run an
experiment?”
– Recording response time and quality of answers from
Google and Aardvark
• Experimental results
– Aardvark: 71.5% answered; rating: 3.93 (σ=1.23)
– Google: 70.5% answered; rating: 3.07 (σ=1.46)
• Aardvark is more suitable for subjective questions
What Do People Ask Social Networks?
Meredith Ringel Morris, MSR
Jaime Teevan, MSR
Katrina Panovich, MIT
http://www.yelp.com
good restaurants in atlanta
Questions About People’s Questions
• What questions do people ask?
– How are the questions phrased?
– What are the question types and topics?
– Who asks which questions and why?
• Which questions get answered?
– How is answer speed and utility perceived?
– What are people’s motivations for answering?
What Is Known About Question Asking
•
•
•
•
Collaborative search [Morris & Teevan]
Searching v. asking [Evans et al.; Morris et al.]
Expertise-finding [vark.com; White et al.; Bernstein et al.]
Online question answering (Q&A) tools
– Question type [Harper et al.: conversational v. informational]
– Response rate [Hseih & Counts: 80%]
– Response time [Zhang et al.: 9 hours; Hseih & Counts: 3 hours]
– Motivation [Raban & Harper; Ackerman & Palen; Beenan et al.]
Survey of Asking via Status Messages
• Survey content
– Used a status message to ask a question?
• Frequency of asking, question type, responses received
• Provide an example
– Answered a status message question?
• Why or why not?
• Provide an example
• 624 participants
– Focus on Facebook and Twitter behavior
Questions About People’s Questions
• What questions do people ask?
– How are the questions phrased?
– What are the question types and topics?
– Who asks which questions and why?
• Which questions get answered?
– How is answer speed and utility perceived?
– What are people’s motivations for answering?
Questions: Phrasing
• Questions short (75 characters, 1 sentence)
• 18.5% of phrased as a statement
I need a recommendation on a good all purpose pair of sandals.
• Often scoped
– 1 out of 5 directed to “anyone”
Anyone know of a good Windows 6 mobile phone that won’t
break the bank?
– Network subset
Hey Seattle tweeps: Feel like karaoke on the Eastside tonight?
Questions: Types
Type
%
Example
Recommendation
29%
Building a new playlist – any ideas for good running
songs?
Opinion
22%
I am wondering if I should buy the Kitchen-Aid ice
cream maker?
Factual
17%
Anyone know a way to put Excel charts into LaTeX?
Rhetorical
14%
Why are men so stupid?
Invitation
9%
Who wants to go to Navya Lounge this evening?
Favor
4%
Need a babysitter in a big way tonight… anyone??
Social connection
3%
I am hiring in my team. Do you know anyone who
would be interested?
Offer
1%
Could any of my friends use boys size 4 jeans?
Questions: Topics
Topic
%
Example
Technology
29%
Anyone know if WOW works on Windows 7?
Entertainment
17%
Was seeing Up in the theater worth the money?
Home & Family
12%
So what’s the going rate for the tooth fairy?
Professional
11%
Which university is better for Masters? Cornell or
Georgia Tech?
Places
8%
Planning a trip to Whistler in the off-season.
Recommendation on sites to see?
Restaurants
6%
Hanging in Ballard tonight. Dinner recs?
Current events
5%
What is your opinion on the recent proposition that
was passed in California?
Shopping
5%
What’s a good Mother’s Day gift?
Philosophy
2%
What would you do if you had a week to live?
Questions: Who Asks What
Type
Recommendation
Topic
old
men
Opinion
Twitter
Entertainment
women
Factual
Rhetorical
Invitation
Technology
Home & Family
Professional
young
Places
Favor
Restaurants
Social connection
Current events
Offer
Shopping
Philosophy
Facebook
Questions: Motives for Asking
Topic
%
Example
Trust
24.8%
I trust my friends more than I trust strangers.
Subjective
21.5%
Search engine can provide data but not an opinion.
Thinks search
would fail
15.2%
I’m pretty search engine couldn’t answer a question of
that nature.
Audience
14.9%
Friends with kids, first hand real experience.
Connect
12.4%
I wanted my friends to know I was asking the question.
Speed
6.6%
Quick response time, no formalities.
Context
5.4%
Friends know my tastes.
Tried search
5.4%
I tried searching and didn’t get good results.
Easy
5.4%
Didn’t want to look through multiple search results.
Quality
4.1%
Human-vetted responses.
Questions About People’s Questions
• What questions do people ask?
– How are the questions phrased?
– What are the question types and topics?
– Who asks which questions and why?
• Which questions get answered?
– How is answer speed and utility perceived?
– What are people’s motivations for answering?
Answers: Speed and Utility
• 94% of questions received an answer
• Answer speed
– A quarter in 30 minutes, almost all in a day
– People expected faster, but satisfied with speed
– Shorter questions got more useful responses
• Answer utility
– 69% of responses helpful
Answers: Speed and Utility
Type
Topic
Recommendation
Opinion
Technology
Fast
Factual
Rhetorical
Entertainment
Home & Family
Unhelpful
Professional
Invitation
Places
Favor
Restaurants
Social connection
Current events
Offer
Shopping
Philosophy
Answers: Motives for Answering
Motive
%
Example
Altruism
37.0
Just trying to be helpful.
Expertise
31.9
If I’m an expert in the area.
Question
15.4
Interest in the topic.
Relationship
13.7
If I know and like the person.
Connect
13.5
Keeps my network alive.
Free time
12.3
Boredome/procrastination.
Social capital
10.5
I will get help when I need it myself.
Obligation
5.4
A tit-for-tat.
Humor
3.7
Thinking I might have a witty response.
Ego
3.4
Wish to seem knowledgeable.
Answers About People’s Questions
• The questions people ask
– Short, directed to “anyone”
– Subjective questions on acceptable topics
– Social relationships important motivators
• The questions that get answered
– Fast, helpful responses, related to length and type
– Answers motivated by altruism and expertise
QUESTIONS?
Meredith Ringel Morris
Jaime Teevan
Katrina Panovich
M. R. Morris, J. Teevan, and K. Panovich. What Do People Ask Their Social
Networks, and Why? A Survey Study of Status Message Q&A Behavior. CHI
2010.
M. R. Morris, J. Teevan, and K. Panovich. A Comparison of Information
Seeking Using Search Engines and Social Networks. ICWSM 2010 (to appear).
Download