42
©2009 Carnegie Mellon University : 1 Smartening the Crowds: Computational Techniques for Improving Human Verification to Fight Phishing Scams Gang Liu Wenyin Liu Department of Computer Science City University of Hong Kong Guang Xiang Bryan A. Pendleton Jason I. Hong Carnegie Mellon University

Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

Embed Size (px)

DESCRIPTION

We looked at how to augment crowdsourcing techniques to improve coverage, accuracy, and timeliness in identifying phishing attacks. We used relatively simple clustering algorithms to group phish together, as well as weighting votes based on previous correct answers.

Citation preview

Page 1: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

00

9 C

arn

eg

ie M

ello

n U

niv

ers

ity :

1

Smartening the Crowds: Computational Techniques for Improving Human Verification to Fight Phishing Scams

Gang LiuWenyin LiuDepartment of Computer ScienceCity University of Hong Kong

Guang XiangBryan A. PendletonJason I. HongCarnegie Mellon University

Page 2: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

2

Page 3: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

3

Detecting Phishing Websites

• Method 1: Use heuristics– Unusual patterns in URL, HTML, topology– Approach is favored by researchers– High true positives, some false positives

• Method 2: Manually verify– Approach used by industry blacklists today

(Microsoft, Google, PhishTank)– Very few false positives, low risk of liability– Slow, easy to overwhelm

Page 4: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

4

Page 5: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

5

Page 6: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

6

Page 7: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

7

Wisdom of Crowds Approach

• Mechanics of PhishTank– Submissions require at least 4 votes

and 70% agreement– Some votes weighted more

• Total stats (Oct2006 – Feb2011)– 1.1M URL submissions from volunteers– 4.3M votes– resulting in about 646k identified phish

• Why so many votes for only 646k phish?

Page 8: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

8

PhishTank StatisticsJan 2011

Submissions 16019

Total Votes 69648

Valid Phish 12789

Invalid Phish 549

Median Time 2hrs 23min

• 69648 votes max of 17412 labels– But only 12789 phish and 549 legitimate identified– 2681 URLs not identified at all

• Median delay of 2+ hours still has room for improvement

Page 9: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

9

Why Care?

• Can improve performance of human-verified blacklists– Dramatically reduce time to blacklist– Improve breadth of coverage– Offer same or better level of accuracy

• More broadly, new way of improving performance of crowd for a task

Page 10: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

10

Ways of Smartening the Crowd

• Change the order URLs are shown– Ex. most recent vs closest to completion

• Change how submissions are shown– Ex. show one at a time or in groups

• Adjust threshold for labels– PhishTank is 4 votes and 70%– Ex. vote weights, algorithm also votes

• Motivating people / allocating work– Filtering by brand, competitions,

teams of voters, leaderboards

Page 11: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

11

Ways of Smartening the Crowd

• Change the order URLs are shown– Ex. most recent vs closest to completion

• Change how submissions are shown– Ex. show one at a time or in groups

• Adjust threshold for labels– PhishTank is 4 votes and 70%– Ex. vote weights, algorithm also votes

• Motivating people / allocating work– Filtering by brand, competitions,

teams of voters, leaderboards

Page 12: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

12

Ways of Smartening the Crowd

• Change the order URLs are shown– Ex. most recent vs closest to completion

• Change how submissions are shown– Ex. show one at a time or in groups

• Adjust threshold for labels– PhishTank is 4 votes and 70%– Ex. vote weights, algorithm also votes

• Motivating people / allocating work– Filtering by brand, competitions,

teams of voters, leaderboards

Page 13: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

13

Ways of Smartening the Crowd

• Change the order URLs are shown– Ex. most recent vs closest to completion

• Change how submissions are shown– Ex. show one at a time or in groups

• Adjust threshold for labels– PhishTank is 4 votes and 70%– Ex. vote weights, algorithm also votes

• Motivating people / allocating work– Filtering by brand, competitions,

teams of voters, leaderboards

Page 14: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

14

Ways of Smartening the Crowd

• Change the order URLs are shown– Ex. most recent vs closest to completion

• Change how submissions are shown– Ex. show one at a time or in groups

• Adjust threshold for labels– PhishTank is 4 votes and 70%– Ex. vote weights, algorithm also votes

• Motivating people / allocating work– Filtering by brand, competitions,

teams of voters, leaderboards

Page 15: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

15

Overview of Our Work

• Crawled unverified submissions from PhishTank over 2 week period

• Replayed URLs on MTurk over 2 weeks– Required participants to play

2 rounds of Anti-Phishing Phil– Clustered phish by html similarity– Two cases: phish one at a time, or in a

cluster (not strictly separate conditions)– Evaluated effectiveness of vote weight

algorithm after the fact

Page 16: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

16

Anti-Phishing Phil

• We had MTurkers play two rounds of Phil [Sheng 2007] to qualify ( = 5.2min)

• Goal was to reduce lazy MTurkers and ensure base level of knowledge

Page 17: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

17

Clustering Phish

• Observations– Most phish are generated by toolkits and

thus are similar in content and appearance– Can potentially reduce labor by labeling

suspicious sites in bulk– Labeling single sites as phish can be hard

if unfamiliar, easier if multiple examples

Page 18: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

18

Clustering Phish

• Motivations– Most phish are generated by toolkits and

thus similar– Labeling single sites as phish can be hard,

easier if multiple examples– Reduce labor by labeling suspicious sites

in bulk

Page 19: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

19

Clustering Phish

• Motivations– Most phish are generated by toolkits and

thus similar– Labeling single sites as phish can be hard,

easier if multiple examples– Reduce labor by labeling suspicious sites

in bulk

Page 20: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

20

Most Phish Can be Clustered

• With all data over two weeks, 3180 of 3973 web pages can be grouped (80%)– Used shingling and DBSCAN (see paper)– 392 clusters, size from 2 to 153 URLs

Page 21: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

21

Page 22: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

22

MTurk Tasks

• Two kinds of tasks, control and cluster– Listed these two as separate HITs– MTurkers paid $0.01 per label– Cannot do between-conditions on MTurk– MTurker saw a given URL at most once

• Four votes minimum, 70% threshold– Stopped at 4 votes, cannot dynamically

request more votes on MTurk– 153 (3.9%) in control and 127 (3.2%) in

cluster not labeled

Page 23: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

23

MTurk Tasks

• URLs were replayed in order– Ex. If crawled at 2:51am from PhishTank

on day 1, then we would replay at 2:51am on day 1 of experiment

– Listed new HITs each day rather than a HIT lasting two weeks (to avoid delays and last minute rush)

Page 24: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

24

Summary of Experiment

• 3973 suspicious URLs– Ground truth from Google, MSIE, and

PhishTank, checked every 10 min– 3877 were phish, 96 not

• 239 MTurkers participated– 174 did HITs for both control and cluster– 26 in Control only, 39 in Cluster only

• Total of 33,781 votes placed– 16,308 in control– 11,463 in cluster (17473 equivalent)

• Cost (participants + Amazon): $476.67 USD

Page 25: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

25

Results of Aquarium

• All votes are the individual votes• Labeled URLs are after aggregation

Page 26: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

26

Comparing Coverage and Time

Page 27: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

27

Voteweight

• Use time and accuracy to weight votes– Those who vote early and accurately

are weighted more– Older votes discounted– Incorporates a penalty for wrong votes

• Done after data was collected– Harder to do in real-time since we don’t

know true label until later

• See paper for parameter tuning– Of threshold and penalty function

Page 28: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

28

Voteweight Results

• Control condition best scenario– Before-after– 94.8% accuracy, avg 11.8 hrs, median 3.8– 95.6% accuracy, avg 11.0 hrs, median 2.3

• Cluster condition best scenario– Before-after– 95.4% accuracy, avg 1.8 hrs, median 0.7– 97.2% accuracy, avg 0.8 hrs, median 0.5

• Overall: small gains, potentially more fragile and more complex though

Page 29: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

29

Limitations of Our Study

• Two limitations of MTurk– No separation between control and cluster– ~3% tie votes unresolved (more votes)

• Possible learning effects?– Hard to tease out with our data– Aquarium doesn’t offer feedback– Everyone played Phil– No condition prioritized over other

• Optimistic case, no active subversion

Page 30: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

30

Conclusion

• Investigated two techniques for smartening the crowd for anti-phishing– Clustering and voteweight

• Clustering offers significant advantages wrt time and coverage

• Voteweight offers smaller improvements in effectiveness

Page 31: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

31

Acknowledgments

• Research supported by CyLab under ARO DAAD19-02-1-0389 and W911NF-09-1-0273

• Research Grants Council of the Hong Kong Special Administrative Region, China [Project No. CityU 117907]

Page 32: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

32

Page 33: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

33

Individual Accuracy

Page 34: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

34

4. Framework

Page 35: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

35

4.1 whitelists include 3208 domains

From Google safe browsing (2784)

http://sb.google.com/safebrowsing/update?version

=goog-white-domain:1:1 From millersmiles (424)

http://www.millersmiles.co.uk/scams.php

Reduce false positive Save human effort

Page 36: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

36

4.2 Clustering Content similarity measurement (Shingling method)

S(q),S(d) denote the set of unique n-grams in page q and d The threshold is 0.65 The average time cost : 0.063 microseconds (SD=0.05)

calculating similarity of two web pages on a laptop with 2GHz dual core CPU with 1 GB of RAM

DBSCAN Eps=0.65 and MinPts=2. The time cost of clustering over all 3973 pages collected

was about 1 second.

dSqS

dSqSdqr

,

Page 37: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

37

4.2 Clustering Incremental update of the data

If there is no similar web page, we create a new cluster for the new submission.

If the similarity is above the given threshold and all similar web pages are in the same cluster, we assign the new submission to this cluster (unless the cluster is at its maximum size).

If there are many similar web papes in different clusters, we choose the largest cluster that is not at its maximum size.

After a new submission is grouped in a cluster It has zero votes and does not inherit the votes of any other

submissions in the same cluster.

Page 38: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

38

4.3 VoteweightThe core idea behind voteweight is that participants

who are more helpful in terms of time and accuracy are

weighted more than other participants.

It measures how powerful of users’ votes impact on the final status of the suspicious URLs.

Its value comes from the accuracy of users’ historical data a correct vote should be rewarded and a wrong one should

be penalized recent behavior should be weighted more than past behavior

Page 39: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

39

4.3 Voteweight

In our model, we use y {t,+∞} y {-t,-∞}∈ ∪ ∈ to label the status of a URL, where y is the sum of voteweight of a given URL, t is the threshold of voteweight, y≥ t means a URL has been voted as a phishing

URL y≤-t means voted as legitimate.

Page 40: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

40

4.3 Voteweight

M

k k

ii

v

vv

1

'

otherwise

RVifRVv iii 0

0

iii PRRV

N

jLC

ji jI

TT

TTR

jij1 0

0 )(1

N

jLC

ji jI

TT

TTP

jij1 0

0 )(1

Axif

AxifxI A 0

1

K

iitit Cvl

1

'

otherwise

phishasvotedifCit 1

1

(5)

(2)

(3) (4)

(1)

(6)

(7) (8)

Page 41: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

41

7. Investigating VoteweightTuning Parameters in Control Condition

Voteweight achieves its best accuracy 95.6% and time cost of 11

hours with t=0.08 and =2.5 in the control condition Average time cost drops to 11 hours (11.8 hours without

voteweight) Median time cost drops to 2.3 hours (3.8 hours without woteweight)

Page 42: Smartening the Crowd: Computational Techniques for Improving Human Verification, at SOUPS 2011

©2

01

1 C

arn

eg

ie M

ello

n U

niv

ers

ity :

42

7. Investigating VoteweightTuning Parameters in Cluster Condition

Voteweight achieves its best accuracy of 97.2% and time cost of

0.8 hours with t=0.06 and = 1 in the control condition Average time cost drops to 0.8 hours (1.8 hours without

voteweight) Median time cost drops to 0.5 hours (0.7 hours without

woteweight)