URL filtering: Difference between revisions

From ASRG
Jump to navigationJump to search
No edit summary
Line 12: Line 12:
As long as URLs found in spams change very frequently, the maintenance of this kind of blacklist is a hard work and, most of the time, use a lots of spamtraps to collect spams.
As long as URLs found in spams change very frequently, the maintenance of this kind of blacklist is a hard work and, most of the time, use a lots of spamtraps to collect spams.


Efficiency of URL filtering is usually something between 50 % and 70 %.
Efficiency of URL filtering is usually something between 50 % and 70 %. False positive rate can be as low as 0.1 %, but some lists are more agressive than others.

Revision as of 15:27, 29 May 2009

Anti-spam technique: URL filtering
Date of first use: early 2000
Effectiveness: Medium
Popularity: High
Difficulty of implementation: Medium
Where implemented: MTA
Harm: Low


The idea behind URL filtering is that most spams contains URLs to redirect people to some web site. So, it's enough to extract all URLs present in the body of the message and check them against a blacklist. Primitive filters can static flat file blacklists, and the efficiency and drawbacks are the same of a static list of keywords. Most modern filters use URL blacklists stored in DNS zones, as this is an easier way to distribute these lists.

As long as URLs found in spams change very frequently, the maintenance of this kind of blacklist is a hard work and, most of the time, use a lots of spamtraps to collect spams.

Efficiency of URL filtering is usually something between 50 % and 70 %. False positive rate can be as low as 0.1 %, but some lists are more agressive than others.