[Chapter-delegates] Internet Filtering

Marcin Cieslak saper at saper.info
Mon Jun 30 09:12:44 PDT 2008


Sivasubramanian Muthusamy wrote:
> I am far from a person who advocate control of the internet

I will refer to the points below because I think they are important to 
improve our understanding of the "malicious activity" on the Internet.
Please do not treat this as a personal response to Sivasubramanian -
I am actually taking opportunity to quote some sentences to engage in
a bit more deeper discussion.

> There is a lot of malice around, that arises out of the power placed
> on the malicious underground (here again I would like to say that the
> "underground" per se is not malicious, but a section of the
> underground)  that brings down computers and whole networks down with
> destructive malware. 

It is important in the discussion we have with the stakeholders that 
various kinds of malice should be treated appropriately. Once cannot put 
  DDoS, hacking, unwanted content into one category. I believe that we 
should stress each and every time that the particular issue should be 
identified and appropriate measures taken and no "catch-all" solutions 
exist.

 > On extreme issues such as child pornography I believe that the 
Internet community
 > needs to find a way to keep these untraceable content completely out 
of the net.

I have trouble understanding why "child pornography" made such a career 
in recent times. Mentioning this over and over as the example of the 
"universally accepted wrongdoing" brings more harm than good; I wonder 
how many people started looking for such content just out of curiosity.

The Internet reflects more and more accurately human activities in the 
real world. It is not the issue to remove the content out of the net.
It is the issue to root-out the particular criminal activity. There has 
been lots of expertise in tracking particular kinds of criminals in the 
real world.

One very often forgotten fact is that Internet brings also new 
possibilities of combating crime; for example it is much easier and 
safer for police agent provocateur to act on the Internet. They do not 
need to risk they lives to engage with the wrongdoers.

> It requires some form of regulatory mechanism. I am not suggesting
> that such powers be vested with the Government. 

There will not be and cannot be a single regulatory mechanism, since we 
are dealing with very different policy issues:

1) Network and protocol-level vandalism
- is handled today to a some successful extent by the network operators 
by managing excess traffic for example.

2) Preventive (ex-ante) censorship - blocking the content before it can 
be seen
- is handled in the best way by the end-user.

Filtering out pornography is no different than popular spam- or 
ad-filtering. I would advocate the use of Privoxy 
(http://www.privoxy.org/) or similar solutions to many end-users. 
Actually, removing pop-under and pop-up advertisements makes it much 
more difficult to reach unwanted web sites.

In my practice since many years I have not been able to reach any 
explicit pornographic content by accident, mainly because I use 
appropriate host security, spam and ad filtering software. Maybe I am 
lucky, but I could access specific content only by explicit request and 
even with some effort.

In general I advocate the use of personal filters because of the very 
fact that every user should be able to define his/her own filtering 
rules. Every user should also be able to temporarily turn off the 
filtering should there be a need to access specific normally unwanted 
content. This way those users' policies (what's unwanted and what's 
wanted) can be personalized and can change over time, in order to become 
more effective.

Whoever takes the responsibility to take care of others (being that ISP, 
the government etc.) by providing a shared policies (global, nation-wide 
or per-site filters) is taking the possibility of free choice off the 
end-users. This, among from free-speech disadvantages mentioned already 
by Oliver, generates also the false sense of security - "we are safe 
because somebody ELSE takes care of us".

We - in Poland  - have analyzed the rules of one the software "blessed" 
to be used at the Polish public schools (unfortunately in Polish: 
http://prawo.vagla.pl/node/6430). The result? It blocked access to many 
popular blogging sites and everything about homosexuality.
The Internet community here has managed to cause that issue to be 
petitioned officially on the parliamentary forum 
(http://prawo.vagla.pl/node/6984).

3) After-fact (post-ante) take down mechanism.

This works only for the best for the most obvious and universally 
regarded as malicious content (for example worms, viruses). 
Traditionally in the early days of the Internet this was handled with 
the so-called LART-mechanism - the informal process of exchanging 
information between network operators. Today this process has been 
largely replaced be some sort of cease-and-desist letters, being 
generated manually or automatically, with limited success.

The issue with #3 is similar to that one of the global filtering - no 
one is able (and I presume will never be able) to provide a global set 
of policy rules stating what's right or wrong.

 > But would there be any harm if such powers are vested with the 
Internet Community
 > - you and me and those from the Internet Community renowned for their 
values of
> freedom and privacy and other rights and values that are
> characteristic of today's internet?

Yes, but finding a universal solution to the wrongdoing on the Internet 
is a special case of solving the general philosophical question "What's 
right and what's wrong?". While it is easy to name few unquestionable 
cases, things are getting really nasty when the edge cases come to play, 
for example as regards to political, religious or sexual content.

Therefore I believe that the general strategy of ISOC should be to 
empower and educate users about making their own explicit choices and 
the possibilities to apply appropriate technology.

-- 
               << Marcin Cieslak // saper at saper.info >>

-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 273 bytes
Desc: OpenPGP digital signature
URL: <https://elists.isoc.org/mailman/private/chapter-delegates/attachments/20080630/9bbde3cc/attachment.asc>


More information about the Chapter-delegates mailing list