P2P Web Publishing
how the web works
- (for non technical readers)
When you type a URL (Uniform Resource Locator) into your browser your computer contacts a DNS (Domain Name System) server and looks up the IP address (Internet Protocol) that belongs with the URL. Contact is established between your computer and the one hosting the website.
The domain name and the central server is owned by some one, this doesn't just mean this person has to pay for it he or she is also responsible for the content being served up, this requires watching the server like a hawk commonly known as moderation, the responsibility entirely justifies the weird level of control such owner has over your online activity. You are doing things on their website after all.
P2P Web Publishing
A p2p web publishing platform is needed. (think: a forum, a wiki, a blog, articles, books etc) Because if you have something to say you should say it on your own domain and/or your own IP address, that way you are responsible for it. No need for any authority to filter everything you say. You can add magnet uri's to pirated content, porn links, be a troll - everything.
- This is how that would work:
(first server to server; later also: client to server, server to client and client to client)
1) the script pulls a record list from either a public server or a private one.
2) the script compares records on the drive with the list
3) the script downloads the records it doesn't have jet
4) the script checks the content against content filters
5) the script logs sources of crap
6) the script adjust future spider interval based on the content quality and quantity then may distribute a crap list.
7) the script combines available source locations for each record
Each node chose to either a) pre moderate messages, b) be the only poster or c) just allow anyone to post. Eventually we can destroy the elegance of the system by in stead of plain text adding html and embedded stuffs, the same way rss and usenet was destroyed.</sarcasm>
Format the data payload
A file type is needed to send data from one client to another.
- XML - Guideline for file types containing and describing data
- RDF - Framework for describing Web resources
- SIOC - Framework allows data to be exported, aggregated and consumed from social websites.
- link lists & txt files
Find target computers
The user domains and IP addresses need to be distributed around the network in a sensible way both for initiation and cancellation preferably in real time. We would rather have slower data propagation than large numbers of unsuccessful attempts to make contact. Existing networks are understandably overly trigger happy in favor of failure.
- peer exchange
- local peer discovery
- tracker (webservers)
Filter the data
This should include spam filters but also means to make higher quality documents travel further and faster around the network. Specific data types (subjects) may be refused or promoted depending on personal interest or location. Such information may be pushed around the network.
There is a need for a way to figure out who wrote what in case this is desirable.
- user registration (forwarding tunnel: site.com > mainsite.com?return=site.com > site.com?user=name&timetoken=4728982)