ACCELER8OR

Jul 17 2011

MondoNet Fights The Internet Power: an Interview with Aram Sinnreich

Share

If you’ve been paying attention to the news lately, I’m sure you’ve heard about the “Declaration of War” by Lulzsec and Anonymous. Regardless of what you think about their actions, it should be obvious that there has been a growing effort by large corporations and governments to recreate the internet in an image completely alien to its original intentions – that of a robust decentralized network which no amount of damage could bring down. DARPA’s original designs have been sidetracked by various groups trying desperately to eliminate a “free and open” network model in favor of centrally controlled, corporate owned, authoritarian “walled gardens” in which your every move is tracked, your every file is subject to the central authorities approval, and your every action with your own possessions is monitored to ensure you’re not doing something a company disapproves of. The corporations have no desire to allow the masses a future in which information is not monitored, metered, charged for, and always in the control of the corporations.

In order to do this, they need the networks to be centralized, run entirely through systems in which every bit of data transmitted back and forth from the end user to the internet is subject to their scrutiny and approval. Due to the manner in which the net has developed using existing infrastructure, there are many bottlenecks which have made this effort possible, from cell phone towers that ensure that smartphones are routed through bandwidth limiting servers that meter per second access and looks for ways to charge customers for “overuse” to broadband services who inspect every packet of information to prevent you from downloading movies and songs. Even the FBI wants the ability to search your computer remotely.

So while I am uncertain of the utility of “cyberterrorism,” I can certainly see the need for methods to prevent the “digital frontier” from becoming a “virtual prison” and to encourage a return to that robust decentralized vision created by DARPA.

That’s pretty much what these guys at MondoNet are about and I think their words say it pretty clearly:

“Although the Internet is highly decentralized in its communication and social patterns, its technical and regulatory foundations are extremely hierarchical, due to centralized control by organizations like ICANN and the oligopolistic ownership of the access business by a handful of broadband ISPs and wireless carriers (Wu, 2010). As a result of this centralization, digital communications are compromised by a degree of surveillance and censorship that would be unthinkable in traditional social arenas, threatening our cyberliberties and “e-speech” rights (Sinnreich & Zager, 2008).

“Seemingly disparate issues like network neutrality, intellectual property treaties and national security measures, taken in combination, threaten to produce a communications environment in which innovation is stifled and normative cultural behaviors are criminalized and punished by censorship, fines and/or imprisonment. One potential solution to this problem would be to create a new communications platform based on existing Internet protocols, but with a decentralized infrastructure free of the bottlenecks and chokepoints that plague the current system. Specifically, this new infrastructure would use mesh networking technologies to produce a stable, ad hoc global wireless network in which each peer is a router, server and client combined, and in which no single state or organization can effectively censor or surveil the population on a massive scale.”

If you’ve never heard of a mesh network, or don’t understand the technical jargon, it basically means that every “device” — be it a smartphone, computer, or other internet connected device — “talks” to every other device, instead of to a “tower” or to a “server.” In other words, instead of talking to an “ISP” or “Carrier” who stands between you and the internet, your device will simply be part of the internet.

In the current infrastructure, if you want to call your friend, your device can’t just call up your friend. First it has to call up to a centralized, corporate controlled network. Then it has to confirm that it is allowed on that network and receive approval to use that network. Then it has to ask permission to create a connection to your friend, and verify that your friend is allowed access to the network, has not been banned from the network, and is connected to the network instead of a different network. Then, it will finally be allowed to create a connection, subject to monitoring from the central network. In a mesh network, your device would simply go to the next nearest device and then on to the next, and so on until it made a connection to your friend, over hundreds of different paths that it would turn into a “virtual private network” in which everything you and your friend say is inaccessible to anyone but the two of you because your data would be being sent along too many different paths to intercept.

Mondo.net again does a fantastic job of summarizing the principles of such a network:
1. Decentralized
The network should not be operated, maintained, or in any way reliant upon a single or minimally differentiated set of entities or technologies. No individual, entity or group should be central to the network to the extent that their absence would measurably impact its functionality or scope. Network participation should not require access to fixed, physical infrastructure of any sort.
2. Universally Accessible
The requisite technology and expertise required to participate in the network should be available at minimal cost and effort to every human being on the planet. Furthermore, all users should be able to extend the network’s content and functionality to suit their own needs, or those of others. No aspect of the network’s functioning should be reliant upon proprietary technologies, information or capital.
3. Censor-proof
The network should be resistant to both regulatory and technical attempts to limit the nature of the information shared, restrict usage by given individuals or communities, or render the network, or any portion of it, inoperable or inaccessible.
4. Surveillance-proof
The network should enable users to choose exactly what information they share with whom, and to participate anonymously if they so desire. Users should only have access to information if they are the designated recipients, or if it has been published openly.
5. Secure
The network should be organized in a way that minimizes the risk of malicious attacks or engineering failure. Information exchanged on the network should meet or exceed the delivery rate and reliability of information exchanged via the Internet.
6. Scalable
The network should be organized with the expectation that its scale could reach or even exceed that of today’s Internet. Special care should be taken to address to the challenge of maintaining efficiency without the presence of a centralized backbone.
7. Permanent
The network’s density and redundancy should be great enough that, despite its ad hoc nature, it will persistently operate on a broad scale, and be available in full to any user within range of another peer.
8. Fast (enough)
The network should always achieve whatever speed is required for a “bottom line” level of social and cultural participation. At present, we assert that the network’s data transfer rate should, at a minimum, be enough for voice-over-IP (VoIP) communications, and low-bitrate streaming video.
9. Independent
While the network will have the capacity to exchange information with Internet users and nodes, it should be able to operate independently, as well. A large-scale failure or closure of Internet infrastructure and content should have minimal effect on the network’s operations.
10. Evolvable
The network should be built with future development in mind. The platform should be flexible enough to support technologies, protocols and modes of usage that have not yet been developed.

The question then is this: Why should you care?

And I’ll be honest here. I see mesh networks naturally evolving to become the dominant form of network over the next few decades, because it’s the most practical solution to a number of problems that will have to be solved in order to build the VR web as well as to connect the entire world to the internet. Centralized networks are only possible in highly developed countries with existing infrastructures like power and telephone grids, as well as roads. You can’t build a tower where you don’t have either power or access. For vast areas of the world, mesh networks will be the only feasible solution. As handheld devices get cheaper, smaller and use less power, and batteries become able to store weeks or month’s worth of power for them, they will become the world’s primary means to access the internet. As billions of devices begin attaching to the net, they will overwhelm any centralized system. At that point, it will be much simpler to use them in a mesh than it will be to try and build sufficient infrastructure to meet demand. A mesh network can grow as fast as you add a new device to it. And unlike traditional networks, it auto-updates itself as users discontinue using older devices and switch to new ones. It also will eliminate bandwidth issues as thousands of paths will allow data streaming at the limits of the devices own hardware. As we move past the multimedia age and into the VR age, the need for vast amounts of data to be transferred will force the abandonment of centralized systems that simply cannot handle the load for robust multispectrum wireless networks that are more akin to P2P torrents than today’s cellular networks. Technology advancement itself is going to ensure we will move to mesh architectures in the very near future. So why, really, should you care?

Because the sooner we begin working towards developing these peaceful, innovative, and practical solutions to the threats of authoritarian control of the worlds developing nervous system, the less need we will have for “Cyberterrorists” like Lulzsec and Anonymous to cause disruptive attacks that hurt the innocent and the guilty alike in a fight to preserve our freedom to access information. And that is a goal I think every side can agree on

Valkyrie Ice

RU SIRIUS:  John GIlmore famously said, “The Internet interprets censorship as damage and routes around it? True or not? And what’s missing from the Internet as it is, structurally, that requires an alternative creation for mass p2p activity?

ARAM SINNREICH: Gilmore was really talking about TCP/IP, and node-level routing. While it’s true that these protocols are inherently decentralized and therefore very resistant to censorship, the problem comes at the infrastructural level, which is far more centralized. The Internet may have a billion nodes now, but only a handful of companies control the Internet’s “backbone,” the broadband ISP market, and the wireless data services market (and two companies, AT&T and Verizon, are the dominant players in each of these sectors in the US). That means that no two Internet users in the US can communicate with one another without their data passing through the hands of one of these companies.

Unless the data is encrypted (in which case, it can be dumped if the companies choose), it is vulnerable to packet-sniffing or other forms of inspection. This is not a hypothetical situation; these companies have long histories of inspecting packets for the purposes of (a) commercially preferred treatment, e.g. non-neutral network operations; (b) commercial exploitation, e.g. consumer profiling; and (c) political censorship and surveillance, either at the behest of the federal government (think NSA wiretaps) or for reasons of corporate ideology (e.g. AT&T censoring an Eddie Vedder webcast critical of President Bush; Verizon blocking text messages between pro-choice groups and their members).

Thus, only a network that replicates the decentralization of nodes at an infrastructural level — or, to put it another way, one in which the nodes themselves are the backbones — can eschew the Internet’s vulnerability to censorship.

RU: There seems to be a lot of activity going on in a similar direction to MondoNet. Are you aware of other efforts and how does your plan differ?

AS: MondoNet was a germ of an idea in my mind from about 2004, when I first read Tim Wu’s work on Net Neutrality, and the idea of mesh networking as a democratizing force was something I discussed with my USC doctoral professor Francois Bar at the time. In fact, a video about mesh networking that Bar produced with my contributions in 2005-6 is available here (see the automata video). So this has been percolating for a while. However, it wasn’t until I started my tenure-track position at Rutgers in 2010 that I had the support to begin doing something with the idea. I taught a doctoral course in the fall called “Visions and Revisions of Cyberspace,” and two of my students in that class, Nathan Graham and Aaron Trammell, shared my enthusiasm for addressing the social and political dimension of network technologies. So at the end of the fall semester, we applied for a small grant to start MondoNet, and have been working on it ever since.

It’s been a gratifying surprise to see how much the concept of mesh networking has taken hold across disciplines and even in the mainstream press in the seven or eight months since we started our work. In January, 2011, OpenMesh launched their initiative, and we started sharing our work on MondoNet.org. In February, Hillary Clinton gave her “Internet freedom” speech, and a few days later, the Freedom Box initiative announced itself in the pages of the New York Times. In April, I announced MondoNet at my TEDxUSC talk.

Now, it seems every day we read a new article or hear about a new initiative along these lines. And what’s really cool is that all of these networking initiatives have started to network as well. We’ve gotten code-sharing offers from other projects, been invited to Google Groups uniting researchers in this field, and are even planning our own Rutgers mini-conference on the subject later this year.

Where I think we differ, and can offer some vital perspective, is in our theoretical orientation. Unlike most of the other initiatives out there, we’re not engineers or policy wonks. We’re critical information scholars, bringing perspectives from social science, political economy and even cultural studies into the mix. This is why my TEDx talk and our soon-to-be-published article in The Information Society begin with what we call “social specifications,” emphasizing the qualities that free society requires from a network, rather than the capacities that given technologies can offer us. Before we even start thinking about protocols and feature sets, we want to be perfectly clear about what we’re trying to accomplish. So today, wireless ad hoc networks might be the best solution to address these social specifications, but ten years from now, there may be other options. Either way, we’re wedded to the principles, not to the tech.

Of course, we’re just as interested in making free and open networked communications a reality as anyone else out there. Our immediate plan is to test a “virtual” version of MondoNet in three different types of community, and to operationalize our social specifications through a variety of different data measurements. Once we are confident that MondoNet will actually move the needle on these target goals (e.g. accessibility or resistance to censorship), we’ll start building actual MondoNet software to spec. And the more we can make use of other projects’ openly-licensed code in the process, the happier we’ll be.

RU: I wonder if this thing scales, or if you want it to.  For example, for all the horrors of Facebook, the charm is in the fact that there are like a billion people there.  I can go find my old high school buddy or my great aunt. And while smaller decentralized alternative networks might be an advantage to, say, revolutionaries in Egypt or wherever, in that the government would find it harder to shut it down, there’s the chance that it won’t reach a lot of the people.

AS:  We definitely want it to scale — not just in terms of growth, but in terms of applicability within social milieus of any size. As you point out, it needs to work for small groups of dissenters within oppressive environments, but it also needs to provide a large-scale platform for an uncorruptible public sphere. In my ideal future, the entire globe will be covered with a stable, decentralized, peer-to-peer communications mesh, which can be used as a platform for public, closed-group, and person-to-person information exchanges.

As to Facebook’s “charm,” you’re certainly right that size matters when it comes to networks. I’m sure you’re familiar with Metcalfe’s law, which states that the value or power of a communications platform grows exponentially as the number of peers grows incrementally. And with MondoNet, this applies to an even greater extent, because users will rely on one another not only as senders and recipients of information, but as components of the network itself. All that being said, we’re not trying to replace or rival Facebook. In fact, we’d be delighted if Facebook chose to mirror its servers on MondoNet peers one day. We’re aiming to be pure infrastructure, simply a reliably secure and open alternative to the increasingly draconian and expensive broadband and wireless commercial networks.

RU:  I’m amazed that Rutgers is supporting this (so far.)  Certainly the government takes very seriously their abilities to surveil communications (going all the way back to the fight over the Clipper Chip in the mid-90s).  Do you expect a visit from some friendly folks at Homeland Security if this becomes viable?

AS: Yes, I’m sure that once we move from the talking phase to the doing phase, someone with national security concerns will come a-knocking. They’ll probably ask us to build in a “back door” to allow wiretaps and other forms of surveillance, just like they have for all the other network service operators. But the beauty of the technology we’re building is, such a back door would be impossible from an engineering standpoint. There’s no central backbone or other point of presence through which the majority of bits will flow. Furthermore, the platform will enable native peer-to-peer encryption (like PGP), which means that each individual node will have the ability to determine the visibility or obscurity of the information it sends. And, because the code is all open-source, even if we did create some kind of workaround back door, other developers could simply engineer it back out, and release an improved, higher-security version.

I’m not surprised that Rutgers is being supportive, though I am gratified — after all, free speech, social equality and technological innovation are key aims of our school, so we’re pretty much in line with the mission statement. However, it is a state university, so I suppose if federal regulators got a bee in their bonnet about the project, they could probably exert some political force to get the university to kibosh us. Hopefully, by the time that happens, we’ll have enough research, code, and project inertia to continue independently.

RU:  What sort of time frame do you think you’re looking at before this starts testing and how long after that do you think it might become viable for lots of people?

AS: As social scientists, we ask questions first and act second. So our first order of business will be to test the premises of MondoNet by creating a “virtual” version for field research. Will a peer-to-peer mobile mesh network actually address the social and political flaws of the existing Internet? Will it increase accessibility? Decentralize communications? Prove resistant to censorship and surveillance? We’re currently in talks with a mobile software developer to create this virtual MondoNet, and hopefully we will test it over the next year.

Once we determine whether our strategy actually does what we hope it does, then we can begin to develop the software itself. The good news is, many of the components (e.g. mesh networking protocols) are already developed or in development under open license elsewhere, so we don’t need to reinvent the wheel. We see our role more as integrating these back-end technologies, creating an easy-to-use and widely accessible user interface, and then shepherding the project development within the open source/free software community. Given that we will probably need to rely on grant money to do this, a realistic timeframe would be about 3 years till public release.

Coming from the business world, I realize that this might seem like a long development cycle, but we want to get it right the first time, and build a strong foundation for future development by ourselves and other coders and communities. If we do it right, the project should take on a life of its own, and the code will be adapted for uses and sociopolitical contexts we can’t predict at this point.

RU: Broadly speaking, there has been a lot of controversy over the years about the liberatory or revolutionary potential of the internet, ranging from technotopian imaginings to scathing indictments.  What’s your view?

AS: Great question. I actually teach a doctoral course on this topic (Nathan and Aaron — both project participants — took it last year). The short answer is, all technologies are inherently neutral. They can be used toward both emancipatory and totalitarian ends, and usually both apply. Through laws, regulations, architecture and social norms, different interests work to redefine the role of technology to achieve the social outcomes they seek. Right now, many regulatory and architectural developments are pushing the platform towards informatic totalitarianism, though emerging social norms continue to explore its emancipatory potential. We see MondoNet as an architectural intervention, reversing the “cable-ization” of the network and undermining the power of regulators to centralize control over information flow.

As a final word, I would also like to mention that this process of social reorganization through evolving communication technologies will never end. It will never reach a happy medium, a comfortable resting place, or a peaceable stasis. As the pace of technological innovation continues to accelerate relative to more organic human processes, communications networks will continue to play an increasingly central role in our politics, and the stakes will grow ever greater. Even if MondoNet is wildly successful, and we achieve our dream of a decentralized, universally distributed global mesh in 20 years, we can’t expect the story to end there.

Nanotechnology, quantum computing, and genetic science are just a few of the emerging fields that have far-reaching political and communication implications, and thus far these implications have been primarily addressed within the discourse of science fiction rather than research and advocacy. That’s why each of us needs to be aware of the power dynamics surrounding technological innovation, and to continually ask ourselves how we can intervene to help shape a future we’d want to live in.

Share
Jun 20 2011

Apple, Google, and the Future of the Cloud

Share

A packed house of over 5,000 software developers watched as Apple’s Steve Jobs introduced the latest operating systems for the Mac computer and for iPhone, iPad, and iPod touch at the recent Apple World Wide Developers’ Conference 2011. He also talked about a new service, iCloud, which stores content on remote servers and makes synchronization of multiple Apple devices, Macs, and PCs as easy as using Apple’s ubiquitous iTunes.

Here Steve Jobs introduces the concept of iCloud (video courtesy of CNET):

Cloud computing, where documents, data and more are stored remotely so you can access them from a connected machine like an iPad or a computer, is not a new concept.  It was most probably derived from the diagrams of clouds used to represent the Internet in textbooks and resulted when telecommunications companies made a radical shift from point-to-point data circuits to Virtual Private Network (VPN) services in the 1990s. By optimizing computer resource utilization through load balancing, they could get their work done more efficiently and inexpensively.

But leave it to Apple to bring the Cloud to the masses in a big way (although it was Google Docs in 2006 which really brought cloud computing to the forefront of public consciousness). While the jury is still out on whether Google docs is a serious competitor to Microsoft’s Office suite — researchers from Pennsylvania State University recently determined that the cost of running small workloads though may actually be more expensive for larger compute jobs, compared to the costs of running such work in-house — there are still many Google Docs success stories. And Google continues to enhance Google Docs with a recent upgrade including a new feature that lets users discuss shared documents in real-time.

Unlike Google’s cloud-based service, Apple’s iCloud uses the cloud “to orchestrate data streams rather than control them” according to a recent ZDNet article. The cloud is used as a central repository for apps, music, media, documents, messages, photos, backups, settings, and more. Both Apple and Microsoft have been exploring the idea of a “central hub” of our digital life and work for over 10 years, with a variety of devices relying on it to coordinate content.

For Apple, the packaging of the this hub in the form of the iCloud service is also a sound business strategy. The New York Times reports that Apple’s iCloud has the potential to wipe out some existing web services and entire businesses with the integration of iCloud, Mac OS X Lion, and iOS 5. These businesses include: Instapaper, Red Pop hardware iPhone camera button, BlackBerry Messenger and GroupMe, Amazon Cloud Player and Google Music, Google Docs and Google Chromebook, and Dropbox.

While Apple wants your data now, Google’s entire strategy and approach to the cloud is based on a future vision of the Internet with low-cost, ubiquitous Internet access — including fiber connections in offices and homes and super-fast mobile broadband all over the world. Here is a fun video explaining the concepts behind Google’s Chrome OS:

Despite the cloud’s potential for cost savings and “reducing the hassles of running in-house computer servers,” the LA Times reports that  it may not yet be as safe as advertised. Because data from hundreds or thousands of companies can be stored on large cloud servers, hackers can theoretically gain control of huge stores of information through a single attack — a process called “hyperjacking.” Security professionals said the many attacks recently in the news reflect both an increase in hacking activity and new pressure on companies to quickly disclose when they’ve been attacked.

A recent study by Baseline shows that while many organizations are still holding back, fearful of security hacks, lost data, or lost control, a growing number of enterprises are driving their business to the cloud. They interviewed several companies including Kelly Services, Lionsgate Entertainment, HarperCollins Publishers, GWR Medical, Imperial Sugar, WhitePages, Suncorp, and Dubset.  Their  findings reinforce the opinions of the CEO panelists at the MIT Sloan CIO Symposium, held in Boston in May. The panelists at the keynote address, “Opportunities and Strategies in the Digital Business World,” spoke about the cloud as a way to free up capital so companies could take advantage of new business opportunities.

Microsoft’s Steve Ballmer implicitly acknowledged the importance of the cloud services in a recent interview — Microsoft is set to announce their answer to Google Docs: the Office 365 online service. And Jan Muehlfeit, European chairman of Microsoft Corp., sees the future in cloud computing.

Apple, Google, Microsoft, as well as Facebook (where many people already store the major events and interactions of their lives on the cloud) seem to be betting their billions on a cloud-based future.

What might this future look like?  SF writer Rick Moss poses some fascinating possibilities and questions in his new novel, Ebocloud.  The “ebocloud” is an internet social network subdivided into tribes formed by shared affinities (think Facebook). The cloud addresses, as one character says, “the primordial urge for belonging” and the end of aloneness. The cloud learns from the actions and interactions of each of the members (called cousins) who are connected to the cloud with “digital tattoos” using brain-stimulating and brain-computer interfacing nanotubes.

Moss asks what happens if the interaction with the cloud became two-way? What if the cloud could begin to affect emotions? “The objective of the human-cloud collective is to facilitate a feedback loop whereby human sensory data and biometrics are uploaded to the cloud to be aggregated, analyzed and used in various applications, then redistributed back to the human participants,” says Moss in a recent interview for H+ Magazine.  He sees this interaction resulting in what he calls a “social singularity,” a transhuman superintelligence emerging from “the cloud” network.

While Moss concedes that the technology is not yet in place for such a social singularity, he considers his novel a warning to future application developers: “The likes of Apple, Google and Facebook, with their billions to burn on wild-eyed notions, could very likely spur technological leaps in these areas. It would be nice if these developers were at least aware of the possibility of a social singularity so that they don’t stumble into the phenomenon blindly. That could be tragic.”

Share