Post Commons

wiki?journal

Thursday, October 2nd, 2014

The first wiki[pedia]2journal article has been published: Dengue fever Wikipedia article, peer-reviewed version (PDF). Modern medicine comes online: How putting Wikipedia articles through a medical journal’s traditional process can put free, reliable information into as many hands as possible is the accompanying editorial (emphasis added):

As a source of clinical information, how does Wikipedia differ from UpToDate or, for that matter, a textbook or scholarly journal? Wikipedia lacks three main things. First, a single responsible author, typically with a recognized academic affiliation, who acts as guarantor of the integrity of the work. Second, the careful eye of a trained editorial team, attuned to publication ethics, who ensure consistency and accuracy through the many iterations of an article from submission to publication. Third, formal peer review by at least one, and often many, experts who point out conflicts, errors, redundancies, or gaps. These form an accepted ground from which publication decisions can be made with confidence.

In this issue of Open Medicine, we are pleased to publish the first formally peer-reviewed and edited Wikipedia article. The clinical topic is dengue fever. It has been submitted by the author who has made the most changes, and who has designated 3 others who contributed most meaningfully. It has been peer reviewed by international experts in infectious disease, and by a series of editors at Open Medicine. It has been copy-edited and proofread; once published, it will be indexed in MEDLINE. Although by the time this editorial is read the Wikipedia article will have changed many times, there will be a link on the Wikipedia page that can take the viewer back to the peer-reviewed and published piece on the Open Medicine website. In a year’s time, the most responsible author will submit the changed piece to an indexed journal, so it can move through the same editorial process and continue to function as a valid, reliable, and evolving free and complete reference for everyone in the world. Although there may be a need for shorter, more focused clinical articles published elsewhere as this one expands, it is anticipated that the Wikipedia page on dengue will be a reference against which all others can be compared. While it might be decades before we see an end to dengue, perhaps the time and money saved on exhaustive, expensive, and redundant searches about what yet needs to be done will let us see that end sooner.

I love that this is taking Wikipedia and commons-based peer production into a challenging product area, which if wildly successful, could directly challenge and ultimately destroy the proprietary competition. The editorial notes:

Some institutions pay UpToDate hundreds of thousands of dollars per year for that sense of security. This has allowed Wolters Kluwer, the owners of UpToDate, to accrue annual revenues of hundreds of millions of dollars and to forecast continued double-digit growth as “market conditions for print journals and books … remain soft.”

See the WikiProject Medicine collaborative publication page for more background on the process and future developments. Note at least 7 articles have been published in journal2wiki[pedia] fashion, see PLOS Computational Biology and corresponding Wikipedia articles. Ideally these 2 methods would converge on wiki?journal, as the emphasized portion of the quote above seems to indicate.

Peer review of Wikipedia articles and publication in another venue in theory could minimize dependencies and maximize mutual benefit between expert authoring (which has historically failed in the wiki context, see Nupedia and Citizendium) and mass collaboration (see challenges noted by editorial above). But one such article only demonstrates the concept; we’ll see whether it becomes an important method, let alone market dominating one.


One small but embarrassing obstacle to wiki?journal is license incompatibility. PLOS journals use CC-BY-4.0 (donor-only relative to following; the version isn’t important for this one) and Wikipedia CC-BY-SA-3.0 (recipient-only relative to previous…and following) and Open Medicine CC-BY-SA-2.5-Canada (donor-only relative to the immediately previous) — meaning if all contributors to the Dengue fever Wikipedia article did not sign off, the journal version is technically not in compliance with the upstream license. Clearly nobody should care about this second issue, except for license stewards, who should mitigate the problem going forward: all previous versions (2.0 or greater due to lack of a “later versions” provision in 1.0) of CC-BY-SA should be added to CC-BY-SA-4.0’s compatibility list, allowing contributions to go both ways. The first issue unfortunately cannot be addressed within the framework of current licenses (bidirectional use could be avoided, or contributors could all sign off, either of which would be outside the license framework).

Daniel Mietchen (who is a contributor to the aforementioned journal2wiki effort, and just about everything else relating to Wikipedia and Open Access) has another version of his proposal to open up research funding proposals up at the Knight News Challenge: Libraries site. Applaud and comment there if you like, as I do (endorsement of previous version).

Near the beginning of the above editorial:

New evidence pours in to the tune of 12 systematic reviews per day, and accumulating the information and then deciding how to incorporate it into one’s practice is an almost impossible task. A study published in BMJ showed that if one hoped to take account of all that has been published in the relatively small discipline of echocardiography, it would take 5 years of constant reading―by which point the reader would be a year behind.

A similar avalanche of publishing can be found in any academic discipline. It is conceivable that copyright helps, providing an incentive for services like UpToDate. My guess is that it gets in the way, both by propping up arrangements oriented toward pumping out individual articles, and by putting up barriers (the public license incompatibility mentioned above is inconsequential compared to the paywalled, umitigated copyright, and/or PDF-only case which dominates) to collaborative — human and machine — distillation of the states of the art. As I wrote about entertainment, do not pay copyright holders, for a good future.

Do not pay copyright holders, for a good future

Sunday, September 28th, 2014

The Unrepentant Bootlegger profiles Hana Beshara, a founder of NinjaVideo, who spent 16 months in prison for defying censorship. Cut to the logic of censorship (emphasis added):

People watch more paid, legal content than ever, but they also continue to download huge amounts of illegal content. “Piracy is putting pressure on antiquated business models, which isn’t necessarily a bad thing,” said Brett Danaher, an economics professor at Wellesley College who studies Internet piracy. “But the prevalence of piracy shows that people are growing up in a culture of free, and that is not good for the future of entertainment, either.”

That we should be concerned for the future of entertainment, at all, is itself bizarre. Freedom and equality should absolutely trump incentivizing a surfeit of entertainment. If we must choose between spectacle and communications, spectacle should be destroyed. We do not need to choose. We can destroy the censorship regime, but entertainment, including for better or worse some of the spectacle variety, will continue to exist and be produced in vastly greater quantities and quality than it is feasible for anyone to even begin to fully appreciate in a lifetime. If the spectacle portion does not include projects with budgets of hundreds of millions of dollars, that is OK — we will love what culture does get produced, as that love and cultural relevance is largely based on being immersed in the culture that exists — we love the culture we’re in. If that culture is less dominated by U.S.-based high investment productions, so much the better for the U.S. and the world.

Another policy significant quote from the article:

Peter Eckersley, technology projects director at the Electronic Frontier Foundation [...] said the law should shift its focus to making sure that copyright holders are paid for their work, rather than trying to stymie how people gain access to it. [...] He suggested a legal framework to retire the “exclusive rights” aspect of copyright law that requires permission to publish ― and that allows copyright holders to seek exorbitant damages from infringers ― and move toward a system that requires sites and people who make money from another’s work to share any profits. Solutions like these, Mr. Eckersley says, would create different priorities that go beyond chasing small-time pirates like Ms. Beshara and her colleagues.

No, copyright holders should not be paid. Any payment by virtue of holding copyright only makes the censorship regime self-perpetuating. Funding of entertainment should be completely decoupled from the censorship regime of copyright. I understand the appeal of paid speech over permissioned speech (of course a tax is usually better than a prohibition, and that applies to privatized regimes as well), but neither is free speech. The paid speech approach would indeed create priorities that go beyond chasing small-time pirates (note Beshara earned $210k over 3 years; note also existing paid speech regimes which involve monitoring and shakedown of small-time restaurants) — it would invite further pervasive and destructive surveillance of communications in the interest of ensuring copyright holders get paid. It is appalling that EFF is still willing to invite sacrifice of everything they fight for at the alter of paying copyright holders. I don’t blame the EFF specifically; this just shows how deeply intellectual parasitism has burrowed in general. Intellectual parasites (which includes most reformers, including me often) need to fully shift to being commons policy advocates (and scholars).

Regarding people and projects like Hana Beshara and NinjaVideo, I’m ambivalent. Performing unpaid marketing and price discrimination services for the censorship industry is distasteful and harmful. But sharing culture (putting the regime aside) is tasteful and helpful. There is too little known about informal circulations and their effects, this lack of knowledge itself a collateral damage of the regime (compare being able to study cultural flows and surveillance required for paid speech; they are of different orders) and far, far, far too little direct competition for the regime.

Proprietary profitability as a key metric for open access and open source

Thursday, August 7th, 2014

Glyn Moody in Beyond Open Standards and Open Access:

Like open source, open access is definitely winning, even if there is some desperate rearguard action by the publishers, who are trying to protect their astonishing profit margins – typically 30-40%.

No doubt open source and open access have progressed, but the competition maintaining astonishing profit margins contradicts “definitely winning.” For publishing, see Elsevier, £0.8b profit on £2.1b revenue, and others. For software most pertinent to Moody’s post (concerning Open Document Format), see Microsoft’s business division, $16b profit on $24b revenue.

These profits coupled with the slow relative progress of open source and open access give proprietary vendors huge range to not only take “desperate rearguard action” but also to create new products and forms of lock-in with which the commons is continually playing catch-up.

We know what the commons “definitely winning” looks like — Linux (server software) and Wikipedia (encyclopedias) — and it includes proprietary vendor profit margins being crushed, most going out of business, and those remaining transitioning to service lines of business less predicated on privatized censorship.

When libraries begin mass cancellation of toll access journal subscriptions and organizations of all sorts cancel Microsoft, Adobe, and similar software subscriptions, then we can consider whether open access and open source are definitely winning. Until then the answer is definitely no.

As for what’s next for open standards and open access (Moody suggests further ODF mandates, which would be fine), the obvious answer is open source. It’s what allows realization of the promise of open standards, and the cancellation of Microsoft subscriptions. It’s also what’s next for academic publishing and everything else — what is not software will be obsolete — though cancellation of those toll access subscriptions is going to require going back to basics.

Free/open/commons advocates should consider destruction of proprietary competition profitability a key aim and metric of success or lack thereof, for both open products and policy. This metric has several benefits:

  • Indicates relative progress. Any non-moribund project/movement can make seeming progress, blind to different and potentially much greater progress by competition.
  • Implicates role of knowledge economy and policy in increasing or decreasing equality (of income and wealth, not just access).
  • Hard numbers, data readily available.
  • It’s reasonable to multiply destruction of proprietary profits when characterizing gains (so as to include decrease in deadweight loss).

Generative acknowledgement

Thursday, July 31st, 2014

Robin Sloan, The secret of Minecraft: And its challenge to the rest of us

In the 2010s and beyond, it is not the case that every cultural product ought to be a generative, networked system.

It is, I believe, the case that all the really important ones will be.

Nathan Matias, Designing Acknowledgement on the Web:

A system which acknowledges the beauty of cooperative relationships can’t be based on the impersonal idea of hypertext or the egocentric notion of authorship. It can’t rely on licenses to threaten people into acknowledging each other.

Via 1 2 3 and confirmation bias about which I can’t think of anything smart to say, so I’ll include a fun word: contextomy. Neither of the above reaches that bar, but I’ll try harder next time.

Posts on the ought of generative, networked production and intellectual parasite debasement of acknowledgement.

Edit Oakland wiki events

Wednesday, July 9th, 2014

Saturday, July 12, there’s a big open streets event in my obscure flats neighborhood where Oakland, Emeryville, and Berkeley meet. A small stretch of San Pablo Avenue will be closed to cars (sadly not only human-driven cars, which would momentarily meet my suggestion). E’ville Eye has a comprehensive post about the event and its origins.

There will be an Oakland Urban Paths walk in the neighborhood during the event, during which obscurities will be related. Usually these walks are in locations with more obvious scenery (hills/stairs) and historical landmarks; I’m looking forward to seeing how they address Golden Gate. Last month they walked between West Oakland and downtown, a historic and potentially beautiful route that currently crosses 980 twice — edit it out!

Monday, July 14 18:00-19:30 there’s a follow-on event at the Golden Gate Branch Library — an OaklandWiki edit party. I haven’t edited Oakland Wiki much yet, but I like the concept. It is one of many LocalWikis, which relative to MediaWiki and Wikipedia have very few features or rules. This ought greatly lower the barrier to many more people contributing information pertinent to their local situation; perhaps someone is researching that? I’ve used the OaklandWiki to look up sources for Wikipedia articles related to Oakland and have noticed several free images uploaded to OaklandWiki that would be useful on Wikipedia.

Saturday, July 19 11:00-16:00 there’s a Wikipedia edit event at Impact Hub in Oakland and online: WikiProject Open Barn Raising 2014 which aims to improve Wikipedia articles about open education — a very broad and somewhat recursive (Wikipedia is an “open educational resource”, though singular doesn’t do it justice, unless perhaps made singular the open educational resource, but that would be an overstatement). If you’re interested in OER, Open Access, open policy and related tools and organizations, or would like to learn about those things and about editing Wikipedia, please participate!

Tangentially, OpenHatch (my endorsement) got a nice writeup of its Open Source Comes to Campus events at WIRED. I view these as conceptually similar to introduction to Wiki[pedia] editing events — all aim to create a welcoming space for newcomers to dive into participating in commons-based peer production — good for learning, careers, communities, and society.

Open policy for a secure Internet-N-Life

Saturday, June 28th, 2014

(In)Security in Home Embedded Devices Jim Gettys says software needs to be maintained for decades considering where it is being deployed (e.g., embedded in products with multi-decade lifetimes, such as buildings) and the criticality of some of that software, an unpredictable attribute — a product might become unplanned “infrastructure” for example if it is widely deployed and other things come to depend on it. Without maintenance, including deployment of updates in the field, software (and thus systems it is embedded in) becomes increasingly insecure as vulnerabilities are discovered (cites a honeymoon period enjoyed by new systems).

This need for long-term maintenance and field deployment implies open source software and devices that users can upgrade — maintenance needs to continue beyond the expected life of any product or organization. “Upgrade” can also mean “replace” — perhaps some kinds of products should be more modular and with open designs so that parts that are themselves embedded systems can be swapped out. (Gettys didn’t mention, but replacement can be total. Perhaps “planned obsolescence” and “throwaway culture” have some security benefits. I suspect the response would be that many things continue to be used for a long time after they were planned to be obsolete and most of their production run siblings are discarded.)

But these practices are currently rare. Product developers do not demand source from chip and other hardware vendors and thus ship products with “binary blob” hardware drivers for Linux kernel which cannot be maintained, often based on kernel years out of date when product is shipped. Linux kernel near-monoculture for many embedded systems, increasing security threat. Many problems which do not depend on hardware vendor cooperation, ranging from unintentionally or lazily not providing source needed for rest of system, to intentionally shipping proprietary software, to intentionally locking down device to prevent user updates. Product customers do not demand long-term secure devices from product developers. There is little effort to fund commons-oriented embedded development (in contrast with Linux kernel and other systems development for servers, which many big companies fund).

Gettys is focused on embedded software in network devices (e.g., routers) as network access is critical infrastructure much else depends on, including the problem at hand: without network access, many other systems cannot be feasibly updated. He’s working on CeroWrt a cutting edge version of OpenWrt firmware, either of which is several years ahead of what typically ships on routers. A meme Gettys wishes to spread, the earliest instance of which I could find is on cerowrt-devel, a harsh example coming the next week:

Friends don’t let friends run factory firmware.

Cute. This reminds me of something a friend said in a group discussion that touched on security and embedded in body (or perhaps it was mind embedded in) systems, along the lines of “I wouldn’t run (on) an insecure system.” Or malware would give you a bad trip.

But I’m ambivalent. Most people, thus most friends, don’t know what factory firmware is. Systems need to be much more secure (for the long term, including all that implies) as shipped. Elite friend advice could help drive demand for better systems, but I doubt “just say no” will help much — its track records for altering mass outcomes, e.g., with respect to proprietary software or formats, seems very poor.

In Q&A someone asked about centralized cloud silos. Gettys doesn’t like them, but said without long-term secure alternatives that can be deployed and maintained by everyone there isn’t much hope. I agree.

You may recognize open source software and devices that users can upgrade above as roughly the conditions of GPL-3.0. Gettys mentioned this and noted:

  • It isn’t clear that copyright-based conditions are effective mechanism for enforcing these conditions. (One reason I say copyleft is a prototype for more appropriate regulation.)
  • Of “life, liberty, and pursuit of happiness”, free software has emphasized the latter two, but nobody realized how important free software would be for living one’s life given the extent to which one interacts with and depends on (often embedded) software. In my experience people have realized this for many years, but it should indeed move to the fore.

Near the end Gettys asked what role industry and government should have in moving toward safer systems (and skip the “home” qualifier in the talk title; these considerations are at least as important for institutions and large-scale infrastructure). One answer might be in open policy. Public, publicly-interested, and otherwise coordinated funders and purchasers need to be convinced there is a problem and that it makes sense for them to demand their resources help shift the market. The Free Software Foundation’s Respects Your Freedom criteria (ignoring the “public relations” item) is a good start on what should be demanded for embedded systems.

Obviously there’s a role for developers too. Gettys asked how to get beyond the near Linux kernel monoculture, mentioning BSD. My ignorant wish is that developers wanting to break the monoculture instead try to build systems using better tools, at least better languages (not that any system will reduce the need for security in depth).

Here’s to a universal, secure, and resilient web and technium. Yes, these features cost. But I’m increasingly convinced that humans underinvest in security (not only computer, and at every level), especially in making sure investments aren’t theater or worse.

“Open policy” is the most promising copyright reform

Thursday, June 26th, 2014

Only a few days (June 30 deadline) for applications to the first Institute for Open Leadership. I don’t know anything about it other than what’s at the link, but from what I gather it involves a week-long workshop in the San Francisco area on open policy and ongoing participation in an online community of people promoting open policies in their professional capacities, and is managed by an expert in the field, Timothy Vollmer. Read an interview with Vollmer (wayback link to spare you the annoying list-gathering clickthrough at the original site, not least because its newsletter is an offender).

The institute and its parent Open Policy Network define:

Open Policy = publicly funded resources are openly licensed resources.

(Openly licensed includes public domain.)

Now, why open policy is the most promising knowledge regulation reform (I wrote “copyright” in the title, but the concept is applicable to mitigating other IP regimes, e.g., patent, and pro-commons regulation not based on mitigating IP):

  • Most proposed reforms (formalities can serve as an example for each mention following) merely reduce inefficiencies and embarrassments of freedom infringing regimes in ways that don’t favor commons-based production, as is necessary for sustainable good policy. Even if not usually conceptualized as commons-favoring, open policy is strongly biased in that direction as its mechanism is mandate of the terms used for commons-based production: open licenses. Most proposed reforms could be reshaped to be commons-favoring and thinking of how to do so a useful exercise (watch this space) but making such reshaping gain traction, as a matter of discourse let alone implementation, is a very long-term project.
  • The concept of open policy is scalable. There’s no reason as it gains credence to push for its expansion to everything receiving public or publicly interested support, including high and very low culture subsidy. At the extreme, the only way to avoid being subject to some open policy mandate would be to create restricted works in an IPer colony, isolated from the rest of humanity.
  • In order to make open policy gain much more credence than it has now, its advocates will be forced to make increasingly sophisticated public policy arguments to support claims that open policy “maximizes public investment” or to shift the object of maximization to freedom and equality. Most proposed reforms, because they would only reduce inefficiency and embarrassment, do not force much sophistication, leaving knowledge regulation discourse rotting in a trough where economists abandoned it over a century ago.
  • Open policy implementation has the potential to destroy the rents of freedom infringing industries. For sustainable good policy it is necessary to both build up the commons as an interest group and diminish interest groups that depend or think they depend on infringing freedom. It is possible for open policy to be gamed (e.g., hybrid journal double dipping). As troubling as that is, it seems to me that open policy flips which side is left desperately clawing for loopholes contrary to the rationale of policy. Most reform proposals at least implicitly take it as a given that public interest is the desperate side.
  • Open policy does not require any fundamental changes to national law or international treaties, meaning it is feasible, now. Hopefully a few reformists have generally grasped the no-brainer concept that a benefit obtained today is more valuable than one obtained in the future, e.g., in 95 years. It also doesn’t mean that open policy is merely a “patch” in contrast the “fixes” of most proposed reforms — which aren’t fixes anyway, but rather mitigations of the worst inefficiencies and embarrassments of freedom infringing regimes. If open policy is a patch, it is a one that helps the body of knowledge regulation to heal, by the mechanisms above (promoting commons production and discourse, diminishing freedom infringing interests).

In my tradition of critical cheering, consider the following Open Policy Network statement:

We have observed that current open policy efforts are decentralized, uncoordinated and insular; there is poor and/or sporadic information sharing.

As illustrated by the lack of the Open Source Definition or any software-centric organizations on Open Policy Network lists of its guiding principles and member organizations. Fortunately software is mentioned several times, for example:

If we are going to unleash the power of hundreds of billions of dollars of publicly funded education, research, data, and software, we need broad adoption of open policies.

Hopefully if the Open Policy Network is to become an important venue for moving open policy forward, people who understand software will get involved (by the way, one of the ways “publicly funded” is scalable is that it properly includes procurement, not only wholly funded new resources), e.g., FSFE and April. I know talking about software is scary — because it is powerful and unavoidable. But this makes it a necessity to include in any serious project to reform the knowledge economy and policy. Before long, everything that is not software or suffused with software will be obsolete.

API commons

Thursday, May 29th, 2014

Notes for panel The API Copyright Emergency: What’s Next? today at API Con SF. The “emergency” is the recent decision in Oracle v. Google, which I don’t discuss directly below, though I did riff on the ongoing case last year.

I begin with and come back to a few times Creative Commons licenses as I was on the panel as a “senior fellow” for that organization, but apart from such emphasis and framing, this is more or less what I think. I got about 80% of the below in on the panel, but hopefully still worth reading even for attendees.

A few follow-up thoughts after the notes.

Creative Commons licenses, like other public licenses, grant permissions around copyright, but as CC’s statement on copyright reform concludes, licenses “are not a substitute for users’ rights, and CC supports ongoing efforts to reform copyright law to strengthen users’ rights and expand the public domain.” In the context of APIs, default policy should be that independent implementation of an API never require permission from the API’s designer, previous implementer, or other rightsholder.

Without such a default policy of permission-free innovation, interoperability and competition will suffer, and the API community invites late and messy regulation at other levels intending to protect consumers from resulting lock-in.

Practically, there are things API developers, service providers, and API consumers can do and demand of each other, both to protect the community from a bad turn in default policy, and to go further in creating a commons. But using tools such as those CC provides, and choosing the right tools, requires looking at what an API consists of, including:

  1. API specification
  2. API documentation
  3. API implementations, server
  4. API implementations, client
  5. Material (often “data”) made available via API
  6. API metadata (e.g, as part of API directory)

(depending on construction, these could all be generated from an annotated implementation, or could each be separate works)

and what restrictions can be pertinent:

  1. Copyright
  2. Patent

(many other issues can arise from providing an API as a service, e.g., privacy, though those are usually not in the range of public licenses and are orthogonal to API “IP”, so I’ll ignore them here)

1-4 are clearly works subject to copyright, while 5 and 6 may or may not be (e.g., hopefully not if purely factual data). Typically only 3 and 4 might be restricted by patents.

Standards bodies typically do their work primarily around 1. Relatively open ones, like the W3C, obtain agreement from all contributors to the standard to permit royalty-free implementation of the standard by anyone, typically including a patent license and permission to prepare and perform derivative works (i.e., copyright, to extent such permission is necessary). One option you have is to put your API through an existing standards organization. This may be too heavyweight, or may be appropriate yet if your API is really a multi-stakeholder thing with multiple peer implementations; the W3C now has a lightweight community group venue which might be appropriate. The Open Web Foundation’s agreements allow you to take this approach for your API without involvement of an existing standards body?. Lawrence Rosen has/will talk about this.

Another approach is to release your API specification (and necessarily 2-4 to the extent they comprise one work, ideally even if they are separate) under a public copyright license, such as one of the CC licenses, the CC0 public domain dedication, or an open source software license. Currently the most obvious choice is the Apache License 2.0, which grants copyright permission as well as including a patent peace clause. One or more of the CC licenses are sometimes suggested, perhaps because specification and documentation are often one work, and the latter seems like a “creative” work. But keep in mind that CC does not recommend using its licenses for software, and instead recommends using an open source software licenses (such as Apache): no CC license includes explicit patent permission, and depending on the specific CC license chosen, it may not be compatible with software licenses, contrary to goal of granting clear permission for independent API implementation, even in the face of a bad policy turn.

One way to go beyond mitigating “API copyrightability” is to publish open source implementations, preferably production, though reference implementations are better than nothing. These implementations would be covered by whatever copyright and patent permissions are granted by the license they are released under — again Apache 2.0 is a good choice, and for software implementation CC licenses should not be used; other software licenses such as [A]GPL might be pertinent depending on business and social goals.

Another way to create a “thick” API commons is to address material made available via APIs, and metadata about APIs. There, CC tools are likely pertinent, e.g., use CC0 for data and metadata to ensure that “facts are free”, as they ought be in spite of other bad policy turns.

To get even thicker, consider the architecture, for lack of a better term, around API development, services, and material accessed and updated via APIs. Just some keywords: Linked Open Data, P2P, federation, Lots of Copies Keep Stuff Safe, collaborative curation.

The other panelists were Pamela Samuelson, Lawrence Rosen, and Annette Hurst, moderated by David Berlind.

I’m fairly familiar with Samuelson’s and Rosen’s work and don’t have comments on what they said on the panel. If you want to read more, I recommend among Samuelson’s papers The Strange Odyssey of Software Interfaces and Intellectual Property Law which shows that the “API copyright emergency” of the panel title is recurrent and intertwined with patent, providing several decades of the pertinent history up to 2008. Contrary to my expectation in the notes above, Rosen didn’t get a chance to talk about the Open Web Foundation agreements, but you can read his 2010 article Implementing Open Standards in Open Source which covers OWF.

Hurst is a lawyer for Orrick representing Oracle in the Oracle v. Google case, so understandably advocated for API copyright, but in the process made several deeply flawed assertions could have consumed the entire duration of the panel, but Berlind did a good job of keeping the conversation moving forward. Still, I want to mention two high level ones here, my paraphrases and responses:

Without software copyright the software economy would go away. This is refuted by software development not for the purposes of selling licenses (which is the vast majority of it), especially free/open source software development, and services (e.g., API provision, the source of which is often never published, though it ought be, see “going beyond” recommendations above). Yes the software economy would change, with less winner-take-all monopoly and less employment for Intellectual Parasite lawyers. But the software economy would be huge and very competitive. Software is eating the world, remember? One way to make it help rather than pejoratively eat the world is to eject the parasites along for the ride.

Open source can’t work without software copyright. This is refuted by 1) software source sharing before software copyright; 2) preponderance of permissively licensed open source software, in which the terms do not allow suing downstream developers who do not share back; 3) the difficulty of enforcing copyleft licenses which do allow for suing downstream developers who do not share back; 4) the possibility of non-copyright regulation to force sharing of source (indeed I see the charitable understanding of copyleft as prototyping such regulation; for perspective on the Oracle v. Google case from someone with a more purely charitable understanding of copyleft, see Bradley Kuhn); and 5) demand and supply mechanisms for mandating sharing of source (e.g., procurement policies, distribution policies such as Debian’s).

These came up because Hurst seemed to really want the audience to conflate software copyright in general (not at issue in the case, settled in a bad place since the early 1980s) and API copyright specifically. Regarding the latter, another point which could have been made is the extent to which free/open source software has been built around providing alternatives to proprietary software, often API-compatible. If API copyright could prevent compatible implementation without permission of any sort, open source, competition, and innovation would all be severely hampered.

There is a recent site called API Commons, which seems to be an API directory (Programmable Web, which ran the conference, also has one). My general suggestion to both would be to implement and facilitate putting all elements of APIs listed above in my notes in the commons. For example, they could clarify that API metadata they collect is in the public domain, publish it as Linked Open Data, and encourage API developers and providers they catalog to freely license specifications, documentation, implementations, and data, and note such in the directories.

In order to get a flavor for the conference, I listened to yesterday morning’s keynotes, both of which made valiant attempts to connect big picture themes to day to day API development and provision. Allow me to try to make connections back to “API commons”.

Sarah Austin, representing the San Francisco YMCA, pointed out that the conference is near the Tenderloin neighborhood, the poorest in central San Francisco. Austin asked if kids from the Tenderloin would be able to find jobs in the “API economy” or would they be priced out of the area (many tech companies have moved nearby in the last years, Twitter perhaps the best known).

Keith Axline claimed The Universe Is Programmable. We Need an API for Everything, or to some extent, that learning about the universe and how to manipulate it is like programming. Axline’s talk seemed fairly philosophical, but could be made concrete with reference to the Internet of Things, programmable matter, robots, nanobots, software eating the world … much about the world will indeed soon be software (programmable) or obsolete.

Axline’s conclusion was in effect largely about knowledge policy, including mourning energy wasted on IP, and observing that we should figure out public support for science or risk a programmable world dominated by IP. That might be part of it, but keeps the focus on funding, which is just where IP advocates want it — IP is an off-the-balance-sheets, “free” taking. A more direct approach is needed — get the rules of knowledge policy right, put freedom and equality as its top goals, reject freedom infringing regimes, promote commons (but mandating all these as a condition of public and publicly interested funding is a reasonable starting place) — given these objectives and constraints, then argue about market, government, or other failure and funding.

Knowledge policy can’t directly address the Austin’s concerns in the Tenderloin, but it does indirectly affect them, and over the long term tremendously affect them, in the Tenderloin and many other places. As the world accelerates its transition from an industrial to a knowledge dominated economy, will that economy be dominated by monopoly and inequality or freedom and equality? Will the former concentrations continue to abet instances of what Jane Jacobs called “catastrophic money” rushing into ill-prepared neighborhoods, or will the latter tendencies spread the knowledge, wealth, and opportunity?

Without Intellectual Property Day [edit]

Saturday, April 26th, 2014
Without Intellectual Property Day by Parker Higgins of the EFF is quite good, and released under CC-BY. Clearly deserving of adaptation. Mine below, followed by a diff.

April 26 is the day marked each year since 2000 by the World Intellectual Property Organization (WIPO) as “World Intellectual Property Day”, in which WIPO tries to associate its worldwide pushes for more enclosure with creativity.

Celebrating creativity is a good thing, but when you’re a hammer, everything looks like a nail. For the World Intellectual Property Organization, it may seem like creativity and “intellectual property” are inextricably linked. That’s not the case. In the spirit of adding to the conversation, let’s honor all the creativity and industry that is happening without a dependence on a system intellectual property.

There’s an important reason to encourage and promote creativity outside the bounds of increasingly restrictive laws: to the extent such creativity succeeds, it helps us re-imagine the range of desirable policy and reduces the resources available to enclosure industries to lobby for protectionism — in sum shifting what is politically possible. It’s incumbent on all of us who want to encourage creativity to continue to explore and utilize structures that reward creators without also restricting speech.

Comedy, Fashion, Cooking, Magic, and More

In the areas in which intellectual freedom is not typically infringed, there is tremendous innovation and consistent creativity outside of the intellectual property system. Chefs create new dishes, designers imagine new styles, comedians write new jokes, all without a legal enforcement mechanism to restrict others from learning and building on them.

There may be informal systems that discourage copying―the comedy community, to take one example, will call out people who are deemed to be ripping off material―but for the most part these work without expensive litigation, threats of ruinous fines, and the creation of systems of surveillance and censorship.

Contributing to a Creative Commons

The free software movement pioneered the practice of creating digital media that can legally and freely be shared and expanded, building a commons. The digital commons idea is being pushed in more areas than ever before, including culture, education, government, hardware design, and research. There are some projects we’re all familiar with — Wikipedia is perhaps the most prominent, creating an expansive and continuously updated encyclopedia that is freely accessible under permissive terms to the entire world.

Focusing on this year’s World IP Day theme of movies, there have been some impressive contributions the commons over the years. Nina Paley’s feature animation Sita Sings The Blues, which she released into the public domain, has spread widely, inspired more work, and earned her money. The short films from the Blender Foundation have demonstrated cutting-edge computer graphics made with free software and, though they’ve sometimes been on the receiving end of bogus copyright takedowns, have been watched many millions of times.

Kickstarting and Threshold Pledges

Finally, crowdfunding platforms like Kickstarter and Indie-Go-Go have made a major splash in the last few years as another fundraising model that can complement, or even replace, copyright exclusivity. These platforms build on theoretical framework laid out by scholars like John Kelsey and Bruce Schneier in the influential “Street Performer Protocol” paper, which set out to devise an alternative funding system for public domain works. But most crowdfunded works are not in the commons, indicating an need for better coordination of street patrons.

Looking at movies in particular: Kickstarter alone has enabled hundreds of millions of dollars of pledges, hundreds of theatrical releases, and seven Oscar-nominated films (including Inocente, winner of the Best Documentary Short category). Blender Foundation is currently crowdfunding its first feature length film, Gooseberry.

***

The conceit of copyright and other “intellectual property” systems is that they can be calibrated to promote the progress of science and the useful arts. But the reality of these systems is corruption and rent seeking, not calibration. The cost is not just less creativity and innovation, but less freedom and equality.

It’s clear from real world examples that other systems can achieve the goal of promoting creativity, progress, and innovation. We must continue to push for both practice and policy that favors these systems, ultimately rendering “intellectual property” a baffling anachronism. In a good future, a policy-oriented celebration of creativity and innovaion would be called World Intellectual Freedom Day.

wdiff -n eff-wipd.html eff-wipd-edit.html |colordiff |aha -w > eff-wipd-diff.html
[-<p>Today, April 26,-]{+<p>April 26+} is the day marked each year since 2000 [-as "Intellectual Property Day"-] by the <a href="https://www.eff.org/issues/wipo">World Intellectual Property Organization [-(WIPO)</a>. There are many areas where EFF has not historically agreed with WIPO,-] {+(WIPO)</a> as "World Intellectual Property Day", in+} which [-has traditionally pushed-] {+WIPO tries to associate its <a href="https://www.eff.org/deeplinks/2013/03/ustr-secret-copyright-agreements-worldwide">worldwide pushes+} for more [-restrictive agreements and served as a venue for <a href="https://www.eff.org/deeplinks/2013/03/ustr-secret-copyright-agreements-worldwide">domestic policy laundering</a>, but we agree that celebrating-] {+enclosure</a> with creativity.</p>+}
{+<p>Celebrating+} creativity is a good [-thing.</p>-]
[-<p>As the saying goes, though:-] {+thing, but+} when you're a hammer, everything looks like a nail. For the World Intellectual Property Organization, it may seem like creativity and <a href="https://www.eff.org/issues/intellectual-property/the-term">"intellectual property"</a> are inextricably linked. That's not the case. In the spirit of adding to the conversation, [-we'd like to-] {+let's+} honor all the creativity and industry that is happening <i>without</i> a dependence on a system intellectual property.</p>
<p>There's an important reason to encourage {+and promote+} creativity outside the bounds of increasingly restrictive [-laws, too. As Ninth Circuit Chief Justice Alex Kozinski eloquently explained in <a href="http://notabug.com/kozinski/whitedissent">a powerful dissent</a> some 20 years ago, pushing only for more IP restrictions tips a delicate balance against creativity:</p>-]
[-<blockquote><p>Overprotecting intellectual property is as harmful as underprotecting it. Creativity is impossible without a rich public domain. Nothing today, likely nothing since we tamed fire, is genuinely new: Culture, like science and technology, grows by accretion, each new creator building on-] {+laws: to+} the [-works-] {+extent such creativity succeeds, it helps us re-imagine the range+} of [-those who came before. Overprotection stifles the very creative forces it's supposed-] {+desirable policy <i>and</i> reduces the resources available+} to [-nurture.</p></blockquote>-]
[-<p>It's-] {+enclosure industries to lobby for protectionism -- in sum shifting what is politically possible. It's+} incumbent on all of us who want to encourage creativity to continue to explore {+and utilize+} structures that reward creators without also restricting speech.</p>
<h3>Comedy, Fashion, Cooking, Magic, and More</h3>
<p>In the areas [-known as copyright's "negative spaces,"-] {+in which intellectual freedom is not typically infringed,+} there is tremendous innovation and consistent creativity outside of the intellectual property system. Chefs create new dishes, designers imagine new styles, comedians write new jokes, all without a legal enforcement mechanism to restrict others from learning and building on them.</p>
<p>There may be informal systems that discourage copying―the comedy community, to take one example, <a href="http://www.slate.com/articles/arts/culturebox/features/2014/the_humor_code/joke_theft_can_a_comedian_sue_if_someone_steals_his_material.html">will call out people</a> who are deemed to be ripping off material―but for the most part these work without expensive litigation, threats of ruinous fines, and the creation of systems [-that can be abused to silence lawful non-infringing speech.</p>-] {+of surveillance and censorship.</p>+}
<h3>Contributing to a Creative Commons</h3>
<p>The free software movement [-may have popularized-] {+pioneered+} the [-idea-] {+practice+} of creating digital media that can legally and freely be shared and expanded, [-but the free culture movement has pushed the-] {+building a commons. The digital commons+} idea [-further-] {+is being pushed in more areas+} than ever [-before.-] {+before, including culture, education, government, hardware design, and research.+} There are some projects we're all familiar [-with―Wikipedia-] {+with -- Wikipedia+} is perhaps the most prominent, creating an expansive and continuously updated encyclopedia that is freely accessible under permissive terms to the entire world.</p>
<p>Focusing on this year's World IP Day theme of movies, there have been some impressive contributions the commons over the years. Nina Paley's feature animation <i><a href="http://www.sitasingstheblues.com/">Sita Sings The Blues</a></i>, which she released into the public domain, has spread widely, inspired more work, and earned her money. The <a href="http://www.techdirt.com/articles/20101002/20174711259/open-source-animated-movie-shows-what-can-be-done-today.shtml">short films from the Blender Foundation</a> have demonstrated cutting-edge computer graphics made with free software and, though they've sometimes been on <a href="http://www.techdirt.com/articles/20140406/07212626819/sony-youtube-take-down-sintel-blenders-open-source-creative-commons-crowdfunded-masterpiece.shtml">the receiving end of bogus copyright takedowns</a>, have been watched many millions of times.</p>
<h3>Kickstarting and Threshold Pledges</h3>
<p>Finally, crowdfunding platforms like Kickstarter and Indie-Go-Go have made a major splash in the last few years as another fundraising model that can complement, or even replace, [-traditional-] copyright exclusivity. These platforms build on theoretical framework laid out by scholars like John Kelsey and [-EFF board member-] Bruce Schneier in <a href="https://www.schneier.com/paper-street-performer.html">the influential "Street Performer Protocol" paper</a>, which set out to devise an alternative funding system for public [-works.</p>-] {+domain works. But most crowdfunded works are not in the commons, indicating an need for better <a href=http://gondwanaland.com/mlog/2013/08/10/street-patrons-missing-coordination-protocol/">coordination of street patrons</a>.</p>+}
<p>Looking at movies in particular: Kickstarter alone has <a href="https://www.kickstarter.com/blog/a-big-day-for-film">enabled hundreds of millions of dollars of pledges</a>, hundreds of theatrical releases, and seven Oscar-nominated films (including <i>Inocente</i>, winner of the Best Documentary Short category). [-Along with other-] {+Blender Foundation is currently+} crowdfunding [-sites, it has allowed the development of niche projects that might never have been possible under the traditional copyright system.&nbsp;</p>-] {+its first feature length film, <em><a href="http://gooseberry.blender.org/">Gooseberry</a></em>.</p>+}
<h3>***</h3>
[-<p>As the Constitution tells us,-]
{+<p>The conceit of+} copyright and other "intellectual property" systems [-can, when-] {+is that they can be+} calibrated [-correctly,-] {+to+} promote the progress of science and the useful arts. [-We continue to work pushing for a balanced law that would better achieve that end.</p>-]
[-<p>But it's also-] {+But the reality of these systems is corruption and rent seeking, not calibration. The cost is not just less creativity and innovation, but less freedom and <a href="http://gondwanaland.com/mlog/2014/01/30/tech-wealth-ip/">equality</a>.</p>+}
{+<p>It's+} clear from [-these-] real world examples that other systems can achieve [-that-] {+the+} goal [-as well. Promoting-] {+of promoting+} creativity, progress, and [-innovation is an incredibly valuable mission―it's good to know that it doesn't have-] {+innovation. We must continue+} to [-come through systems-] {+push for both practice and <a href="http://gondwanaland.com/mlog/2014/02/09/freedoms-commons/#regulators">policy+} that [-can-] {+favors these systems</a>, ultimately rendering "intellectual property" a baffling anachronism. In a good future, a policy-oriented celebration of creativity and innovaion would+} be [-abused to stifle valuable speech.</p>-] {+called World Intellectual Freedom Day.</p>+}

Sum of all questions

Saturday, April 19th, 2014

I thoroughly enjoyed memesteader Gordon Mohr’s Quora & Wikipedia: Might one ever bail out the other? Futures of ‘Qworum’ or ‘WiQipedia’ which posits two futures in which the sites respectively decline mostly due to internal failure — essentially not adequately dealing with spam and unscrupulous behavior in both cases, though the spam and behavior is different for each.

Both futures seem plausible to me, inclusive of the decline and bail out in each. I also take the medium term absolute decline and death of Quora and steep relative decline of Wikimedia as likely. This relative assessment isn’t a knock on Quora — it and many others waiting in the wings can get big or fail — commons-based projects don’t have much experience in trying to do that (but need to, or find some other way to maintain long-term competitiveness).

Of course “waiting in the wings” is an understatement: I suspect the decline of both Quora and Wikimedia will be less due to internal failure than to being outcompeted by new entrants. Mohr has long been rumored to be working on one, but I imagine there must be many entrepreneurs dreaming of taking a chunk of Wikipedia traffic. I enjoy the Kill Hollywood request for startups, but Kill Wikipedia seems like a more plausible target for VC-term investment. (My preference is to target proprietary monopolies for destruction through competition, replacing them with commons; long ago I even imagined a financially leveraged/risk-seeking approach, but more feasible ones badly needed still.)

Go read and enjoy Mohr’s post, take it at least semi-seriously, and reflect on the future. Doing so makes me pine for something which does not yet exist: combinatorial prediction markets for everything.

I hadn’t looked at Quora in some time. I note that it still requires logging in to read, but has added Google — previously Facebook login (or not) was the only choice. There have been at least semi-serious explorations of a Wikimedia general Q&A sister project, but I’m not sure if any of them are listed in project proposals.