Thursday, 26 August 2010

Open Data: The Panton Discussions

If you are interested in the Open Data (OD) movement but unclear about the issues, or what scientists can do to support the movement, what better way of finding out than by talking to leading OD advocates Peter Murray-Rust of the University of Cambridge and Jordan Hatcher of Open Data Commons.

That was what I did last Tuesday as part of a new initiative called the Panton Discussions. The first in a planned series, the event lasted around two hours and took place in the Panton Arms in Cambridge.

Below is a sample of the kind of questions discussed:

* What is Open Data and why is the OD movement important? What is the problem it aims to fix?

* Amongst the OD tools available there is the Public Domain Dedication and Licence (PDDL), a process of Public Domain Dedication and Certification (PDDC), and Creative Commons Zero (CC0). What are these tools, how do they work, and how do they differ?

* Likewise, there is the Science Commons Protocol for Implementing Open Access Data and the Open Knowledge/Data Definition. How do these differ? Why do we need two similar initiatives?

* More recently we have also seen the introduction of The Panton Principles? What do this initiative provide that was not available before?

* Where does Open Data fit with Open Access (OA)?

* Where does Open Science fit in?

* What about Open Notebook Science (ONS)? Where does OD fit with ONS?

* How should scientists go about making their data open? What pitfalls do they need to avoid?

Help sought

Peter hopes to crowdsource the creation of a transcript of the discussion. Jamaica Jones and Graham Steel have both kindly offered to help, but more volunteers would make the task easier, and quicker. Peter can be contacted by email here.

Thursday, 12 August 2010

Preserving the Scholarly Record: Interview with digital preservation specialist Neil Beagrie

One of the many challenges of our increasingly digital world is that of establishing effective ways of preserving digital information — which is far more fragile than printed material. What are the implications of this for the scholarly record, and where does Open Access (OA) fit into the picture?

In a 1999 report for the Council on Library and Information Resources Jeff Rothenberg , a senior research scientist at the RAND Corporation, pointed out that while we were generating more and more digital content each year no one really knew how to preserve it effectively. If we didn't find a way of doing it soon, he warned, "our increasingly digital heritage is in grave risk of being lost."

In launching the UK Web Archive earlier this year British Library chief executive Dame Lynne Brindley estimated that the Library would only be able to archive about one per cent of the 8.8 million .co.uk domains expected to exist by 2011. The remaining 99 per cent, she said, was in danger of falling into a "digital black hole".

In the context of Rothenberg's earlier warning Brindley's comment might seem to suggest that very little has changed in the past eleven years so far as digital preservation is concerned. But that would be the wrong conclusion to reach. Rather, it draws attention to the fact that digital preservation is not just a technical issue.

As it happens, many of the technical issues associated with digital preservation have now been resolved. In their place, however, a bunch of other issues have emerged — including legal, organisational, social, and financial issues.

What concerns Brindley, for instance, are not the technical issues associated with archiving the Web, but the undesirable barrier that today's copyright laws imposes on anyone trying to do so. Since copyright requires obtaining permission from the owner of every web site before archiving it the task is time consuming, expensive, and quite often impossible.

Clearly there are implications here for the research community.

State of play

So what is the current state of play so far as preserving the scholarly record is concerned?

First we need to distinguish between two different categories of digital information. There is retro-digitised material, which in the research context consists mainly of data created as a result of research libraries digitising their print holdings — journals, books, theses, special collections etc. Then there is born-digital material — which includes ejournals, eBooks and raw data produced during the research process.

It is worth noting that the quantities of raw data generated by Big Science can be mind-boggling. In the case of the Large Hadron Collider, for instance, CERN expects that it will generate 27 terabytes of raw data every day when it is running at full throttle — plus 10 terabytes of "event summary data".

To cater for this deluge CERN has created a bespoke computing grid called the WLCG. While the costs associated with the WLCG will be shared amongst 130 computing centres around the world, the personnel and materials costs to CERN alone reached 100 million Euros in 2008, and CERN's budget for the grid going forward is 14 million Euros per annum.

Of course, these figures by no means represent preservation costs alone, and they are not typical — but they provide some perspective on the kind of challenges the science community faces.

So how is the research community coping with the challenges? With the aim of finding out the Alliance of German Science Organisations recently commissioned a report (which was published in February).

What were the main findings?

So far as retro-digitisation is concerned, the Report points out that funding is limited and "the quantity of non-digitised material is huge". Even so, it adds, there is general concern about "the sustainability of hosting" the data that has been generated from digitisation. This is a particular concern for small and medium-sized institutions.

With regard to born-digital material the Report found that the largest gaps are currently in the "provision for perpetual access for e-journals".

The situation with regard to eBooks and databases is less clear since, as the Report points out, "experience in digital preservation with these content types is currently more limited."

While the Report focused on the situation in Germany the international nature of today's research environment suggests the situation will be similar in all developed nations (Although Germany does have two unique mass digitisation centres).

We should not be surprised that the German Report found the largest gap to be in the preservation of journal content. As we shall see, the migration from a print to a digital environment has disrupted traditional practices and responsibilities, and led to some uncertainty about who is ultimately responsible for preserving the scholarly record.

We should also point out that one important area that the German Report did not look at is the growing trend for scholars to make use of blogs, wikis, open notebooks and other Web 2.0 applications. Should this data not be preserved? If it should, whose responsibility is it to do it, and what peculiar challenges does it raise? As we have seen, for instance, preserving web content is not a technical issue alone. Amongst other things there are copyright issues. (Although as the research community starts to use more liberal copyright licences these difficulties should ease somewhat).

Another recently published report did look at the issue of web-created scholarly content, but reached no firm conclusion. Produced by the Blue Ribbon Task Force, this Report concluded: "[I]n scholarly discourse there is a clear community consensus about the value of e-journals over time. There is much less clarity about the long-term value of emerging forms of scholarly communication such as blogs, products of collaborative workspaces, digital lab books, and grey literature (at least in those fields that do not use preprints). Demand may be hypothesised — social networking sites should be preserved for future generations — but that does not tell us what to do or why."

Open Access

One issue likely to be of interest to OA advocates is whether institutional repositories should be expected to play a part in preserving research output.

Evidence cited by the German Report suggests that repositories are not generally viewed as preservation tools. It pointed out, for instance, that the Dutch National Library's KB e-Depot currently archives the content hosted in 13 institutional repositories in the Netherlands.

The Blue Ribbon Report, by contrast, appears to believe that repositories do have a long-term archiving role. It suggests, for instance, that self-archiving mandates should always be accompanied by a "preservation mandate".

The Report goes on to suggest that the inevitable additional costs associated with repository preservation should be taken out of the institution's Gold OA fund (where such a fund exists).

##

If you wish to read the rest of this introduction, and the interview with preservation specialist Neil Beagrie, please click on the link below. I am publishing it under a Creative Commons licence, so you are free to copy and distribute it as you wish, so long as you credit me as the author, do not alter or transform the text, and do not use it for any commercial purpose.

If you would like to republish the interview on a commercial basis, or have any comments on it, please email me at richard.poynder@btinternet.com.

To read the rest of the introduction and the interview with Neal Beagrie (as a PDF file) click here.

Monday, 2 August 2010

University of Ottawa Press launches OA book initiative

Last week the University of Ottawa Press (UOP) announced a new open access (OA) book initiative. This, it says, will provide "free and unrestricted access to scholarly research". But what does it mean in practice? And what issues arise?

UOP's new initiative is part of a wider open access strategy first unveiled last December. Initially it will consist of making 36 French-language and English-language in-print titles in the arts, humanities and social sciences freely available online via the University of Ottawa's institutional repository (IR), uO Research.

The UOP news is of interest for a couple of reasons.

First, until relatively recently open access was seen as an issue of relevance only to scholarly journals, not books, and for the sciences rather than the humanities.

It is only in the last few years, for instance, that new OA publishers like Bloomsbury Academic, Open Humanities Press (OHP), and re.press have appeared on the scene; and only recently that traditional publishers and university presses have started to introduce OA book initiatives — e.g. The University of Michigan Press' digitalculturebooks project and Penn State University Press' Romance Studies.

Second, unlike Bloomsbury Academic, OHP, re.press, and the University of Michigan, UOP has not released its OA books under creative commons licences, but simply placed the text in a PDF file with the original "all rights reserved" notice still attached to it. (E.g. in this 24 MB file).

UOP's move suggests that traditional presses can no longer afford to ignore the rising OA tide — despite the fact that there is still no tried and trusted business model for OA books. It also demonstrates that there is as yet no consensus on how best to go about it, or what to do about copyright.

The latter issue could prove a source of some confusion for readers of UOP's books.

Libre/Gratis

For instance, anyone who read UOP's announcement that it is providing its books on a "free and unrestricted access" basis who then downloaded one of the books would surely scratch their head when they saw the all rights reserved notice attached to it.

While they could be confident that they were free to read the book, they might wonder whether they were permitted to forward it to a colleague. They might also wonder whether they were free to print it, whether they could cut and paste text from it, or whether they were permitted to create derivative versions.

Free and unrestricted access would seem to imply they could do all those things. All rights reserved suggests quite the opposite — indeed, a copyright lawyer might argue that even downloading a book infringes an all-rights licence.

It does not help that there appears to be no terms and conditions notice on the UOP web site clarifying what readers can and cannot do with the books — as there is, for instance, on PSU's Romance Studies site.

In fact, UOP is only granting permission for people to read, download and print the books.

But it need not be that confusing. OA comes in different flavours, and what UOP is offering is what OA advocates call Gratis OA (that is, it has removed the price barriers); it is not offering Libre OA (which would require removing permission barriers too — i.e. relaxing the copyright restrictions).

Gratis OA is a perfectly legitimate way of providing OA, so long as you make it clear that that is what you are offering. Some, however, might argue that there is a contradiction between what UOP says it is offering and the true state of affairs — that the publisher is claiming to offer something that it is not.

"While there's nothing deceptive in using the term 'OA' for work that is Gratis OA, there is something deceptive in using language suggesting Libre OA for work that is Gratis OA," the de facto leader of the OA movement Peter Suber commented when I asked for his views. Stressing that he has not yet looked at the details of the UOP initiative, Suber added: "The phrase 'unrestricted access' suggests Libre OA."

There is no reason to doubt UOP's motives: It believes that using the term free and unrestricted is accurate given that the OA books do not come with DRM, and "any user with a computer can access the books, download them and read them freely".

Nevertheless, it does seem to be sending out a confusing message. And when putting content online publishers should really aim to be as precise as possible in the terms they use, and the claims they make — particularly in light of the many copyright controversies that have arisen in connection with digital content. UOP has surely failed to do this.

Suber would perhaps agree. "I realise that most people aren't familiar with the Gratis/Libre distinction", he emailed me. "But at the same time, people who do understand the distinction should use it, and could help everyone by describing the Ottawa position accurately. If it's Gratis and not Libre (which I haven't had time to check), then it should be described as Gratis."

We might however add that some OA advocates believe Gratis OA to be an inadequate way of making research available online. And it is noteworthy here that in his definition of OA, Suber assumes Libre OA to be the default. Open access literature, he states, is "digital, online, free of charge, and free of most copyright and licensing restrictions."

Some would doubtless claim that worrying about such matters is a non-issue. After all, they might say, aside from reading it, what more could you possible want to do with a book? So why does it matter whether you make it available online as Gratis OA or Libre OA?

But a few years ago exactly this issue led to some heated debates in connection with making scholarly papers OA, with many insisting that worrying about such matters was a complete irrelevancy — until Peter Murray-Rust pointed out that in a Web 2.0 environment there are very good reasons for providing re-use rights to scholarly work.

Indeed, it was as a result of that long-running debate that the movement eventually hammered out the Gratis/Libre distinction.

It is, of course early days for book publishers, who are still in experimental mode vis-à-vis OA. But they would surely benefit from reviewing some of the debates that have taken place in connection with providing OA to refereed papers.

In order to get UOP's views on these matters I contacted the publisher's eBook Coordinator Rebecca Ross, who kindly agreed to an email interview. Below are her answers.

The good news is that UOP does hope to adopt creative commons licences in the future!

clip_image002[4]

Rebecca Ross, UOP eBook Coordinator

RP: I understand that UOP has made 36 of its books available on an OA basis and these can be accessed via the University's institutional repository. Is there a list of these books you can point me to?

RR: You can browse the books by title here.

RP: Where can people obtain more information about UOP, and its activities?

RR: Unfortunately the UOP website is in a state of transition with a new website launching very soon. To give you a bit of background about UOP, we are Canada's oldest French-language university press and the only fully bilingual (English-French) university press in North America.

RP: How many books does UOP publish each year, and what kinds of books does it publish?

RR: UOP was founded in 1936 and has published over 800 titles. We currently publish 25-30 books annually in four main subject areas: social and cultural studies, translation and interpretation, literature and the arts, and political and international affairs.

RP: How did you choose which books to make OA?

RR: The books were chosen based on input from Michael O'Hearn (UOP Director) Eric Nelson (Acquisitions Editor), Marie Clausén (Managing Editor), Jessica Clark (Marketing Manager) and myself as a collaborative process to determine a collection of books that are diverse in terms of language, date published, and subject matter.

This will help UOP best determine the books that work effectively as open access. For example, we want to test questions like: does an 800 page collected work about social policy work better as OA than a monograph about Canadian literature?

We also wanted to test if an electronic open access version gives a second life to the print edition or generates interest in a second edition. In this sense, open access is also a marketing tool for us to reach a wider audience than traditional marketing.

In our decision process we also made sure to include books whose authors would be amiable to licensing their work open access (we have several authors who are very excited by having their work as open access!), and to include topics that are relevant, timely and even timeless (for example a reappraisal of Stephen Leacock's work).

Free and unrestricted access

RP: The UOP press release says that the books are being made available on a "free and unrestricted access" basis. What does that mean?

RR: All of the books included in the open access collection are protected by copyright. UOP does not support DRM or restrictive access to our eBooks, whether they are part of the open access collection or for sale.

RP: The books are not being made available under Creative Commons licences are they?

RR: The books are not under Creative Commons. The authors granted a non-exclusive distribution license to the Press for providing access via uO Research.

RP: Many OA advocates might argue that OA implies using creative commons licensing. You don't agree?

RR: Where possible, UOP is very interested in moving forward with creative commons licensing. We're learning from our colleagues at Athabasca University Press, who publish, where possible, using a Creative Commons license: (Attribution-Noncommercial-No Derivative Works 2.5 Canada).

The decision to use the current licensing model was made to best align UOP with the University of Ottawa Library and the University’s institutional repository uO Research.

RP: I do not think it says anywhere on your site exactly what users can do with the books. Anyone downloading the files will see a traditional "all rights reserved" notice attached. The UOP announcement, however, says that the works are available on a free and unrestricted access basis. Readers might therefore wonder what exactly they are permitted to do with the text — whether, for instance, they can print them out, whether they can freely copy and distribute them, whether they can cut and paste text from them, and whether they can they create derivative versions. What exactly can they do?

RR: So far as the open access collection is concerned users can read them, download them and print them.

Without DRM we are unable to control what exactly users do with the books, but as I said, they are protected by copyright.

In the end, we are pleased that users are accessing our content and our authors are pleased that their research is reaching a wider audience.

RP: Would you agree that you are offering the books Gratis OA rather than Libre OA? That is, you have removed the price barriers, but not the permission barriers?

RR: Describing UOP's open access collection (as it is now) as Gratis OA rather than Libre OA is both fair and accurate. We've removed the price barrier as a first step; the next step will be working with our authors and editors to remove the permission barriers.

RP: Do you not think that saying the books are being offered on a free and unrestricted access basis might be a slight overstatement. Does not "unrestricted access" imply the removal of both price and permission barriers?

RR: When compared to print books offered at sometimes very high and restrictive prices and made available only in certain parts of the world, I don't think the description of "free and unrestricted access" is an overstatement.

UOP's open access books are free, and their access is unrestricted, any user with a computer can access the books, download them and read them freely. At this stage UOP open access books are protected by copyright: this is partly for us and partly for our authors.

Once UOP's open access program has been fully defined and the level of support we will receive from our host institution is determined, we will be in a better position to remove permission barriers.

In conceptualising UOP's open access program the first objective was to provide a wider reach for our authors and books. Most of our authors write, not to make a living, but to further scholarship and research in their fields; allowing their work to be distributed for free is an excellent way to do so.

As I said, the next step for UOP's open access program will be working in collaboration with our authors and the University of Ottawa Library to remove the remaining permission barriers. We are looking into Creative Commons and defining what it means to offer UOP books as open access.

Right now we are very excited to be involved with open access and looking forward to the next steps of the project.

RP: Would you say you were offering the books as Green OA or Gold OA, or do such distinctions only make sense in the context of journals?

RR: As it stands right now, these distinctions seem appropriate only in the context of journals.

If I had to make the distinction I would say we fall into Green OA because of our participation in uO Research the University of Ottawa's institutional repository.

When preparing and researching for the open access program we found that much of the literature is about open access for journals and many university presses, both in Canada and the United States, are just starting to think about how open access can work for books.

Still at an early stage

RP: Does UOP believe that OA is an inevitable development for scholarly monographs?

RR: The University of Ottawa announced its open access program in late 2009. This includes support to UOP in publishing a collection of OA books. Although there is much research surrounding open access in academic journals, open access book publishing is still at an early stage.

UOP launched this open access collection to determine the effects of open access on our publishing program, to eventually determine what kind of support we require to become an open access press.

It would be difficult to say with certainty that OA in now an inevitable way for scholarly monographs to be published in the future but it does appear that way and UOP is interested in testing and researching this notion.

At this stage, it is UOP's assumption that open access will only suit certain books, for example we are not including any textbooks in the open access collection. However, this assumption is based on previously published books and going forward open access will be an important aspect of UOP's acquisition procedure and publishing program.

RP: When I interviewed Northwestern University Dean Sarah Pritchard about Northwestern University Press earlier this year I suggested that the model many advocate see for OA books is that of making the text freely available online but selling the print version. Pritchard replied that she saw that as a very logical model, and one that she envisages NUP adopting before it moves to a totally OA environment. Is that your view too?

RR: Absolutely. We are in the business of publishing books both in print and electronically. At the moment we are borrowing models and ideas from many of our colleagues within Canada and the US, including Athabasca University Press and the International Development Research Centre.

Sarah Pritchard brings forward many important issues that are relevant to us at UOP. I do believe that electronic versions of print books will drive print sales — that assumption is the backbone of our open access program.

The wider the distribution an author or a publisher has, the better the chance for course adoptions, sales and even translation rights. The model UOP has adopted is a hybrid model: we are doing a bit of everything right now, and we will continue experimenting to see what fits best.

RP: Does UOP pay its way today, or is it subsidised by the University? Can you see OA affecting the current state of affairs?

RR: UOP is subsidised by the University. Our publications are too specialised to make the best-seller list; alas, we will never become a cash cow for our home institution! This is a bit of an experiment.

We don't know if OA will have a negative effect on the sales of the print version or if it will encourage people to buy the print version, especially in the case of single-authored volumes containing long and complex arguments — books like these are likely easier to read in the traditional paper format than on a computer screen.

In either case the level of support the University provides its Press will change accordingly.

Friday, 2 July 2010

Creature from the Black Lagoon


Was having some fun yesterday just sketching and this is what came of it. I'll ink it up this weekend and then color it next week. Enjoy.

Monday, 28 June 2010

Free our data: For democracy's sake

The open data movement is growing apace. What better demonstration of this than news that the UK coalition government is making its Combined Online Information System (COINS) freely available on the Internet, inviting people not only to access the data but to re-use it too?

COINS, The Guardian newspaper points out, is one of the world's biggest government databases and provides, "the most detailed record of public spending imaginable. Some 24m individual spending items in a CSV file of 120GB presents a unique picture of how the [UK] government does its business."

For The Guardian the release of COINS marks a high point in a crusade it began in March 2006, when it published an article called "Give us back our crown jewels" and launched the Free Our Data campaign. Much has happened since. "What would have been unbelievable a few years ago is now commonplace," The Guardian boasted when reporting on the release of COINS.

Why did The Guardian start the Free Our Data campaign? Because it wanted to draw attention to the fact that governments and government agencies have been using taxpayers' money to create vast databases containing highly valuable information, and yet have made very little of this information publicly available.

Where the data has been made available access to it has generally been charged for. Moreover, it has usually been released under restrictive copyright licences prohibiting redistribution, and so preventing third parties from using it to create useful new services.

The end result, The Guardian believes, is that the number and variety of organisations able to make use of the data has been severely curtailed and innovation stifled. As the paper explains, "Making that data available for use for free — rather as commercial companies such as Amazon and Google do with their catalog and maps data — would vastly expand the range of services available."

And it is this argument that has become the rallying cry of the burgeoning open data movement.

But is it true? Is there evidence to demonstrate that by keeping publicly-funded data under wraps governments have been stifling innovation? And does the free availability of government data inevitably lead to a flowering of new information products and services?

The economic argument

In the hope of answering these questions Italian open data advocate Marco Fioretti will shortly be undertaking a research project in conjunction with the Laboratory of Economics and Management (LEM) at the Sant'Anna School of Advanced Studies in Pisa.

The project will form part of an EU-funded work package entitled "Open Data for an Open Society, itself part of a larger initiative called Dynamics of Institutions and Markets in Europe (DIME). DIME is sponsored by the 6th Framework Programme of the European Union.

Explaining the background to Fioretti's project the manager of DIME's open data package, and Associate Professor of Economics at the Sant'Anna School, Dr Giulio Bottazzi says: "We ran an open call for the assignment of part of the activity, specifically the design and realisation of a survey to help assess the present situation concerning the 'openness' of data management in public institutions. Dr Fioretti applied for and obtained the contract — essentially because of his CV and his publication record in the area of open source software and open document standards."

Importantly, Fioretti has a special interest in how better use can be made of public sector information (PSI). And he believes that simply making public data freely available is insufficient. It also needs to be made available using open standards and open licenses. In other words, making digital information freely available is only half the task. Unless it is released in non-proprietary formats, and licensed in a way that allows adaptation and re-use, its usefulness is significantly curtailed.

Fioretti plans to restrict his research to local government data, since he believes it is more likely to be consistent. As he explained recently to O'Reilly Media editor Andy Oram, "the structure of government projects and costs are more similar from one city to another — even across national EU borders — than from one national government to another."

The project will consist of three main phases. First, Fioretti will produce a report discussing the role of fully accessible and reusable digital raw data in an open society. This will be based on examples taken both from the European Union and the rest of the world.

Second, during the summer he will conduct an online survey — which will be hosted on the LEM website. The survey will aim to establish how many EU municipalities and regions are already making their raw data and procedures available by means of open standards and open licenses.

Finally, he will write a further report analysing the results of the survey, and providing some guidelines and best practices for improving full access to digital data.

The grant made available to Fioretti is a modest one — around €12,000 ($14,767), plus an additional €6,000 discretionary fund — but it would seem to be further evidence that the open data movement is gaining mindshare.

In the meantime Fioretti is very keen to hear about real-world examples involving local businesses — both those who have built a successful business (anywhere in the world) by exploiting the free availability of local government data, and those (in Europe) who have struggled to create a viable business as a result of the inaccessibility of this data.

Fioretti can be contacted on mfioretti@nexaima.net.

As noted, Fioretti's research will test the thesis of The Guardian's Free Our Data campaign — i.e. if governments make their data freely available will it encourage businesses to develop new value-added products? And will this in turn spur innovation and create new jobs? Essentially it is an economic argument.

Transparency

But we should not overlook the fact that there are other reasons for governments to open their databases to citizens. Indeed, some might argue that the economic argument should be seen as of secondary importance only. A far more compelling reason for "freeing our data", they might add, is that it increases government transparency, and so is an inherently democratic step to take.

To do him justice, while he appears to have prioritised the economic case in his project, Fioretti is not blind to the transparency argument. As he points out on his web site, in addition to providing access to valuable government-created information and encouraging innovation, open data makes it easier for citizens to monitor their government's activities and how their tax dollars are being spent. "Modern software technologies and data networks make it both possible and relatively inexpensive to publish online tenders, regulations, documents, procedures and many other raw public data, from digital maps to pollution measurements," he says. "Making this information really accessible, that is online, in open formats and under open licenses, can both improve transparency in government and foster local economical and cultural activities."

In putting the transparency case to O'Reilly's Andy Oram, Fioretti cited the planned construction of the Strait of Messina Bridge (at an estimated cost of €6.1 billion). When the government announces how many tax dollars it plans to spend on a project like this, Fioretti asked, how can the public know that the costs are reasonable if it does not have access to all the data?

Evidently in agreement with Fioretti, Oram expanded on the argument: "[C]ontracts must be very specific about the delivery of data that [a government] commissions — and not just the data, but the formulas and software used to calculate results. For instance, if a spreadsheet was used in calculating the cost of a project, the government should release the spreadsheet data and formulas to the public in an open format so that experts can check the calculations."

Certainly it was in the interests of transparency that the UK government released COINS. As the H M Treasury web site puts it, "The coalition agreement made clear that this Government believes in removing the cloak of secrecy from government and throwing open the doors of public bodies, enabling the public to hold politicians and public bodies to account. Nowhere is this truer than in being transparent about the way in which the Government spends your money. The release of COINS data is just the first step in the Government's commitment to data transparency on Government spending."

Coming in the wake of the so-called "MPs' expenses scandal" in Britain, the UK government's decision is far from surprising. What that scandal clearly demonstrated was that Freedom of Information (FoI) legislation alone cannot provide adequate transparency about the way in which governments spend taxpayers' money.

Indeed, it was only after several legal challenges that the UK government eventually conceded the need to be more transparent about MPs' expenses in any case. Even then, it continued to prevaricate and it was only after an insider leaked data to the Daily Telegraph that the public become fully apprised of how irresponsibly British politicians had been spending taxpayer's money for their own personal gain — including using public money to help them fund unnecessary second homes, to buy toilet seats, duck houses etc. etc.

This suggests that it would be a serious mistake to limit the case for open data to the economic argument alone. Governments also need to open their databases so that the public can know exactly how its money is being spent, and exactly how well the government is managing the country on its behalf.

Participation

But there is a third — and arguably even more important — reason why governments should open their databases: Doing so will help citizens participate more directly in the government of their country, and at an important moment in history.

As the publisher Tim O'Reilly points out, if governments were to exploit the digital network effectively they could engage the public in the political process in ways never previously possible in a modern society. Citizens could, for instance, "crowdsource" social, political and economic problems, take over responsibility for tasks traditionally seen as the prerogative of the state (by means of what O'Reilly calls "citizen self-organisation"), and even play a direct role in political decision-making.

This would be a radical change. The traditional top-down model of government assumes that public participation in the political process is necessarily circumscribed, since in large analogue societies collective decision making does not scale. This has seen citizens consigned to simply casting their votes and then sitting back while politicians govern in their name. One consequence of this has been the emergence of a professional political class that now views itself as standing over and above citizens, who tend to be treated as inferior, or simply as children, rather than fellow citizens. Essentially the electorate has been infantilised.

Yet as O'Reilly points out, "government is, at bottom, a mechanism for collective action." And in the age of the Web it is possible to make the political process more horizontal and egalitarian. Rather than viewing government as a hierarchical system in which only elected officials (with the help of professional civil servants) come up with solutions, make decisions, and set the rules, government should now be viewed as a platform to facilitate collective action — much as the Web itself has become a platform across which multiple and diverse applications can run without fear or favour (as advocates of net neutrality like to put it, "all bits are equal").

For this reason, says O'Reilly, the key question for governments today should be, "How do you design a system in which all of the outcomes aren't specified beforehand, but instead evolve through interactions between government and its citizens, as a service provider enabling its user community?"

Clearly this would not be possible unless everyone had access to all the relevant data, much as those building web-based services need to know the protocols of the Web in order to participate in the network economy. This suggests that open data should not be viewed as the endgame, but a necessary precondition for something far more radical.

The necessary transition, however, is unlikely to occur naturally: just as the Web is only open because its designers created an open model, so creating an open platform for democratic governance would need to be a deliberate decision. And evidence suggests that politicians will resist greater openness and transparency, since it would require them to give up some of their traditional power and authority.

In this regard perhaps open government advocates have been somewhat naïve. When in 2008 Obama won the US presidential election, for instance, they assumed the battle had been won. Arguing that Obama's success was because his campaign had adopted the open source approach to electioneering pioneered by Howard Dean in 2004 (under the guiding hand of Joe Trippi) they anticipated that Obama introduce the same openness in the White House, and so revolutionise the way in which presidents govern in the US.

In fact, Obama appeared to promise as much. As he put it during the election campaign: "We must use all available technologies and methods to open up the federal government, creating a new level of transparency to change the way business is conducted in Washington, and giving Americans the chance to participate in government deliberations and decision making in ways that were not possible only a few years ago."

With that end in mind, Obama promised to post pending legislation online for comments. And on entering the White House he put his weekly addresses up on YouTube, oversaw the creation of a White House blog, and the launch of several federal web sites — including Data.gov — where the public could access "high value, machine readable datasets generated by the Executive Branch of the Federal Government".

Clearly open data was viewed as a given in a Government 2.0 environment.

Caught up in the excitement, Newsweek called it Government 2.0: "Instead of a one-way system in which government hands down laws and provides services to citizens, why not use the Internet to let citizens, corporations and civil organisation work together with elected officials to develop solutions?" it asked.

Within weeks of Obama's arrival in the White House, however, the promised transparency began to give way to traditional top-down government, secrecy, and back-room deals.

Earlier this year, for instance, Obama was accused of reneging on a promise to make his health care negotiations freely available on C-SPAN (The Cable-Satellite Public Affairs Network, also freely available on the Internet).

And a year after the launch of Data.gov the site was being derided for offering little more than "thoughtless data dumps".

In reality, concluded cynics, Obama had simply paid lip service to openness and transparency in order to gain power.

However, it is surely more complicated than that. As O'Reilly points out, "we can be misled by the notion of participation to think that it's limited to having government decision-makers 'get input' from citizens … It's a trap for outsiders to think that Government 2.0 is a way to use new technology to amplify the voices of citizens to influence those in power, and by insiders as a way to harness and channel those voices to advance their causes."

In the meantime, the gap between promise and delivery in the Obama administration appears to be widening. Earlier this month, for instance, the New York Times ran an article suggesting that the Obama administration is actually proving more, not less, secretive than previous ones — and cracking down on government leaks in an unprecedented manner. A recent case [in which an intelligence bureaucrat faces 10 felony charges for handing over classified documents to a blogger], the paper suggested, "epitomizes the politically charged debate over secrecy and democracy."

It added, "In 17 months in office, President Obama has already outdone every previous president in pursuing leak prosecutions. His administration has taken actions that might have provoked sharp political criticism for his predecessor, George W. Bush."

Dangerous times

What has gone wrong? In pondering on this question earlier this year the online magazine Slate listed a number of possible reasons — including the White House's use of poor technology, an inability to let go of the traditional top-down approach and, citing an article by Micah Sifry — the co-founder of the annual tech/politics conference Personal Democracy Forum — the possibility that much of the hype surrounding the Obama campaign's web savvy was hype and nothing else.

The disappointment of open government advocates was palpable at the 2010 Personal Democracy Forum earlier this month. Indeed, some have become sufficiently disenchanted that they have come to conclude that, rather than facilitating greater democracy, the Web is acting against it. For this reason, suggested co-founder of the Electronic Frontier Foundation (EFF) John Perry Barlow, those disappointed in Obama should blame the Internet, not the president.

"The political system has partly broken because of the Internet," he argued. "It's made it impossible to govern anything the size of the nation-state". As a consequence, he predicted, "We're going back to the city-state. The nation-state is ungovernably information-rich."

Those not steeped in the culture of American libertarianism will doubtless view Barlow's conclusion as little more than the dystopian fantasy of a cyber-guru who had expected so much more of the Internet. Certainly it runs counter to everything we have learned about the ability of the Web to cope with scale.

So what do we conclude? Is open data really only ever about job creation and innovation? Should we give up on the democratic potential of the Internet? Can we afford to?

Undoubtedly it is possible that governments could release more of their data without making a significant change to the traditional political system. Certainly we cannot rely on them to act in the best interests of their electorate; and so we cannot rely on them to become more transparent, or at least not without a great deal of pressure — much as British politicians had to be forced to change the way in which they managed their expenses system.

But as public cynicism grows the need for citizens to be brought into the political process in a more meaningful way could become pressing. Today's growing disenchanted with governments and politicians is a dangerous development — and comes at a critical historical moment.

For as governments around the world are forced to cut spending and raise taxes in response to the global financial crisis, citizens could become sufficiently alienated that political unrest begins to destabilise democratic governments.

Already we have seen pressure placed on the political stability of Greece, signs of incipient civil unrest in a number of other European countries (e.g. Germany, Spain, Portugal, Italy, and doubtless soon in the UK in the wake of the new government's emergency budget). And the now routine protests at international forums like G-20 are a further reminder of how more and more citizens are becoming alienated from the political process.

The threat that the financial crisis poses to European democracies was highlighted recently by the President of the European Commission José Manuel Durão Barroso — who warned that crisis-hit countries in southern Europe could fall victim to military coups or popular uprisings as interest rates soar and public services collapse because their governments run out of money.

In short, these are dangerous times. So we cannot afford to retreat into Barlowesque libertarian fantasies. Rather we should be pointing out that opening up government databases is just the first necessary step for engaging citizens more effectively in the political decision-making process. After all, when people are involved in decision making they are more likely to buy into the proposed solutions — and that surely is the raison d'être for collective decision making.

Moreover, any solution emerging from citizens themselves is more likely to be a robust one. As Trippi points out in his book The Revolution will not be Televised (describing the Dean Campaign) some of the best ideas adopted by Dean came not from campaign insiders, but from ordinary citizens. "It became pretty obvious quickly that a couple of dozen sleep-deprived political junkies couldn't possibly match the brainpower and resourcefulness of six hundred thousand Americans," explains Trippi. "We couldn't see every hole and every flaw that they could see."

Likewise, describing Dean's radical decision to post details of the campaign's fund-raising on the Web, Trippi explains that the logic was to, "[I]nvite people in and open up the books. Give them the knowledge and information — how much money we wanted to raise — and they'd take the responsibility for it." And, he adds, they did take responsibility — as people do when you stop infantilising them.

Yes, governments should make their databases freely available to citizens. Yes, new businesses and services will doubtless be created as a result. Yes, that will surely benefit society at large. But open data should be viewed as merely a preliminary to a more far-reaching change in the way democracies are governed. The more important task is to re-engage citizens in the political process, and empower them to take part in collective decision making. That is not possible without open data, but open data alone will not make it happen.

In short, open data is a good cause to advocate, and Fioretti's initiative is both important and timely. But we need to broaden the discussion. It is not a topic that should be viewed exclusively through the lens of economics and entrepreneurship.

As such, the message to governments should be "Free our data: For democracy's sake".

Sunday, 23 May 2010

Is nano-technology something to get very excited about with regard to it's use in medicine?


Answer:
Good question. Check out :
http://www.zyvex.com/nanotech/nanotechan...
As long as it's not used to Borgify us all!
Well, to me everyone is excited about nanotech, but it really has some interesting application. However, I'm not sure the potential in Medcine.

is no side effect if i take multi vitamins at the same time i'm taking pills?

i'm taking pills but at the same time i want to take multi vitamins, is there any side effect for this?
Answer:
There might be an interaction. Certain medicines and certain supplements can interfere with each other. Tell your doctor or pharmacist what you are taking.
It depends on the actual pills and the actual vitamins. Ask your pharmacist and read the prescription info you were given!
it would really depend on what kind of pills i cant think of any but talk to your doctor or a pharmacists he will he able to help you out.
you can take both at the same time., there are no side effects for this.But dont take the vitamin pills continuously; take the vitamin pills for 15 days then stop taking the vitamin pills for 15 days , go on like this.
water sluble vitamins as ( C and B complex ) have no side effect
but others do