links for 2009-05-21

links for 2009-05-20

Talk: The Future of Science Publishing

Picture 2 I think about the fusion of mobile and Web all the time. And I’ve been talking about and thinking about designing services and software for years. But I also was an academic researcher for about 10 years, with 18 co-authored papers.

All this converged about a year and half ago when I met Matt Cockerill from BioMedCentral, an Open Access publisher of scientific papers.

He had a sort of embarrassment of riches – servers full of papers, videos, info. The problem was how to take all that info and make it work, derive relevance, give value back to the scientists.

That got me thinking. I framed it as a problem – how to make it easy to find-navigate-recombine-share? Suddenly, I saw this as one of the big challenges for the Web.

Now, I see it everywhere in other areas, but science publishing catches my attention, mostly due to my recent focus back on science.

The Rise of the Scientific Paper
Scientific papers arose about 450 years ago as a way to distribute, between scientists, public letters and correspondence on findings and reports. The natural scarcity of publication and distribution made this a necessity.

From this arose lead publishers (for example, Nature and Science) and all that science publishing entails – star editors, reputation, authority, impact factors, and so on.

But that’s so Web 1.0.

Waves of the Web
OK, so I try really hard not to use the Web 1.0, Web 2.0 etc terminology. I view the Web more in waves than labels. Each of these waves take the cycle of create, consume, connect to another level.

For me Wave 1 was the Age of the Hyperlinked Document.  The first wave was characterized by a rush to digitize traditional publishing assets, such as databases, newspapers, encyclopedias. This wave also saw the rise of Web indexes (Yahoo), search (AltaVista), email, and the browser wars. But in the end, the creators were traditional publishers and indexers. Regular folk just “browsed” stuff, without any contribution.

Wave 2 was the Age of the Fragmentation of the Web. This wave saw the coming of micropublishing (blogs, wikis), emergent (crowd-sourced) indexes (wikis, delicious), social networks, and new ways to search (Google, Technorati). And expectations of interactions with people and content was heavily influnced by IM (rapid morsels of conversational text) and rich interfaces (through flash, video, and AJAX). But the biggest change (at least in this story) was that everyone became a publisher

Publishing, therefore, had gone from static monoliths to morsels of info free to socialize. This has caused the collapse of traditional publishing (witness the record and newspaper industries). Furthermore, there has been an explosion of morsels of data on the Web. Everything has become search-able, comment-able, link-able, embed-able, feed-able. Data and people mix in a social, living, Web.

In short, Wave 1 weakened traditional publishing that used to be based on scarcity. Wave 2 made everyone a source of info, everyone an annotator of data, everyone a publisher; it took hyperlinked documents and morselized the web.

How have scientific publishers fared in this Wave 2?
They’ve basically kept the status quo. Online. Stuck in Wave 1.

As with many other traditional publishers, science publishers replicated their closed subscription-based model on the Web, republishing their content online.

Open Access has been battling the status quo for 10 years (at least in terms of access). Only now are they getting strong recognition, impact factors, authority, and a little respect. But they are predicated mostly on and restricted heavily by the traditional model of science publishing (for example, stuck to impact factors).

Recently science publishers have been experimenting with comments and annotations. But with little traction (and I have a few ideas as to why). And, granted, the non-paper publishing part of traditional publishers have embraced the Web, but I am speaking of the core product here.

So many similarities…
The irony is that Tim Berners-Lee actually envisioned the Web as a way to share science information and publications. Openness and sharing are at the heart of science. And the core cultural structures replicate well online. Wave 2 behaviors are the same as in research: find, navigate, recombine, share.

And the Web also has structures found in traditional publishing, such as ways to deal with authority and primacy.

In short, science publishing as it should be mirrors the Web.

If there is a Wave 1 and Wave 2, is there a Wave 3?
My view is that we are entering the true era of (and need for) the Semantic Web. Context is about relevancy is about meaning is about semantics. I claim that the semantic Web has not advanced in the past many years because the focus has been on what I call “librarian” tasks of formatting data and manually building ontologies and so on.

What we know now (from Wave 2 behavior) is that emergent semantics, created through data-mining, but especially via people just using the Web, will be key in helping us navigate the sea of data. In short, the next wave of the Web will require a mix of data mining, librarian tasks, and people to make sense of it all.

How do I see science publishing taking advantage of the Web?
I mapped out behaviors and how it could be on the Web.

 

Traditional_vs_social_publishing

Culture vs tech
Risking sounding dramatic, I think more changes are inevitable, despite publishers wishes to hold on to traditional structures. But the sad irony is that the future of science publishing depends on culture not tech. All the tech is here, and it’s evolving, mixing Web, mobile, context, semantics and other wonder, whether the scientific publishers want to or no.

But will scientists lead the way?

This post was written from the notes of my talk at the 3rd WLE Symposium in London back in March (presentation below).

Changing the journal impact factor through real-time transparent statistics

744px-PageRanks-Example I've mentioned Mendeley before. They refer to themselves as a Last.fm for science papers, but I think it'll be much more.

One thing they realize they are changing, as a side effect, is the impact factor (sort of like a Page Rank of science papers, based on incoming links (citations) to the paper and the journal).

Link: Changing the journal impact factor | Mendeley Blog:

At a higher level then, Mendeley’s significance isn’t just about real-time impact factors and article-level metrics. It’s about using technology for the first time to crowd source data and forever change how research is done. That is why I’m crazy enough to move half-way around the world. Mendeley literally isn’t just another “Silicon Valley” start-up.

Spot on. When I heard Victor (one of the founders) talk about this at Next09 I practically jumped out of my seat.

Thompsons was set up in an age when you needed someone to manually go through references and such and report to the community. That's probably part of the reason it takes three years to establish an impact factor. [I pointed this out already a while back.]

PLoS and BMC, who imported the broken authority model from the print world, missed an opportunity in the past 10 years to upturn Thomspons world. So, it's good to hear that PLoS is starting to be transparent in their traffic and links, providing the start of a new way to look at authority.

One thing: being a bit publisher-minded, I, myself, missed the other side effect of opening up stats that could show authority – basically, such transparency might be able to highlight a high-impact paper from an obscure journal. In the traditional world, that paper would have been buried by the journal's own impact factor.

Yeah, we need to open up these stats on a real-time paper level. There's no reason not to do it.

(and go read the rest of the article on Mendeley's site)

Image from wikipedia, on Page Rank

links for 2009-05-17

links for 2009-05-16

When the Central Dogma is not enough – microbial small RNAs

Centraldogma_nodetails One thing that has always bugged me was a sort of pendatic repetition of what's called the Central Dogma of molecular biology – that DNA goes to RNA goes to protein.

What bothered me was that it way oversimplified the complexity of information transfer and control in organisms. And for me, the excitement has been in all the exceptions to this Dogma.

For example, I had a sort of Rip van Winkle gap between when I left science to when I re-engaged years later (missed the deep stuff, while keeping up lightly with the superficial stuff). Back in 1999 we were talking about some weird things going on in nematodes, where you could control gene expression simply by adding some small RNAs to cells. Fast forward to 2006 and I find out that these small RNAs have been found everywhere as a control mechanism.

Now mix that with the resurgence of microbiology (or at least it looks like a resurgence to me) and folks are starting to use small RNAs as a way to read gene expression patterns in micro-organisms. The idea is that it's a quick readout before the organism starts responding to the effects of collection and removal from its native environment.

"If we think of marine bacteria and their proteins as tiny factories performing essential biogeochemical activities — such as harvesting sunlight to create oxygen and synthesize sugar from carbon dioxide — then the sRNAs are the internal switches that turn on and off the factories' production line. Their discovery in the ocean samples opens the way to learning even more detailed information in the lab: the researchers can now conduct lab experiments to look at the effects of environmental perturbation on microbial communities. These new sRNAs also expand our general knowledge of the nature and diversity of these recently recognized regulatory switches." [apologies to the person I got the link of, as I have forgotten who it was] 

Cool.

Image from wikipedia

Mendeley, the future of science publishing

I was at Next09 this week, giving a talk with Rudy de Waele on "Mobile 2.0". Next09 was a good conference, a mix of talks in German and English, covering things like social media, Web tech, and mobile. And, of course, I caught up with a ton of interesting folks I know, and met a ton of interesting folks I was pleased to meet.

There were some startups also pitching there. I met Jon Froda and Ezra Goldman who were pitching their companies, working in helping corporations capture processes and manage change processes through more social Web services, respectively. I also met Renato Valdes Olmos who had a cool NFC social gizmo.

Bringing citation software to the Living Web
One particular startup I almost missed (Rudy pointed it out to me) was Mendeley. I had heard about it from some of the folks I follow on Twitter (as @molecularist) and had it on my "must check" list, but had no idea what it really was. Rudy summarized it succinctly as "Think of Last.fm for scientific papers".

That was enough to grab my interest, as it seemed to touch on key aspects of the future of scientific publishing that I have been thinking and thinking and thinking and thinking of over the past 18 months or so.

Mendeley is a combination Web and desktop app service for managing your bibliography of science papers. A desktop client helps you extract metadata, annotate, and share the scientific papers you add to your bibliographic database. You can also backup your database online and fill the database ("one-click") with papers from other publication databases.

One of the key upcoming features is a "recommend" feature that helps you find other papers related to the ones you are collecting.

I suppose one could say it has features from delicious (social bookmarking) and StumbleUpon and Last.fm (social discovery), with a twist of sematics and data-mining.

I got a demo of the desktop app and Website and am really impressed. It seemed simple and useful and all the right elements were there.

Kindred spirit
Victor Henning, who is the Founder and Director of Mendeley was kind enough to sit down and talk to me after his talk. He indulged me in my excitement to share thoughts with him regarding what they were doing.

He told me that the idea for Mendeley arose from his and a friend's general frustration in using citation tools that were basically industry standard. Like all great services, Mendeley was something that they built because they needed something like it.

They've been brewing the service for a few years, and have been in a beta for about 4 months. Already they have thousands of users from some of the top research institutions in the world, and are growing at a great clip. Based on the papers placed in the system (over a million), the largest groups of users are from life sciences and computer sciences.

Another cool story is that they reveal a lot of the usage stats, and saw an emergent version of an impact factor as an article from PLoS rose to the top among the most added paper in their database (I think the Web is so well suited to track emergent authority and such).

Good foundation
I shared some thoughts I had about science publishing, and it seemed that some were issues Victor was thinking about. He's quite excited about the service and feels like he could always do more. We touched upon a ton of cool potential and upcoming features. And like always, ideas are more plentiful than one can implement. But, the core is solid, they have a grate foundation that they can build upon, and their position enables them to offer valuable services that folks would pay for.

Furthermore, he mentioned some designers and developers who are working with him and it seems he has an amazingly strong team to make this happen. What's more, some folks from Last.fm, Warner, and Skype have put in 1.5M Euros into Mendeley. So they are going to be moving along for some time still.

And something tells me I'll be cheering for them all along.

links for 2009-05-06