Friday, April 15, 2011
Sunday, April 10, 2011
Academic e-resources in the UK: promoting discovery and use
Vic Lyte and Sophia Jones from Mimas, The University of Manchester, presented on The UK Institutional Repository Search (IRS), which is a Mimas project commissioned by JISC in partnership with UKOLN and SHERPA. The project was completed in July 2009 and the service has been running continuously since then.
Content stored in institutional and academic repositories is growing and they recognise that there are limited ways to access this information. This project has taken cross-search and aggregation to the next level, creating a visionary platform that pulls together disparate content, making it easier to search and discover in ways that meet personal or contextual needs.
They demonstrated how the search works including an impressive 3D visulisation option.They gave an overview of the JISC Historic Books and JISC Journal Archives products. They also talked about the JISC Collections e-platform enabling cross-agregated search of a unique resource from the British Library. Content (300,000 books) previously inaccessible will be searchable on the platform. Features include three types of search (exact; detailed and serendipitous) and tabbed filters; Google-style listings and search clouds. It all looked very impressive.
Labels: breakout session 18
Driving usage - what are publishers and librarians doing to evaluate and promote usage?
Sarah Pearson from the University of Birmingham kicked off this breakout session and outlined her experience of collection development analysis at her institution. She went on to explain that while they have been doing this for some time, usage alone doesn't tell the whole story. They have been looking increasingly at how users get access to content and what path they take.
Sarah highlighted the numerous ways they promote usage at her university. These include news feeds about new acquisitions and trials; making content available in resource discovery interfaces; activating in link resolvers (SFX); integrating with Google Scholar/A&I services; making authentication as seamless as possible and embedding in apps on other sites.
There is a Mylibrary tab on the institutions' portal page and a library news section, which are widely used. Users can search the library catalogue direct from the portal page of the university rather than go to the library pages.
They are also about to user Super Search on Primo Central, which will be embedded in the virtual learning environment and Facebook.
To analyse usage they use a number of services including in-house templates that compare and contrast big deal usage with subscription analysis; JUSP (Jisc) and SCONUL Returns. They look at JR1 reports and evaluate cost per use. They pay particular attention to those resources with low or zero use. They also look at DB1 searches & sessions and compare archive with frontfile usage.
With budgets under threat librarians are looking at cancelling poorly performing content and big deals, for example, have to demonstrate overall good value.
The University of Birmingham approach, Sarah explained, is to activate online access content everywhere and let the user decide.
Google Analytics is being used to look at user behaviour now and to help understand more about their journey to access content. They know that the Institution's portal page is the number one access point but the OPAC and Google are still high referrer sites. There is a low number for access via mobile devices but they expect that to increase.
Evaluating usage is still very manual and it is labour intensive to measure the ROI of resources. It is important with increased pressure on budgets to ensure librarians are making the right decisions about which content to subscribe to and purchase. Evaluating usage is an important step in doing this.
Christian Box from IOP Publishing followed on with an interesting presentation about the work they are currently doing at the Institute of Physics. By sharing data between publishers and librarians, he said, we can make the industry more efficient.
I was particularly interested to hear more about the video abstracts they launched in February this year. Authors can now submit video abstracts and so far they have had over 10,000 video views. The human factor is important in engaging with students and researchers and helps to humanise the journal by conveying the inspiration and enthusiam of the author or editor.
Publishers can learn a lot from evaluating the data they have such as seeing which research areas are growing. Web analytics; train of thought analysis; traffic dashboards including social media indexes and extended metrics such as A&I services are all important.
Platform development and ensuring connectedness is key. SEO is still vitally important here.
Social Networking/Media activity and how it impacts on usage is difficult to track. Physics World has 8,510 follows on Twitter.
Local language sites (Japan; Latin America and China) have moderate but growing traffic so far.
Access via mobile devices including iphones and ipads is growing and publishers need to operate in this space to ensure users can access content wherever they are.
Challenges for publishers and librarians alike include creating new and meaningful metrics to cope with the rate of industry change; niche areas of resarch and primtive metrics.
As Christian stated at the beginning of his presentation, it is important for librarians and publishers to work together as much as possible and share data to increase efficiency wherever possible.
Friday, April 08, 2011
William Gibson and the future of libraries
On day one of the UKSG 2011 Conference, John Naughton (The Open University and Cambridge University Library) paraphrased William Gibson, 'The future has already arrived...stop trying to predict it.'
'We are living through a revolution and we have no idea where it is going,' he suggested. He used the term 'information bewilderment' to explain further.
Capitalism, he argued, relies on the creative destruction of industries in waves of activity. This is exciting for those on the creative side but scary for those on the destructive (ie newspaper and music industries) side.
Obsolete business models are at threat and everyone at the conference is affected, he warned. In the digital age, 'disruptive innovation' is a feature and a way of cutting out the 'middle man' to create profit.
He cited Amazon Kindle Singles as an example, whereby they invite authors (previously published or unpublished) to publish shorter articles (longer than a magazine or journal article but shorter than a novel) as an e-book on the Amazon Kindle platform.
Prediction is futile but you can measure changes. Complexity is the new reality and the rise and rise of user-generated content offers numerous opportunites for end users to 'cut out the middle man' (ie publishers).
In the old ecosystem there were big corporations while the new ecosystem relies on everything being available in smaller chunks on content (tracks not albums, articles not journals etc).
What's it got to do with libraries?
There is an intrinsic belief that libraries and librarians do good work but a wave of 'creative disruption' doesn't care. Libraries have traditionally taken a physical form and one of the debates has been about how to maintain the idea of a 'library' when users are increasingly accessing content online. When all academic activity takes place in a digital environment (soon?) how will libraries justify their existence (from place to space)?
John Naugthon ended his presentation by suggesting librarians could add value by building services around workflows (social media; rss feeds etc) as the everyday avalanche of data crys out for the skills of the librarian to create order.
'The best way to predict the future is to invent it.'
Sounds like good advice for those of us in publishing too.
Labels: Plenary session 1
Thursday, April 07, 2011
Collections 2021: The Future of the Collection Is Not a Collection
- PDA is the new assumption, although it's not the only way. The collection service will be mainly a conduit service, building only limited permanent collections.
- The smart phone is the killer delivery app. While few people want to sit down for two hours and read a novel on a small device, many of us have lots of blocks of 10-15 minutes in which we will happily read something that's conveniently available.
- Most academic print acquisition is print on demand, this avoiding the major waste of print runs based on guesswork.
- Most search is done on primary documents rather than proxies such as the library catalogue.
- It is difficult to distinguish library services from other educational services.
- Big collecting libraries such as Oxford, Harvard, LoC will maintain "monuments to Western civilisation"
- Local research institutions will have smaller and more specialist collections based on curricula
- Less well funded liberal arts and community colleges will be conduits and will rely heavily on Google Book Search and just-in-time delivery.
- Sclerotic librarians. There are difficult conversations to be had about change.
- Traditional accreditation structures - counting books on shelf to assess the worth of the library.
- Fainthearted publishers - justifiably so. PDA will put some publishers out of business. You can't make as much money selling just what people want as you can by selling them content they don't want bundled with that which they do.
- Customer-focused competitors for patrons time such as Google and Amazon. These competitors aren't interested in helping people find good information in the way that libraries are, but they're quick and convenient.
Wednesday, April 06, 2011
Squatting in the Library? Visitors and Residents at UKSG
An attempt to capture the OA debate between Alma Swan and Steve Hall
Tuesday, April 05, 2011
Goodbye Serials, Hallo Insights!
Shhhh, Turn of Your Phone!
The first mobile phone call (from a car) was made in June 1946! This is not a new technology, we have lived with them for a long time. The first proper handheld mobile phone call was made on April 3rd 1973. It has only been in the last couple of years that we have really begun to exploit the potential of mobile devices. The i-pad sold 14 million units in the first 9 months of its life - mobile is everywhere.
James goes on to show how all new developments have been treated with scepticism (and btw gives an excellent way of how to actually give a 'history of' at a conference). The 'evil' slate, pen, calculator have all been criticised. We do things because they have always been done that way - resistance to change is normal. So we put up signs that say don't use your phones (or don't swim) to try and control change - but what problems to phones *actually* cause in a library?
The Culture of NO is a big problem in libraries today. When you see a big sign saying DO NOT, it is human nature not to respect it. It is much better to talk to learners and help them respect their environment rather than dictate and direct. A YES culture is a much better place to be.
So what CAN we do with mobile devices in the library?? My contributions are some fantastic ideas from thewikiman and Jo Alcock. Other ideas:
- Use the web. Sounds obvious, but very necessary. Unfortunately, there are very few journal or eBook platforms that are well developed for browsing on phones, or even small notepads.
- Collaboration. AudioNote is a great example of tools that can be used in this way.
- QRCodes - dotted around the library to provide extra information.
- Augmented Reality - layering information over images within the library. It's a great way to UpSell the resources.
- Barcode Scaners - scanning a barcode of a book in WHSmiths to see if it is in the library.
- Making Notes - such as tools like EverNote.
- Using tools like Google Googles to find more information about a statue, a picture, a resource.
Cost is a real issue. Mobile should be about enhancing the service you already provide, it should not be exclusive and discriminatory to those who cannot afford expensive devices. James also bravely states that eBooks will never replace real books :-)
The digital divide is a real issue (a la Andy Powell's talk this morning). As is connectivity (conference wifi anyone?). The pace of change also makes it difficult for libraries to keep up with changing devices, skilling staff etc. Prioritising is an important focus here.
Despite all of these issues, James is a clear believer in using mobile devices in libraries - and he encourages us all to think about just one way in which we could too.
An introduction to ORCID
- Thomson donated codebase for its researcher ID to help jumpstart ORCID
- Various functions were added to this for ORCID's alpha prototype - Thomson's system was based on personal "claims", so the organisational layer had to be added
- Now working out last details for licensing the codebase to build a phase I version of the system
- And planning for future sustainability (funding / staff)
- Hoping to have something that people can use, next year
- Q: Authors are allowed to create profiles - how can IDs remain unique?
A: Authors cannot change the identifier, only the information associated with it.
- Q: The contributor ID could become increasingly complex - how do we define where 'contribution' begins and ends?
A: We will studiously avoid defining that - it's going to evolve. But the answer is essentially that people will record what they think is important, and if it's not important, it won't be counted for much. [Given that people will have to take the time to enter this data, they will likely only claim credit for things that are useful / important]
- Q: How will this fit with the requirements of REF?
A: It's not clear where REF responsibilities will sit but hopefully ORCID will make the process of gathering information easier.
- Q: Pseudonymity?
A: A lot of this information is public already, but in aggregation it's more powerful. What if it becomes too easy to find details about stem cell researchers in Alabama or animal sci researchers in Oxford. People do have good reasons to want to hide information - even just if you want to be credited for peer reviewing without it being public. ORCID will allow any or all information except the identifier itself to be hidden.
- Q: What is happening with the development of IDs in different countries?
A: It would be a bad idea to think "ORCID's coming, let's stop working on our system". Other systems will continue to exist and be important. At minimum, ORCID will be able to include information about other relevant identifiers.
- Q: What work will be involved for publishers?
A: A classic example: a researcher submitting a manuscript currently fills in all the information each time, and that information quickly becomes stale (e.g. contact data). In future, they will upload their ORCID, and publishers can query and recheck information as necessary.
- Q: Who will be the arbiter of who will be attached to a work as a contributor?
A: For example, the corresponding author will have more credibility in saying who else contributed.
- Q: Disambiguity of affiliations?
A: We may integrate with e.g. Ringgold to create a controlled vocabulary for organisations.
- Q: What are the data protection issues?
A: We are transparent about what is being revealed, to whom, and we give authors control - they can make anything except the identifier private.
- Q: What's the long term funding plan?
A: Exactly. The technology doesn't matter if we can't sustain an organisation to keep it running. We are looking at future models, from related service provision to membership.
Curated tweetstream: what our audience said about Charles B. Lowry on the economic crisis
- bookstothesky Telling wordcloud from Lowry re library budgets: key words emphasized are budget, reduction(s), cut, reduced :-/ #uksg
- jharvell Does that mean that 10 universities in the us had library budgets of over 40 million dollars before the cuts? #uksg #didimisssomething?
- chriskeene @jharvell and the lowest category was 'libraries with budget less than $20million'. different world!
- jharvell Don't get me wrong those big budgets are brilliant. Brilliant. But my gob hasn't closed for the last 5 mins. #ineverknew #uksg
- jharvell With the amount of money available in US budgets why are publishers even bothering listening to us in the uk #uksg
- charlierapple Decreasing budgets are the new norm, not an aberration, with consequences for teaching and research internationally #uksg Lowry
- ORourkeTony @charlierapple #uksg I heard someone say recently that flat was the new up!
- MelindaKenneway Time to head to Canada by the looks of things - they seem to be the only libraries left with budget. #uksg
- antet Not sure I like the detached phrase "reduced commitment to human resources" #uksg
Nurturing innovation, or why we need to kiss more frogs
The future is open (thanks to metadata)
Rufus Pollock from Open Knowledge Foundation tells us how metadata can and will be more open in the future, and why we should care.
Libraries and publishing used to be mainly about reproduction of the printed word. Access and storage also but reproduction mainly which once upon a time reproduction was very costly; people needed to club together and form societies in order to afford reproduction.
Now we're matching, filtering and finding, but there's too much info and every password you have to enter slows you down, and slows down innovation and innovators. Matching is king in a world of too much info - Google's aim is to match people with information and it all relies on humans making the links and building sites. Imagine if they'd had to ask permission of every single person - we would have missed out on something big.
Of course people have to be paid, machines have to run etc. BUT much of this production is already paid for i.e. via academia itself: instead of using the same few favourite books, why not ask friends? Or create our own journals?
Data and content are not commodities to sell but platforms to build on... there are plenty of ways to make money without going closed (although it might be different people making the money of course!)
And why does metadata matter so much? It's the easy way in; everything attaches to it: purchasing services; wikipedia; analytics such as who wrote it, how many people bought it etc.
Data is like code and so the level of re-use, and the number of applications we can create is huge.
One such project is JISC OpenBib which has three million open records provided by the British Library. It integrates with wikipedia, and includes a distributed social bibliography platform so that users can contribute and correct and enhance. We need to harness the users to help us make much better catalogues, to enrich catalog data.
So metatada is the skeleton and right now we have the chance to make a significant change for the better. Metadata and content WILL all be free one day... it may take some time but it will happen. The day is coming when there won't be a choice. There will be enough people with open data to make it happen.
Do the math: PDA not the answer?
- 40% of Liverpool's e-resource usage is e-books - yet 95% of the budget goes on journals.
- Liverpool's usage is typical - approx 40% of titles in a collection are used in the first year; approximately 60% have been used by the second year
- Some subjects (e.g. mathematics, at Liverpool) seem to perform badly - is this a factor of how/when information is used in different disciplines? need to be careful before making collection development decisions based on this data
- All types of books get used at least a bit, but some content (e.g. conference proceedings) is used more than other content (e.g. monographs)
- Pareto principle applies! 80% of downloads from top 21% of ebooks - Terry doesn't think this should be a factor in how collections are purchased / priced. (Looking more closely, 35% of usage on one platform came from one title! - doesn't tell you anything about the broader collection, just that some books are heavily used)
- Even on aggregator platforms (where there's a greater level of individual title selection than a publisher package), a third of ebooks have had only 1 or 2 accesses during 2 years
- With patron-driven acquisition, all ebooks are used (because you don't buy them unless they are) - so should be better value? Terry used ebrary model (purchase triggered by 10 page turns / 10 minutes in a title / copy & pasting / printing) to analyse Liverpool's Springer ebook usage stats and calculated that PDA costs would overtake package costs in just one year in most cases (even when cheaper backfiles were excluded from analysis).
- Evidence from elsewhere (e.g. U Iowa ebrary pilot) also shows that PDA budgets run out quickly - libraries who started trials had to resort to buying packages after all
- ... other PDA models are available ... (and may show different results) but Terry found that a PDA model would have to allow for "6 chapters free" before it would be comparable to package pricing.
- Some packages are better value than others, and libraries should prioritise these in collections
- Aggregated databases give cheap critical mass
- Single title selections are important for core texts
- PDA can fill the gaps, but not form the foundations
- Need to centralise book budgets - stop fragmenting by formats etc - a hard sell for lots of faculty / librarians
- Rapid move to e-only book acquisition - implications for logistics / staffing
- Packages need to be at least 50% discount for it to be worth it for the library - make it a "no brainer" (70% discount) for the library to purchase, and you'll solve the budget crisis
- Offer combined books / journals packages with appropriate cost weighting / discounting.
Monday, April 04, 2011
The Gatekeeper is Dead. Long Live the Gatekeeper!
What scientists really want from digital publishing.
This section of the conference allowed librarians and publishers to hear directly from scientific researchers; first up is Philip Bourne from University of San Diego who is a computational biologist among other things (e.g. open access advocate).
Bourne starts by explaining his big hope for scientists’ relationship with publishers in the future:
“as a scientist I want an interaction with a publisher that does not begin when the scientific process ends but begins at the beginning of the scientific process itself”
The current situation is:
- 1. Ideas
- 2. Experiments
- 3. Data gathering
- 4. Conclusions - it’s at this stage that the publisher comes in
But why couldn't the publisher come in at the data stage? They could help store it for our group. Or even earlier, at the ideation stage: The moment I jot down a few ideas, the publisher could control access to that information and then at some point down the line when the access is opened up – that’s when it becomes ‘published’.
There are movements in that direction. For example in Elsevier’s ScienceDirect (and some others too) you can click on a figure/image and move it around and manipulate it – the application is integrated on the platform because a publisher and a data provider has cooperated. But this is just the beginning; when you click on the diagram in the article, you’re getting some data back but it’s generic and it might not be organised in the way that you want. It’s generic – the figure is being viewed separately from the article text and related data – now you have to figure out what that metadata means to the article. So this is a good step but it’s not capturing all of the knowledge that you might want. It needs more cooperation, more open and interactive apps. And it needs:
- Integrated rich media that improves comprehension, viewable in different ways. A video of an experiment actually being done, delivered to me alongside the text from the article.
- The ability to review and interact with data on the mobile platform. Should have apps not just to read but also interact with data.
- Mashups with content from other articles and data, must be at the point of capture, not post- anything.
- Semantic linking of data that can lead to new knowledge discovery. To find all references to that piece of data – that data itself is probably not cited – would like to know how the actual data is being used discover relationships that other people have found between your data and other sets of data.
So Bourne wants publishers to become more involved with his work as – he confesses – some of the work is less than organised. He thinks scientists need help with management of data in general, and specifically:
- · Project management. They use e.g. basecamp for project management but email folders are primary – this in an unhealthy ‘hub and spoke’ situation
- · Content management. It’s a mess with content stored all over such as on slides, posters, lab notebooks etc.
- · Manage negative data. They generate way more negative data than positive – Negative data is important. But you can’t find it – it stays hidden. This needs to change.
- · Software. All the software they create is open source but when the grad student that wrote it leaves, it’s lost.
Bourne’s ‘Beyond The PDF’ workshop has generated discussion and ideas. He says “the notion of a journal is just dead – sorry. The concept of a journal is lost to me; its components and objects and data are what I think about. Research articles are useful but the components could be seen as a nanopublication.”
We need more:
- · Semantic tagging of PDFs and beyond
- · Citation ontologies
- · Scholarly html – to write these workflows
- · Authoring tools
Microsoft are looking at some of these things already and Bourn’es group has written plugins for Word – e.g. as you type you auto-check various ontologies that may suggest you change a common name to a standard name. You can tag that at the point of authoring.
All of this is not yet a huge success but it’s coming. Right now there is not much incentive but if publishers can help fast-track the development of these applications then authors will start using them. There’s no use talking about it but it’s only on researchers’ radar when they see science done in a way where this process has made a difference. For example Bourne’s group is running a test to look at spinal muscular atrophy (designated by the NIH as treatable). They will coalesce a set of disparate tools, engage the publishers (Elsevier have opened up ths), in order to address a specific problem that could change lives.
If this works then that would get the kind of attention that scientists would take notice. Only when they see thus process succeeding will they start adopting it. The tipping point will come when the tenure ‘reward system’ starts to change for the next generation, the way science is researched will improve.
From Tortoise Shells to Tweets - The Future of the Book
The earliest know form of writing is some scratchings on a tortoise shell from 6000BC, found in Northern China. From there Prichard takes us on a rapid tour of the familiar names from the history of the book - the Art of War, Guttenburg, Caxton, the Penny Dreadful. All of these developments had a framwork of publication around them that we can reflect on in the same way when we think about digital books - and Ingram Digital has an invested interest in thinking about this. Some of these examples may confuse form and function somewhat, but I think the main point is that we have been consuming 'writing' in various ways for a long time.
So, what is the definition of 'book' in today's age? Is the move from print to digital any different than the move from scroll to bound text? Prichard highlights some trends around the digital move:
1. Shifting Market.
This describes the move from physical stores to online sales. This reminds me of a recent (personal) blog post I wrote about the recent closure of libraries. This is being supported by the growing use of appropriate devices. The predications for 2011 are that there will be 14.7 million e-readers 44.6 million tablets in use.
Prichard also predicts that academic libraries will be 80% e-only by 2020 in US (seems quite slow to me!).
2. Generational Shift.
Schools are using a mixture of modern devices like the i-pad with traditional books that have been in use in the classroom for years. There is a significant change in language - text speak is effecting teenagers learning across the board.
3. Enhanced eBooks.
Moving beyond merely trying to deliver the print version of a book in digital format. The lonely planet's new travel guides are an interesting example of this - I'd note they have to be, as they fight for their market against user-generated content on places like TripAdvisor. The books container is changing, the book itself must change to keep pace with this.
Prichard poses some ideas of where we might go with this:
- Could we use biometrics to change the ending of a book based on your mood?
- Could your car remember where you were in a book and start reading to you when you start a journey again?
- Could locations used in a book change based on where you physically are?
- Could books interact with each other more, e.g. viewing other people's underlinings on Kindle?
Prichard closes by saying that print on demand has to be the future of publishing - it reinvigorates the supply chain, its green, and its user appropriate. He does not see print as vanishing - and reflects on the failure of the 'paperless office' as an example of why printed books will not vanished. There was quite a bit of disagreement on Twitter about this...but I wonder if we think about students printing articles / photocopying book chapters as part of the 'print' process. We might not BUY print, but print consumption will always be a personal choice.
Naturally, the audience is not going to get a publisher get away scot-free with giving a presentation without identifying some of the ways in which publishers are NOT helping the shift to digital. The poor business models for ebooks were highlighted, with prices often being higher than print making them inaccessible. The JISC eBook Observatory project has carried out some interesting work around this concept.
Following a question from Peter Burnhill, Prichard notes that the solution will not come from one part of the industry - we should put the pavements where the students chose to walk.