LIBLICENSE-L@LISTSERV.CRL.EDU


View:

Message:

[

First

|

Previous

|

Next

|

Last

]

By Topic:

[

First

|

Previous

|

Next

|

Last

]

By Author:

[

First

|

Previous

|

Next

|

Last

]

Font:

Proportional Font

LIBLICENSE-L Home

LIBLICENSE-L Home

LIBLICENSE-L  December 2014

LIBLICENSE-L December 2014

Subject:

Re: Question: Modifications to Titles in Journal Packages

From:

LIBLICENSE <[log in to unmask]>

Reply-To:

LibLicense-L Discussion Forum <[log in to unmask]>

Date:

Mon, 22 Dec 2014 20:36:57 -0500

Content-Type:

text/plain

Parts/Attachments:

Parts/Attachments

text/plain (176 lines)

From: "Hinchliffe, Lisa W" <[log in to unmask]>
Date: Mon, 22 Dec 2014 02:21:17 +0000

A few more thoughts on why surveys may be mis-matched to documenting
use and offering an attempt at clarity on my access log and
self-perception bias comments...

Whether you can determine the kind of person who is accessing the
content via access logs will be determined by how you have set-up
access and what you track. If you require logins and capture patron
type, you will be able to determine use by patron group. If not, yes -
you'll have everyone's use. I'm probably on the side of thinking that
postdocs and graduate students are pretty important users of research
content and, honestly, undergraduates too. But, even if not
undergrads, I hope we'd find middle ground on thinking that ALL of
those with research responsibilities (i.e., driving research
productivity at the University) are key constituents here.

I'm skeptical on the response rate you'd get - I think it would be
much lower than you'd like. But, let's say it is high ... I think the
data itself would be really difficult to rely upon. The
self-perception bias phrase I used is probably more accurately labeled
"set of cognitive biases"
(http://en.wikipedia.org/wiki/List_of_cognitive_biases). In this case,
a number of memory biases seem likely to be an issue. Remembering
accurately how often one has done anything over the past year, for
example, is not easy much less remembering at the level of journal
title when presented with a long list.  And, there is also the
complexity here that there would be motivation for a faculty member to
over-report use because of the perception he/she has of the importance
of the journal in the field ("of course I use this journal - it is an
important one!").

Having said all this - it would be wonderful if someone had empirical
data to test this with. A relatively straightforward design would be
to compare a faculty members reports of use with the sources they cite
in publications. This design has obvious limitations but it would
allow comparison of faculty member reported use with documented use.
Better of course would be correlating access log use with faculty
member reported use, but that would require tracking not just patron
type but actual individual patrons. In the US, I suspect few libraries
would have that data readily available but it might be possible in
Australia I believe if I correctly understand their access logging
practices.

Lisa

Lisa Janicke Hinchliffe
Professor/Coordinator for Strategic Planning
Coordinator for Information Literacy Services and Instruction
University Library, University of Illinois at Urbana-Champaign
[log in to unmask]

________________________________________
From: Sandy Thatcher <[log in to unmask]>
Date: Fri, 19 Dec 2014 09:22:57 -0600

Sure, if you got a very low response rate, but I'm guessing enough
faculty would find this a very important survey to respond to. And it
would reveal a lot more than usage counts, which presumably include
uses by everyone, not just faculty. (I'm not sure what access logs
are, but do they reveal who is doing the accessing? And how does
"self-perception bias," whatever that is, enter into the equation?) If
faculty regularly use a journal instead of just very occasionally
using an article from it, isn't that a very important piece of
information?   It seems like a pretty straightforward question to ask,
particularly if you define what "regularly" means.

Sandy Thatcher



> From: "Hinchliffe, Lisa W" <[log in to unmask]>
> Date: Fri, 19 Dec 2014 04:10:38 +0000
>
> A survey seems mis-matched to this. Response rates, self-perception
> bias, etc. Why not just use the access logs?
>
> On the original question - I would question why libraries would have
> to review/compare listings themselves? Seems like a pretty standard
> thing to expect in making a purchase would be "here's a list of what
> you are buying and here's how it differs from your last list"? Doesn't
> seem like something libraries should have to devote staff time to
> compiling? Now, as to whether such lists get reviewed ... well, I hope
> so and annually!
>
> Lisa
>
> Lisa Janicke Hinchliffe
> Professor; Coordinator for Strategic Planning;
> Coordinator for Information Literacy Services and Instruction
> University Library, University of Illinois at Urbana-Champaign
> [log in to unmask]
>
>
>
> From: Sandy Thatcher <[log in to unmask]>
> Date: Wed, 17 Dec 2014 20:56:58 -0600
>
> Would it be all that difficult to do a survey of faculty and ask them
> which of the journals in a package they use regularly, or not at all?
> Does any library now do this?
>
> Sandy Thatcher
>
>
>>  From: Karin Wikoff <[log in to unmask]>
>>  Date: Wed, 17 Dec 2014 07:35:58 -0500
>>
>>  For us, we'd only be likely to notice if something our users use
>>  regularly disappears.  20 journals no one uses could disappear, and we
>>  probably wouldn't notice.  But if the one journal some faculty member
>>  uses all the time disappeared, then there'd be a big problem.  It's a
>>  good question, and not one I'd given a lot of thought.  I'll be
>>  interested to read other folks' replies.
>>
>>  Karin
>>
>>  Karin Wikoff
>>  Electronic and Technical Services Librarian
>>  Ithaca College Library
>>  953 Danby Rd
>>  Ithaca, NY 14850
>>
>>  Phone: 1-607-274-1364
>>  Fax: 1-607-274-1539
>>  Email: [log in to unmask]
>>
>>
>>
>>  On 12/16/2014 8:17 PM, LIBLICENSE wrote:
>>
>>  From: Ann Shumelda Okerson <[log in to unmask]>
>>  Date: Tue, 16 Dec 2014 20:15:48 -0500
>>
>>  Dear liblicense-l readers.  Your listowner/moderator (me) has a
>>  question for you.  I would very much welcome the views of anyone on
>>  this list, whether publisher or librarian or someone in the scholarly
>>  communications chain.  There's no right answer; in fact, I'm not sure
>>  there is even an answer, but I was in a group that started discussing
>>  this matter and we felt caught short.  And we felt we should have a
>>  reasoned opinion, when we did not.  Please read on.
>>
>>  Most many big deal journal packages contain language [such as that
>>  below] re. modification to "portions of the Licensed Materials."  The
>>  contracts say that if any of the changes make the materials less
>>  useful, the institutions may seek to terminate this agreement for
>>  breach.  And, there will likely be language of this sort: "If any such
>>  withdrawal renders the Licensed Materials less useful to Licensee or
>>  its Authorised Users, Licensor shall reimburse XX for the withdrawal
>>  in an amount proportional to the total Fees owed."
>>
>>  My question is this:  if my library has a "big [or medium] deal,"
>>  let's pretend it's 300 or 500 or 1000 or 2000 titles, what is a
>>  reasonable expectation for the numbers or percentage of content that
>>  will leave the package before the library or consortium would either
>
>  > seek reimbursement (more likely) or total termination (less likely)?
>>
>>
>>  Do libraries (or consortia) review the big-deal lists each year to look
>>  for changes?  Every 3 years? If there were a loss of previous titles
>>  in the amount of 5%, would it be a concern? How about 10%?
>>
>>  Of if not a percentage "bright line," then what would cause a review
>>  of the list and a concerned conversation with the big deal publisher?
>>  Would it be the loss of a couple of absolutely key titles?  the loss
>>  of a particular smaller publisher's journals list?  a disciplinary
>>  impact? a dollar impact?  If "it depends," what does it depend on?
>>
>>  Do libraries care very much about what's actually in these large
>>  packages, or are we too busy to pay attention to their changes? What
>>  would it take to get libraries' attention?
>>
>>  Thank you, Ann Okerson

Top of Message | Previous Page | Permalink

Advanced Options



Archives

April 2019
March 2019
February 2019
January 2019
December 2018
November 2018
October 2018
September 2018
August 2018
July 2018
June 2018
May 2018
April 2018
March 2018
February 2018
January 2018
December 2017
November 2017
October 2017
September 2017
August 2017
July 2017
June 2017
May 2017
April 2017
March 2017
February 2017
January 2017
December 2016
November 2016
October 2016
September 2016
August 2016
July 2016
June 2016
May 2016
April 2016
March 2016
February 2016
January 2016
December 2015
November 2015
October 2015
September 2015
August 2015
July 2015
June 2015
May 2015
April 2015
March 2015
February 2015
January 2015
December 2014
November 2014
October 2014
September 2014
August 2014
July 2014
June 2014
May 2014
April 2014
March 2014
February 2014
January 2014
December 2013
November 2013
October 2013
September 2013
August 2013
July 2013
June 2013
May 2013
April 2013
March 2013
February 2013
January 2013
December 2012
November 2012
October 2012
September 2012
August 2012
July 2012
June 2012
May 2012
April 2012
March 2012
February 2012
January 2012
December 2011
November 2011

RSS1