But Were We Successful?
Using Online Asynchronous Focus Groups to Evaluate
Library Services
Rebekah Cummings
rebekah.cummings@utah.edu
Lindsay Ozburn
lindsay.ozburn@usu.edu
Andrea Payant
andrea.payant@usu.edu
Mike Shelton
michael.shelton@usu.edu
Ryan Bushman
ryan.bushman@usu.edu
Utah Library Association Conference - May 20, 2021
Why Was RDMS Created?
Why did Utah State University create a research data management services
program?
RDMS was created to comply with federal mandates stemming from the
2013 OSTP Memo on “Increasing Access to the Results of Federally
Funded Scientific Research.”
2013 OSTP Memo
In 2013 the Office of Science and Technology Policy issued a memorandum
titled:
“Increasing Access to the Results of Federally Funded Scientific Research”
• “The Administration is committed to ensuring that, to the greatest extent
and with the fewest constraints possible and consistent with law and the
objectives set out below, the direct results of federally funded scientific
research are made available to and useful for the public, industry, and the
scientific community. Such results include peer-reviewed publications and
digital data.”
• Stipulated that federal agencies with over $100M in R&D had to develop
plans to make publications and data that originated from federally funded
research freely available to the public.
What This Meant for Researchers:
● The need to maintain and make their data publicly
accessible, generally through data repositories.
● The need for their publications to be open access
and to generally have them listed in the funder’s
public access repository.
● The need to include information about any data
and publications created with their annual reports
to their funders.
● The inclusion of a data management plan with the
award applications.
Utah State University recognized several needs necessary to help Primary Investigators (PI)
maintain compliance with funder mandates:
USU’s Response to the OSTP Memo:
• The need to create a Data Librarian position to assist researchers with their publications and data: Fulfilled in
Summer 2015.
• Bring campus stakeholders together to access needs and create a compliance program: In Autumn 2015 a cross-
campus group with members from the Office of Research and Graduate Studies (RGS), the Office of Information
Technology (IT), and the Merrill-Cazier Library started meeting to access how to best monitor and assist USU
researchers’ compliance.
• The needs to create and implement a University Research Data Policy and a process to create auditable records to
facilitate access to data and publications: By Autumn 2016 a plan was implemented that included the process for
award compliance tracking
USU’s Approach to Compliance:
USU decided to take an approach that facilitates the researchers’ ability to be in
Funder Compliance
USU worked to develop a compliance process that would:
• Demonstrate to Funders that the University is making good faith effort to comply with mandates.
• Emphasis to researchers that complying with funder mandates is an institutional priority.
• Demonstrate to researchers that resources are available to support them in meeting funder mandates to
make data openly accessible.
But Were We Successful: Using Online Asynchronous Focus Groups to Evaluate Library Services
Continued Support for RDMS from Administration
Both the Library and University administration recognizes that compliance to funder
mandates is an institutional priority. They continue to support RDMS in a variety of
ways:
• Monthly meetings of the University’s Research Data Management Committee, which is made up members from the Library
and the Office of Research. These meeting serve to discuss the needs of both RDMS and the researchers, news in the data
management field, and to develop instructional opportunities with the campus community.
• Once a semester the Library (via RDMS’s supervising Associate Dean) sponsors Data Club, a working lunch where PIs are
invited to visit with the RDMS staff and a representative from the Research Office to discuss their data concerns.
• In February 2020, the Library and Research office co-sponsored RDMS in conducting a half-day workshop called
“Datapalooza” geared toward teaching graduate students good data management practices. This was followed up in April
2021 with a second Zoom based Datapalooza event.
Establishing the Need for Program Assessment
•Association of American Universities (AAU) and the Association of Public & Land-Grant Universities
(APLU) working group conference
•Two members of RDMS, as well as the Interim Vice President of Research represented USU.
•Outcome: recognizing the need to assess the quality and effectiveness of the University’s
compliance program via RDMS’s efforts.
Assessing the compliance of federal mandates:
the library’s involvement
Early 2010’s…
• Federal organizations started requiring Data Management Plans to ensure research reproducibility
• Research reproducibility becomes a hot topic in LIS, with libraries examining what role they might play in
assisting their campus’ researchers with the reproducibility and accessibility of empirical research (Vitale,
2016)
o This led to a shift in academia = libraries become increasingly more engaged in campus research efforts and support.
• Emergence of RDMS programs and technical infrastructure as a support mechanism
Assessing the compliance of federal mandates:
the library’s involvement
Late 2010’s…
• Librarians and professional library orgs are more involved than ever in shaping policy and best practices for
RDMS in support of researchers
o Sparsely explored topic in LIS literature
• This research addresses the gap in literature and assessment of RDMS efforts
Our Assessment Methods
• Traditional survey coupled with an online asynchronous focus group (OAFG)
• Why two methods? Data triangulation
o Survey gave us a broad look at what works and what could be improved. OAFG dug
deeper into these questions to give us specifics for improvement
• OAFG discussion questions modeled after survey questions and results
Asynchronous Focus Groups as a Methodology: What are they?
• Focus groups conducted through online mediums where participants do not need
to be in the same place or online at the same time.
o Typically, they are given a time frame (2 day, for example) in which to participate
• Who uses them?
o Marketing- generating consumer insights that inform response to advertising
o Health - discussing healthcare issues with professionals across the globe (Han et al.,
2019; Ramo et al., 2019)
Pros and Cons of Asynchronous Focus Groups
Pros:
● Not bound in time and space as much; as
people can assemble to participate in
whatever time frame you give them
● Often, they have the potential to be
anonymous, which is helpful for sensitive
topics
● A lack of geographic boundaries helps increase
the diversity of a participant pool for specific
types of research projects
● In academia, busy faculty schedules are less of
a barrier for participation
Cons:
• Still dealing with the influence of social cues in
focus groups but in a much more presumptive
way (i.e., lack of vocal tone and eye contact can
lead to assumptions about someone’s answer).
• Previous forum posts can influence future posts
Why? While this is the case when speaking,
participants get more time to craft their answers
in online forums, leading to a higher potential
for influencing factors
• Ethical considerations of cyberspace: anonymity
could lead to people misrepresenting who they
are and/or sharing other participant’s responses
with others.
How did we use OAFG to assess our program?
• We used Canvas, an online learning platform that both USU and U of U had access
to.
• Of the team, we used the librarian at U of U as an outside moderator to encourage
honesty and to scrub data of identifiable information.
o Outside moderators are important with focus groups, in general, to prevent
unintentional biases in questions and to encourage honest answers from participants.
Hold Up.
You were assessing participants at one location
Why did you use an OAFG?
Pros and Cons of Asynchronous Focus Groups
Pros:
● Not bound in time and space as much; as
people can assemble to participate in
whatever time frame you give them
● Often, they have the potential to be
anonymous, which is helpful for sensitive
topics
● A lack of geographic boundaries helps increase
the diversity of a participant pool for specific
types of research projects
● In academia, busy faculty schedules are less of
a barrier for participation
Cons:
• Still dealing with the influence of social cues in
focus groups but in a much more presumptive
way (i.e., lack of vocal tone and eye contact can
lead to assumptions about someone’s answer).
• Previous forum posts can influence future posts
Why? While this is the case when speaking,
participants get more time to craft their answers
in online forums, leading to a higher potential
for influencing factors
• Ethical considerations of cyberspace: anonymity
could lead to people misrepresenting who they
are and/or sharing other participant’s responses
with others.
Qualities of Survey
• Population: 78 Researchers
• Sample: 28 Researchers
o Initial response of 32 researchers ( 41.03% of the population)
o Data cleaning determined 28 of respondents provided usable data (34.6% of population)
• Introduction of Bias
o We opted for a convenience sample rather than a randomized sample
The Survey
Two styles of question:
• Questions designed to determine the frequency of a behavior of the researcher often used the following
scale
o Never
o 0-3 times (a redundancy we would like to fix in further research projects)
o 4-7 times
o 8-10 times
o More than 10 times
• Questions designed to determines a researcher's response to USU’s efforts used the following scale
o Strongly Disagree
o Somewhat Disagree
o Neither Agree Nor Disagree
o Somewhat Agree
o Strongly Agree
Key Statistics
• 19 of the 27 (70.4%) respondents indicated they had three or fewer awards that
required deposit of publications into a public access repository
• 18 of the 27 (66.67%) respondents indicated they had three or fewer awards that
required deposit of data into a publicly accessible repository.
• “Could Sponsored Programs or the Library better facilitate efforts to meet agency
mandates?”
o 13 of 27 (48.15%) responded yes
o 14 (51.85%) responded no.
o When examined by years as a researcher, a higher percentage of respondents in the range of 11-20
years as a researcher responded “yes” to this question.
Key Statistics: Behavior of the Researcher
• Motive for making data publicly available
o 84.6% to comply with journal publishing requirements
o 92.6% to satisfy a funder
o 61.5% with the desire for their data to be public
• 92.59% of respondents indicated that they anticipated writing journal publications
up to 5 years after the close of their award
• 88.46% of respondents anticipated needing up to 4 years after the close of their
award to have final data ready for deposit in a public repository
Key Statistics: Response to USU’s Efforts
• The email from SPO was helpful
o 20 (74.07%) strongly agree
o 5 (18.52%) somewhat agree
• Researchers understand how to respond to the email
from SPO
o 13 (48.15%) strongly agree
o 7 (25.93%) somewhat agree
o 5 (18.52%) somewhat disagree
• The email from SPO explains how they could find more
help
o 16 (59.26%) strongly agree
o 8 (29.63%) somewhat agree
o 2 (7.41%) somewhat disagree
• The email from DL was helpful
o 11 (64.71%) strongly agree
o 5 (29.41 % somewhat agree
o 1 (5.88%) somewhat disagrees
• Researchers understand how to respond to email from
DL
o 11 (64.71%) strongly agree
o 4 (23.53%) somewhat agree
o 2 (11.76%) neither agree nor disagree
• The email from DL explains how they could find more
help
o 11 (64.71%) strongly agree
o 5 (29.41%) somewhat agree
o 1 (5.88%) somewhat disagree
Online Asynchronous Focus
Groups (OAFG)
SOCIALLY DISTANCED BEFORE IT
WAS COOL
• SECURE
• FAMILIAR
• ALLOWED FOR DISCUSSION
• RESPOND IN LINE
• TECHNICAL SUPPORT
Technology
Five Discussions
Introductions Response to funder
mandates
Researcher data
practices &
publication deposit
Research
Office and Library
program
What would help?
First Attempt
• Selected 15 participants at random
• VP for Research invited the fifteen participants
to join the OAFG at their convenience the
following day
• With only a single day’s notice, only two of the
fifteen participants responded, and they,
inexplicably, couldn’t get into the Canvas site.
Second Attempt
• Moved Canvas site from University of Utah to USU
• Gave participants several weeks of advance notice
• Confirmed attendance with five participants, but only three
showed up
• Good news! The three participants gave excellent, detailed
feedback, interacted with one another, and didn’t have any
technical issues
Qualitative Data Analysis
1. Praise for the Data Librarian, the library, and the research office (16
statements)
2. Data sharing challenges (6 statements)
3. Data resources (such as DMP Tool, FigShare, Digital Commons,
Metavist,Morpho) (10 statements)
4. Appropriate time to make data available (6 statements)
5. Open Access (7 statements)
6. Actionable steps USU can do to help researchers (6 statements)
Findings – Participants don’t love data sharing
“Honestly, when I first heard that Federal agencies required
researchers to make all the raw data available to the public I
thought: “Really? This is useless and opens the door for a lot of
plagiarism.” I still feel that way a little bit. It’s one thing to have open
access journals where everybody has the chance to download the
paper, but making data available (even after the paper has been
published) is a little bit too risky… I don’t make any data available
until after the paper has been published.”
Findings – Participants appreciate current services
“What helps me? Betty!!! She is great! As
[another participant] mentioned, when I have
to upload anything, Betty is my first point of
contact. I know that I get excellent information
from her.”
Findings – What could help?
• Meet with the Data Librarian after receiving an award
• Checking to see what publications can be uploaded and updating PI records
accordingly
• Depositing data and publications on behalf of researchers
• Custom support for any agency requirements
• Sending reminders beyond the close of the award, until the PI notified the Data
Librarian that all data and publications related to the award have been deposited
• Help paying open access fees.
Summary
• Pioneering the USU compliance tracking program
• Identifying the need for a program assessment
o Goals
 Understanding researcher behavior
 Evaluating response to USU’s compliance tracking program
o Methods
 Survey
 OAFG
• Qualitative data analysis and findings
Lessons Learned
• Survey
o Data representative of a portion of researchers
 Convenience vs. Randomized – taking a randomized sample makes the data more representative of
the population
o Work with someone that has statistical expertise
• OAFG
o Timing and buy in
o Platform needs
How our results will shape future services
• Clarification or requirements expected by individual funders
o Continue updating to data management services website
o Continue to tailor reminder emails to include links to resources for each PI’s agency information about
requirements​​
o The library is currently changing the workflow to include additional help post close-out of awards
o Improved communication from the USU Libraries and Sponsored Programs
• Further investigation and research
o Improving interaction with researchers less inclined to think Sponsored Programs/USU Libraries can help
them meet agency requirements to deposit publications and data
Future of the Project
• Developing a culture of responsible data stewardship
o Understanding research hesitancy to share data
o Further sessions
 When data should be released
 Data citation
 Data licensing
 Effective methods for describing research data
Read our article:
Rebekah Cummings, Lindsay Ozburn, Andrea Payant, Betty Rozum, Michael Shelton & Ryan Bushman
(2020) Assessing Research Compliance for Federally Funded Projects: The Good, the Bad, and the
Publicly Accessible, Journal of Library Administration, 60:7, 726-751, DOI:
10.1080/01930826.2020.1786985
Learn More
Thanks!
Do you have any questions?
Rebekah Cummings
rebekah.cummings@utah.edu
Lindsay Ozburn
lindsay.ozburn@usu.edu
Andrea Payant
andrea.payant@usu.edu
Mike Shelton
michael.shelton@usu.edu
Ryan Bushman
ryan.bushman@usu.edu

More Related Content

PPTX
Can You Scan This For Me? Making the Most of Patron Digitization Request in t...
PPTX
Charting Communication: Assessment and Visualization Tools for Mapping the Co...
PPTX
Wisdom of the Crowd: Successful Ways to Engage the Public in Metadata Creation
PPTX
Assessment and Visualization Tools for Technical Services
PPTX
Where's the Data?
PPTX
Using Data to Help Tell Your Library's Story
PPTX
Outside In: Retooling Cataloging Outreach Efforts
Can You Scan This For Me? Making the Most of Patron Digitization Request in t...
Charting Communication: Assessment and Visualization Tools for Mapping the Co...
Wisdom of the Crowd: Successful Ways to Engage the Public in Metadata Creation
Assessment and Visualization Tools for Technical Services
Where's the Data?
Using Data to Help Tell Your Library's Story
Outside In: Retooling Cataloging Outreach Efforts

What's hot (20)

PPTX
MARC-y MARC and the Coding Bunch
PPTX
Crowdsourcing Metadata Practices at USU
PPTX
Relationship status: Libraries and linked data in Europe
PDF
Oehrli - Creating Data Literate Students
PDF
Read Surkis Facilitating Development of Research Data Services
PDF
Alexander - Education in the Internet of Everything
PPTX
NISO/BISG Changing Standards Landscape: EBook Discovery and Requirements for ...
PDF
Goldman "Collaboratively Build Data Science Services and Skills"
PPTX
Enhancing DMPTool: Further Streamlineing Data Mangement Planning Process
PPTX
Using the research lifecycle to design technology and support services
PPTX
Cook "Using Library Assessment Data in Senior Decision Making: A Perspective ...
PPTX
PPTX
Publisher of the Community: We're All Doomed
PDF
Emery - Buying Openly - NISO Sept 7
PPTX
Data Driven Libraries
PDF
Webb -- Libraries Supporting Media and Data Literacy
PPTX
Expert engagement: practical researcher digital literacy provision at City - ...
PPTX
Reveal Digital: innovative library crowdfunding model for open access digita...
PDF
Warren & Rauh Creating a Culture of Research Reputation
MARC-y MARC and the Coding Bunch
Crowdsourcing Metadata Practices at USU
Relationship status: Libraries and linked data in Europe
Oehrli - Creating Data Literate Students
Read Surkis Facilitating Development of Research Data Services
Alexander - Education in the Internet of Everything
NISO/BISG Changing Standards Landscape: EBook Discovery and Requirements for ...
Goldman "Collaboratively Build Data Science Services and Skills"
Enhancing DMPTool: Further Streamlineing Data Mangement Planning Process
Using the research lifecycle to design technology and support services
Cook "Using Library Assessment Data in Senior Decision Making: A Perspective ...
Publisher of the Community: We're All Doomed
Emery - Buying Openly - NISO Sept 7
Data Driven Libraries
Webb -- Libraries Supporting Media and Data Literacy
Expert engagement: practical researcher digital literacy provision at City - ...
Reveal Digital: innovative library crowdfunding model for open access digita...
Warren & Rauh Creating a Culture of Research Reputation
Ad

Similar to But Were We Successful: Using Online Asynchronous Focus Groups to Evaluate Library Services (20)

PDF
Slides | Targeting the librarian’s role in research services
PPTX
The benefits and challenges of open access: lessons from practice - Helen Bla...
PDF
Data-Informed Decision Making for Digital Resources
PDF
Data-Informed Decision Making for Libraries - Athenaeum21
PPTX
Case sStudy
PPTX
Stepping stones to ‘big data’: supporting quantitative methods teaching with ...
PDF
Okraku_Research Prospectus Outline_042016
PPT
Survey development
DOC
It resource needsassessment
PPTX
Ps rwebinar january2019final
PPTX
Social science research methods for libraries
PDF
SLIDES | High-impact library services and outreach: Student success * Systema...
PDF
RDAP 15 Co-circular RDM: A Pilot service for Graduate Students at the Univers...
PDF
Chu and Hensley "Strategic Planning and the Experience of the University of I...
PPTX
6 oct15 writing kmb plan edited
PDF
Trends in-connecting-research-sgd-2013
PPTX
Reviewing the role of teaching librarians in supporting student's digital cap...
PDF
RDAP14 Poster: Samantha Guss Data management planning and responsible conduct...
Slides | Targeting the librarian’s role in research services
The benefits and challenges of open access: lessons from practice - Helen Bla...
Data-Informed Decision Making for Digital Resources
Data-Informed Decision Making for Libraries - Athenaeum21
Case sStudy
Stepping stones to ‘big data’: supporting quantitative methods teaching with ...
Okraku_Research Prospectus Outline_042016
Survey development
It resource needsassessment
Ps rwebinar january2019final
Social science research methods for libraries
SLIDES | High-impact library services and outreach: Student success * Systema...
RDAP 15 Co-circular RDM: A Pilot service for Graduate Students at the Univers...
Chu and Hensley "Strategic Planning and the Experience of the University of I...
6 oct15 writing kmb plan edited
Trends in-connecting-research-sgd-2013
Reviewing the role of teaching librarians in supporting student's digital cap...
RDAP14 Poster: Samantha Guss Data management planning and responsible conduct...
Ad

More from Andrea Payant (14)

PPTX
Avoiding a Level of Discontent in Finding Aids: An Analysis of User Engagemen...
PPTX
On Your MARC, Get Set, Code!
PPTX
Let's Get Digital!
PPTX
The Missing Link: Metadata Conversion Workflows for Everyone
PPTX
Mitigating the Risk: identifying Strategic University Partnerships for Compli...
PPTX
Just Keep Cataloging: How One Cataloging Unit Changed Their Workflows to Fit ...
PPTX
Research Data Management at USU
PPTX
liwalaawiiloxhbakaa (How We Lived): The Grant Bulltail Absáalooke (Crow Natio...
PPTX
Homeward Bound: How to Move an Entire Cataloging Unit to Remote Work
PPTX
Memes of Resistance, Election Reflections, and Voices from Drug Court: Social...
PPTX
Giving Credit Where Credit is Due: Author and Funder IDs
PPTX
VOCAB for Collaboration: How “Work Language” Can Help You Win at Teamwork
PDF
Retooling Your Story: Using Visualizations to Demonstrate Your Impact
PPTX
ARK de Triumph: Linking Finding Aids & Digital Libraries Using a Low-Tech App...
Avoiding a Level of Discontent in Finding Aids: An Analysis of User Engagemen...
On Your MARC, Get Set, Code!
Let's Get Digital!
The Missing Link: Metadata Conversion Workflows for Everyone
Mitigating the Risk: identifying Strategic University Partnerships for Compli...
Just Keep Cataloging: How One Cataloging Unit Changed Their Workflows to Fit ...
Research Data Management at USU
liwalaawiiloxhbakaa (How We Lived): The Grant Bulltail Absáalooke (Crow Natio...
Homeward Bound: How to Move an Entire Cataloging Unit to Remote Work
Memes of Resistance, Election Reflections, and Voices from Drug Court: Social...
Giving Credit Where Credit is Due: Author and Funder IDs
VOCAB for Collaboration: How “Work Language” Can Help You Win at Teamwork
Retooling Your Story: Using Visualizations to Demonstrate Your Impact
ARK de Triumph: Linking Finding Aids & Digital Libraries Using a Low-Tech App...

Recently uploaded (20)

PDF
Journal of Dental Science - UDMY (2021).pdf
PDF
Uderstanding digital marketing and marketing stratergie for engaging the digi...
PPTX
What’s under the hood: Parsing standardized learning content for AI
PDF
Empowerment Technology for Senior High School Guide
PDF
Race Reva University – Shaping Future Leaders in Artificial Intelligence
PDF
FORM 1 BIOLOGY MIND MAPS and their schemes
PDF
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
PDF
Skin Care and Cosmetic Ingredients Dictionary ( PDFDrive ).pdf
PDF
English Textual Question & Ans (12th Class).pdf
PDF
LIFE & LIVING TRILOGY - PART - (2) THE PURPOSE OF LIFE.pdf
PPTX
ELIAS-SEZIURE AND EPilepsy semmioan session.pptx
PDF
MICROENCAPSULATION_NDDS_BPHARMACY__SEM VII_PCI .pdf
PDF
Hazard Identification & Risk Assessment .pdf
PPTX
Education and Perspectives of Education.pptx
PDF
Environmental Education MCQ BD2EE - Share Source.pdf
PPTX
A powerpoint presentation on the Revised K-10 Science Shaping Paper
PDF
Mucosal Drug Delivery system_NDDS_BPHARMACY__SEM VII_PCI.pdf
PDF
Paper A Mock Exam 9_ Attempt review.pdf.
PDF
Complications of Minimal Access-Surgery.pdf
PPTX
Core Concepts of Personalized Learning and Virtual Learning Environments
Journal of Dental Science - UDMY (2021).pdf
Uderstanding digital marketing and marketing stratergie for engaging the digi...
What’s under the hood: Parsing standardized learning content for AI
Empowerment Technology for Senior High School Guide
Race Reva University – Shaping Future Leaders in Artificial Intelligence
FORM 1 BIOLOGY MIND MAPS and their schemes
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
Skin Care and Cosmetic Ingredients Dictionary ( PDFDrive ).pdf
English Textual Question & Ans (12th Class).pdf
LIFE & LIVING TRILOGY - PART - (2) THE PURPOSE OF LIFE.pdf
ELIAS-SEZIURE AND EPilepsy semmioan session.pptx
MICROENCAPSULATION_NDDS_BPHARMACY__SEM VII_PCI .pdf
Hazard Identification & Risk Assessment .pdf
Education and Perspectives of Education.pptx
Environmental Education MCQ BD2EE - Share Source.pdf
A powerpoint presentation on the Revised K-10 Science Shaping Paper
Mucosal Drug Delivery system_NDDS_BPHARMACY__SEM VII_PCI.pdf
Paper A Mock Exam 9_ Attempt review.pdf.
Complications of Minimal Access-Surgery.pdf
Core Concepts of Personalized Learning and Virtual Learning Environments

But Were We Successful: Using Online Asynchronous Focus Groups to Evaluate Library Services

  • 1. But Were We Successful? Using Online Asynchronous Focus Groups to Evaluate Library Services Rebekah Cummings rebekah.cummings@utah.edu Lindsay Ozburn lindsay.ozburn@usu.edu Andrea Payant andrea.payant@usu.edu Mike Shelton michael.shelton@usu.edu Ryan Bushman ryan.bushman@usu.edu Utah Library Association Conference - May 20, 2021
  • 2. Why Was RDMS Created? Why did Utah State University create a research data management services program? RDMS was created to comply with federal mandates stemming from the 2013 OSTP Memo on “Increasing Access to the Results of Federally Funded Scientific Research.”
  • 3. 2013 OSTP Memo In 2013 the Office of Science and Technology Policy issued a memorandum titled: “Increasing Access to the Results of Federally Funded Scientific Research” • “The Administration is committed to ensuring that, to the greatest extent and with the fewest constraints possible and consistent with law and the objectives set out below, the direct results of federally funded scientific research are made available to and useful for the public, industry, and the scientific community. Such results include peer-reviewed publications and digital data.” • Stipulated that federal agencies with over $100M in R&D had to develop plans to make publications and data that originated from federally funded research freely available to the public.
  • 4. What This Meant for Researchers: ● The need to maintain and make their data publicly accessible, generally through data repositories. ● The need for their publications to be open access and to generally have them listed in the funder’s public access repository. ● The need to include information about any data and publications created with their annual reports to their funders. ● The inclusion of a data management plan with the award applications.
  • 5. Utah State University recognized several needs necessary to help Primary Investigators (PI) maintain compliance with funder mandates: USU’s Response to the OSTP Memo: • The need to create a Data Librarian position to assist researchers with their publications and data: Fulfilled in Summer 2015. • Bring campus stakeholders together to access needs and create a compliance program: In Autumn 2015 a cross- campus group with members from the Office of Research and Graduate Studies (RGS), the Office of Information Technology (IT), and the Merrill-Cazier Library started meeting to access how to best monitor and assist USU researchers’ compliance. • The needs to create and implement a University Research Data Policy and a process to create auditable records to facilitate access to data and publications: By Autumn 2016 a plan was implemented that included the process for award compliance tracking
  • 6. USU’s Approach to Compliance: USU decided to take an approach that facilitates the researchers’ ability to be in Funder Compliance USU worked to develop a compliance process that would: • Demonstrate to Funders that the University is making good faith effort to comply with mandates. • Emphasis to researchers that complying with funder mandates is an institutional priority. • Demonstrate to researchers that resources are available to support them in meeting funder mandates to make data openly accessible.
  • 8. Continued Support for RDMS from Administration Both the Library and University administration recognizes that compliance to funder mandates is an institutional priority. They continue to support RDMS in a variety of ways: • Monthly meetings of the University’s Research Data Management Committee, which is made up members from the Library and the Office of Research. These meeting serve to discuss the needs of both RDMS and the researchers, news in the data management field, and to develop instructional opportunities with the campus community. • Once a semester the Library (via RDMS’s supervising Associate Dean) sponsors Data Club, a working lunch where PIs are invited to visit with the RDMS staff and a representative from the Research Office to discuss their data concerns. • In February 2020, the Library and Research office co-sponsored RDMS in conducting a half-day workshop called “Datapalooza” geared toward teaching graduate students good data management practices. This was followed up in April 2021 with a second Zoom based Datapalooza event.
  • 9. Establishing the Need for Program Assessment •Association of American Universities (AAU) and the Association of Public & Land-Grant Universities (APLU) working group conference •Two members of RDMS, as well as the Interim Vice President of Research represented USU. •Outcome: recognizing the need to assess the quality and effectiveness of the University’s compliance program via RDMS’s efforts.
  • 10. Assessing the compliance of federal mandates: the library’s involvement Early 2010’s… • Federal organizations started requiring Data Management Plans to ensure research reproducibility • Research reproducibility becomes a hot topic in LIS, with libraries examining what role they might play in assisting their campus’ researchers with the reproducibility and accessibility of empirical research (Vitale, 2016) o This led to a shift in academia = libraries become increasingly more engaged in campus research efforts and support. • Emergence of RDMS programs and technical infrastructure as a support mechanism
  • 11. Assessing the compliance of federal mandates: the library’s involvement Late 2010’s… • Librarians and professional library orgs are more involved than ever in shaping policy and best practices for RDMS in support of researchers o Sparsely explored topic in LIS literature • This research addresses the gap in literature and assessment of RDMS efforts
  • 12. Our Assessment Methods • Traditional survey coupled with an online asynchronous focus group (OAFG) • Why two methods? Data triangulation o Survey gave us a broad look at what works and what could be improved. OAFG dug deeper into these questions to give us specifics for improvement • OAFG discussion questions modeled after survey questions and results
  • 13. Asynchronous Focus Groups as a Methodology: What are they? • Focus groups conducted through online mediums where participants do not need to be in the same place or online at the same time. o Typically, they are given a time frame (2 day, for example) in which to participate • Who uses them? o Marketing- generating consumer insights that inform response to advertising o Health - discussing healthcare issues with professionals across the globe (Han et al., 2019; Ramo et al., 2019)
  • 14. Pros and Cons of Asynchronous Focus Groups Pros: ● Not bound in time and space as much; as people can assemble to participate in whatever time frame you give them ● Often, they have the potential to be anonymous, which is helpful for sensitive topics ● A lack of geographic boundaries helps increase the diversity of a participant pool for specific types of research projects ● In academia, busy faculty schedules are less of a barrier for participation Cons: • Still dealing with the influence of social cues in focus groups but in a much more presumptive way (i.e., lack of vocal tone and eye contact can lead to assumptions about someone’s answer). • Previous forum posts can influence future posts Why? While this is the case when speaking, participants get more time to craft their answers in online forums, leading to a higher potential for influencing factors • Ethical considerations of cyberspace: anonymity could lead to people misrepresenting who they are and/or sharing other participant’s responses with others.
  • 15. How did we use OAFG to assess our program? • We used Canvas, an online learning platform that both USU and U of U had access to. • Of the team, we used the librarian at U of U as an outside moderator to encourage honesty and to scrub data of identifiable information. o Outside moderators are important with focus groups, in general, to prevent unintentional biases in questions and to encourage honest answers from participants.
  • 16. Hold Up. You were assessing participants at one location Why did you use an OAFG?
  • 17. Pros and Cons of Asynchronous Focus Groups Pros: ● Not bound in time and space as much; as people can assemble to participate in whatever time frame you give them ● Often, they have the potential to be anonymous, which is helpful for sensitive topics ● A lack of geographic boundaries helps increase the diversity of a participant pool for specific types of research projects ● In academia, busy faculty schedules are less of a barrier for participation Cons: • Still dealing with the influence of social cues in focus groups but in a much more presumptive way (i.e., lack of vocal tone and eye contact can lead to assumptions about someone’s answer). • Previous forum posts can influence future posts Why? While this is the case when speaking, participants get more time to craft their answers in online forums, leading to a higher potential for influencing factors • Ethical considerations of cyberspace: anonymity could lead to people misrepresenting who they are and/or sharing other participant’s responses with others.
  • 18. Qualities of Survey • Population: 78 Researchers • Sample: 28 Researchers o Initial response of 32 researchers ( 41.03% of the population) o Data cleaning determined 28 of respondents provided usable data (34.6% of population) • Introduction of Bias o We opted for a convenience sample rather than a randomized sample
  • 19. The Survey Two styles of question: • Questions designed to determine the frequency of a behavior of the researcher often used the following scale o Never o 0-3 times (a redundancy we would like to fix in further research projects) o 4-7 times o 8-10 times o More than 10 times • Questions designed to determines a researcher's response to USU’s efforts used the following scale o Strongly Disagree o Somewhat Disagree o Neither Agree Nor Disagree o Somewhat Agree o Strongly Agree
  • 20. Key Statistics • 19 of the 27 (70.4%) respondents indicated they had three or fewer awards that required deposit of publications into a public access repository • 18 of the 27 (66.67%) respondents indicated they had three or fewer awards that required deposit of data into a publicly accessible repository. • “Could Sponsored Programs or the Library better facilitate efforts to meet agency mandates?” o 13 of 27 (48.15%) responded yes o 14 (51.85%) responded no. o When examined by years as a researcher, a higher percentage of respondents in the range of 11-20 years as a researcher responded “yes” to this question.
  • 21. Key Statistics: Behavior of the Researcher • Motive for making data publicly available o 84.6% to comply with journal publishing requirements o 92.6% to satisfy a funder o 61.5% with the desire for their data to be public • 92.59% of respondents indicated that they anticipated writing journal publications up to 5 years after the close of their award • 88.46% of respondents anticipated needing up to 4 years after the close of their award to have final data ready for deposit in a public repository
  • 22. Key Statistics: Response to USU’s Efforts • The email from SPO was helpful o 20 (74.07%) strongly agree o 5 (18.52%) somewhat agree • Researchers understand how to respond to the email from SPO o 13 (48.15%) strongly agree o 7 (25.93%) somewhat agree o 5 (18.52%) somewhat disagree • The email from SPO explains how they could find more help o 16 (59.26%) strongly agree o 8 (29.63%) somewhat agree o 2 (7.41%) somewhat disagree • The email from DL was helpful o 11 (64.71%) strongly agree o 5 (29.41 % somewhat agree o 1 (5.88%) somewhat disagrees • Researchers understand how to respond to email from DL o 11 (64.71%) strongly agree o 4 (23.53%) somewhat agree o 2 (11.76%) neither agree nor disagree • The email from DL explains how they could find more help o 11 (64.71%) strongly agree o 5 (29.41%) somewhat agree o 1 (5.88%) somewhat disagree
  • 23. Online Asynchronous Focus Groups (OAFG) SOCIALLY DISTANCED BEFORE IT WAS COOL
  • 24. • SECURE • FAMILIAR • ALLOWED FOR DISCUSSION • RESPOND IN LINE • TECHNICAL SUPPORT Technology
  • 25. Five Discussions Introductions Response to funder mandates Researcher data practices & publication deposit Research Office and Library program What would help?
  • 26. First Attempt • Selected 15 participants at random • VP for Research invited the fifteen participants to join the OAFG at their convenience the following day • With only a single day’s notice, only two of the fifteen participants responded, and they, inexplicably, couldn’t get into the Canvas site.
  • 27. Second Attempt • Moved Canvas site from University of Utah to USU • Gave participants several weeks of advance notice • Confirmed attendance with five participants, but only three showed up • Good news! The three participants gave excellent, detailed feedback, interacted with one another, and didn’t have any technical issues
  • 28. Qualitative Data Analysis 1. Praise for the Data Librarian, the library, and the research office (16 statements) 2. Data sharing challenges (6 statements) 3. Data resources (such as DMP Tool, FigShare, Digital Commons, Metavist,Morpho) (10 statements) 4. Appropriate time to make data available (6 statements) 5. Open Access (7 statements) 6. Actionable steps USU can do to help researchers (6 statements)
  • 29. Findings – Participants don’t love data sharing “Honestly, when I first heard that Federal agencies required researchers to make all the raw data available to the public I thought: “Really? This is useless and opens the door for a lot of plagiarism.” I still feel that way a little bit. It’s one thing to have open access journals where everybody has the chance to download the paper, but making data available (even after the paper has been published) is a little bit too risky… I don’t make any data available until after the paper has been published.”
  • 30. Findings – Participants appreciate current services “What helps me? Betty!!! She is great! As [another participant] mentioned, when I have to upload anything, Betty is my first point of contact. I know that I get excellent information from her.”
  • 31. Findings – What could help? • Meet with the Data Librarian after receiving an award • Checking to see what publications can be uploaded and updating PI records accordingly • Depositing data and publications on behalf of researchers • Custom support for any agency requirements • Sending reminders beyond the close of the award, until the PI notified the Data Librarian that all data and publications related to the award have been deposited • Help paying open access fees.
  • 32. Summary • Pioneering the USU compliance tracking program • Identifying the need for a program assessment o Goals  Understanding researcher behavior  Evaluating response to USU’s compliance tracking program o Methods  Survey  OAFG • Qualitative data analysis and findings
  • 33. Lessons Learned • Survey o Data representative of a portion of researchers  Convenience vs. Randomized – taking a randomized sample makes the data more representative of the population o Work with someone that has statistical expertise • OAFG o Timing and buy in o Platform needs
  • 34. How our results will shape future services • Clarification or requirements expected by individual funders o Continue updating to data management services website o Continue to tailor reminder emails to include links to resources for each PI’s agency information about requirements​​ o The library is currently changing the workflow to include additional help post close-out of awards o Improved communication from the USU Libraries and Sponsored Programs • Further investigation and research o Improving interaction with researchers less inclined to think Sponsored Programs/USU Libraries can help them meet agency requirements to deposit publications and data
  • 35. Future of the Project • Developing a culture of responsible data stewardship o Understanding research hesitancy to share data o Further sessions  When data should be released  Data citation  Data licensing  Effective methods for describing research data
  • 36. Read our article: Rebekah Cummings, Lindsay Ozburn, Andrea Payant, Betty Rozum, Michael Shelton & Ryan Bushman (2020) Assessing Research Compliance for Federally Funded Projects: The Good, the Bad, and the Publicly Accessible, Journal of Library Administration, 60:7, 726-751, DOI: 10.1080/01930826.2020.1786985 Learn More
  • 37. Thanks! Do you have any questions? Rebekah Cummings rebekah.cummings@utah.edu Lindsay Ozburn lindsay.ozburn@usu.edu Andrea Payant andrea.payant@usu.edu Mike Shelton michael.shelton@usu.edu Ryan Bushman ryan.bushman@usu.edu