Article

 

Evidence Based Practice Using Formative Assessment in Library Research Support

 

Jackie Wolstenholme

Research Services Librarian

Information & Research Services

Eddie Koiki Mabo Library

James Cook University

Townsville, Australia QLD 4811

Email: jackie.wolstenholme@jcu.edu.au

 

Received: 28 Jan. 2015    Accepted: 17 Jun 2015

 

 

cc-ca_logo_xl 2015 Wolstenholme. This is an Open Access article distributed under the terms of the Creative CommonsAttributionNoncommercialShare Alike License 4.0 International (http://creativecommons.org/licenses/by-nc-sa/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly attributed, not used for commercial purposes, and, if transformed, the resulting work is redistributed under the same or similar license to this one.

 

Abstract

 

Objective – The purpose of this study was to develop and review the effectiveness of a new evidence-based approach for teaching library research support.

 

Methods – Formative assessment, through two variations of the One Minute Paper model, is used to poll the experiences of university researchers in library research support sessions. Prior to a session, Polling One Minute Papers (POMPs) assess what researchers know about topics that will be covered in the session. After a session, Reflective One Minute Papers (ROMPs) review whether university researchers achieved the intended learning outcomes of the session. POMPs were used for 16 sessions and ROMPs were used for a subset of 11 of these sessions. Examples of responses from the POMPs and ROMPs were presented to describe and analyse the effectiveness of this approach for library support of research.

 

Results – POMP and ROMP responses were remarkably informative given their simplicity and the little effort required on the part of the instructing librarian or researchers. The completion rate of POMPs was 72.7%. They gave researchers the opportunity to self-assess their current level of knowledge or skills about the topic to be covered in the upcoming session. The librarian could then tailor the session content to this level of knowledge. POMP responses were shared as part of the session content, enabling researchers to benchmark themselves against their peers. Completion rate of ROMPs was 20.9%, with the level of reflection in the individual researchers’ responses varying from shallow to insightful. Deeper responses stated how the researchers would use what they learned or pose new questions which emerged from their learning.

 

Conclusion – Polling One Minute Papers (POMPs) and Reflective One Minute Papers (ROMPs) are an effective and efficient approach for guiding the learning of researchers and closing the feedback loop for librarians. These tools extend the opportunity for librarians to engage with researchers and, through tailoring of session content, assist to maximise the benefit of library research support sessions for both librarians and researchers. Sharing of POMP and ROMP responses can assist librarians to coordinate the teaching of the researchers that they support. At an institutional level, evidence in POMPs and ROMPs can be used to demonstrate the value that the library has contributed to improving awareness and performance of its researchers.


Introduction

 

Researchers in universities are working in an increasingly complex and competitive environment (e.g., Frances, Fletcher, & Harmer, 2011; Kennan, Corrall, & Afzal, 2014; Richardson, Nolan-Brown, Loria, & Bradbury, 2012). Factors driving these changes include Internet and digital technologies and greater accountability through performance management and institutional benchmarking. These changes are requiring researchers to adapt faster than most would achieve through their traditional discipline-based networks, including information sharing among colleagues.

 

The Internet and digital technologies have transformed scholarly communication. Research outputs, although still published as books and journals, are now also made available in an array of other digital options including blogs and other social media, multimedia formats, and data files which may be displayed through sophisticated visualization tools. The numbers of research outputs have vastly increased and are distributed through a growing range of publishing models, many offering some form of Open Access. Researchers, as creators of research outputs, need to consider copyright and licensing for managing their rights, in balance with maximizing accessibility to their research outputs. The quality of publishers also needs to be assessed, to ensure that researchers avoid unethical publishers (e.g., see Beall, 2014).

 

To measure and benchmark performance, researchers and their institutions rely on citation ranking metrics. Researchers need to understand how these metrics are calculated and how citation indexes (e.g., the h-Index) are calculated. Researchers are also expected to have an online presence, ideally as a professional profile to promote their research interests and achievements. Altmetrics are emerging as an additional measure of impact, by measuring the online activity of a researcher or their outputs (Priem, Taraborelli, Groth, & Neylon, 2010).

 

University libraries can assist researchers to work in this environment and make the most of emerging opportunities. To provide this support, university libraries are moving their core business from provision of information resources to provision of information services and information solutions (Association of College and Research Libraries, 2010; Kaufman, 2009; Parsons, 2010). Information resources have traditionally involved the development and management of collections. In contrast, information services and solutions include infrastructure such as repositories (Simons & Richardson, 2013) and instructional support on topics such as scholarly and open access publishing, managing research data, maximizing research visibility, and measuring research performance (e.g., Auckland, 2012; Haddow, 2012; Kennan, et al., 2014).

 

Effective communication skills are essential for building a rapport with researchers and providing a valued service (Auckland, 2012; Creaser & Spezi, 2013; Parker, 2012). Research support librarians need to be confident in talking about the range of topics that researchers need to learn, match the information they provide with the skill level of the researcher, and explain the information in a way that is understandable for the researcher. This study investigates a teaching and learning approach which can assist librarians to achieve this.

 

Literature Review

 

Teaching and Learning in Library Research Support

 

Teaching and learning for researchers is best suited to the learning theory of andragogy. This theory is based on the assumption that adults are self-directed learners who are interested in immediate application of knowledge (Merriam, 2001). According to this theory, adults take the initiative in diagnosing their learning needs, including formulating objectives, identifying resources, implementing strategies, and evaluating outcomes (Knowles, 1975). Other learning theories of relevance to library research support include problem-based learning (Knowles, 1975), experiential learning (Kolb, 1984), and informed learning (Hughes & Bruce, 2012). In problem-based and experiential learning, learners draw on their prior knowledge and experience (Brodie, 2012), enabling learning to be built on a researcher’s existing practices. Informed learning describes how learners develop flexibility and confidence to use information in constantly evolving information environments, shifting the focus of information literacy education from mastering skills to learning to use information critically, ethically, and creatively (Hughes & Bruce, 2012).

 

In addition to self-directed learning, another key aspect of researcher learning is that researchers frequently learn from their peers. The peer is a defining figure in research practice. For example, it is implicit in the institution of “peer review” (Boud & Lee, 2005). As described for higher degree research students, research learning can be usefully construed in terms of entry into communities of practice, where peer learning becomes a powerful tool for describing and developing a rich understanding of the learning resources available (Boud & Lee, 2005). Peer interaction can enhance learning by stimulating the production of deeper thought through the desire to know what a colleague knows, prompting self-assessment and clarification of uncertainties (Draper, 2009).

 

Assessment in Library Research Support

 

Library research support needs to contribute to improving research performance without adding additional burden to a researcher’s workload. Researchers operate in a constant environment of research performance assessment (Parker, 2012), e.g., through funding or promotion applications, performance management acquittal, or as part of institutional assessment exercises such as those in Australia and the United Kingdom (Australian Research Council, 2014; REF2014, 2014). Research librarians need to be acutely aware of this research assessment landscape (Parker, 2012). The learning needs of researchers must form the core content of library research support material, in terms of what researchers need to know as well as their current status of understanding a particular topic.

 

Within a teaching and learning framework, assessment of the performance of researchers is analogous to summative assessment, and library support is analogous to formative assessment. Summative assessment tasks focus a student’s learning on “what counts,” while formative assessment provides a fine tuning mechanism which guides the learner’s learning progress (Boud, 2000). Summative assessment occurs after the learning process for the purpose of certification (Sadler, 1989). In contrast, formative assessment occurs as part of the learning process. Through formative assessment, the learner gains feedback which is intended to shape and improve their learning, leading to independent learners who are able to self-monitor their learning needs (Sadler, 1989). To implement this analogy of formative assessment in library research support, research librarians will be most effective if they develop teaching materials which incorporate the established practices of researchers’ self-directed and peer learning.

 

Learning is a cyclical process, as explained in experiential learning theory (Kolb, 1984). In the context of library research support, there is no starting point in the researchers’ learning process, but rather, they build on what they already know or have experienced. Within the learning cycle of assessment (Figure 1, Crisp, 2009), there may be many feedback loops between the phases of diagnostic, learning, and formative assessment (Sadler, 1989). Often, formative assessment leads to summative assessment, as researchers take on tasks to advance their careers or to meet institutional requirements.

 

Research Framework: Formative Assessment of Researchers

 

The One Minute Paper (OMP) is a formative assessment tool that has been successful in improving the teaching of, and learning by, undergraduate students (Bartlett & Morrow, 2001; Chizmar & Ostrosky, 1998). OMPs are effective for gaining student feedback in return for a modest amount of student and instructor effort (Bartlett & Morrow, 2001; Chizmar & Ostrosky, 1998; Drummond, 2007; Stead, 2005). The OMP is a questionnaire which asks:

 

1.     What was the most important thing you learned today?

2.     What was the most confusing point in today’s lecture?

3.     What was the most interesting fact that you learned today?

 

 

Figure 1

Relationship between diagnostic, formative, and summative assessments (redrawn from Crisp, 2009)

The addition of a third question was recommended by Bartlett and Morrow (2001:

 

 

The OMP benefits both instructors and students, regardless of their teaching or learning ability (Chizmar & Ostrosky, 1998). OMPs can provide specific and immediate feedback to the instructor about student learning, helping to set the pace and content of future instruction. This is useful for inexperienced instructors or instructors of new material (Stead, 2005), as is often the case in library research support. Instructors can also use the feedback to identify and then address misconceptions (Bartlett & Morrow, 2001). This closing of the feedback loop demonstrates that the instructor values student opinion and encourages students to actively contribute to their own learning experience (Stead, 2005). Class discussion of issues raised in OMP’s have reassured students by enabling them to benchmark their learning against their classmates, often revealing that the problems that others are experiencing are the same as their own (Bartlett & Morrow, 2001).

 

The simplicity of the OMP makes it an ideal tool for identifying the learning needs and learning outcomes of researchers. Content of library research support sessions can then be tailored accordingly. The OMP is typically assigned at the end of a class, but could also be adapted for implementation prior to a class (Stead, 2005). Pre-class formative quizzes encourage students to think critically about course content prior to a session (Dobson, 2008), offering the benefits of identifying current learning needs or learning gaps, providing an indication of what will be covered in the upcoming session, and creating an opportunity for self-assessment.

 

Aims

 

This study describes a method, adapted from formative assessment in teaching and learning, to assist research support librarians to develop an evidence-based foundation to support their teaching. Two variations of the One Minute Paper (OMP), Polling OMPs and Reflective OMPs, are developed in this study. A case study approach, from a series of multiple workshops on a range of topics, is used to investigate whether the two variations of the OMP are an effective and efficient approach for guiding the learning of researchers and closing the feedback loop for librarians.

 

Questions asked are:

 

1.       Do POMPs stimulate researcher engagement and interest?

2.       Can POMPs identify learning needs of researchers?

3.       Are POMPs or ROMPs effective tools for gaining feedback about researcher learning?

 

Methods

 

This study reports on the outcomes of a series of case studies, to explore the effectiveness of using the One Minute Paper (OMP) model for the purpose of library research support sessions. Two variations of the OMP were developed in this study: POMPs, i.e., Polling One Minute Papers, and ROMPs, i.e., Reflective One Minute Papers.

 

Both the POMPs and ROMPs are intended as tools which guide the formative learning of researchers. POMPs were distributed prior to a session and ROMPs were distributed after a session. Sessions were organized in response to specific requests from researchers on behalf of a research group, rather than according to a specific schedule.

 

Overview of the James Cook University Research Profile

 

The OMPs described in this study were developed for library research support sessions at James Cook University (JCU). The Strategic Intent of JCU is to create a brighter future for life in the tropics world-wide (James Cook University, 2015).The number, distribution and turnover of JCU researchers make it challenging to identify and meet their evolving library research support needs. There are approximately 2600 academic staff and more than 600 Higher Degree Research students (James Cook University, 2014) across Townsville, Cairns, Singapore and other smaller, regional centres.

 

Research needs vary with disciplinary research practices and career stage. At JCU, the largest and fastest growing area of research is in the medical disciplines, with many of these researchers having a strong applied knowledge but limited research experience. In contrast, internationally recognized researchers in the biological and environmental sciences tend to have metrics-driven library support needs. Humanities and social sciences, including Law and Creative Arts, have the most discipline-centric research needs. Career stage also influences library research support requirements: postgraduate students and early career researchers need to develop their research skills, mid-career researchers may be concerned about keeping up with technological changes, and senior researchers may be under pressure to maintain their high research standing.

 

Polling One Minute Papers (POMPs)

 

POMPs are a self-assessment tool. The questions asked in a POMP were structured around the topic of a library research support session, polling researchers to gauge their understanding of the topic. Session content was then tailored for this level of understanding. The questions and response options in POMPs are listed in the Appendix. POMPs were also intended to promote a session and stimulate interest about the content that would be covered in that session.

 

Sessions were organized in collaboration with research leaders, e.g., key researchers or research managers. This strategy helped to increase attendance and facilitate discussion because participants shared common research interests and usually knew each other prior to the session. POMPs were distributed approximately one week prior to a session in an email. This email was sent by the research leader to all researchers who he felt should attend the session. POMPs were voluntary, with participants being encouraged to submit their responses prior to the session. A summary of the POMP results were shared during the corresponding session to enable peer benchmarking and therefore further self-assessment. POMP responses were presented as descriptive data in tables, histograms or pie charts.

 

Reflective One Minute Papers (ROMPs)

 

ROMPs are a feedback tool which encouraged voluntary reflection about a session. The same three questions were asked in all ROMPs:

 

1.       What was the most important thing you learnt?

2.       What was the most confusing thing I covered?

3.       What was the most interesting thing you learnt?

 

ROMPs were completed on a voluntary basis. By responding to these three questions, researchers were able to provide feedback about the session and their learning to the librarian. ROMPs were developed partway through this project in order to close the formative assessment loop. ROMPs were used for 11 sessions, as listed in Table 1. A link to the three ROMP questions was usually distributed on the last slide of a session presentation or immediately after a session. In one exception (Session 16), the ROMP link was sent out 2 weeks after the session. ROMP responses were in an unstructured, free text format. Thematic analysis, also known as analytic coding, was used to interpret the responses and quantify them according to themes (Guest, MacQueen, & Namey, 2012; Richards, 2015).

 

Data Collection

 

POMPs and ROMPs were created in Google Forms (Google, 2014). The selection of Google Forms was based on a number of criteria: Google Forms is a free service with no limit on the number of questions that can be asked; checklists of multiple options per question can be selected, enabling quick standardized responses; and, the OMPs could be shared through an online link, enabling participation by both local and remote researchers.

 

POMPs and ROMPs were anonymous, a decision based on the presumption that anonymity would make researchers more likely to provide honest responses and therefore enable more realistic benchmarking amongst peers (Dillman, Smyth, & Christian, 2009). Respondents were also not required to log in using Google Forms, building researcher confidence that the OMPs were anonymous.

 

Results

 

The results report on a series of case studies to explore the effectiveness of POMPs and ROMPs to respond to the three questions stated in the Aims.

 

Response Rates of POMPs and ROMPs

 

The numbers of researchers attending the sessions for which POMPs and ROMPs were distributed are summarized in Table 1.

 

The completion rate of POMPs was high, at 72.7%, i.e., 136 from a pool of 187 researchers who attended the 16 sessions. A likely factor contributing to this high rate of completion was the simplicity of the POMP form. In one click from a link in an email, it was immediately evident what the researcher needed to respond to. The entire POMP could be viewed on a desktop screen without scrolling, visually emphasizing that the form would be quick to complete, with the format of all or most responses being checklists. Promoting the POMP as a “1 minute quiz” with endorsement from a research leader was also effective. For example, the Director of a research centre for one of the Life and Environmental Science sessions included the following statement of support in his email:

 

As part of the prep for the planning day, and to help with our understanding of the use of Research Profiles, can I ask you to take this 1 minute quiz? I took it and it took even less than 1 minute.

 

The completion rate of ROMPs was lower, at 20.9%, i.e., 23 from a pool of 110 researchers who attended the 11 sessions for which ROMPs were distributed. A major factor contributing to this low response rate may have been that the response format was free text.

 

Researchers were from the disciplines of Life and Environmental Science, Health Science, Social Science and Humanities, and Mixed Disciplines (Table 1). The librarian was present in the same room with participants for sessions 1-10, 15 and 16. Sessions 11-14 were conducted remotely via videoconference (Table 1). Sessions 1-15 were presented by the author. Session 16 was presented by another research support librarian who provided peer feedback about the value of OMPs.

 

 

Table 1

Summary of Attendance, POMPs and ROMPs for each Session

Session #

Discipline / Client Group

Session Topic

Remote*

Session date

Attendance

# POMPs submitted

% POMPs submitted

# ROMPs submitted

% ROMPs submitted

1

Life and Environmental Sciences: Post-Docs

Altmetrics

No

9 October 2013

23

17

73.9

----

----

2

Life and Environmental Sciences

Managing research profiles

No

11 October 2013

28

28

100.0

----

----

3

Life and Environmental Sciences

Research profile update and management

No

12 August 2014

8

9

112.5

6

75.0

4

Life and Environmental Sciences

Research profile update and management

No

29 August 2014

8

6

75.0

1

12.5

5

Life and Environmental Sciences

Research profile update and management

No

19 September 2014

5

4

80.0

1

20.0

6

Life and Environmental Sciences

Research profile update and management

No

9 October 2014

4

4

100.0

0

0.0

7

Life and Environmental Sciences

Research profile update and management

No

10 October 2014

4

2

50.0

0

0.0

8

Life and Environmental Sciences

Research profile update and management

No

24 November 2014

9

8

88.9

0

0.0

9

Health Sciences

Publishing academic research

No

23 October 2013

11

6

54.5

----

----

10

Health Sciences

Research and social media

No

27 August 2014

14

15

107.1

4

28.6

11

Health Sciences

Quality publishing

Yes

28 August 2014

3

2

66.7

3

100.0

12

Social Sciences and Humanities

Researcher identifiers

Yes

9 September 2014

2

3

150.0

2

100.0

13

Social Sciences and Humanities

Altmetrics

Yes

18 September 2014

5

4

80.0

4

80.0

14

Social Sciences and Humanities

Promoting and maximising research impact

Yes

2 December 2013

2

6

300.0

----

----

15

Mixed disciplines: Academic teaching staff

Research impact and publishing

No

13 February 2014

13

9

69.2

----

----

16

Mixed disciplines: Higher Degree Research students

Intellectual Property & Copyright

No

2 and 9 October 2014**

48

13

27.1

2

4.2

Total Attendance and number of POMPs/ROMPs submitted for all sessions

187

136

 

23

 

 

 

Figure 2

Display of results for the POMP question "Have you used any of the following to promote or discuss your research?" (Responses from 78 submitted POMPs distributed to 89 researchers, compiled progressively and presented in sessions 1-8).

 

 

Aim 1: Do POMPs Stimulate Researcher Engagement and Interest?

 

POMPs enhanced the engagement of researchers. Completion of a POMP and then the sharing of POMP results provided researchers with two opportunities for self-assessment. Firstly, POMPs could provide additional incentives to go to a session if researchers identified gaps in their knowledge or skills based on questions asked in the POMP. In at least one case, a researcher started to use some of the tools listed in the POMP prior to the session. Secondly, in sessions, researchers showed great interest in the activities and tools being used by their peers. Figure 2 displays the combined responses for eight sessions on the same topic (Managing research profiles) for researchers from the same discipline (Life and Environmental Sciences). The cumulative responses were shared in each of the eight sessions. The opportunity to benchmark against peers and resulting discussion gave endorsement to the advice provided by the librarian. For example, one researcher reported that he now has a new international collaborator with substantial funding through a LinkedIn™ connection. Hearing how their peers were using such tools gave context-relevant evidence, making it easier to sell the concept of using social media tools in a research context. Researchers were also able to see which tools were widely used by their peers, giving an indication of where to get started, or a confidence boost if they were already using those tools. Learning about lesser used tools such as The Conversation (http://theconversation.com/au) gave researchers ideas for how they could increase awareness of their research.

 

Aim 2: Can POMPs Identify Learning Needs of Researchers?

 

POMPs were effective and efficient in identifying current learning needs of researchers and what the focus of the session content should be prior to a session. POMP responses indicated what researchers understood or were doing well and also gaps in their current knowledge or activities.

 

Responses from Life and Environmental Science researchers attending one of the eight sessions represented in Figure 2 (Session 2, as listed in Table 1) indicated that participants in this session were using some online tools and almost half had used traditional media to communicate and promote their research (Figure 3). This suggested that they were generally aware of the importance of communicating and promoting their research. Despite this apparent awareness, the moderate to nil use of more than half the listed tools (Figure 3) also highlighted areas for further instruction.

 

 

Figure 3

Display of results for the POMP question "Have you used any of the following to promote or discuss your research?" (Responses from 28 submitted POMPs distributed to 28 researchers, presented in session 2).

 

 

As a further example of how POMPs were useful for identifying learning needs, Health Science and Life and Environmental Science researchers were mostly aware of whether an article processing charge (APC) had been paid for their article, but were less clear about whether they had signed copyright ownership over to the publisher. Approximately one third of responding researchers from each discipline had authored a paper in which an APC had been paid and less than 10% were not sure (Figure 4). For authors who had paid an APC, a serious issue for discussion was highlighted: eight of the Life and Environmental Science researchers (Figure 5) and the four Health Science researchers who indicated they had paid an APC were not sure if they had retained copyright ownership of their work. This finding highlighted the need to focus on the importance of understanding the conditions of a publisher copyright agreement in the session. This was pertinent given that ResearchGate™, a site which facilitates the sharing of research publications, was a frequently used social media tool (Figures 2 and 3).

 

 

Figure 4

Relative proportions of researchers, based on POMP responses, who have paid an Article Processing Charge (APC) to make an article Open Access; Health Science (15 responses, Session 10) and Life and Environmental Science (33 responses, Sessions 3-8)

 

 

Figure 5

Display of results for the POMP questions 5a: "Have you or a co-author paid an Article Processing Charge to make any of your articles open access?" and 5b "If you have paid an Article Processing Charge, do you know if you retained copyright ownership of your article?”. (Responses from 38 submitted POMPs distributed to 38 researchers, compiled progressively and presented in sessions 3-8).

 

 

Despite the anonymity of POMP responses, examining answers to each question from individual responses was also useful for identifying learning needs. In relation to scholarly publishing, researchers attending session 9 each had some publishing experience, mostly with journals and conferences (Table 2). However, only three researchers were familiar with the Australian Government Higher Education Research Data Collection (HERDC) (Department of Education, 2014) and Excellence in Research for Australia (ERA) (Australian Research Council, 2014). Although this group of researchers was aware of the importance of publishing, most were not using the HERDC and ERA specifications, a form of summative assessment for researchers in Australian universities, to guide their publishing decisions. Session content was therefore adapted to explain the specifications at an introductory level. Similarly, of the 12 researchers in Figure 3 who had used traditional media to promote their research, only 6 identified ResearchOnline@JCU, the institutional repository, as a tool for communicating and promoting their research. This identified a point for discussion in the session, revealing that some researchers had only considered the institutional repository as an administrative reporting tool and not an avenue for communicating and promoting their research.

 

 

Table 2

POMP Responses from Health Science Researchers about Their Experience in Publishing Academic Research*

Have you published any of the following types of journal articles?

Have you published any of the following types of conference works?

Have you published any of the following types of books or book chapters?

Have you co-authored other types of works that are derived from your research?

Familiarity with HERDC or ERA

Use of journal ranking tools

Peer reviewed article

Peer reviewed paper, Poster, Abstract or summary

 

 

ERA: Excellence in Research for Australia

 

Peer reviewed article, Short note or commentary

Abstract or summary

 

 

 

Journal Citation Reports

Peer reviewed article, Non-refereed article, Case study

Peer reviewed paper, Non-refereed paper, Poster

 

 

 

 

Peer reviewed article, Non-refereed article, Short note or commentary

Poster, Abstract or summary, Edited a conference proceedings

Teaching material

 

HERDC: Higher Education Research Data Collection, ERA: Excellence in Research for Australia

Journal Citation Reports

Peer reviewed article, Non-refereed article, Short note or commentary

Abstract or summary

Non-commercial

 

HERDC: Higher Education Research Data Collection

Journal Citation Reports

Non-refereed article

 

 

 

 

Journal Citation Reports, Scopus analytics

 

 

Aim 3: Are OMPs an Effective Tool for Gaining Feedback about Researcher Learning?

 

Polling One Minute Papers (POMPs)

 

POMPs created an opportunity to learn about the audience prior to an upcoming session, freeing up time in a session for discussion and teaching additional content. POMPs also created the opportunity to ‘hear’ from researchers who would normally not speak up in a session or who were only prepared to share information anonymously. Even when the number of responses was low, due to a small number of researchers attending a session, POMPs were still useful because all or most of the audience responded.

 

POMP responses were useful for refining how sessions were taught. For example, only a small proportion (14.7%) of researchers indicated that they had a good understanding of the term altmetrics, but all who selected this option provided an appropriate description of the term (Table 3). In contrast, of the 75 researchers who responded to this question, 42% indicated they had no understanding and 22% indicated they had some understanding of altmetrics. Given this range, discussion was encouraged in sessions to facilitate learning from peers who provided explanations that other session participants could easily relate to.

 

POMPs created an unexpected insight from one group of remote researchers (session 14). Enquiry as to why only one researcher had responded to the POMP revealed cultural differences as the underlying issue. The group’s research manager explained that the researchers were embarrassed that they had no or few publications and so were reluctant to respond to the POMP. This information was useful in itself as it indicated that the session content needed to be directed at getting the researchers started with publishing. Further explanation about the POMP boosted the response rate to 6, although only 2 researchers attended the session. These insights will be considered in the planning of future sessions with this group of researchers.

 

Feedback from a colleague who trialed the use of a POMP (session 16) reported that it was very interesting and useful to see the range of disciplines and prior knowledge of the researchers who had registered for the session. In this POMP, respondents were asked to table any prior questions, which gave the librarian time to prepare for complex questions.

 

Reflective One Minute Papers (ROMPs)

 

ROMP responses gave an insight into what researchers gained from their session. They were used by the librarian to determine whether intended learning outcomes were achieved, and consider how the teaching or content of future sessions could be improved. ROMPs were particularly effective in facilitating reflective feedback from remote sessions with small numbers of participants. In discussions at the end of sessions with 2-5 participants (Sessions 11-13), responses were received from all or most participants. In some cases, immediate clarification was given in the session by the librarian. For more complex issues, future sessions were offered as a response. Although response rates were lower for larger sessions, the ROMPs were still beneficial for encouraging reflection, receiving feedback, and continuing the conversation with session participants. ROMPs were least effective for Sessions 4-8, which were hands-on computer sessions with substantial discussion and feedback throughout the sessions. When asked to complete the ROMP, the researchers repeated comments they had made during the session, but only 2 of 30 recorded their feedback in a ROMP. Feedback from these hands-on sessions was mostly positive, with the exception of one researcher who was frustrated with the work she needed to do to manage her online presence.

 

Most researchers responding to the ROMPs gave a response for each of the three questions. Within the 23 ROMP responses, 17 researchers responded to all 3 questions, 4 responded to 2 questions, and 2 responded to only 1 question. Using thematic analysis, responses were grouped into one of five categories (Table 4). The number of responses to each of the three questions is presented in Table 5. The categories “topic named” and “positive statement” suggest a relatively shallow level of learning and limited engagement with the session content. “Reflective statement” suggests some level of engagement, while responses coded as “reflective statement with further insight” and “reflective question” each provide evidence of deeper learning. Most responses to the question about the most important things learned were reflective statements. The majority of responses to the question about the most confusing thing covered were positive, indicating that respondents felt that they understood the session content. Responses to the question about the most interesting thing learned indicated a range of levels of engagement from positive statements or a reflective statement with some demonstrating further insight.

 

 

Table 3

POMP Responses from Researchers Who Indicated that They Had Some or a Good Understanding of Altmetrics

Discipline

Understanding of altmetrics

If you have heard of the term altmetrics, briefly describe what you understand it to mean?

Health Sciences

Good

stats of research acknowledgement

Health Sciences

Good

It appears to be a count of twitter mentions (although it might include more than that)

Health Sciences

Some

heard it but not sure of it's meaning

Health Sciences

Some

No much I understand it

Health Sciences

Good

A measure of the social media impact of your paper

Health Sciences

Good

means of measuring research impact

Life and Environmental Sciences

Good

Non-traditional metrics, number of mentions on websites, social media, media etc

Life and Environmental Sciences

Good

proposed/potential alternative to Impact Factor, as a measure of influence

Life and Environmental Sciences

Good

Another way of measuring research 'impact'

Mixed Disciplines

Good

A number that represent the amount of attention an article receives from blogs, twitter, etc.

Mixed Disciplines

Good

measure of attention an article has received relative to 'lifespan'

 

 

Table 4

Categories Used to Group ROMP Responses

Response category

Explanation of category

Example responses from this study

Topic named

Simple listing of an aspect covered in the session, with no insight to the researcher's learning

·  open access

·  altmetrics

Positive statement

Indicates session was worthwhile, with no insight to the researcher's learning

·  Nothing was confusing - excellently done

·  All of it. Informative and interesting presentation. Thank you. No negative feedback was recorded in any responses.

Reflective statement

Repeats content from the session, highlighting specific aspects

·  labouring the points about "dodgy" journals. I liked the tips to improve your Altmetric score.

·  I also didn't know that Twitter can be so useful.

Reflective statement with further insight

What the researcher learnt and how it applies to their personal situation or how they will use what they have learnt

·  The most interesting thing I learn was about the importance of open access. I am going to bring some of this information to my lab group and postgrads. Overall, thank you so much for taking the time and going over these things - you are so knowledgeable and kind, it was a wonderful workshop!

·  The importance of twitter to academics. I always thought it was meant for teenagers. I will definitely sign up for a twitter account after this session.

Reflective question

Indicates deeper thinking by new issues that the session content raised for the researcher

·  Probably not covered entirely, but I am interested how publications in journal with lower impact, but receive higher citations, might influence the indices we looked at during the session

·  Nothing was confusing. However would have liked an example of how to tweet a publication. I have tweeted a publication but there are no doughnuts associated with the corresponding author publication list in Research Portfolio.

 

Table 5

Number of Responses to Each of the Three ROMP Questions

 

ROMP Question and number of responses

Response category

What was the most important thing you learnt?

What was the most confusing thing I covered?

What was the most interesting thing you learnt?

Topic named

2

1

1

Positive statement

1

10

6

Reflective statement

13

3

8

Reflective statement with further insight

4

2

5

Reflective question

0

4

1

 

 

Discussion

 

This study describes a new approach to providing library support for researchers. The defining feature of this approach is that it is simple yet informative. The approach adapts and combines two strategies derived from a formative assessment framework. The first strategy is polling researchers using the One Minute Paper concept to (1) identify their learning needs and (2) increase levels of engagement (e.g., Hoppenfeld, 2012). The Polling One Minute Paper (POMP) is designed to be quick for researchers to complete, and easy for librarians to interpret and gain a snapshot of current learning needs of the target group of researchers. Completing the POMP prior to the session gives responding researchers an indication of the session content, raising their interest and allowing them to self-assess their understanding. During the session, a summary of the anonymous POMP responses is presented.

 

Sharing the POMP responses is effective in that it provides context relevant information against which session participants can benchmark themselves, and prompts peer to peer discussion within the context of the participants’ discipline. Increased levels of discussion provide informal evidence of the effectiveness of this strategy, which would be unlikely to occur through didactic delivery of the session content. At the end of the session, participants are asked to complete a Reflective One Minute Paper (ROMP), encouraging researchers to reflect on the session content and provide feedback which enables the librarian to review what the researchers gained from the session. This second strategy is adapted from the original One Minute Paper (Chizmar & Ostrosky, 1998), which was first used in library instruction by Choinski and Emmanuel (2006).

 

The POMP-ROMP approach can improve the value of library research support sessions for researchers. It offers a responsive approach to the current learning needs of researchers. Variation in undergraduate students’ competencies represents a significant challenge in information literacy pedagogy (Dunaway & Orblych, 2011). Researcher learning needs may be more diverse than that of undergraduate students, varying with career stage, discipline, current research priorities or activities, and previous training. The increasingly complex and competitive environment that researchers now work in (Richardson, et al., 2012) also makes it difficult to identify and track researcher learning needs. Given this variability, the two complementary OMP tools described in this study are an effective means of rapidly obtaining a snapshot of the current learning status of a specific research group, immediately prior to and following a library support training session. Librarians can use this snapshot to determine the level of detail that they teach in a session, with the aim of meeting the learning needs of researchers and providing the right amount of challenge to spark engagement in the topic. Using a tailored approach also creates proactive rather than passive library services, providing “just-in-time” and “just-for-me” assistance (Association of College and Research Libraries, 2010).

 

The POMP-ROMP approach is designed to slot into researchers’ workflow. The simplicity of both the POMPs and ROMPs mean that they can be prepared at short notice and used for small, large, remote, face-to-face, lecture, or hands-on sessions. The flexibility of these tools enables the library to contribute to improving researcher learning without adding additional burden to a researcher’s workload (Parker, 2012). The POMPs and ROMPs also extend the opportunity for discussion with researchers beyond the defined period of a library research support session, creating further opportunity to build and strengthen the researcher-librarian relationship (Auckland, 2012; Parker, 2012). In this study, ROMP feedback received in discussion at the end of sessions was responded to directly, with the offer of future sessions where relevant. Further consideration is needed for how to respond to written responses. One option would be to ask researchers to include their name on the ROMP if they would like further information. Another option would be to provide a single response, shared with all session participants, which responds to all issues raised in each ROMP from the corresponding session.

 

Many academic libraries are now developing or offering support programs for their researchers (Auckland, 2012; Richardson, et al., 2012). Maximizing the benefit for researchers attending library research support sessions will also benefit libraries by efficiently using the time that librarians spend in supporting researchers. Efficient use of the time of librarians is important because it is likely that libraries will develop research support services with no or little additional resources (Kennan, Cole, Willard, Wilson, & Marion, 2006; Kennan, et al., 2014). Therefore, as more librarians become involved in supporting researchers in their institution, a coordinated approach will be necessary to make efficient use of limited library resources. Aggregating POMP and ROMP data from all library research sessions could be used as a professional development tool for research support librarians, as also described for peer review of teaching data (Drew & Klopper, 2014). From the perspective of academic libraries, aggregated POMP and ROMP data could be used for identifying learning gaps, sharing evolving perspectives in researcher feedback, and optimizing content in library research support programs. Such a community of practice could fast track the development of research librarians (Drew & Klopper, 2014). As more librarians become involved, the opportunity for peer review of library instruction could also lead to new insights (Drew & Klopper, 2014) which may improve POMPs, ROMPs and other strategies for teaching library support to researchers.

 

One of the most important but often overlooked parts of the assessment cycle is for teachers or instructors to close the loop by reflecting on results and making appropriate changes, such as adjusting teaching methodologies or changing the structure of a program (Oakleaf, 2009; Swoger, 2011). Ultimately, the goal of all instruction and assessment efforts is to engage in reflective practice (Oakleaf, 2014). Teaching programs can nearly always be improved (Swoger, 2011) and should evolve to keep pace with new teaching strategies and content. POMPs and ROMPs are a viable means of gaining regular, systematic feedback from researchers to assist with developing and improving library research support programs. If implemented as a routine practice, the POMP-ROMP model could support a cyclical process of quality control and improvement.

 

Academic libraries and librarians must demonstrate their value. Libraries and librarians can no longer rely on an assumed belief by stakeholders that they are important (Association of College and Research Libraries, 2010). This is particularly true for the context of library research support. In this relatively new enterprise of research support, libraries are trialing various models of operation and entering spaces that were previously the domain of other sections of the university, e.g., research offices. Assessing impact, to demonstrate value, is made more complicated because academic libraries operate in a changing environment in which people, services, and needs are constantly evolving (Association of College and Research Libraries, 2010). It is also difficult to prove that actions taken by the library contributed to improvements in the performance of researchers (Association of College and Research Libraries, 2010). Implementing the POMP-ROMP approach as standard operating practice for library research support will develop a data source which provides evidence of a library’s value and how its contribution to the support of the institution’s researchers has changed over time.

 

Limitations

 

The strength but also limitation of the POMP-ROMP model is that collected information applies to a specific situation, with respect to a topic and point in time. POMP-ROMP responses should not be extrapolated to other contexts.

 

Each POMP-ROMP dataset provides a snapshot of the status of the learning of a specific group of researchers. No control groups are used and sample sizes are often small and not randomly selected. All responses are analyzed and samples are comprised of researchers with similar interests or skill levels. Questions asked in POMPs are tailored to the topic of an upcoming session, so are not intended to provide a detailed or comprehensive insight to library research support issues.

 

It should also be recognised that POMP and ROMP responses may not reflect the learning or experience of all researchers in a corresponding session. The needs of researchers who do not submit a POMP may not be considered in the preparation of session content. Non-responses are most likely due to the researcher having other priorities, not seeing a personal benefit in submitting a response, or being reluctant to share his thoughts. In the most extreme situation, non-response could be due to a lack of understanding of the question. Given these assumptions, an unexpected finding was that the number of POMP responses for sessions 3, 10, 12 and 14 was higher than session attendance. As a result, the needs of the responding researchers may be addressed but not actually apply to the researchers who attended the session. In the current study, non-response to ROMPs seemed mainly due to researchers not having anything to report that was not already expressed during the session. Unfortunately, this verbal feedback was not recorded for longer term analysis and comparison. In future sessions, the value of completing a ROMP will be emphasized, to encourage researchers to reflect on the session content and to record feedback to the librarian.

 

Conclusions

 

Polling One Minute Papers (POMPs) and Reflective One Minute Papers (ROMPs) offer a new approach for librarians to guide the learning process of the researchers they support. This study demonstrates that POMPs and ROMPs have the potential to be remarkably informative, despite requiring little effort on the part of the researchers or librarians. POMPs and ROMPs are adaptations of conventional formative assessment tools which extend the opportunity for librarians to engage with researchers, both before and after a library research support session. POMPs allow researchers to benchmark the status of their learning needs and assist librarians to identify learning gaps. ROMPs encourage researchers to reflect on what they learned in library research support sessions and assist librarians to determine whether intended learning outcomes were achieved. The simplicity of POMPs and ROMPs enable them to be slotted into researchers’ workflow. As librarians take on research support duties, these tools can be used to share recorded evidence of the evolving learning needs of researchers. Responses to POMPs and ROMPs also document evidence of the value that a library has contributed to supporting its researchers.

 

References

 

Association of College and Research Libraries. (2010). The value of academic libraries: a comprehensive research review and report. Researched by Megan Oakleaf. Chicago: Association of College and Research Libraries. Retrieved from http://www.acrl.ala.org/value/page_id=21

 

Auckland, M. (2012).  Re-skilling for research - An investigation into the role and skills of subject and liaison librarians required to effectively support the evolving information needs of researchers. Retrieved from http://www.rluk.ac.uk/files/RLUK%20Re-skilling.pdf

 

Australian Research Council. (2014). Excellence in Research for Australia (ERA). Retrieved from http://www.arc.gov.au/era/default.htm

 

Bartlett, M. G., & Morrow, K. A. (2001). Method for assessing course knowledge in a large classroom environment: an improved version of the minute paper. American Journal of Pharmaceutical Education, 65(3), 264-267.

 

Beall, J. (2014). Scholarly Open Access: critical analysis of scholarly open-access publishing. Retrieved from http://scholarlyoa.com/

 

Boud, D. (2000). Sustainable assessment: rethinking assessment for the learning society. Studies in Continuing Education, 22(2), 151-167. http://dx.doi.org/10.1080/713695728

 

Boud, D., & Lee, A. (2005). ‘Peer learning’ as pedagogic discourse for research education. Studies in Higher Education, 30(5), 501-516. http://dx.doi.org/10.1080/03075070500249138

 

Brodie, L. (2012). Problem-based learning. In L. Hunt & D. Chalmers (Eds.), University teaching in focus: A learning-centred approach. Hoboken: Taylor and Francis.

 

Chizmar, J. F., & Ostrosky, A. L. (1998). The one-minute paper: Some empirical findings. The Journal of Economic Education, 29(1), 3-10. http://dx.doi.org/10.1080/00220489809596436

 

Choinski, E., & Emmanuel, M. (2006). The one-minute paper and the one-hour class: Outcomes assessment for one-shot library instruction. Reference Services Review, 34(1), 148-155. http://dx.doi.org/10.1108/00907320610648824

 

Creaser, C., & Spezi, V. (2013). Improving perceptions of value to teaching and research staff: The next challenge for academic libraries. Journal of librarianship and information science, 46(3), 191-206. http://dx.doi.org/10.1177/0961000613477678

 

Crisp, G. (2009). Designing and using e-assessments. Milperra, NSW: Higher Education Research and Development Society of Australasia.

 

Department of Education. (2014). Higher Education Research Data Collection (HERDC). Retrieved from http://education.gov.au/higher-education-research-data-collection

 

Dillman, D. A., Smyth, J. D., & Christian, L. M. (2009). Internet, mail, and mixed-mode surveys: The tailored design method. New Jersey: Wiley.

 

Dobson, J. L. (2008). The use of formative online quizzes to enhance class preparation and scores on summative exams. Advances in Physiology Education, 32, 297-302. http://dx.doi.org/10.1152/advan.90162.2008

 

Draper, S. W. (2009). Catalytic assessment: understanding how MCQs and EVS can foster deep learning. British Journal of Educational Technology, 40(2), 285-293. http://dx.doi.org/10.1111/j.1467-8535.2008.00920.x

 

Drew, S., & Klopper, C. (2014). Evaluating faculty pedagogic practices to inform strategic academic professional development: A case of cases. Higher education, 67(3), 349-367.  http://dx.doi.org/10.1007/s10734-013-9657-1

 

Drummond, G. (2007). Programme feedback: It only takes a minute! International Journal of Educational Management, 21(1), 29-36. http://dx.doi.org/10.1108/09513540710716803

 

Dunaway, M. K., & Orblych, M. T. (2011). Formative assessment: transforming information literacy instruction. Reference Services Review, 39(1), 24-41. http://dx.doi.org/10.1108/00907321111108097

 

Frances, M., Fletcher, J., & Harmer, S. (2011). Reshaping and rescoping university libraries to fit changing academic requirements. Paper presented at the 32nd IATUL Annual Conference: Libraries for An Open Environment: strategies, technologies and partnerships, Warsaw, Poland. Retrieved from http://www.bg.pw.edu.pl/iatul2011/proceedings/ft/Frances_M.pdf

 

Google. (2014). Create a survey using Google forms. Retrieved from http://support.google.com/drive/answer/87809?hl=en

 

Guest, G., MacQueen, K. M., & Namey, E. E. (2012). Applied thematic analysis. Los Angeles: SAGE Publications.

 

Haddow, G. (2012). Research support in a research assessment environment: The experience of 'new' universities. Library and information research, 36(113), 62-80.

 

Hoppenfeld, J. (2012). Keeping students engaged with web-based polling in the library instruction session. Library Hi Tech, 30(2), 235-252. http://dx.doi.org/10.1108/07378831211239933

 

Hughes, H., & Bruce, C. (2012). Informed learning in online environments: Supporting the higher education curriculum beyond Web 2.0. In P. P. Godwin & J. Parker (Eds.), Information literacy beyond library 2.0 (pp. 65-79). London: Facet.

 

James Cook University. (2014). Facts & figures 2014: Data snapshot. Retrieved from http://www.jcu.edu.au/about/public/groups/everyone/documents/annual_report/jcu_140345.pdf

 

James Cook University. (2015). About JCU: Strategic intent. Retrieved from http://www.jcu.edu.au/about/strategic-intent/

 

Kaufman, P. T. (2009). Carpe Diem: Transforming services in academic libraries. In IDEALS. Retrieved from http://www.ideals.illinois.edu/handle/2142/12032

 

Kennan, M. A., Cole, F., Willard, P., Wilson, C., & Marion, L. (2006). Changing workplace demands: What job ads tell us. Aslib proceedings: New information perspectives, 58(3), 179-196. http://dx.doi.org/10.1108/00012530610677228

 

Kennan, M. A., Corrall, S., & Afzal, W. (2014). "Making space" in practice and education: Research support services in academic libraries. Library Management, 35(8/9), 666-683. http://dx.doi.org/10.1108/LM-03-2014-0037

 

Knowles, M. S. (1975). Self-directed learning: a guide for learners and teachers. Englewood Cliffs, N.J.: Cambridge Adult Education.

 

Kolb, D. A. (1984). Experiential learning: experience as the source of learning and development. Upper Saddle River, NJ: Prentice-Hall.

 

Merriam, S. B. (2001). Andragogy and self-directed learning: Pillars of adult learning theory. New directions for adult and continuing education, 2001(89), 3-13. http://dx.doi.org/10.1002/ace.3

 

Oakleaf, M. (2009). The information literacy instruction assessment cycle: A guide for increasing student learning and improving librarian instructional skills. Journal of documentation, 65(4), 539-560. http://dx.doi.org/10.1108/00220410910970249

 

Oakleaf, M. (2014). A roadmap for assessing student learning using the new framework for information literacy for higher education. The Journal of Academic Librarianship, 40(5), 510-514. http://dx.doi.org/10.1016/j.acalib.2014.08.001

 

Parker, R. (2012). What the library did next: Strengthening our visibility in research support. Paper presented at eM-powering eFutures, the 16th Biennial VALA Conference and Exhibition, Melbourne, Victoria, Australia.  Retrieved from http://hdl.handle.net/1959.3/214486

 

Parsons, A. (2010). Academic liaison librarianship: Curatorial pedagogy or pedagogical curation? Ariadne, 65, 1-11. Retrieved from http://www.ariadne.ac.uk/issue65/parsons/

 

Priem, J., Taraborelli, D., Groth, P., & Neylon, C. (2010). Altmetrics: A manifesto. Retrieved from http://altmetrics.org/manifesto/

 

REF2014. (2014). REF2014: Research Excellence Framework. Retrieved from http://www.ref.ac.uk/

 

Richards, L. (2015). Handling qualitative data: A practical guide (3rd ed.). Los Angeles: SAGE Publications.

 

Richardson, J., Nolan-Brown, T., Loria, P., & Bradbury, S. (2012). Library research support in Queensland: a survey. Australian Academic & Research Libraries, 43(4), 258-277. http://dx.doi.org/10.1080/00048623.2012.10722287

 

Sadler, D. R. (1989). Formative assessment and the design of instructional systems. Instructional Science, 18(2), 119-144. http://dx.doi.org/10.1007/BF00117714

 

Simons, N., & Richardson, J. (2013). New content in digital repositories: The changing research landscape. Oxford: Chandos.

 

Stead, D. R. (2005). A review of the one-minute paper. Active Learning in Higher Education, 6(2), 118-131. http://dx.doi.org/10.1177/1469787405054237

 

Swoger, B. J. M. (2011). Closing the assessment loop using pre- and post-assessment. Reference Services Review, 39(2), 244-259. http://dx.doi.org/10.1108/00907321111135475

 

 

Appendix

Session #

Discipline / Client Group

Questions

Response Options

 

1

Life and Environmental Sciences: Post-Docs

How would you rate your understanding of altmetrics?

None; Low; Good

 

1

Life and Environmental Sciences: Post-Docs

Which of the following do you use to do your research, or to talk about your research interests?

Academia.edu; Blogging; CiteULike; Comments on other sites; Facebook; Mendeley; Newspaper interviews; Radio interviews; ResearchGate; Slideshare; Twitter; Wikipedia; YouTube; Other

 

1

Life and Environmental Sciences: Post-Docs

Which of the following sites do you use to manage your research profile?

JCU Research Portfolio; ResearchOnline@JCU; ResearcherID; Scopus; ORCID; Google Scholar; Other

 

2

Life and Environmental Sciences

Is the information in your JCU Research Portfolio profile up-to-date?

Yes; No; Other

 

2

Life and Environmental Sciences

Which of the following do you use to do your research, or to talk about your research interests?

ResearchOnline@JCU; ResearcherID; Scopus; ORCID; Google Scholar; Other

 

2

Life and Environmental Sciences

Which of the following sites do you use to manage your research profile?

Academia.edu; Blogging; CiteULike; Comments on other sites; Facebook; Mendeley; Newspaper interviews; Radio interviews; ResearchGate; Slideshare; Twitter; Wikipedia; YouTube; Other

 

3, 4, 5, 6, 7, 8

Life and Environmental Sciences

Have you used any of the following to promote or discuss your research?

ResearcherID; Scopus; Google Scholar; ORCID; The Conversation; Traditional media - newspaper, radio, TV; Blogging; Comments on the blogs or articles by others; Twitter, ResearchGate; YouTube; LinkedIn; Mendeley; Other

 

3, 4, 5, 6, 7, 8

Life and Environmental Sciences

Have you heard of the term altmetrics?

Yes; No; Not sure

 

3, 4, 5, 6, 7, 8

Life and Environmental Sciences

If you have heard of the term altmetrics, briefly describe what you understand it to mean?

[Free text]

 

3, 4, 5, 6, 7, 8

Life and Environmental Sciences

If you have paid an Article Processing Charge, do you know if you retained copyright ownership of your article?

Yes; No; Not sure

 

3, 4, 5, 6, 7, 8

Life and Environmental Sciences

Have you or a co-author paid an Article Processing Charge to make any of your articles Open Access?

Yes; No; Not sure

 

9

Health Sciences

Have you published any of the following types of journal articles?

Peer reviewed article; Non-refereed article; Case study; Short note or commentary; Other

 

9

Health Sciences

Have you published any of the following types of conference works?

Peer reviewed paper; Non-refereed paper; Poster; Abstract or summary; Edited a conference proceedings; Other

 

9

Health Sciences

Have you published any of the following types of books or book chapters?

Research; Non-research; Teaching material; Non-commercial; Reference; Later edition; Report

 

9

Health Sciences

Tick the box if you are you familiar with the following government reporting or assessment exercises:

HERDC: Higher Education Research Data Collection; ERA: Excellence in Research for Australia; Other

 

9

Health Sciences

Tick the box if you have used any of the following for ranking the value of a journal:

Journal Citation Reports; Scopus analytics; Beall's list of predatory publishers; Other

 

10

Health Sciences

Have you used any of the following to promote or discuss your research?

ResearcherID; Scopus; Google Scholar; ORCID; The Conversation; Traditional media - newspaper, radio, TV; Blogging; Comments on the blogs or articles by others; Twitter, ResearchGate; YouTube; LinkedIn; Mendeley; Other

 

10

Health Sciences

Have you heard of the term altmetrics?

Yes; No; Not sure

 

10

Health Sciences

If you have heard of the term altmetrics, briefly describe what you understand it to mean?

[Free text]

 

10

Health Sciences

Have you or a co-author paid an Article Processing Charge to make any of your articles Open Access?

Yes; No; Not sure

 

10

Health Sciences

If you have paid an Article Processing Charge, do you know if you retained copyright ownership of your article?

Yes; No; Not sure

 

11

Health Sciences

Have you published any of the following types of journal articles?

Peer reviewed article; Non-refereed article; Case study; Short note or commentary; Other

 

 

 

 

 

 

11

Health Sciences

Have you published any of the following types of conference works?

Peer reviewed paper; Non-refereed paper; Poster; Abstract or summary; Edited a conference proceedings; Other

 

11

Health Sciences

Have you published any of the following types of books or book chapters?

Research; Non-research; Teaching material; Non-commercial; Reference; Later edition; Report

 

11

Health Sciences

Would you consider that you have a fair understanding of:

HERDC: Higher Education Research Data Collection; ERA: Excellence in Research for Australia; Other

 

11

Health Sciences

Are any of your publications Open Access:

In ResearchOnline@JCU?; On the publisher's website?; Other

 

12

Social Sciences and Humanities

Are you the author of a peer reviewed journal article?

No; Yes

 

12

Social Sciences and Humanities

Are you the author of a peer reviewed conference paper?

No; Yes

 

12

Social Sciences and Humanities

Are you the author of a commercially published book or book chapter about a research topic?

No; Yes

 

12

Social Sciences and Humanities

Do you have publications available in ResearchOnline@JCU?

No; Yes

 

12

Social Sciences and Humanities

Do you have a Google Scholar profile that lists your research outputs?

No; Yes

 

12

Social Sciences and Humanities

If you have answered yes to Questions 1,2 or 3, have you searched for your publications in Web of Science?

No; Yes

 

12

Social Sciences and Humanities

If you have answered yes to Questions 1,2 or 3, have you searched for your publications in Scopus?

No; Yes

 

12

Social Sciences and Humanities

Is there any topic or question that you would like me to talk about in the session?

[Free text]

 

13

Social Sciences and Humanities

Have you used any of the following to promote or discuss your research?

ResearcherID; Scopus; Google Scholar; ORCID; The Conversation; Traditional media - newspaper, radio, TV; Blogging; Comments on the blogs or articles by others; Twitter, ResearchGate; YouTube; LinkedIn; Mendeley; Other

 

13

Social Sciences and Humanities

Have you heard of the term altmetrics?

Yes; No; Not sure

 

13

Social Sciences and Humanities

If you have heard of the term altmetrics, briefly describe what you understand it to mean?

[Free text]

 

13

Social Sciences and Humanities

Have you or a co-author paid an Article Processing Charge to make any of your articles Open Access?

Yes; No; Not sure

 

13

Social Sciences and Humanities

If you have paid an Article Processing Charge, do you know if you retained copyright ownership of your article?

Yes; No; Not sure

 

14

Social Sciences and Humanities

How many HERDC eligible publications do you have? This includes peer reviewed publications or commercially published books or book chapters.

[give number]

 

14

Social Sciences and Humanities

Have you heard of the term altmetrics, and how would you rate your understanding of altmetrics?

I have not heard of the term; I have some understanding; I have a good understanding

 

14

Social Sciences and Humanities

Which of the following do you use to manage your research profile?

JCU Research Portfolio; ResearchOnline@JCU; ResearcherID; Scopus; ORCID; Google Scholar; Other

 

14

Social Sciences and Humanities

Which of the following do you use to do your research, or to talk about your research?

Academia.edu; Blogs; Comments on any sites; Facebook; Mendeley; Newspaper interviews; Radio interviews; ResearchGate; Slideshare; Twitter; Wikipedia; YouTube; Other

 

15

Mixed disciplines: Academic teaching staff

Have you published any of the following outputs about your research?

Peer reviewed journal article; Non-peer reviewed journal article; Non-peer reviewed conference paper; Book or book chapter; Report; Other

 

15

Mixed disciplines: Academic teaching staff

Have you published any of the following outputs about your teaching?

Peer reviewed journal article; Non-peer reviewed journal article; Non-peer reviewed conference paper; Book or book chapter; Report; Other

 

15

Mixed disciplines: Academic teaching staff

Do you use any of the following social media tools to talk about or promote your teaching and learning?

Blogs; Comments on other people's posts; Twitter; Facebook; Academic.edu or ResearchGate; Slideshare; Wikipedia; YouTube; (Your) JCU Research Portfolio; (Your) Google Scholar profile; Other

 

16

Mixed disciplines: Higher Degree Research students

What is the discipline or subject area of your research?

 

16

Mixed disciplines: Higher Degree Research students

Are you a:

Research Masters student; PhD student; Early Career Researcher i.e. are you within 5 years of starting your post-PhD research career?; Other

16

Mixed disciplines: Higher Degree Research students

Are you using (or planning to use) material in your thesis for which the copyright may have expired? Select Other if you would like to provide more information.

Yes; No; Not sure; Other

16

Mixed disciplines: Higher Degree Research students

Are you using (or planning to use) material in your thesis which is still under copyright? Select Other if you would like to provide more information.

Yes; No; Not sure; Other

16

Mixed disciplines: Higher Degree Research students

Have you ever requested copyright permission to use someone else's work in your research?

Yes; No; Not sure; Other

16

Mixed disciplines: Higher Degree Research students

Are you planning to publish your research in:

Journal article(s); Conference paper(s); book or book chapter(s); Other

16

Mixed disciplines: Higher Degree Research students

Have you heard of Creative Commons?

Yes; No; Not sure

16

Mixed disciplines: Higher Degree Research students

Do you have a copyright question that you would like answered in the workshop? Please provide more details here.

[Free text]