header
published quarterly by the University of Borås, Sweden

vol. 23 no. 4, December, 2018



Proceedings of ISIC: the Information Behaviour Conference, Krakow, Poland, 9-11 October, 2018: Part 1.

The "user turn" in practice: information behaviour researchers’ constructions of information users


Heidi Julien, Lynne McKechnie, Sarah Polkinghorne and Roger Chabot


Introduction. This study examines how "users" (i.e., research participants) are constructed in the empirical research reported at the ISIC conference between 1996 and 2016 (n=194).
Method. The study uses content analysis to examine variables relating to authorship, research methods, theoretical orientation, participant group, and reporting practices. Qualitative analyses are used to examine construction of research participants through writing practices.
Analysis. Data are analysed quantitatively and qualitatively.
Findings. Most authors of empirical papers are academics in information science, the largest proportion of authors work in the U.S., interviews continue to be the most commonly used method, and professionals are the most studied group. The ISIC oeuvre includes many fine examples of respectful construction of research participations, and some examples of less respectful constructions.
Conclusion. There is significant opportunity for information behaviour scholars to improve the ways in which they write about participants.

Introduction

More than three decades have passed since the inception of what can be described as a user turn in information science, a shift commonly attributed to the spark provided by Dervin and Nilan in their 1986 review article (Tabak, 2014; Talja and Hartel, 2007). The ‘user turn’ is the cumulative change in focus and direction within information behaviour research during which users have been espoused as an increasingly central concern. Similarly, Talja and Hartel (2007) call this narrative of change "the user-centred turn," first characterising it as a turn akin to the many other "turns" that have taken place within social science disciplines.

Today, while information behaviour researchers frequently claim user-centrality, little is known about whether or not, and how, they go about centering users within their research. This means that we do not know the extent to which a user turn has actually taken place. We are not readily able to characterize the position of users in information behaviour research, as instantiated in research practices, in detail. If information behaviour scholars are to improve their research practices, we need to understand where opportunities for improvement are evident.

One way to better understand the present positioning of users is to examine how researchers construct their study participants in published literature. This approach involves studying how researchers actively position participants through practices including, but not limited to, inclusion of descriptive details, assignment of pseudonyms, and incorporation of participants´ own words. It also involves uncovering assumptions about participants that are present but may be implicit. This paper reports a study that uses this approach to investigate the constructions of users in empirical research reported in the Information Seeking in Context (ISIC) conference proceedings between 1996, the inaugural ISIC, and 2016.

Literature review

There are few recent information science publications that examine either the user as a concept or the user construction practices of researchers. We discuss recent examinations of the user concept below, after summarizing how users are presently constructed in information behavior literature. User construction is evident in the language chosen by researchers to describe study participants. Publications, as observed by Hedemark, Hedman, and Sundin (2005), are an ‘important arena for the production, reproduction and mediation of different views of the user’ (para. 1).

In recent literature, study participants are predominantly, though not exclusively, constructed in relation to places they encounter or technologies they use. Participants are positioned as users of physical locations or institutions, such as academic and public libraries, as well as digital or metaphorical spaces, such as the digital information environment (Nel and Fourie, 2016;Yi, 2015; Connaway, White, Lanclos and Le Cornu, 2013).

Participants are also positioned as users of specific technologies, systems, or formats. Examples include ‘smartphone users’, ‘Internet users’, and ‘image users’ (Zhitomirsky-Geffet and Blau, 2017; Chen, Lee, Straubhaar and Spence, 2014; Matusiak, 2017). Less predominantly, study participants are constructed as being involved in specific activities or processes, such as social commerce (Farivar, Turel and Yuan, 2017), or as members of demographic or vocational groups, such as traditional farmers (Meyer, 2009).

The information behaviour literature continues to contain little discussion of people who could be described as non-users. Lwoga and Chigona (2017) and Mutshewa, Grand, Totolo, Zulu and Jorosi (2010) are exceptions. They characterize non-users of the Internet in Tanzania and non- users of libraries in Botswana, respectively.

The breadth of user constructions in the literature is expanding. This is demonstrated by trends moving in opposing directions, toward more atomistic constructions on the one hand, and more holistic ones on the other. For example, psychometric and biomedical tests and techniques appear in the recent literature. Studies employing such techniques analyse particular attributes of users, such as their personality traits or cognitive styles, in order to correlate these attributes with facets of information behaviour such as query formulation and fact-seeking (Al-Samarraie, Eldenfria and Dawoud, 2017; Kinley, Tjondronegoro, Partridge and Edwards, 2014). At the same time, studies employing ethnographic techniques emphasize users’ social surroundings and increasingly, their embodied experiences, in an effort to examine information behaviour in ways that represent users as whole people (Cox, Griffin and Hartel, 2017; Lloyd and Olsson, 2017; Lueg, 2015).

The above examples illustrate the variety of user constructions evident in recent information behaviour literature. Conceptual examinations of the user, whether theoretical, critical, or empirical, are rare by comparison. One significant example is Talja (1997). Presenting at the first ISIC conference, Talja challenges the conventional cognitive view of users as ‘uncertain people who need help’ and, as an alternative, presents a discourse analytic perspective that enables ‘the conceptualization of users as knowing subjects in the practical-discursive context of everyday life’ (pp. 77-78). Julien (1999) also laments the standard construction of users as ‘bungling fools who ‘lack’ appropriate skills to effectively find what they are seeking’ (p. 207).

Subsequent researchers present divergent analyses. Booth (2008) argues that it is futile to try to define a ‘typical’ or ‘representative’ library user. Olsson (2009) argues that prominent information behaviour models tend to foreground user uncertainty, lack, or need, and thus remain grounded in systems priorities and assumptions. Miksa (2009) examines user characterizations from 1500 to today, and concludes that the concepts of ‘information use’ and ‘information user’ remain ‘for all practical purposes as mysterious as they have always been’ (p. 364). Fleming-May, analysing the concept of "use," observes how common definitions of usage serve to conceptually divorce use from users (2014). McKechnie, Julien, Pecoskie and Dixon (2006) observe a contradiction between the espoused centrality of research participants and their often-peripheral framing, through labels and other language use, within research publications.

Most recently, Day (2011) and Tabak (2014) provide notable critiques. Day (2011) links the concept of the user to deterministic assumptions that can be identified within information behaviour research. He argues that the user reflects information behaviour’s reliance on the conduit metaphor for communication, rooted in Shannon’s information theory, which has led to ‘viewing human expressions and their understandings in terms of determinative causes and effects’ (p. 80). The conduit metaphor and its mechanistic causality are ‘inadequate for describing agency and the relationship between subjects and objects’ (p. 78). Drawing on a variety of psychoanalysts and philosophers, Day proposes an understanding of people’s encounters with information that emphasizes the social, cultural, and physical affordances, and interbody affects, which together mediate subjects and objects (p. 85).

Tabak (2014) identifies tension within information behaviour research between an individual focus on the cognitive user and a social focus on context. He argues that

[Information behaviour] models and theories frequently shift between the two poles in an attempt to address challenges from other positions. However, by jumping between user and context or placing a priori actors in a context, IB models fail to trace actors’ own contextualization. Therefore, IB research needs models that give actors a space to perform their own positioning. (p. 2227)

Tabak describes how actor-network theory can help, as it argues ‘that individual and collective are merely moments in the circulation of information’ (p. 2230).

Tabak and Day’s papers push the user concept theoretically. The challenges they identify also circulate within research design and practice. Whether recent information behaviour studies reflect a nuanced and respectful relationship with users is a question pursued by the study presented in this paper. The authors are making the assumption that users and research participants should be constructed as whole human beings and discussed respectfully in published work. There have been no content analyses of how users are constructed within research and writing since McKechnie et al. (2006). This paper updates and extends that work, with a focus on the research question: how are research participants constructed in information behaviour studies, as represented in empirical papers in the ISIC proceedings?

Method

To address our research question we conducted a content analysis (Krippendorff, 2013) of the 194 full papers reporting empirical research studies that investigated human information behaviour found in the published proceedings of the biennial ISIC conferences from 1996 to 2016. While human information behaviour researchers communicate their work through a variety of publications and conferences, we chose ISIC as being one of the major international venues for dissemination of human information behaviour research. While this may be regarded as a limitation to the study, it resulted in a sample which made data collection and analysis feasible and which, we argue, constitutes a reasonable representation of the information behaviour research conducted by key scholars in the area. It is clear, however, that the findings of this study do not necessarily apply to information behaviour studies not included in the ISIC proceedings. Another limitation is that we did not have access to the context of the papers included in the sampling frame; thus, findings are restricted to practices apparent in these papers.

The ISIC proceedings include a variety of types of papers, such as theory papers, method papers and reports of empirical research studies. Some years of the proceedings include both full and short papers. We did not analyse the short papers, as the length restriction is likely to have limited the reporting and discussion. Our first step was to look carefully at the full papers to identify the reports of empirical research that investigated human information behaviour, the corpus examined in this study. We defined empirical research as that which is based on observed or measured phenomena, rather than strictly from theory or belief. This distinction is based on categories used in previous research (e.g., Pettigrew and McKechnie, 2001). Our next step was to compile a list of types of people commonly investigated by information behaviour researchers. This list was developed from earlier studies of the characteristics of Information Behaviour and other Library and Information Science publications that included the type of people studied in their data collection and reported results (Julien, 1996; Julien and Duggan, 2000; Julien, Pecoskie and Reed, 2011; Julien and O’Brien, 2014).

Papers were coded for quantitative data as follows:

We also used qualitative open coding (Corbin, 2015) of the articles to explore how research participants were constructed in the reports, what assumptions were made about these people, what conclusions were being drawn about the groups represented and information behaviour more generally, and anything else of interest. During the analysis process, we kept reflective notes to document theoretical insights and keep track of particularly cogent examples related to the authors’ reporting (and not reporting) about the groups of people studied.

To provide a preliminary assessment of the validity of the coding instrument, all authors coded a small sample of three randomly chosen papers as a basic test for inter-coder reliability. The unit of analysis for this was the individual decision point; in other words, every time a decision was made (e.g., were groups of people studied explicitly reported or not reported) comprised an instance of the unit of analysis. Our coding scheme included thirty-nine decision points. Frequencies and percentages were calculated for the number of agreements. While it is agreed there are problems associated with this measure, research indicates percentage agreement is the most commonly used measure, applied in 69% of studies (Lombard, Snyder-Duch and Bracken, 2010). For our purpose of getting a rough sense of the validity of our instrument, it worked reasonably well. The rate of agreement was 98.0%, suggesting that the coding scheme was likely valid and reliable.

The data were analysed quantitatively with frequencies and percentages calculated for all categories. Although our sample size was relatively large (n=194), the data did not meet the criterion of having five or more data points in each cell to be able to calculate tests of Chi Square to look for relationships between variables.

Results and discussion

Our sample includes 194 papers. When discussing our findings, we refer to specific papers in our sample, which we cite as examples of particular practices by the surname of the first author and the year of the ISIC conference in which the paper was delivered. In all instances our critiques are offered with respect and collegiality. We did not access publications of authors outside of the ISIC conference proceedings; therefore, our findings reflect only the papers published within this sampling frame, rather than necessarily reflecting work published elsewhere.

Quantitative results

Authors are mostly LIS academics (Table 1).


Table 1: Author affiliation.
Affiliation of first Number Percent
Library and Information Science Academic16183%
Other (non-academic, non-professionals)168%
Academics in non-LIS disciplines105%
Professional42%
Research center21%

Others were 15 students, 1 research assistant

First authors come from universities and other institutions in a large number of countries (Table 2).


Table 2 – First Author’s Country.
CountryNumberPercent
USA4925%
UK2513%
Finland2412%
Sweden168%
Australia147%
Canada147%
Singapore63%
Japan53%
Israel42%
South Africa42%
Croatia32%
Lithuania32%
Denmark21%
France21%
Germany21%
Iceland21%
Ireland21%
Netherlands21%
New Zealand21%
Spain21%
Uruguay21%
Brazil11%
Estonia11%
India11%
Malaysia11%
Poland11%
Slovakia11%
Slovenia11%
Taiwan11%
Uganda11%

Methods used in the papers are varied, and differ in some respects from those methods found in previous studies of information behaviour work (Table 3). For example, Julien, Pecoskie and Reed (2011), analyze information behaviour research published from 1999 to 2008, but use a different sampling frame that includes a significant proportion of work published by professional librarians, which is a subset of the information behaviour literature not typically included in the ISIC conferences. They find that empirical work in the area largely uses surveys/questionnaires, and fewer interviews. A follow-up study examining information behaviour literature published from 2009 to 2013 finds that interviews are the most-used method (Julien and O’Brien, 2014).


Table 3 – Methods used.
Methods used Number Percent
Interviews14675%
Surveys and/or questionnaires6634%
Observations4825%
Content Analysis3619%
Diaries1910%
Focus groups126%
Experiments105%
Transaction Log Analysis95%
Think aloud protocols95%
Case studies63%
Social network analysis32%
Secondary data analysis11%

Our analyses confirm that method is very much tied to construction of participants. Surveys tend to be distancing methods. Those methods analyzing meaning-making tend to construct participants as individuals. We also found that often the participants have a secondary role in the paper – the primary interest is in the method or the theory (even though the paper is clearly empirical), and the participants are vehicles for discussing method and/or theory. In some cases, we found a discrepancy between authors’ intentions and their practices. For example, some authors emphasize the user turn, or critique information science for inadequately focusing on users, while simultaneously neglecting to include context around their own study participants, or otherwise reflect a user-centred approach.

The study participants in this sample are categorized in Table 4. Once again, the results differ from those identified by Julien, Pecoskie and Reed (2011), suggesting that scholars presenting at the ISIC conference differ from researchers doing information behaviour work more generally. In the 2011 study, the largest identifiable participant group is students. In the follow- up study (Julien and O’Brien, 2014), non-professional workers is the most-studied group, followed by scholars.


Table 4 – Study participants.
Who was studiedNumberPercent
Professionals13771%
General public (adults)2714%
Students2312%
Non-professional workers21%
Scholars21%
Youth or children11%
Unspecified11%

Qualitative results

In many papers authors refer to those studied by both their demographic category (e.g., student) and their role in the study (e.g., participant, respondent, subject). Most often the role in the study approach occurs in the Methods section of the paper but it also appears elsewhere. The ‘users’ in the study are participants in the empirical work analyses, and also belong to a user group of interest to the authors. Most authors tell us little about their participants; an exception is when data collection includes demographic data, which provides additional detail. We find some use of distancing language, such as the use of terms such as subjects, or respondents, or when participants are identified alphanumerically rather than given pseudonyms. Many authors use multiple synonyms for research participants (e.g., participants, respondents, users, etc.), which may be rhetorical, to avoid redundancy. We like to think that use of subject is done without deep consideration for the connotations of that term. Some authors construct their participants as individuals with agency, personality, and context; others are constructed as generic people in generic contexts. For example, it seems that when using activity theory, the participants (subjects) as individuals are rendered less important than the case study context and the activities of the participants.

One interesting finding is that the particular people studied often simply disappear or are absent from both the preliminary material (literature search) and the findings and discussion. The particular group (the sample) is identified and described in the method, but after that authors speak about users. This acts to imply generalization to a much broader group (in most cases to all people). This generalization is sometimes not justified in any way. When it is justified it is listed as a limitation to the study. Examples of this type of over-generalization include Wang (1996), Erdelez (1996), Allen (2000), Enochsson (2000), Light (2000), Erdelez (2000), Martinivic (2016), Faletar-Tanackovic (2016), and Lee (2016). Good examples where this type of over-generalization is not apparent include Williamson (1996), Enwald (2016), Huvila (2016), and Juric (2016). Hersberger (2000) specifically notes that her participants are ‘not homogenous’ (p. 120), and Julien (1996) warns that ‘[t]he results may be limited in their general applicability, since the participants derived essentially from the same milieu’ (pp. 383- 384).

There are studies that give relatively little space to participants’ own words, which is particularly noticeable in qualitative research, which is expected to give voice to research participants (McGinn, 2008). These studies are also unlikely to use pseudonyms or other indicators enabling the reader to tell participants apart and think of them as fully-fledged people. In these studies, pseudonyms are rarely used. This is surprising but understandable in studies that revolve around practices, activities, and processes. For example, Sairanen (2010) offers an elegant use of pseudonyms, whereas Bruce (2010) uses alphanumeric codes for participants, which are not as readable as pseudonyms but still serve the purpose of distinguishing people. Examples in which participants’ own voices are given space, even though participants are not distinguished from one another with pseudonyms, include Veinot (2010), Bowler (2010), and Hartel (2006).

Anderson (2006) provides a good example of a study in which participants are given pseudonyms, and are assumed to be capable individuals. Their struggles are characterized as natural or inherent to their work, rather than ascribed to personal deficiencies. The participants’ own self-characterizations are included and affirmed. This is respectful stance, and is demonstrated by this quote: ‘Both informants describe themselves as curious, avid explorers of ideas. Encounters with them during the course of this study confirm their self-assessments’ (Section ‘Study design‘, para. 1).

Another good example of a careful and concise construction of participants is offered by Harris (2006). Her participants are constructed as self-reliant and active seekers of health information. While they are not presented as individuals, their own words are used to elaborate the quantitative findings. Heinstrom (2006) is another an example of a paper that manages to discuss differences among people in a measured, detailed way without painting anyone as either deficient or superior. The participants are constructed respectfully and their own words are closely integrated into the reporting of results. Wathen (2006) also demonstrates respect for her participants, who are constructed as complex people, and who are given voice in lengthy passages in their own words. This enables the reader to get an authentic sense of the participants as people, with their ways of thinking and speaking. This paper also provides very detailed contextual background for the study participants. Ross (1998) gives her participants pseudonyms and includes several pages of paragraph-length quotations. Given (2002) also provides detail about her participants, and gives them voice through lengthy quotations. McKenzie’s (2002) analysis is framed around the words of her participants; their words are given prominence in this research. Fourie (2008) takes great care in the study design and implementation to attend to the emotional needs of her participants. This study is exemplary in its respect for informants.

An example of work which is less focused on participants as individuals is evident in Baxter (2010). In this study, people are constructed only in terms of their role within their organization. No space is given to characterizing them, and through metonymic slippage, the participants are actively erased, because they are assumed to be interchangeable with their organizations. Primarily, participants’ responses are referred to as organizational responses. Participants are called group or organization continuously, for example, ‘two organizations requested a copy of the interview schedule and submitted their responses by email’ and ‘[w]ithout wishing to sound cynical, the authors suspect that many of the other organizations were less honest at this stage of the interviews, despite guaranteed anonymity’ (Section Gathering information for a consultation response, para. 9). Another example is Davenport (1996), which refers to participants as ‘households’, not people, even though people are the ones supplying the data. It is the case that household could be considered a collective noun for a group of people, but it still has characteristics of a thing, rather than a person. As the authors put it, ‘[t]he household, in our study, is presented as an example of a managed group, whose concerns are as much collective as individual’ (p. 390). We find near-total metonymic slippage in some papers. This is where participants are referred to as their group or organization, rather than as people, e.g., ‘Without wishing to sound cynical, the authors suspect that many of the other organizations were less honest at this stage of the interviews, despite guaranteed anonymity’ (Baxter, 2010). This practice muddies the results because it obscures who precisely is speaking or being discussed.

Metonymic slippage also erases the individuality and agency of participants (e.g., Durrance, 2006).

Bruce (2010) constructs participants sympathetically, in part by framing the research problem in ‘we’ terms:

[w]e are all limited by our abilities to accurately predict the information that we will need over time for any tasks that occupy our interest. Our memories fail us when we encounter a need for information and we forget that we have the right information for this need in a personal information collection. (Section "Conclusion," para. 2)

Yeh’s (2008) participants are also constructed sympathetically, with concern for their situation. Some authors use the acknowledgement sections of their papers to indicate sympathy and respect for their participants. For example, Genuis (2014) writes: ‘[s]incere gratitude to the interviewed women and health professionals who so generously shared their thoughts and experiences.’ Similarly, Rutter (2014) writes: ‘[t]hank you to all who participated from a Sheffield (UK) primary. Firstly, to the children who generously and eloquently shared their information-seeking experiences.’ These gestures demonstrate respect and appreciation for participants.

There are excellent examples of authors constructing their participants as having agency, as being actively engaged, and as being skilled at articulating their own experiences (e.g. Sairanen, 2010). In that study, student participants are given pseudonyms, and they are constructed as complex individuals. The authors provide thick description, e.g. describing students’ health challenges and family circumstances, to contextualize the participants’ statements. Berryman (2008) also constructs her participants as people with agency and whose meaning-making is valued. Olander (2008) gives participants pseudonyms and includes many quotations, ascribed to individuals. In this study, meaning-making is relevant and participants are true individuals. Sholam (2008) also uses pseudonyms to identify speakers of quotations. These participants clearly have agency, and their information needs are important, despite the assumption of a clear deficit model. Asla (2014) includes multiple quotations ascribed to named individuals who are clearly respected, and whose meaning-making is valued. Paul’s (2014) participants are rendered generously; these are individuals with clear motivations, interests, and concerns.

Others, despite displaying obvious sympathy for their participants, give little space to participants’ own words through quotations (e.g., Veinot, 2010). Keane (1998) offers two vignettes of participants who are given pseudonyms, with thick description of their situations and behavior, but includes no direct quotations; and, the remaining eight study participants are invisible. Limberg (1998) provides a table of participant names, which helps to emphasize their individuality, but the experiences of these individuals are not included. Lomax (1998) provides demographic data for participants, and lengthy direction quotations are provided, but these are identified with an identification number, rather than a pseudonym. Marton (2002) quotes extensively from open-ended survey questions, giving voice to her respondents, although they are not distinguished in any way. Anderson (2004) is a good example of a paper where the approach is deeply qualitative and the participants are respected, but the author is so focused on her central question (relevance judgements) that the participants as individuals fade into the background. These are examples of a large middle ground of approaches, where authors may demonstrate respect for participants, but do not employ expected strategies to flesh that out. For example, in Cooper (1998), the participants demonstrate agency, but are not ascribed any personality characteristics. Dixon (1998) constructs participants with a fair deal of agency, but the literature review constructs the individuals as incompetent, and lacking in experience, skills and/or knowledge to make effective decisions. The paper includes some direct quotations from observation participants, but the speakers are not identified in any sort of way. The results section provides a vignette of one participant’s information experience, but that person is assigned no descriptive characteristics outside of the task that is being studied.

Some authors use detailed context to construct participants, giving the reader an in-depth picture of participants’ backgrounds and situations. Good examples of this approach are Sundin (2002), Courtright (2004) and Fisher (2004). Fisher (2004) is an excellent example of how including thick description of context portrays a three-dimensional picture of the participants, even when specific demographics or other descriptions of the specific participants are sparse.

Some studies do not bother constructing participants in any way, or only in minimal ways. For example, the participants in Huotari (1998) are visible only through their actions and not their words. Individuals are referred to by their title as a ‘basic part’ of the organization in which they work. Even less flattering constructions are evident in other studies. For example, Nicholas (1998) calls some of his participants ‘resentful dinosaurs’ (p. 456). In Wilson (1998), the research participants are invisible, and no quotations are provided. Allen (2008) is an excellent example of a paper with absolutely minimal information provided about participants; the focus is very much on their ‘activity’ and the context of their work.

Also evident in our analyses is evidence for a continuing deficiency mindset, i.e., where participants are constructed as deficient (Julien, 1999). However, for the most part, authors construct users respectfully, as competent and capable of communicating their experiences (e.g., Tramullas, 2010; Martinović, 2016).

We find multiple studies that construct people at a distance. There are examples where, in studies of participants in helping or caring professions, such teachers, librarians, or health workers, the focus in the papers’ findings and/or discussion is on the people served by those professionals, e.g., students. Thus, the researchers provide a platform for professionals’ assessments of the information needs or practices of a group of less powerful people whose own voices are not sought out or included in the research. Studying professional expertise is legitimate, but in these cases the researchers construct their study participants, and the study participants construct their clients and/or students. This secondary user construction is particularly conspicuous when the research participants (the professionals) are barely constructed at all, while by contrast there is significant effort given to represent participants’ constructions of their clients and/or students (e.g., Limberg, 2006; Sabelli, 2016).

Multiple papers construct participants as invisible, generic, or only valuable in terms of the data provided to support a point or a theoretical concept. This occurs in many quantitative and some qualitative studies, where authors construct participants very instrumentally – as suppliers of research data. These studies typically construct participants in minimal, or flat ways, with little or no detail about them as individuals. For example, Benoit (2002) includes quotations focused on the participants’ search experiences, but as individuals they are invisible. Miwa’s (2002) participants are invisible, and interview results are paraphrased. The participants in Lilley (2008) are of interest only because of their ethnicity; no other aspects of these participants’ lives or situations appears relevant. Sakai (2014) provides an example of ethnographic research that fails to flesh out participants, suggesting little interest in them as individuals, or in their meaning-making.

Ishimura (2012) is one of the few studies that explicitly uses the language of operationalization to talk about a sample population. The authors write: ‘Canadian students are defined as students who use English as their mother tongue and were primarily educated in Canada’ (Methods, para. 1).

Study participants are mentioned throughout the articles analysed (Table 5).


Table 4 – Study participants.
Article SectionNumberPercent
Title10152%
Abstract13871%
Introduction/literature review13871%
Results18997%
Discussion/conclusion16786%

We find that in methods sections, when describing the sample/population of the study, authors frequently are vague. Authors may mention that they chose a sample that had wide variation in sociodemographic factors, but then never give a clearer picture than that. In addition, titles seem to be poor indicators of people studied; often, authors are not very specific about the people they studied until their methods sections, and will refer to the populations more generally in the introductions/abstracts, etc. This practice has implications for people searching the literature for certain types of research participants. It is possible that the conference format does not encourage authors to supply significant detail, a concern identified in earlier research focused on methodological rigour (McKechnie, Dalmer, Chabot, Julien and Mabbott, 2016). However, the structured abstracts required in the later ISIC papers encourage identification of participants in the methods section.

Theory was mentioned in seventy papers (36%). A very wide range of theories was included. It is important to note that just because a study does not use theory does not mean it does not use concepts (e.g., credibility; information overload), or ground the study in a solid literature review. In some cases, authors name-drop many information behaviour theories, without ever identifying a single theory as the foundation for the study.

Limitations of the study

Our study instrument did not take into account multiple groups that were studied at the same time, or when the information behaviour of one group was constructed via the insights of another group (e.g. asking teachers and librarians about students’ behaviour). In addition, an aspect of this analysis that would require more study is the cultural context that influences each paper. ISIC’s status as a global conference is powerfully illustrated in these papers and there may be cultural differences in the ways in which researchers are trained with respect to research methods (e.g., views on how participants are understood and positioned), and to writing research papers. For many, writing academically in a language (English) other than one’s first language may be challenging, and some nuances of meaning may be lost. We also find that authors sometimes choose to refer the reader to a fuller description of participants in another article or dissertation, to save space. This could limit the value of the conference proceedings because the articles do not stand on their own (e.g., Anderson, 2004). A limitation of this sample is the fact that it draws from conference proceedings, and authors may not be as careful or as thorough as they might otherwise be in constructing and representing their study participants.

Conclusion

Our analyses highlight many encouraging practices, which position "users" or research participants with agency, individuality, and full-fleshed out contexts. However, we also identify significant opportunities in the information behaviour scholarship examined here to improve our representations of those whose actions, thoughts, feelings, and motivations form the basis of our scholarship. We anticipate that these findings will encourage scholars contributing to the ISIC conference, and also to the wider information behaviour field, to reexamine their research design and research writing practices, with a renewed determination to reflect the respect that our research participants deserve.

About the authors

Heidi Julien, Professor, State University of New York at Buffalo, 526 Baldy Hall, Buffalo, NY 14260, USA. She received her PhD in 1997 from the University of Western Ontario. Her research interests are information behaviour, digital literacy, and research methodology. She can be contacted at heidijul@buffalo.edu.
Lynne (E.F.) McKechnie, University of Western Ontario, Faculty of Information and Media Studies, London, Ontario, Canada N6A 5B7. She received her PhD in 1996 from the University of Western Ontario. Her research focuses on the intersection between children, reading and public libraries. She can be contacted at mckechnie@uwo.ca.
Sarah Polkinghorne is a doctoral candidate at Swinburne University of Technology, Australia and librarian at the University of Alberta, Canada. Sarah received her MLIS in 2007 from the University of Alberta. Sarah's doctoral research focuses on the nature and role of information in ubiquitous everyday activities, such as those connected to food. She can be contacted at sarah.polkinghorne@ualberta.ca.
Roger Chabot, University of Western Ontario, Faculty of Information and Media Studies, London, Ontario, Canada N6A 5B7. He received his MLIS from the University of Western Ontario in 2012. His research interests are religious information behaviour and the application of Buddhist philosophy and psychology to topics in LIS . He can be contacted at rchabot2@uwo.ca.

References


How to cite this paper

Julien, H., McKechnie, L., Polkinghorne, S. & Chabot, R. (2018). The "user turn" in practice: information behaviour researchers’ constructions of information users In Proceedings of ISIC, the Information Behaviour Conference, Krakow, Poland, 9-11 October: Part 1. Information Research, 23(4), paper isic1804. Retrieved from http://InformationR.net/ir/23-4/isic2018/isic1804.html (Archived by WebCite® at http://www.webcitation.org/74IB54LvF)

Check for citations, using Google Scholar