MJA
MJA

Improving communication when seeking informed consent: a randomised controlled study of a computer-based method for providing information to prospective clinical trial participants

Asuntha S Karunaratne, Stanley G Korenman, Samantha L Thomas, Paul S Myles and Paul A Komesaroff
Med J Aust 2010; 192 (7): 388-392. || doi: 10.5694/j.1326-5377.2010.tb03561.x
Published online: 5 April 2010

Abstract

Objective: To assess the efficacy, with respect to participant understanding of information, of a computer-based approach to communication about complex, technical issues that commonly arise when seeking informed consent for clinical research trials.

Design, setting and participants: An open, randomised controlled study of 60 patients with diabetes mellitus, aged 27–70 years, recruited between August 2006  and October 2007 from the Department of Diabetes and Endocrinology at the Alfred Hospital and Baker IDI Heart and Diabetes Institute, Melbourne.

Intervention: Participants were asked to read information about a mock study via a computer-based presentation (n = 30) or a conventional paper-based information statement (n = 30). The computer-based presentation contained visual aids, including diagrams, video, hyperlinks and quiz pages.

Main outcome measures: Understanding of information as assessed by quantitative and qualitative means.

Results: Assessment scores used to measure level of understanding were significantly higher in the group that completed the computer-based task than the group that completed the paper-based task (82% v 73%; P = 0.005). More participants in the group that completed the computer-based task expressed interest in taking part in the mock study (23 v 17 participants; P = 0.01). Most participants from both groups preferred the idea of a computer-based presentation to the paper-based statement (21 in the computer-based task group, 18 in the paper-based task group).

Conclusions: A computer-based method of providing information may help overcome existing deficiencies in communication about clinical research, and may reduce costs and improve efficiency in recruiting participants for clinical trials.

It is well recognised that there are significant deficiencies in the current process for obtaining informed consent for participation in clinical research.1-7 Various attempts have been made to enhance participant understanding, with limited success.1,8-12 Some strategies have resulted in improvements in information transmission to and retention by not only study participants but also patients in general.9,10,13-19 However, complex methods of information provision, such as multimedia methods, may cause confusion and thereby reduce understanding.20,21

Studies of the provision of information in clinical research have usually adopted the perspectives of researchers and regulatory bodies, rather than those of participants.22 There have been a few exceptions;23-26 for example, a comparison of an information statement developed by participants with an information statement developed by researchers showed that the former was associated with greater participant understanding.26 We therefore sought to assess the efficacy of a computer-based method of communicating information to prospective clinical trial participants, with the aim of improving participant understanding.

Methods
Sample size

On the basis of an earlier study,8 the sample size was initially estimated as 100 for a power of 0.8, with an expected difference in means of 5% given a putative population standard deviation of 8.8. However, when variance was verified on the basis of the first 10 participants (SD, 7.6) the sample size was re-estimated as 60, to attain the same difference in means and power.

Measures and end points

Participants’ levels of understanding were assessed using quantitative and qualitative measures. The key quantitative measure of understanding, and primary end point, was the percentage of correct answers to questions in a paper-based questionnaire (in multiple-choice, true/false and yes/no format) that was administered to both groups at completion of the reading task. This consisted of 43 questions: eight demographic questions (personal characteristics such as age and sex), 26 assessment questions, one question about hypothetical participation and eight distracter questions.

Assessment questions covered the purpose of the mock study, the sponsor, study procedures, benefits, risks, privacy and confidentiality, contact details, voluntariness, injury and reimbursement; answers to all these questions could be found verbatim in the computer-based presentation and the paper-based statement. The difference in understanding between the two groups was determined by comparing overall assessment scores as well as performance on individual assessment questions.

The time taken to complete each reading task was also measured. Unlike other studies that have compared a standard form to other procedures,12,27,28 time restrictions were not imposed. This was to respect different reading abilities.

Participants’ levels of understanding and appraisal of methods were measured qualitatively via a semi-structured one-to-one interview, consisting of 11 common questions and three questions that were specific to the particular group. The interviews, which allowed individuals to comment on their experience, were audiotaped and transcribed in de-identified form.

Results
Characteristics of participants

Data were gathered from 60 participants (30 assigned to the computer-based task, and 30 assigned to the paper-based task), whose characteristics are summarised in Box 3. Most were male (42/60), most used computers on a daily basis (50/70), and 40% were working full-time (24/60). The mean age was 52.0 years (range, 27–70 years). In the group that completed the paper-based task, 21 of 30 had completed tertiary education, compared with 15 of 30 in the group that completed the computer-based task. All participants were fluent in spoken English, and all but one were fluent in written English.

The average time taken to read the information for those who completed the computer-based task was 6 minutes longer than that for the group that completed the paper-based task (mean [range], 19 [9–33] minutes v 13 [6–32] minutes; P < 0.001).

Quantitative assessment of participant understanding

The percentage of correct answers used to assess understanding was based on 23 of the 26 assessment questions (three assessment questions that were answered correctly by > 90% of participants were excluded from the analysis). The average percentage of correct answers for the group that completed the computer-based task was significantly higher than that of the group that completed the paper-based task (82% v 73%; P = 0.005 [two-tailed t test]).

Frequencies of percentages of correct answers for both groups are shown in Box 4. These scores were clearly different in the two groups — scores of participants who completed the computer-based task were skewed towards the higher percentages, and scores of participants who completed the paper-based task were lower and more spread out. The group that completed the computer-based task had a highest individual score of 96% (two participants with 22 correct answers) and a lowest of 65% (two participants with 15 correct answers), compared with 91% (5 participants with 21 correct answers) and 17% (one participant with four correct answers), respectively, in the group that completed the paper-based task. The participant who was not fluent in written English achieved the highest assessment score in the group that completed the paper-based task. Multivariate analysis showed no correlation between percentages of correct answers and age or sex.

The group that completed the computer-based task answered several questions significantly better than the group that completed the paper-based task, including questions about procedures (28% v 19% correct answers; P = 0.005), the site of catheter insertion (30% v 25%; P = 0.02), privacy (27% v 16%; P = 0.006), method of disclosing study results (29% v 21%; P = 0.01), contact persons (27% v 18%; P = 0.02) and compensation in the event of possible injury (30% v 24%; P = 0.04). Two questions, regarding possible side effects to the heart and lungs, were answered equally well by both groups. The group that completed the paper-based task performed slightly better in four questions, which were about study sponsor, a minor procedure, benefits of participation and withdrawal from the study.

Qualitative assessment of participant understanding and appraisal of methods

The interviews undertaken after participants completed the reading tasks were 3–18 minutes in duration. A selection of representative quotes from the interviews is presented in Box 5. The computer-based task received positive feedback, especially about its presentation and special features. Participants stated that these characteristics made them feel better informed and better able to make a decision about being involved in the study. Participants in the group that completed the paper-based task stated that they found the information difficult to understand, and made negative comments about the length and presentation of the document. After verbally explaining to participants the nature of the other form of information delivery tested in this study, more participants from both groups stated that they believed they would find a computer-based presentation easier to understand (21 in the computer-based task group, 18 in the paper-based task group).

Discussion

This study has shown that a computer-based approach to communicating information about clinical research to prospective trial participants can improve the consent process, compared with a conventional approach using a paper-based statement. Four key findings related to this improvement.

First, we found a major difference between the groups in the understanding of the more complex details of the study. Other studies examining the efficacy of multimedia consent processes in enhancing understanding of clinical trials have shown limited success.29-31

Second, participants who completed the computer-based task felt more comfortable in making a decision about being involved in the study. Building trust between researchers and participants is a cornerstone of any study, and feeling informed about a study may help improve these relationships and allay participant anxiety about taking part in a study.23 The quizzes within the computer-based task allowed participants to self-assess their understanding and affirm, for themselves, their eligibility to participate in it. This could not only save researchers valuable time in explaining study procedures11 but also facilitate an appropriate emphasis on issues of special concern to individual participants23 without rushing the consent process. This feature may benefit mass screening programs where large numbers of individuals can self-assess their understanding and also self-select themselves as potential participants, in addition to being contacted by researchers to take part in a study.

Third, the overall lower assessment scores in the group that completed the paper-based task raise concerns about participants’ levels of understanding when this conventional system is used. Further, the wide range of these scores in this group suggests variability in understanding among participants enrolling in research studies with paper-based information statements.9 Participants in the group that completed the computer-based task received and understood uniform and complete information presented in an attractive manner. This is likely to be of significant advantage in multicentre trials, where a computer-based approach could be employed to uniformly and reliably communicate with participants at many locations.

Fourth, participants in the group that completed the computer-based task were more likely to indicate a willingness to participate in the mock study (if it were real). This could indicate a benefit in recruiting (and perhaps even retaining) study participants.8

The computer-based approach was associated with three possible drawbacks. First, a computer-based approach is intrinsically more expensive and time consuming to set up and administer than a paper-based approach and it may not always be clear that the advantages will justify the additional costs. However, ethics committees could provide an online template — as with paper-based statements — from which study teams could create multimedia statements in a timely and affordable way. Templates could ensure that the quality of multimedia statements remains high and includes essential information. It may therefore be most appropriate to start implementing the multimedia option in large-scale, multicentre studies — these are more likely to have sufficient resources to implement a multimedia option and, as the recruitment process could be more complex in such studies, a multimedia method may help simplify this process.

Second, not all individuals are computer literate. For some participants, a computer-based approach may be too complicated to understand, and some may not have access to a computer. However, most participants in both arms of our study were computer literate. Further, those in the group that completed the computer-based task spent significantly more time reading the information. However, they were more engaged in the process and did not mind spending the additional time.

Third, there is a theoretical risk that verbal communication between researchers and participants may decrease if researchers become reliant on computers to provide information. Researchers must remain aware that computers cannot replace the trust and understanding that comes from taking the time to talk to study participants.

Our study had some limitations. It was restricted to English-speaking, computer-literate patients with diabetes. Although the computer-based method was successful in our study, further research is necessary to assess its efficacy in other settings and participant groups. Also, we measured participants’ levels of understanding immediately after they completed the reading tasks; this may have demonstrated improvement in information recall rather than understanding,9 but it is more likely that both are improved. In addition, further research is needed to assess whether the findings apply equally to men and women.

In conclusion, we have shown that a strategy for communication which uses the interactive capacity of computers is likely to provide an effective means for overcoming key deficiencies in the conventional, paper-based system of communication about clinical research projects. As access to and familiarity with computer-based approaches to communication increase, it is likely that such methods will become part of a new standard of practice in the clinical research consent process.

5 Quotes from participants regarding the methods of information presentation

General comments about the computer-based presentation

“I don’t think you could get it any easier ...”

“It’s the most clearly written piece of medical information I’ve ever seen. I could actually understand it ...”

Positive aspects of the computer-based presentation

Video

“... at least you can see beforehand what you have to go through ... sometimes when you go to have a procedure you’re not aware of what’s going on [and] you can be very fearful of it.”

Hyperlinks

“Easy to access other missing information if you needed to, if you weren’t sure about what you were reading on the screen.”

Quizzes

“I got one lot wrong and I realised I really hadn’t read that section properly. So it forced me to go back ...”

Shortcomings of the paper-based statement

Too technical

“I wouldn’t say it was easy reading ... I didn’t find it too difficult because I have got a background in anatomy and physiology ... But I think if you didn’t ... that part would have been fairly difficult to follow.”

Too long

“Yeah, well, all documents are lengthy. It reminds me of filling out those bank application forms ... pages and pages and pages.”

Presentation could be improved

“There is some scope of it missing, in terms of presentation.”

Preference for computer-based presentation

“Personally I think a visually based one would be easier for me to understand than a text based one.”

“I tend not to concentrate too much on forms. I sort of glaze over ... I use the computer all day so it’s just a much more natural thing for me.”

Received 28 April 2009, accepted 19 October 2009

Online responses are no longer available. Please refer to our instructions for authors page for more information.