About this Journal Submit a Manuscript Table of Contents
Depression Research and Treatment
Volume 2012 (2012), Article ID 309094, 11 pages
http://dx.doi.org/10.1155/2012/309094
Research Article

Using Self-Guided Treatment Software (ePST) to Teach Clinicians How to Deliver Problem-Solving Treatment for Depression

1Department of Psychiatry, Brigham and Women's Hospital, Harvard Medical School, 1249 Boylston Avenue 3rd floor, Boston, MA 02215, USA
2Harvard Medical School, Boston, MA 02215, USA
3Department of Psychiatry, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
4Department of Psychiatry, Dartmouth-Hitchcock Medical Center, Dartmouth Medical School, Lebanon, NH 03756, USA
5Department of Educational Psychology, University of Texas at Austin, Austin, TX 72705, USA

Received 2 April 2012; Revised 19 August 2012; Accepted 26 August 2012

Academic Editor: Amy Kilbourne

Copyright © 2012 James A. Cartreine et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Problem-solving treatment (PST) offers a promising approach to the depression care; however, few PST training opportunities exist. A computer-guided, interactive media program has been developed to deliver PST electronically (ePST), directly to patients. The program is a six-session, weekly intervention modeled on an evidence-based PST protocol. Users are guided through each session by a clinician who is presented via hundreds of branching audio and video clips. Because expert clinician behaviors are modeled in the program, not only does the ePST program have the potential to deliver PST to patients but it may also serve as a training tool to teach clinicians how to deliver PST. Thirteen social workers and trainees used ePST self-instructionally and subsequently attended a day-long workshop on PST. Participants’ PST knowledge level increased significantly from baseline to post-ePST ( ) and did not increase significantly further after attending the subsequent workshop. Additionally, attending the workshop did not significantly increase the participants' skill at performing PST beyond the use of the ePST program. Using the ePST program appears to train novices to a sufficient level of competence to begin practicing PST under supervision. This self-instructional training method could enable PST for depression to be widely disseminated, although follow-up supervision is still required.

1. Introduction

Effective treatments have been developed for depression, including antidepressant medications and psychotherapies. Many patients prefer counseling or psychotherapy to taking medications, if it is available [1], and matching treatment (medication or therapy) to patient preference has been demonstrated to improve outcomes, independent from depression severity [2, 3]. Unfortunately, many mental healthcare providers lack training in evidence-based treatments [4] or fail to implement them properly [5], making evidence-based psychotherapy unavailable to many persons with depression in the United States. One way to increase access to evidence-based psychotherapies is to develop more cost-efficient and accessible training methods.

Problem-solving treatment (PST; also known as problem-solving therapy) has emerged as an effective treatment for depression [68]. The basis of PST is that enhancing problem-solving skills and attitudes and working to solve concrete problems in one’s life can reduce depression. PST breaks the problem-solving process into steps and teaches participants to go through these steps systematically and effectively, targeting parts of the process that are particularly challenging for depressed patients. PST has been found to produce outcomes comparable to antidepressant medications [9, 10] and better than treatment as usual [11, 12].

PST has great potential for dissemination because it can be delivered by mental health as well as nonmental health specialists in a wide range of healthcare settings, including primary care [13]. Healthcare professionals, such as nurses and social workers are trained in a one- to two-day seminar that includes didactics, demonstrations, and roleplay with feedback [14]. They are then supervised on PST by a recognized expert over the course of three to five cases. For each case, the supervisor listens to and critiques a recording of three PST sessions, sometimes providing this supervision via the telephone.

Automating PST training could improve the ability to further disseminate PST and potentially result in cost savings. Although the supervision portion of the training requires the involvement of a live clinician, it is conceivable that some or all of the workshop could be automated. Self-instructional training is not a new concept; self-instructional books have existed for centuries and self-instructional videos for decades. These could be good starting points to learn PST. However, it has been demonstrated that the more interactive a training experience is, the more effective it is [1518]. Since the skill at hand, PST, is a highly interactive process, developing a highly interactive self-guided training approach should enhance PST skill building.

1.1. ePST Software

An interactive media-based computer program has been developed to provide PST electronically (ePST) in an entirely automated manner [19]. In this self-help program, users are guided through six sessions of PST by a master clinician (M. T. Hegel) who is an expert in the intervention and conveys the “nonspecific” characteristics of effective therapists, such as warmth, genuineness, compassion, and the ability to provide support when the patient experiences setbacks [20]. Users interact with the program by entering text via the keyboard or by clicking on answers to questions, and hundreds of branching audio and video clips are used to tailor the program to users’ inputs. The program tracks the user’s depression via the Patient Health Questionnaire-9 (PHQ-9; [21]) and the on-camera host provides feedback and recommendations. The program reviews the user’s work across sessions and across problems to provide guidance on how to improve his or her problem-solving success, through the use of failure analyses [2224] algorithms. Although the program was originally developed for astronauts to treat their own depression on long-duration space missions, it was designed for use by the public.

Because the program models best practices for providing PST, it may be useful not only as an intervention but also as a teaching tool, to demonstrate how to deliver PST. The purpose of this study was to obtain pilot data on the feasibility and efficacy of using this program to teach PST to therapists unfamiliar with the intervention. Primary research questions addressed were as follows.(1)Is use of ePST associated with change in knowledge of the steps and process of PST?(2)Is live training associated with change in skill at performing PST, for persons who have already been trained via ePST?

Secondary research questions were as follows.(1)Is use of ePST associated with change in self-efficacy for performing PST-specific skills, and for doing counseling in general?(2)How acceptable and easy to use is ePST for training?

2. Methods

2.1. Participants

The subject population consisted of clinical staff and trainees from a human services organization in Framingham, MA, with no prior training in PST. All participants volunteered for the study and received no compensation for participation. Training was offered to all clinicians and trainees working at the site.

2.2. Procedure

The study was approved by the Massachusetts General Hospital’s Institutional Review Board, and consent to participate in the study was implied by voluntary completion of the study questionnaires. Receiving training was not contingent on participation in the evaluation component, and participants were not required to participate as part of their employment.

Training consisted of participants using the ePST program for four sessions as if they were patients. Participants were encouraged to work on real problems in their own lives via ePST as a means of learning how PST works. They were also asked to view the on-camera therapist as a model for delivering PST. Although a full course of treatment via ePST involves six sessions, the investigators judged a priori that using the program for only four sessions would provide sufficient exposure to the model for trainees to learn PST. At no point in this study did the participants interact with actual patients using the PST approach.

Assessments were conducted at three time points and trainings were conducted at two points. At baseline (before beginning training), participants completed a test of their knowledge of PST and a measure of self-efficacy to do counseling (regarding both general and PST-specific skills). Following completion of their training using ePST, participants again completed questionnaires on their knowledge of PST and self-efficacy as well as the acceptability and usability of the computer program. They also participated in audiotaped roleplay sessions as a PST therapist treating a standardized patient (a research assistant trained to portray a depression patient). After the workshop, participants completed the knowledge and self-efficacy questionnaires for the third time and then conducted another taped roleplay session with a different standardized patient. Three standardized patient cases of similar difficulty were written for the roleplay. Subjects were randomly assigned to different cases for their roleplay sessions.

2.3. Measures

Knowledge of PST
This was a primary measure, which consisted of one open-ended essay-format question: “Please describe the process of problem-solving therapy in detail, including all steps and the criteria for successfully completing each one.” Essays were scored using criteria developed for the study (see Appendix A). The possible range of scores was 0 to 120, with higher scores indicating greater knowledge. The essays were rated by M. T. Hegel and J. L. Seville, blind to the time point of each essay.

Skill Implementing PST
The standardized patient roleplay was a primary measure. The audiotaped roleplay sessions were scored using the Problem-Solving Treatment in Primary Care Adherence and Competence Scale (PST-PAC) [25, 26] which evaluates trainees’ performance on nine dimensions of PST, each on a six-point scale (0–5; worst to best performance) (see Appendix B). Because each standardized patient was being seen for an initial PST session, the first rating dimension, “Defining the outcome” (of the previous session’s action plan) was not included. Therefore, participants were rated on dimensions two through eight of the PST-PAC. The possible range of scores was 0 to 40, with higher scores indicating better performance.

Scoring was performed by J. A. Cartreine and J. L. Seville, who were blind to the time point of each roleplay. One of the raters (J. L. Seville) had been trained in the use of the PST-PAC previously; the other (J. A. Cartreine) was trained to rate roleplays by rating audio recordings of sample cases.

Self-Efficacy for Implementing PST and Doing Counseling, in General
The PST and Counseling Self-Estimate Inventory (PCSEI) was created for this study, it consists of 30 items that measure different aspects of a therapist’s perception of his or her ability to perform specific PST skills and general counseling skills (see Appendix C). It was adapted from the Counseling Self-Estimate Inventory (CSEI) [27]. To reduce the length of the measure, some items from the CSEI less relevant to the study were dropped, while several new items pertinent to performing PST-specific skills were added, using similar language. Items were scored on a six-point scale (1 = “strongly disagree”; 6 = “strongly agree”), with a possible range of 30 to 180 for the total measure, a range of 16 to 96 for the “PST-specific” subscale and 14 to 84 for the “General” subscale. Some items are reverse scored to avoid acquiescence bias; higher scores reflect higher self-efficacy.

Acceptability of Using ePST to Learn PST
The Program Acceptability Questionnaire (PAQ) was written for this study and consisted of six questions on various aspects of the acceptability of the PST computer program, such as how much the program helped the trainee learn PST and whether the subject would recommend its use to other prospective trainees (see Appendix D). Items were rated on a six-point scale as above (1 = “strongly disagree”; 6 = “strongly agree”).

Usability of the ePST Program
The System Usability Scale (SUS) [28, 29] was used to assess ease of using the software. The SUS is a ten-item questionnaire that is widely used for assessing software usability, such as whether a system is excessively complex or cumbersome to use, or whether the trainee needs technical support to use it. Statements about the software are rated on a six-point scale (0 = “strongly disagree”; 6 = “strongly agree”).

2.4. Analytic Methods

Score Calculations
Scores for the PST-PAC were obtained using the following procedures: for each standardized patient interview, all eight scores given by each rater were averaged, yielding two scores (one for each rater). For items where the raters’ scores differed by more than one point, the raters discussed the item and arrived at a consensus score for that item. Then, the average was taken of the two raters’ scores, yielding one final score per standardized patient interview.
PST knowledge essays were scored on 27 items that map onto seven subscales, which correspond to the steps of PST. Because the seven subscales contain different numbers of items, an average of the scores for each subscale, for each rater, was calculated. The seven scores were then summed, yielding one score per essay for each rater. Finally, an average was calculated between scores awarded by each rater, to obtain a full-scale score for each participant on each essay written.
Self-efficacy scores were calculated by summing the responses to individual items (with some questions reverse scored). Additionally, subscales were calculated by summing the items measuring general self-efficacy and those measuring PST-specific self-efficacy.

Analyses
Due to the small sample size, nonparametric analyses were used to answer the research questions. The Wilcoxon signed ranks test was used to compare skill levels between Assessment 2 and 3 (no skill test was administered in Assessment 1). The same test was also used to gauge change in knowledge and self-efficacy at Assessment 1 compared to 2, and Assessment 2 compared to 3. Because multiple comparisons were made, a Bonferroni correction was used, which set the cutoff for significance at for the primary outcomes of skill and knowledge. The Bonferroni correction was not used for the secondary outcome of self-efficacy.
Descriptive statistics were used to characterize the sample and results of the SUS and PAQ. Because the PAQ was created for this study, an insufficient quantity of data has been collected to support a factor analysis and the reporting of a composite score. Therefore, results on this questionnaire are reported by item.

3. Results

Thirteen participants (11 female and 2 male) enrolled, with a mean age of 37.5 (±12.9); 11 self-identified as Caucasian and the other 2 as racial minorities. Six were licensed clinical social workers, with an average of 2 years in practice (±1.4, range 2 to 6 years); the other 7 were social work graduate students. No participants who inquired about the study subsequently refused to participate or dropped out.

Interrater reliability for both the essay and roleplay scoring was calculated using percent agreement between raters. For the skill measure, interrater agreement per item (defined as agreement on the rating plus or minus 1 point) between the two raters averaged 79.6%. To increase concordance between raters, consensus scores (as described in the methods section) were obtained for items in which raters differed by >1 point. For the knowledge (essay) measure, the percent agreement (within one point) between the raters across items was 87.5%. Because agreement between raters on the knowledge measure was high from the outset, a consensus process was not needed to improve concordance.

Summary results of the knowledge, skill, self-efficacy, and usability measures are presented in Table 1.

tab1
Table 1: Assessments of knowledge, skill, self-efficacy, and ePST program usability.

3.1. Primary Outcomes

Knowledge scores significantly increased from pretraining to post-using ePST ( , ) but did not increase significantly more after attending the live workshop ( , ). No significant difference was found in skill level after using ePST versus after the live workshop, using either nonparametric (Wilcoxon signed ranks test ; ) or parametric ( ; ; ) statistics. This suggests that the workshop did not substantially increase the knowledge or skill of the participants beyond what they learned using the ePST program. On an item-level analysis of the roleplay ratings, only skill at helping patients identify pros and cons of solutions changed before to versus after attending the live workshop. On this item, before attending the workshop (but after completing ePST) the median score was 2.5 on a scale of 0 to 5 (range .5 to 4.5; mean of ); and after the workshop, the median increased to 3.5 (range .5 to 5.0; mean ; Wilcoxon signed ranks test ; ). Also, across outcome measures and time points, there were no significant correlations between participants’ knowledge, skill, or self-efficacy levels.

To facilitate comparisons between knowledge and skill levels, scaled scores of 0 to 100 were calculated by converting the median to the percentage of the maximum score possible on each scale (see Figure 1). Both knowledge and skill were near the midpoint after using ePST (42.46 and 51.25, resp.) and remained near that point after completing the workshop (45.97 and 65).

309094.fig.001
Figure 1: Scaled scores for skill, knowledge, and self-efficacy. Scaled Score = Median converted to percent of maximum possible score.
3.2. Secondary Outcomes

Self-Efficacy for PST and Counseling, Overall
Five items were unanswered in the PCSEI data set. Values for these missing data points were imputed by calculating the mean of the participant’s responses to the other items from the same subscale on that administration of the PCSEI. Composite self-efficacy scores, which included all items on the PCSEI, remained stable from pretraining to post-ePST training to post-live workshop, although there was a significant increase in composite self-efficacy from pre-training to post-workshop ( ). The Wilcoxon signed ranks test -score for self-efficacy at pretest compared to post-ePST was −1.014; ( ; ; ) and the -score for post-ePST compared to post-live training was −1.751; ( ; ; ). On the PST-specific subscale, there was a significant improvement from post-ePST and post-live workshop training ( -score = −2.033; ). No significant difference was found for the subscale of general counseling self-efficacy between pretest and post-ePST training, or between post-ePST and post-live workshop training.

3.3. Usability and Acceptability

The median SUS score was 67.5 (range 60 to 87.5; mean = ) on a scale of 0 to 100. On the PAQ, the medians for all but one item were 5 (means ranged from 4.54 to 5.23) on a scale of 0 to 6 (see Table 2).

tab2
Table 2: Program Acceptability Questionnaire (scale range 0 to 6).

4. Discussion

PST is an evidence-based intervention for depression; however, few providers have training in it. A novel approach to teaching PST is the use of a computer-automated treatment, ePST. Users are guided through it via branching videos to simulate the experience of being in treatment with a master clinician (M. T. Hegel). As such, best practices for delivering PST are modeled and the program may serve as a training tool for practitioners, in addition to a treatment for patients.

Thirteen persons with no prior experience delivering PST (6 licensed clinical social workers and 7 social work graduate students) used the ePST interactive media program for four sessions over the course of two weeks. They were instructed to work on a real problem in their own lives, to make the training more meaningful. After the self-guided instruction, all 13 attended a standard PST training workshop presented by a veteran PST trainer (also M. T. Hegel), which is the standard method to train new practitioners.

It was found that live training does not add knowledge or skill beyond what ePST provides. Trainees were assessed on knowledge at three points: once at baseline, again after using ePST for four sessions, and again after completing the live day-long workshop. They were also assessed on PST skill via roleplays with standardized patients after completing ePST and again after completing the workshop. Knowledge about how to conduct PST went from a near-zero level (1.5 on a scale of 0 to 35) at baseline to a significantly higher level after using ePST (mean score 14.1) and did not significantly improve following the workshop (mean score 16.7).

PST skill was not assessed at baseline due to logistical constraints; however, it is likely that it would also have been very low before training, since none of the subjects had received prior training. After completing 4 sessions of ePST, mean skill was 19.7 (on a scale of 0 to 40) and was 25.2 following the live workshop. This change was not statistically significant. Composite self-efficacy for performing PST skills and for doing counseling in general gradually increased from baseline to post-live training. This appears to be due to a significant improvement in self-efficacy for performing PST skills after the workshop and one standardized patient interview. Therefore, confidence in one’s ability to conduct PST increased significantly as a result of the live workshop and/or standardized patient interview, even though actual skill and knowledge did not.

The only specific areas of improvement found after completing the workshop were knowledge and implementation of decision-making guidelines (i.e., evaluation of pros and cons in problem solving). Both skill and knowledge regarding decision-making guidelines improved following the workshop, suggesting that ePST could be strengthened in this regard. No significant change in either direction was noted for other content areas following the live workshop.

A key ingredient in clinical training is supervision while treating actual patients, and it is unrealistic to expect that use of a self-instructional training program or a one-day workshop would produce fully competent PST clinicians. However, after using ePST, social workers and social work trainees possessed a sufficient level of skill to begin providing PST under supervision. As such, ePST may be a convenient and low-cost alternative to the in-person PST workshops that are the standard of practice. As the number of PST practitioners increases, the number of persons with sufficient experience to serve as supervisors will also increase. Moreover research has already supported PST supervision via telephone [12], meaning the supervisor need not be in the same room—or the same time zone—as the trainee. Together, ePST plus telephone-based supervision could support the diffusion of PST throughout the medical and mental health communities.

Regarding the usability of ePST, an analysis of all published studies that used the SUS to evaluate software of all types found that the mean score was 70.14 [28]. The mean SUS score for ePST was 69.23; however, this may underestimate the program usability. Because ePST is designed for weekly sessions, to run the program more than once per week (as was done in this study), participants had to manually advance the system date on their computers, which can be challenging. A training version of ePST could readily overcome this limitation.

Acceptability of ePST appeared high. That said results of the Program Acceptability Questionnaire, written for this study, may have been subject to response set, as the means and standard deviations were very similar between items. For example, on a scale of 1 to 6, the mean of “I would recommend the program to a colleague” was 4.54 (±1.13); similarly, “I would rather do training in a live workshop than with the computer” was also 4.54 (±1.05).

ePST appears to be only the second computer program to evaluated the potential of a computer-automated treatment to train clinicians, and this appears to be the first study to use a behavioral outcome measure (i.e., standardized patient roleplays). The researchers identified one other computer-automated treatment program that has been used to teach clinicians how to perform a skill. FearFighter, a UK program for these treatment of panic and phobias [30], was used to teach the process of exposure therapy in two studies [31, 32]. Both studies used written measures to assess instructional gains, and both concluded that educational gains were equivalent to classroom training.

This study advances the literature on self-instructional training of therapists, and of clinicians in general, by using an e-therapy program as a training tool to teach clinicians an evidence-based treatment. Research on self-directed/computer-automated/online training has reported mixed training outcomes, likely because the pedagogies of the tools vary widely, from passively viewed didactics to written manuals to interactive programs [33]. Highly interactive training experiences, such as using ePST, have been demonstrated to produce superior outcomes regarding therapists’ skill and actual behavior in the implementation of evidence-based treatments [33].

4.1. Limitations and Future Directions

This study has several limitations, including the mixed educational background of the sample and the lack of a comparison group design, which limit generalizability. Without a comparison group, it is impossible to say for certain what trainees learned from ePST versus from the live workshop. This study also had a relatively small sample size; however, we have noted that training studies often enroll fewer participants than treatment studies. Additionally, the use of standardized patients is a well-accepted measure of clinician skill [34, 35]; however, participants’ performance might have been different with actual patients. Confidence in these results would be strengthened if interrater reliability of participant skill (in the roleplays) were higher. Finally, trainees may have learned from participating in the roleplay (the first of which happened after the post-ePST assessment and before the workshop); this effect might have artificially inflated the estimates of learning from the workshop, thus leading to an underestimate of the full effect of ePST training.

Regarding modifications to the ePST software, many trainees suggested the inclusion of a modeling video demonstrating the use of PST with a patient. Such a video has already been recorded and can be incorporated into a training version of ePST. Also, there exists a wealth of print training material for PST that could augment the computer-based training. Finally, supervision of new PST therapists is a necessary component of PST training [25, 26] and is unlikely to be automated.

Several national and statewide initiatives are implementing evidence-based treatments for depression and could benefit from cost-effective training, such as those by the National Network of Depression Centers [36], The University of Washington’s Improving Mood: Promoting Access to Collaborative Treatment (IMPACT) [37] and its Program to Encourage Active, Rewarding Lives for Seniors (PEARLS) [38] programs, and Depression Improvement Across Minnesota, Offering a New Direction (DIAMOND) [39]. ePST opens the door to a new training method that could promote the mass dissemination of PST among healthcare providers in order to improve patient access to care and ultimately reduce the impact of depression in the population.

Appendices

A. PST Knowledge

A.1. Problem-Solving Treatment Knowledge Assessment

Authors
James A. Cartreine, Brigham and Women’s Hospital/Harvard Medical School, Boston, MA.
Janette L. Seville, Geisel School of Medicine at Dartmouth, Hanover, NH.
Trina E. Chang, Massachusetts General Hospital/Harvard Medical School, Boston, MA.
Mark T. Hegel, Geisel School of Medicine at Dartmouth, Hanover, NH.

Instructions
In the space below, please answer the following essay question. You have 10 minutes.

Please describe the process of problem-solving therapy in detail, including all steps and the criteria for successfully completing each one.

Scoring Criteria
Assess whether the written response addresses each element below, on a scale of 0–5.

012345
Very PoorPoorBorderlineSatisfactoryGoodVery Good

NumberElementSubscale
(1)Problem should be observableProblem
(2)Problem should be under client’s controlProblem
(3)Goal should be behavioral (something client does)Goal
(4)Goal should be general (more than one way to reach it)Goal
(5)Goal should be observable (or countable)Goal
(6)Goal should be achievable (or short term, i.e., 2 weeks)Goal
(7)Goal should not just be the converse of the problemGoal
(8)Brainstorming should include multiple ideas/solutionsBrainstorming
(9)Do not prejudge solutionsBrainstorming
(10)Pros are identified for each idea/solution (or what makes each unique)Solutions
(11)Multiple cons are assessed for each idea/solutionSolutions
(12)Effort to implement solution is assessedSolutions
(13)Time to implement solution is assessedSolutions
(14)Cost of solution is assessedSolutions
(15)Need for other people to implement solution is assessedSolutions
(16)Negative impact on others is assessedSolutions
(17)One or more solutions must be chosenSolutions
(18)Action plan can be implemented soonAction plan
(19)Action plan is step by step, detailedAction plan
(20)Action plan includes Who, What, Where, and When are the steps to be takenAction plan
(21)Anticipate obstacles and plan around themAction plan
(22)Plan B’s (backup plans) are createdAction plan
(23)Enjoyable activities should be scheduled for each weekPleasant events
(24)Clinician checks whether action plan was implemented and goal was reachedProgress check
(25)Clinician checks client’s satisfaction with his or her effortProgress check
(26)Clinician helps clients troubleshoot how to improve their problem solvingProgress check

B. PST Skill

B.1. Problem-Solving Treatment in Primary Care Therapist Adherence and Competence Scale

Version 8-13-09

Authors
Mark T. Hegel, Ph.D., Geisel School of Medicine at Dartmouth, Hanover, NH.
Laurence Mynors-Wallis, M.D., University of Southampton, UK.

Rater Instructions
For each item, assess the therapist on a scale of 0–5 and record the rating on the line next to the item number.

012345
Very PoorPoorBorderlineSatisfactoryGoodVery Good
(1) Evaluating the outcomeReview of all current tasksPraise successExploration of failureRate Satisfaction and MoodReinforce PST-PC ModelReview previous problem areas(2)Defining the problem Specific, feasible problem chosenDescribed in objective termsProblem explored, clarifiedIdentify barriers(3)Establishing a realistic goal Goal is objectiveDescribed in behavioral termsGoal is achievableGoal is generalFollows directly from problem statementAddresses barriers identified in problem(4)Generating solutions Prime for brainstormingBrainstorming facilitatedSolutions from patientWithhold judgment(5)Implementing Decision-Making Guidelines and Choosing the Solution(s)Consider “pros” and “cons” for one’s self/othersRate each themeCompare solutionsSolution(s) satisfies the goalsNegative impact is limited(6)Implementing the preferred solution(s)Specific tasks identifiedRealistic behavior requirementsPlan BPlan pleasant activities for the week(7)Process tasksClear demarcation of PST-PC stagesKept session near a 30-minute timeframeCue and review for stagesSummarize process at end of sessionFacilitate a positive problem orientationFacilitate independence in guiding PST process(8)Communication and interpersonal effectivenessFacilitates communication (supportive vocalizations/nonverbals)Use of patient’s own language and phrasesWarm/confident/professionalTactful limiting of peripheral and unproductive discussion(9)Global ratingHow would you rate the problem-solving therapist overall in this session? (does not need to approach a mathematical average of previous eight items)

C. Self-Efficacy

C.1. PST and Counseling Self-Estimate Inventory

Authors
Based on Larson et al. [27]. (PST-specific questions added by the authors.)

Instructions
This is not a test. There are no right or wrong answers. Rather, this is an inventory that attempts to measure how you think you behave doing problem-solving therapy. Please respond to the items as honestly as you can to most accurately portray how you think you behave as a PST therapist. Do not respond with how you wish you could perform each item, or think you might in the future. Rather, answer in a way that reflects your actual estimate of how you perform as a counselor at the present time.

123456
Strongly DisagreeStrongly Agree

ItemSubscale
(1)I can effectively redirect clients who choose to work on problems over which they have limited controlPST
(2)I am likely to impose my values on the client during the interviewGeneral
(3)When I initiate the end of a session, I am positive it is in a manner that is neither abrupt nor brusque, and I end sessions on timeGeneral
(4)I can help clients construct action plans that meet the criteria of problem-solving therapyPST
(5)I feel that I will not be able to respond to the client in a nonjudgmental way with respect to the client’s values, beliefs, and so forthGeneral
(6)I feel that I can respond to the client in an appropriate amount of time (neither interrupting the patient nor waiting too long to respond)General
(7)I anticipate that the type of response I use at a particular time may not fit with the problem-solving therapy approachPST
(8)I can help clients with the brainstorming step of problem-solving therapyPST
(9)I feel confident that I have resolved conflicts in my personal life so that they will not interfere with my counseling abilitiesGeneral
(10)I feel that I have enough fundamental knowledge to do effective problem-solving therapyPST
(11)I am able to respond in a helpful way when clients report that they have not worked on a problem since last sessionPST
(12)I may not be able to maintain the intensity and energy level needed to produce client confidence and active participationGeneral
(13)I am not sure that when doing problem-solving therapy I will express myself in a way that is natural without deliberating over every response or actionPST
(14)I am confident that I can conduct problem-solving therapy, adhering to the guidelines set out in the PST therapy manualPST
(15)My assessments of client problems may not be as accurate as I would like them to bePST
(16)I am confident that I can help patients define their problems in a manner suitable for problem solvingPST
(17)I do not feel I possess a large enough repertoire of techniques to deal with the different problems my client may presentGeneral
(18)I am uncomfortable about dealing with clients who appear unmotivated to work toward their goalsPST
(19)I have difficulty dealing with clients who do not verbalize their thoughts during the counseling sessionGeneral
(20)I am unsure how to deal with clients who appear noncommittal and indecisiveGeneral
(21)I am an effective counselor with clients of a different socioeconomic statusGeneral
(22)I am unsure how to lead my client to work toward concrete goalsPST
(23)I am confident that I can assess my client’s readiness and commitment to work on a given problemPST
(24)When working with ethnic minority clients I am confident that I will be able to bridge cultural differences in the counseling processGeneral
(25)I am confident that I can help my patients evaluate the pros and cons of their solutions and choose a feasible solution in an efficient and helpful mannerPST
(26)I am confident that I can support the client in choosing his or her own problems on which to workPST
(27)I feel I may give adviceGeneral
(28)In working with culturally different clients, I have a difficult time viewing situations from their perspectiveGeneral
(29)I anticipate having difficulty helping clients write goal statements that meet all of the problem-solving therapy criteriaPST
(30)I am afraid that I may not be able to effectively relate to someone of lower socioeconomic status than meGeneral

D. Acceptability

D.1. Program Acceptability Questionnaire

Author
James A. Cartreine, Brigham and Women’s Hospital/Harvard Medical School, Boston, MA.

Instructions
Please read each item carefully and circle a number to show how much you agree or disagree with the statement.

123456
Strongly DisagreeStrongly Agree
(1)I felt comfortable using the PST program for training(2)Doing training using this program was acceptable to me(3)Using the program helped me understand how to do PST(4)I enjoyed using the program to learn PST(5)I would rather do training in a live workshop than with the computer(6)I would recommend this program to a colleague who is interested in learning how to do PST

Disclosures

The ePST software was designed by Cartreine (née Carter). When this study was conducted, the ePST software was owned by Beth Israel Deaconess Medical Center and Dartmouth Medical School. The ePST software is now owned by Cognitive Behavioral Technologies, LLC, which was founded by Cartreine and Hegel. Data analysis was conducted by another author (TC) in consultation with a statistical analyst, both who have no financial interest in ePST.

Acknowledgments

Development of the ePST software was supported by the National Space Biomedical Research Institute through NASA NCC 9–58. The authors wish to thank Ms. Rie Maurer for her statistical assistance and Ms. Lynn Kerner of Advocates, Inc., Framingham Massachusetts, and Mr. Steven Corso of Baypath Elder Services, Marlborough, Massachusetts, for their assistance in the recruitment and administration of this study. They also wish to thank Mr. Fred Connors and the Staff of The Troupe Modern Media (http://www.thetroupe.com/), of Windham, New Hampshire, which produced the ePST software and research assistants Kate Bentley, Elizabeth Dalton, Tracie Goodness, Irene Shyu, Jessica Sousa, Soo Youn, and Carrie Brintz.

References

  1. M. Dwight-Johnson, C. D. Sherbourne, D. Liao, and K. B. Wells, “Treatment preferences among depressed primary care patients,” Journal of General Internal Medicine, vol. 15, no. 8, pp. 527–534, 2000. View at Publisher · View at Google Scholar · View at Scopus
  2. P. Lin, D. G. Campbell, E. F. Chaney et al., “The influence of patient preference on depression treatment in primary care,” Annals of Behavioral Medicine, vol. 30, no. 2, pp. 164–173, 2005. View at Publisher · View at Google Scholar · View at Scopus
  3. R. Mergl, V. Henkel, A. K. Allgaier et al., “Are treatment preferences relevant in response to serotonergic antidepressants and cognitive-behavioral therapy in depressed primary care patients? Results from a randomized controlled trial including a patients' choice arm,” Psychotherapy and Psychosomatics, vol. 80, no. 1, pp. 39–47, 2010. View at Publisher · View at Google Scholar · View at Scopus
  4. S. R. Woody, J. Weisz, and C. McLean, “Empirically supported treatments: 10 years later,” The Clincical Psychologist, vol. 58, no. 4, pp. 5–11, 2005.
  5. F. Azocar, B. Cuffel, W. Goldman, and L. McCarter, “The impact of evidence-based guideline dissemination for the assessment and treatment of major depression in a managed behavioral health care organization,” Journal of Behavioral Health Services and Research, vol. 30, no. 1, pp. 109–118, 2003. View at Publisher · View at Google Scholar · View at Scopus
  6. A. C. Bell and T. J. D'Zurilla, “Problem-solving therapy for depression: a meta-analysis,” Clinical Psychology Review, vol. 29, no. 4, pp. 348–353, 2009. View at Publisher · View at Google Scholar · View at Scopus
  7. P. Cuijpers, A. van Straten, and L. Warmerdam, “Problem solving therapies for depression: a meta-analysis,” European Psychiatry, vol. 22, no. 1, pp. 9–15, 2007. View at Publisher · View at Google Scholar · View at Scopus
  8. J. M. Malouff, E. B. Thorsteinsson, and N. S. Schutte, “The efficacy of problem solving therapy in reducing mental and physical health problems: a meta-analysis,” Clinical Psychology Review, vol. 27, no. 1, pp. 46–57, 2007. View at Publisher · View at Google Scholar · View at Scopus
  9. L. M. Mynors-Wallis, D. H. Gath, A. Day, and F. Baker, “Randomised controlled trial of problem solving treatment, antidepressant medication, and combined treatment for major depression in primary care,” British Medical Journal, vol. 320, no. 7226, pp. 26–30, 2000. View at Scopus
  10. L. M. Mynors-Wallis, D. H. Gath, A. R. Lloyd-Thomas, and D. Tomlinson, “Randomised controlled trial comparing problem solving treatment with amitriptyline and placebo for major depression in primary care,” British Medical Journal, vol. 310, no. 6977, pp. 441–445, 1995. View at Scopus
  11. T. E. Oxman, M. T. Hegel, J. G. Hull, and A. J. Dietrich, “Problem-solving treatment and coping styles in primary care for minor depression,” Journal of Consulting and Clinical Psychology, vol. 76, no. 6, pp. 933–943, 2008. View at Publisher · View at Google Scholar · View at Scopus
  12. P. A. Arean, M. Hegel, S. Vannoy, M. Y. Fan, and J. Unuzter, “Effectiveness of problem-solving therapy for older, primary care patients with depression: results from the IMPACT project,” Gerontologist, vol. 48, no. 3, pp. 311–323, 2008. View at Scopus
  13. M. T. Hegel and P. A. Arean, “Problem-solving Treatment for Primary Care (PST-PC): A Treatment Manual for Depression,” 2003, http://impact-uw.org/tools/pst_manual.html.
  14. M. T. Hegel, J. Imming, M. Cyr-Provost, P. H. Noel, P. A. Arean, and J. Unutzer, “Role of behavioral health professionals in a collaborative stepped care treatment model for depression in primary care: project IMPACT,” Families, Systems and Health, vol. 20, no. 3, pp. 265–277, 2002. View at Scopus
  15. D. A. Eckerman, C. A. Lundeen, A. Steele, H. L. Fercho, T. A. Ammerman, and W. K. Anger, “Interactive training versus reading to teach respiratory protection,” Journal of Occupational Health Psychology, vol. 7, no. 4, pp. 313–323, 2002. View at Publisher · View at Google Scholar · View at Scopus
  16. R. E. Mayer, “The promise of multimedia learning: using the same instructional design methods across different media,” Learning and Instruction, vol. 13, no. 2, pp. 125–139, 2003. View at Publisher · View at Google Scholar · View at Scopus
  17. R. E. Mayer, G. T. Dow, and S. Mayer, “Multimedia learning in an interactive self-explaining environment: what works in the design of agent-based microworlds?” Journal of Educational Psychology, vol. 95, no. 4, pp. 806–813, 2003. View at Publisher · View at Google Scholar · View at Scopus
  18. H. H. Teo, L. B. Oh, C. Liu, and K. K. Wei, “An empirical study of the effects of interactivity on web user attitude,” International Journal of Human Computer Studies, vol. 58, no. 3, pp. 281–305, 2003. View at Publisher · View at Google Scholar · View at Scopus
  19. J. A. Carter, J. C. Buckey, L. Greenhalgh, A. W. Holland, and M. T. Hegel, “An interactive media program for managing psychosocial problems on long-duration spaceflights,” Aviation Space and Environmental Medicine, vol. 76, no. 6, pp. B213–B223, 2005. View at Scopus
  20. D. F. Peck, “The therapist-client relationship, computerized self-help and active therapy ingredients,” Clinical Psychology and Psychotherapy, vol. 17, no. 2, pp. 147–153, 2010. View at Publisher · View at Google Scholar · View at Scopus
  21. K. Kroenke and R. L. Spitzer, “The PHQ-9: a new depression diagnostic and severity measure,” Psychiatric Annals, vol. 32, no. 9, pp. 509–515, 2002. View at Scopus
  22. T. T. Hewett, “Importance of failure analysis for human-computer interface design,” Interacting with Computers, vol. 3, no. 1, pp. 3–8, 1991. View at Scopus
  23. S. I. Nishida, Failure Analysis in Engineering Applications, Materials and Corrosion, vol. 43, Butterworth-Heinemann, Oxford, UK, 1992.
  24. A. Sutcliffe and G. Rugg, “A taxonomy of error types for failure analysis and risk assessment,” Plastics, Rubber and Composites Processing and Applications, vol. 10, no. 4, pp. 381–405, 1998. View at Scopus
  25. M. T. Hegel, J. E. Barrett, and T. E. Oxman, “Training therapists in problem-solving treatment of depressive disorders in primary care: lessons learned from the ‘treatment effectiveness project‘,” Families, Systems and Health, vol. 18, no. 4, pp. 423–435, 2000. View at Scopus
  26. M. T. Hegel, A. J. Dietrich, J. L. Seville, and C. B. Jordan, “Training residents in problem-solving treatment of depression: a pilot feasibility and impact study,” Family Medicine, vol. 36, no. 3, pp. 204–208, 2004. View at Scopus
  27. L. M. Larson, L. A. Suzuki, K. N. Gillespie, M. T. Potenza, M. A. Bechtel, and A. L. Toulouse, “Development and validation of the counseling self-estimate inventory,” Journal of Counseling Psychology, vol. 39, no. 1, pp. 105–120, 1992. View at Publisher · View at Google Scholar · View at Scopus
  28. A. Bangor, P. T. Kortum, and J. T. Miller, “An empirical evaluation of the system usability scale,” International Journal of Human-Computer Interaction, vol. 24, no. 6, pp. 574–594, 2008. View at Publisher · View at Google Scholar · View at Scopus
  29. J. Brooke, “SUS: a quick and dirty usability scale,” in Usability Evaluation in Industry, P. W. Jordan, et al., Ed., pp. 189–194, Taylor & Francis, London, UK, 1996.
  30. M. Kenwright, S. Liness, and I. Marks, “Reducing demands on clinicians by offering computer-aided self-help for phobia/panic. Feasibility study,” British Journal of Psychiatry, vol. 179, pp. 456–459, 2001. View at Publisher · View at Google Scholar · View at Scopus
  31. L. Gega, I. J. Norman, and I. M. Marks, “Computer-aided vs. tutor-delivered teaching of exposure therapy for phobia/panic: randomized controlled trial with pre-registration nursing students,” International Journal of Nursing Studies, vol. 44, no. 3, pp. 397–405, 2007. View at Publisher · View at Google Scholar · View at Scopus
  32. M. McDonough and I. M. Marks, “Teaching medical students exposure therapy for phobia/panic—randomized, controlled comparison of face-to-face tutorial in small groups vs. solo computer instruction,” Medical Education, vol. 36, no. 5, pp. 412–417, 2002. View at Publisher · View at Google Scholar · View at Scopus
  33. R. S. Beidas and P. C. Kendall, “Training therapists in evidence-based practice: a critical review of studies from a systems-contextual perspective,” Clinical Psychology, vol. 17, no. 1, pp. 1–30, 2010. View at Publisher · View at Google Scholar · View at Scopus
  34. L. A. H. Erby, D. L. Roter, and B. B. Biesecker, “Examination of standardized patient performance: accuracy and consistency of six standardized patients over time,” Patient Education and Counseling, vol. 85, no. 2, pp. 194–200, 2011. View at Publisher · View at Google Scholar · View at Scopus
  35. M. Shirazi, M. Sadeghi, A. Emami, et al., “Training and validation of standardized patients for unannounced assessment of physicians' management of depression,” Academic Psychiatry, vol. 35, no. 6, pp. 382–387, 2011.
  36. The National Network of Depression Centers, “About Us, The National Network of Depression Centers, Ann Arbor, MI,” 2012, http://www.nndc.org/about-us.
  37. University of Washington, “IMPACT Evidence-based Depression Care, Seattle, WA: University of Washington,” 2011, http://impact-uw.org/.
  38. Harborview Medical Center, “PEARLS: Together We Create Active and Rewarding Lives, Seattle, WA: Harborview Medical Center,” 2011, http://www.pearlsprogram.org/.
  39. Institute for Clinical Systems Improvement, “DIAMOND, Bloomington, MN: Institute for Clinical Systems Improvement,” 2011, http://www.icsi.org/health_care_redesign_/diamond_35953/.