Globally, and especially in low- and middle-income countries (LMIC), there is a large treatment gap for mental disorders, as indicated by the lack of service for a great majority of people in need of mental healthcare.Reference Thornicroft, Chatterji, Evans-Lacko, Gruber, Sampson and Aguilar-Gaxiola1 Integrating mental healthcare into primary healthcare has been recommended as a way of bridging this treatment gap by offering more accessible, holistic and less stigmatising services to people in need.2 The World Health Organization (WHO) launched the Mental Health Gap Action Programme Intervention Guide (mhGAP-IG) in 2010 (and a second version in 2016), to provide evidence-based clinical guidance for the assessment and management of priority mental, neurological and substance use disorders in LMIC non-specialist settings.3 Building on this development, and in order to facilitate the process of integration, a broader health system strengthening approach involving policy changes, sustainable financing mechanisms and workforce development is required.
One such system requirement is a well-functioning routine health information system that ensures the production, analysis, dissemination and use of reliable and timely information related to service delivery.Reference Aqil, Lippeveld and Hozumi4 Moreover, in order for the much-needed investments in scaling up of mental health services to be efficiently used, comprehensive and accurate information on the availability, utilisation and equity of such services is required. However, health information systems to support actions in mental healthcare are weak in LMIC. Most routine information systems do not have any, or have very limited, indicators related to mental healthcare. When information is collected, it often produces unreliable data.Reference Upadhaya, Jordans, Lund, Alem, Hanlon and Kigozi5 The aim of this study was to evaluate the performance and perceived utility of a new set of indicators for routine monitoring of mental healthcare within primary healthcare settings in five LMIC.
Method
Setting
The study was conducted in five LMIC in Africa and Asia participating in the Emerald programme (India, Nepal, Nigeria, South Africa and Uganda). The Emerald program aimed to improve mental health outcomes by generating evidence and capacity to enhance health system performance in LMIC.Reference Semrau, Evans-Lacko, Alem, Ayuso-Mateos, Chisholm and Gureje6
Emerald programme
As a way of responding to the lack of indicators to monitor mental healthcare, the Emerald programme developed, through a Delphi study, a set of indicators that can be used to appraise mental healthcare needs, utilisation, outcomes and costs.Reference Jordans, Chisholm, Semrau, Upadhaya, Abdulmalik and Ahuja7 From an initial set of 52 generated items, the eight most highly prioritised indicators were selected and operationalised through consultative workshops in each country setting convened to develop data collection formats suitable for use by frontline health workers. The eight indicators included were: diagnosis (as multiple-choice response options, which included depression, alcohol use disorder, schizophrenia and epilepsy, as well as other disorders depending on priorities per country team), exact diagnosis (as open response format), severity, functioning, administered treatment, referral, follow-up and payment for services (the latter six items all had multiple response options, including none or not relevant). These formats were subsequently integrated into district-level mental healthcare programmes, as part of ongoing research programmes evaluating the integration of mental healthcare into primary healthcare.Reference Lund, Tomlinson, De Silva, Fekadu, Shidhaye and Jordans8,Reference Eaton, Gureje, Silva, Sheikh, Ekpe and Abdulaziz9 (See supplementary File 1, available at https://doi.org/10.1192/bjp.2019.22, for the standard form, which was adapted in each setting.)
Instruments
First, a structured questionnaire was developed for the purpose of this study, to assess health workers' perceptions of the utility (i.e. acceptability and feasibility) of using the data collection format for the new indicators. The questionnaire consisted of 14 questions, which were administered through interviews conducted by trained research assistants. Second, a health information case record review form was developed to assess the performance of the newly introduced indicators based on observation of the completed formats. A trained research assistant checked all case records, first assessing whether the completion of each of the indicators (i.e. level of completion (response options: ‘yes/no’)). Subsequently, for each completed indicator, the research assistant assessed the level to which completed information was correctly or incorrectly completed (i.e. level of correctness of completion (response options: ‘correct/incorrect’), with the latter defined as ‘illogical completion/missing information/illegible information’). The form consisted of eight items and the ratings of its completion and correctness of completion were made by trained research assistants.
Sample
See Table 1 for an overview of the number of interviews and case record reviews across sites and time-points. The structured interviews were administered to 272 health workers who had been trained to provide services and to use the new data collection forms. The sample was selected from primary healthcare facilities (all countries) and a few district hospitals (India and Uganda). Selection was done either randomly (in Nepal, among the 205 trained health workers, 177 were still in post, among whom 50 were randomly selected; in Nigeria, two health workers were randomly selected among the 4–6 trained health workers per facility) or purposively (in South Africa, staff were selected based on availability and willingness until the target sample was reached; in Uganda and India, the health facility in charge approached health workers involved in record keeping or mental health services, respectively). The health information records (n = 586) assessed for completion and correctness were selected randomly in all of the countries. Randomisation of the case records was done by: (a) randomly selecting a specified number of facilities and (b) randomly selecting a number of case records within the selected facility.
T1 refers to 2–4 weeks after introduction of the new indicators within primary healthcare settings; T2 refers to 6–9 months after the introduction.
Procedure
The interviews with health workers and the health information record reviews were performed 2–4 weeks after the new forms were introduced and had started to be used in the clinics. We chose the period of 2–4 weeks to allow some time for proper adoption of the procedure post training, including the printing and distribution of the required registers/forms. This time frame was to allow the staff some time to establish familiarity with the new forms before the interviews/assessments were conducted at baseline. A second wave of interviews and record reviews were conducted with the same health workers and in the same health facilities 6–9 months after the introduction of the forms, to assess whether there was improvement, deterioration or constancy in the scores over time. Interviews and record reviews were conducted by research assistants who had received an average of 5 days training at each of the participating sites.
Data analysis and management
Data from questionnaires and record review forms (for both time points; T1 and T2) were entered into SPSS and analysed using descriptive statistics. The calculation of the proportion of completion of indicators was based on the total number of cases; correctness was calculated based on the completed indicators and was determined by whether the information entered in the data form was assessed by the research assistant to be correct. Thus, for example, if the indicator slot for ‘diagnosis’ was filled, a rating was made to indicate that the information was indeed collected (completion) and whether the information filled into the slot was indeed a relevant diagnosis (correctness). Data collection took place between August 2015 and October 2017.
Ethics
Ethical approval for this study was obtained from King's College London, the World Health Organization and the institutional review boards of each of the participating sites. Subsequently, permission was obtained from the appropriate Department/Ministry of Health officials, as well as from the healthcare workers and health facility managers. Written informed consent was sought and obtained from all study participants. The facility managers were approached by research assistants for their participation in the study and for permission to access the health information records.
Results
Across the five sites, 272 health workers were interviewed and 586 case records were reviewed. Table 2 shows the performance ratings of the new indicators with regard to levels of completion and correctness of completion. Overall, high levels of completion and of correctness were achieved for most of the indicators at both time points. The combined completion and correctness ratings for both time points were especially high for diagnosis (94 and 95%, respectively), severity (83 and 99%, respectively) and treatment (92 and 86%, respectively). These scores were slightly lower for exact diagnosis (72 and 84%, respectively), functioning (76 and 83%, respectively) and follow-up (83 and 82%, respectively). To give an indication of the samples of cases for which records were reviewed, the breakdown of diagnoses recorded was as follows (for T1 data): AUD (18% in Nepal, 3% in Uganda, 20% in India, 22% in Nigeria); depression (28% in Nepal, 100% in South Africa, 14% in Uganda, 46% in India, 60% in Nigeria); epilepsy (13% in Nepal, 68% in Uganda, 5% in Nigeria); psychosis (32% in Nepal, 5% in Uganda, 1% in India, 13% in Nigeria); and others (such as anxiety disorders, dementia, development disorders; 9% in Nepal, 10% in Uganda, 33% in India).
a. The response options and the denominators for these items were different from the other countries; these are therefore not included in the summaries of results.
b. Correctness refers to level of completed indicators that was correctly completed (as opposed to ‘incorrect completion’).
The combined completion and correctness for both time points was lowest for referral (60 and 84%, respectively) and payment (58 and 72%, respectively). The lower scores for referral and payments were all clustered within one country setting (i.e. Uganda), where 7/8 measurements points showed results that were <50% positive. The only other trend for such lower scores were for the completion of indicators at T1 in Nepal (4/8 scoring 50% positive). In South Africa scores for completion and correctness were high across indicators and time points with 18/24 measurements showing 100% positives. Furthermore, the T2 correctness measurements for Nepal and Nigeria were all over 90%. These contributed to a cross-country trend for improvement of scores between T1 and T2 for completion of indicators (7/8 showed improvement over time) and for correctness of completion (6/8 showed improvement). There was only a cross-country trend for deterioration over time for completion of the severity indicator.
The utility of the new mental health indicators was assessed through health workers' perceptions of the feasibility and acceptability of using the indicators in routine healthcare settings across the five countries (Table 3). Overall, there were moderate-to-high scores on perceived utility. All but two items had over 80% positive responses (positive responses were defined as the combination of ‘agree’ and ‘strongly agree’) across time and countries. These were especially high for perceived importance of routine mental health indicators (96% positive) motivation to use these indicators (95%), as well as perceived relevance (91%), integration in health workers' daily work (90%), ease of use (86%), confidence to use (84%) and overall satisfaction with the new forms (82%). Perceived usefulness of the indicators in daily practice fell a bit behind, with 74% positive responses. Having insufficient time was the only item where only a minority had a positive response (33%). Indeed, a combined average of 27% of the reports showed that more than 10 min additional time was spent on recording.
Over time, there was an overall reduction of perceived utility. In 15 of the T1–T2 comparisons, there was a reduction of 20% or greater. These were especially seen in South Africa. In only one instance was there an increase of more than 20% in positive evaluation for an item from T1 to T2; this was for the level of confidence in using the formats in India. Finally, Nepal and Nigeria were the two sites with the most positive assessments of feasibility and acceptability by respondents (i.e. both scoring over 80% positive on 14/18 items).
Discussion
A functioning routine mental health information system is required to guide the process of reducing the treatment gap for mental illness globally.Reference Ryan, De Silva, Terver, Ochi and Eaton10 The poor availability of data is a hindrance for the achievement of that goal in LMIC.Reference Eaton, McCay, Semrau, Chatterjee, Baingana and Araya11 A mental health information system aims to improve the effectiveness and efficiency of the mental health service and ensure more equitable delivery. Information that enables managers and service providers to make more informed decisions could aid the delivery of quality care.12 Given that most LMIC do not have any, or only very few, indicators to monitor mental healthcare as part of their existing Health Management Information System (HMIS), this study aimed to assess the performance and utility of a new set of indicators in five LMIC (India, Nepal, Nigeria, South Africa and Uganda). This is in line with WHO's Mental Health Action Plan 2013–2020, which sets a target of 80% of countries to routinely collect and report on a core set of mental health indicators within their national information systems by 2020.13 The new set of indicators go beyond monitoring of diagnoses, commonly the only indicator included in information systems in LMIC,Reference Upadhaya, Jordans, Lund, Alem, Hanlon and Kigozi5 to include subsequent need for, and utilisation of services, as well as related progress and cost.
Across countries we found that a newly developed and introduced set of indicators for routine monitoring of mental health service demonstrated good performance and high levels of perceived utility by non-specialist primary healthcare workers. First, we found high levels of completion (78% when considering all indicators) as well as high levels of accurate completion (87%). This was especially so for indicators on diagnosis (using multiple choice responses), severity of patients' problems and treatment provided. Documentation of functioning and follow-up also had good completion and correctness rates. Items capturing payment and referrals scored lower, even though this was mainly limited to one site (Uganda). Moreover, performance seemed to improve over time. These findings are important given the lack of studies examining the performance of indicators for monitoring mental health services. A systematic review of 106 publications on performance indicators for public mental healthcare found that only two studies explicitly assessed issues of data reliability, even though inadequate completeness and correctness are known risks for utility of data.Reference Lauriks, Buster, de Wit, Arah and Klazinga14
Second, health workers' views about the feasibility of the routine conduct of mental healthcare monitoring using the new forms were generally positive across the five settings (and especially so in Nepal and Nigeria). The majority of measurements across countries and time points had more than 80% positive responses regarding the importance attached to them, the motivation and ease of their use, the usefulness of the collected information, the perceived need for the indicators and level of confidence in using them. Health workers were less positive about the additional time required for the completion of the forms. There was a trend of a downward drift in positive attitudes over time, though these were mostly observed in South Africa.
While the positive results were quite uniform across sites and time-points, the especially high levels of perceived feasibility and acceptability in the first month after training and introduction of the indicators suggests that ongoing training and supervision may contribute to keeping such positive attitudes. At the same time, reduction in positive attitudes over time did not result in reduced performance 6–9 months afterwards. On the contrary, continued use actually seemed to improve the results over time. The concern of health workers with regard to the time required for the additional reporting is understandable given the high workload in many of the understaffed health facilities in LMIC.Reference Ahuja15 It is nevertheless instructive that this concern did not seem to translate to poor performance or negative attitudes towards the importance or feasibility of implementing the surveillance system.
The set of indicators tested included those measuring needs, utilisation and outcomes. The combination of these would provide information necessary to assess coverage. Evaluating coverage of mental health programmes is essential in order to track efforts to scale up services for people with mental disorders.Reference De Silva, Lee, Fuhr, Rathod, Chisholm and Schellenberg16 It is therefore reassuring that these indicators performed well in this study. The addition of an indicator for financial protection for mental health services would provide valuable data to help monitor the level of universal health coverage, an important goal within the United Nations Sustainable Development Goals. However, with only three countries (Nepal, Nigeria and Uganda) adopting the indicator in this study and with poorer performance compared with other indicators, it would appear that such an addition might be a step too far.
There are a number of limitations that should be considered in interpreting the findings of this study. First, the structured interviews with health workers might have introduced a bias towards social desirability, with a tendency for respondents to over-report use. Still, we saw low ratings on a number of questions, indicating that critical perspectives were also voiced. Second, not all countries used the same or the full set of indicator ratings. The variance between sites is a result of a process of in-country consultations and decision-making with regards to selection and formulation of the indicators. Third, this study did not include the assessment of (a) the quality of information that is reported using these new indicators (i.e. it did not include any validation of the completed information, for example, assessing the accuracy of the diagnosis given), or (b) the use of the information subsequent to collection (i.e. uptake within the larger HMIS system). The assessment of correctness of completion therefore only refers to whether and how the newly introduced formats were completed – a first threshold for adequate use in practice. Fourth, different countries utilised different selection methods for recruiting respondents (random, purposive and convenience), which may detract from the uniformity of the findings. Finally, this study focused exclusively on health facility level indicators directly related to service provision. Other health system indicators (e.g. number of trained personnel, availability of medications) are equally important for adequate monitoring of mental health services.Reference Jordans, Chisholm, Semrau, Upadhaya, Abdulmalik and Ahuja7
This study has several implications. First, efforts to close the mental health treatment gap should be matched with efforts to close the information gap associated with mental healthcare in LMIC.Reference Ryan, De Silva, Terver, Ochi and Eaton10 This is especially pertinent, as meaningful planning for scaling up mental health services cannot take place in the absence of reliable mental health information to serve as a reliable benchmark of service utilisation and trends over time. Based on the results of this study, a limited set of indicators can be recommended for adoption into routine mental healthcare surveillance, which will be particularly useful when rolling out mhGAP-IG programmes.
Second, in addition to improved planning, the integration of a set of mental health indicators in HMIS will be an important ingredient for maintaining or improving the quality of care. Health workers and/or their supervisors, can use the information to better monitor the adequacy of initiated treatments following diagnosis, treatment adherence or drop-out, as well as patients' improvements over time, or lack thereof.
Third, if the set of indicators that were part of this study were integrated in HMIS and adequately implemented to collect data within routine primary healthcare settings, they could potentially be used to monitor coverage for mental healthcare through the following combination of indicators; diagnosis, (change in) severity, (change in) functioning, (frequency of) treatment utilisation, and also in low-resource health systems. This is especially salient as most mental health services information is limited to tertiary care settingsReference Ahuja, Shidaye, Semrau, Thornicroft and Jordans17 and most focuses on health system inputs rather than processes or outcomes.12 Monitoring the level of financial protection may require a different method, for example, household surveys. Moreover, for calculation of actual coverage, reliable prevalence data are needed (i.e. the denominator), which may be unavailable in many settings.
Fourth, perceived utility is especially high shortly after training. Even though across the board introduction after training seems sufficient to achieve good performance, ongoing supervision and booster trainings may be needed to sustain positive attitudes towards use.Reference Ahuja15
Fifth, the next steps should involve the evaluation of using this set of indicators outside of a research programme. While the study took place in a routine care setting, a non-research setting will more closely reflect real-life implementation. Also, future research should evaluate the use of mental healthcare indicators within the HMIS process subsequent to data collection, i.e. processing, analyses, dissemination and use at facility levels, as well as beyond.
In conclusion, this study set out to explore the feasibility of monitoring mental healthcare when provided at scale by non-specialist health workers. The study demonstrates high levels of performance and utility across time for a set of indicators that could ultimately be used to monitor the coverage of mental healthcare in primary healthcare settings in five LMIC and contribute to maintaining or improving the quality of care. We recommend that these indicators be included in existing health information systems, and adopted as part of systems strengthening interventions necessary to facilitate the rolling out of the mhGAP-IG package, and to allow for evidence-based and data-driven decision-making and planning.
Supplementary material
Supplementary material is available online at https://doi.org/10.1192/bjo.2019.22.
Funding
This work was supported by the European Union's Seventh Framework Programme (FP7/2007–2013) under grant agreement no. 305968, and by the PRIME Research Programme Consortium, funded by the UK Department of International Development for the benefit of developing countries. G.T. is supported by the National Institute for Health Research (NIHR) Collaboration for Leadership in Applied Health Research and Care (CLAHRC) South London and by the NIHR Applied Research Centre (ARC) at King’s College London NHS Foundation Trust, and the NIHR Applied Research and the NIHR Asset Global Health Unit award. The views expressed are those of the author(s) and not necessarily those of the NHS, the NIHR or the Department of Health and Social Care. G.T. receives support from the National Institute of Mental Health of the National Institutes of Health under award number R01MH100470 (Cobalt study). G.T. is supported by the UK Medical Research Council in relation the Emilia (MR/S001255/1) and Indigo Partnership (MR/R023697/1). awards. I.P., C.L. and G.T. are funded by the NIHR Global Health Research Unit on Health System Strengthening in Sub-Saharan Africa, King’s College London (GHRU 16/136/54) using UK aid from the UK Government. The views expressed in this publication are those of the authors and not necessarily those of the NIHR or the Department of Health and Social Care. No funding bodies had any role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. The views expressed in this publication are not necessarily those of the funders.
Acknowledgements
We thank Drs Charlotte Hanlon and Atalay Alem and the TPO Nepal Emerald team for their work on this study. We also thank Grace Ryan (London School of Hygiene and Tropical Medicine) for her contributions to the development of the indicators and study design. The views expressed are those of the author(s) and not necessarily those of the NHS, the NIHR or the Department of Health and Social Care. D.C. is a staff member of the World Health Organization. The authors alone are responsible for the views expressed in this publication and they do not necessarily represent the decisions, policy or views of the World Health Organization.
eLetters
No eLetters have been published for this article.