Hostname: page-component-745bb68f8f-f46jp Total loading time: 0 Render date: 2025-01-22T04:16:28.727Z Has data issue: false hasContentIssue false

Semantic search helper: A tool based on the use of embeddings in multi-item questionnaires as a harmonization opportunity for merging large datasets – A feasibility study

Published online by Cambridge University Press:  20 January 2025

Karl Gottfried
Affiliation:
Institute of Applied Medical Informatics, University Hospital Center Hamburg-Eppendorf, Hamburg, Germany
Karina Janson
Affiliation:
Department of Child and Adolescent Psychiatry and Psychotherapy, Central Institute of Mental Health, Medical Faculty Mannheim, University of Heidelberg, Mannheim, Baden-Württemberg, Germany Institute of Medical Psychology and Medical Sociology, University Medical Center Schleswig-Holstein, Kiel University, Preußerstraße 1-9, Kiel, Schleswig-Holstein, Germany
Nathalie E. Holz
Affiliation:
Department of Child and Adolescent Psychiatry and Psychotherapy, Central Institute of Mental Health, Medical Faculty Mannheim, University of Heidelberg, Mannheim, Baden-Württemberg, Germany German Center for Mental Health (DZPG), Partnersite Mannheim-Heidelberg-Ulm, Germany
Olaf Reis
Affiliation:
Department of Child and Adolescent Psychiatry, Neurology, Psychosomatics and Psychotherapy, Rostock University Medical Centre, Rostock, Germany
Johannes Kornhuber
Affiliation:
Department of Psychiatry and Psychotherapy, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Erlangen, Germany
Anna Eichler
Affiliation:
Department of Child and Adolescent Mental Health, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Erlangen, Germany
Tobias Banaschewski
Affiliation:
Department of Child and Adolescent Psychiatry and Psychotherapy, Central Institute of Mental Health, Medical Faculty Mannheim, University of Heidelberg, Mannheim, Baden-Württemberg, Germany German Center for Mental Health (DZPG), Partnersite Mannheim-Heidelberg-Ulm, Germany
Frauke Nees*
Affiliation:
Department of Child and Adolescent Psychiatry and Psychotherapy, Central Institute of Mental Health, Medical Faculty Mannheim, University of Heidelberg, Mannheim, Baden-Württemberg, Germany Institute of Medical Psychology and Medical Sociology, University Medical Center Schleswig-Holstein, Kiel University, Preußerstraße 1-9, Kiel, Schleswig-Holstein, Germany
*
Corresponding author: Frauke Nees; Email: [email protected]

Abstract

Background

Recent advances in natural language processing (NLP), particularly in language processing methods, have opened new avenues in semantic data analysis. A promising application of NLP is data harmonization in questionnaire-based cohort studies, where it can be used as an additional method, specifically when only different instruments are available for one construct as well as for the evaluation of potentially new construct-constellations. The present article therefore explores embedding models’ potential to detect opportunities for semantic harmonization.

Methods

Using models like SBERT and OpenAI’s ADA, we developed a prototype application (“Semantic Search Helper”) to facilitate the harmonization process of detecting semantically similar items within extensive health-related datasets. The approach’s feasibility and applicability were evaluated through a use case analysis involving data from four large cohort studies with heterogeneous data obtained with a different set of instruments for common constructs.

Results

With the prototype, we effectively identified potential harmonization pairs, which significantly reduced manual evaluation efforts. Expert ratings of semantic similarity candidates showed high agreement with model-generated pairs, confirming the validity of our approach.

Conclusions

This study demonstrates the potential of embeddings in matching semantic similarity as a promising add-on tool to assist harmonization processes of multiplex data sets and instruments but with similar content, within and across studies.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press on behalf of European Psychiatric Association

Introduction

In large epidemiological and clinical studies in the field of mental health, data not only span a wide array of constructs, including behavior, cognitive patterns, personal attitudes, or beliefs that primarily come from questionnaires but also a wide array of instruments used to capture these constructs, both across and within studies and cohorts [Reference Lydeard1,Reference Kelley, Clark, Brown and Sitzia2]. If researchers aim to merge data from two or more studies for analysis, the gold standard is to use data collected with identical questionnaires or, at the very least, with comparable (sub)scales of these questionnaires, thereby ensuring adherence to standardized values. This can be achieved through the recourse to catalogs that provide an overview of measures used in different cohorts (e.g., https://lifecourse.melbournechildrens.com/cohorts/; https://www.cataloguementalhealth.ac.uk). However, the aspect of comparability already raises questions as it can be assessed and treated in different ways. Moreover, given the large number of existing studies and cohorts in mental health [Reference Bauermeister, Orton, Thompson, Barker, Bauermeister and Ben-Shlomo3Reference Chen, Jutkowitz and Gross6], it is often evident that studies or cohorts of interest have not used consistent questionnaires to measure the same constructs. Additionally, in longitudinal studies or cohorts, different questionnaires are sometimes used at various assessment points over time.

Therefore, it is particularly important to develop methods that enable the utilization of this data across different studies and projects for population-based analyses [Reference Kezios, Zimmerman, Buto, Rudolph, Calonico and Zeki Al Hazzouri7]. In this respect, there is a growing emphasis on expanding individual datasets through ex-post harmonization, which involves merging data to create a unified, comprehensive, and semantically fitting dataset that preserves the validity and reliability of the outcome measures [Reference Aizpurua, Fitzgerald, de Barros, Giacomin, Lomazzi and Luijkx8Reference Dahal, Patten, Williamson, Patel, Premji and Tough12]. This involves standardizing the available questionnaire data down to the item level [Reference Chen, Jutkowitz and Gross6,Reference Nguyen, Khadka, Phan, Yazidi, Halvorsen and Riegler13Reference Chen, Jutkowitz, Iosepovici, Lin and Gross15]. It requires a comprehensive understanding of the significance of each collected variable and methodological consistency across studies, including the definitions, measurement instruments, and data collection protocols used. By thoroughly examining these aspects, researchers can identify comparable variables, mitigate potential biases, and enhance the reliability and validity of harmonized datasets, ultimately facilitating robust cross-cohort analyses and more accurate scientific inferences. When data on common constructs are available only from different questionnaires, this harmonization extends beyond variables to individual questionnaire items to ensure that the items from different questionnaires measure the same and thus can be treated as indicators of the same construct. In this respect, it is important to determine the similarity of items from the questionnaires, which can be done by experts from the field, and thus on a subjective level, or by natural language processing (NLP)-based algorithms. For example, questionnaires like the Alcohol Use Disorders Identification Test (AUDIT) [Reference Chen, Jutkowitz, Iosepovici, Lin and Gross15] and the Michigan Alcohol Screening Test (MAST) [Reference Tomescu-Dubrow and Slomczynski16] have similar questions about alcohol consumption. For example, AUDIT’s question 8 (“How often during the last year have you been unable to remember what happened the night before because you had been drinking?”) inquiries about the frequency of memory lapses due to alcohol intake over a specific period. Similar content is assessed in MAST question 2 (“Have you ever awakened the morning after drinking the night before and found that you could not remember a part of the evening?”), albeit with different wording and structure. In this case, the question texts provide relevant information for assessing the potential for harmonization between the questionnaires. Thus, the question text serves as a form of meta-data description that can be beneficial for identifying semantic similarities across questionnaires [Reference McElroy, Wood and Bond17]. This semantic harmonization is a pre-requisite for any further harmonization processes, such as syntactical harmonization. Syntactical harmonization involves addressing distinct data types and the data granularity of the responses (e.g., AUDIT – numeric variable versus MAST dichotomous variable). This syntactical alignment can only be performed after semantic matching has been established. Finally, it is essential to implement a manual integration step, that captures afore mentioned aspects regarding existing catalogs and frameworks as well as the expert rating(s). This also includes technical interoperability tools such as meta-data repositories and uniform models as well as code systems, terminologies, and classifications that align the data harmonization with the FAIR principles [Reference Wilkinson, Dumontier and IjJ18]. Examples such as LOINC, SNOMED CT, RxNorm, or ICD-10 play a significant role in aligning data and are therefore increasingly used for this task [Reference Bietenbeck, Boeker and Schulz S19-Reference Sathappan, Jeon, Dang, Lim, Shao and Tai22]. These tools are specifically designed for incorporating lists and standardized information, and the related information should be captured also in the context of NLP-based semantic similarity tests. To facilitate the integration of a manual as well as a user-/expert-based perspective also means the use of advanced features to support various display options, resolutions, and value ranges. This integration requires advanced features to support various display options, resolutions, and value ranges. Beyond simply considering the degree of matching or network analysis, this enables subjective evaluation levels, important especially for critical elements or even already those that are set in a relatively high similarity range of 60%–70%, which still raises the question of the appropriateness of the data merging step. This approach could also initiate further processes that lay the groundwork for future studies on item combinations and their assessment, contributing to a more in-depth evaluation. The consideration of merging and comparing instruments in different languages is finally another important feature because often cohorts from different countries are available and can bring corresponding benefits, but standard instruments are not always validated for different languages.

In summary, comparing data from large and complex datasets in terms of constructs, significance, and indication levels is an immense and often unmanageable task, especially regarding the comparability of questionnaire data from different instruments. Given that manual harmonization in this respect is error-prone and resource-intensive [Reference Bandyopadhyay, Tingay, Borja, Griffiths, Akbari and Bedford9,Reference Bauermeister, Phatak, Sparks, Sargent, Griswold and McHugh10,Reference Dahal, Patten, Williamson, Patel, Premji and Tough12,Reference Tomescu-Dubrow and Slomczynski16], the full utilization of available multi-cohort data and missed potential opportunities is hindered [Reference O’Connor, Spry, Patton, Moreno-Betancur, Arnup and Downes23,Reference O’Connor, Olsson, Lange, Downes, Moreno-Betancur and Mundy24]. In this respect, automated methods that can pre-select semantic item pairs based on criteria such as formulation and content using NLP, such as transformer-based embeddings [Reference Vaswani, Shazeer, Parmar, Uszkoreit, Jones, Gomez, Kaiser and Polosukhin25,Reference Bi, Goodman, Kaminsky and Lessler26], may offer a solution. Embedding techniques from advanced language models convert words or sentences into numerical representations, allowing computers to process and compare them. These embeddings represent data in a way that preserves meaning, facilitating the identification of semantically similar items. This offers significant advantages over previously prevalent frequency-based approaches such as the frequency-inverse document frequency (TF-IDF) method [Reference Cer, Diab, Agirre, Lopez-Gazpio, Specia, Bethard, Carpuat, Apidianaki, Mohammad, Cer and Jurgens27Reference Kang, Yoon, Kim, Jo, Lee and Kam29]. Results are represented as distances in a virtual space, where words or sentences with similar meanings are closer together, allowing for a measurable comparison of their similarity. These techniques have shown high performance in tasks involving text-based semantic similarity comparison tasks [Reference Patil, Han and Jadon30Reference Devlin, Chang, Lee and Toutanova32], rendering them appropriate for semantic search within the framework of data harmonization [Reference Bergman, Sherwood, Forslund, Arlett and Westman33]. Platforms such as Hugging Face [34], spaCy [Reference Honnibal and spaCy35], or services like ChatGPT [36,37] enable these techniques. In the present article, we employ automated methods, together with advanced and extensive representation and visualization features, and language-based testings, to detect semantically similar questionnaire items related to mental health, while still making use of available terminology standards such as LOINC and integrating the user perspective. This is based on the aim to target some of the challenges when pooling cohort data, including too imprecise estimates and misalignment between cohorts [Reference O’Connor, Spry, Patton, Moreno-Betancur, Arnup and Downes23]. We evaluate this approach with a prototype application, called “Semantic Search Helper,” which is seen to assist harmonization processes with data from different studies, cohorts and based on multiplex data sets, and thus to “help” the researchers and experts in their decision on whether or not to merge heterogeneous data obtained with different instruments. We present a use case analysis with data from various large-scale health-related cohort studies and expert ratings for further validation.

Methods

Health-related cohort studies and related questionnaire data

For the use case analysis, we used questionnaire data from four heterogeneous cohorts (see Table 1) merged in the context of the IMAC-Mind (Improving Mental Health and Reducing Addiction in Childhood and Adolescence through Mindfulness: Mechanisms, Prevention, and Treatment) consortium [Reference Arnaud, Banaschewski, Nees, Bucholz, Klein and Reis38]: ROLS (Rostocker Längsschnittstudie) [Reference Meyer-Probst and Reis39]; MARS (Mannheim Study of Children at Risk) [Reference Holz, Zabihi, Kia, Monninger, Aggensteiner and Siehl40Reference Laucht, Esser and Schmidt42]; FRANCES (Franconian Cognition and Emotion Studies); [Reference Eichler, Grunitz, Grimm, Walz, Raabe and Goecke43] and POSEIDON (Pre-, Peri-, and Postnatal Stress: Epigenetic Impact on Depression). Within IMAC-Mind, these cohorts have been selected with the aim to increase knowledge about the development of addiction during childhood and adolescence enlarging the population size, enhancing the statistical power of evidence, and thereby magnifying the study’s impact and validity. The cohorts consist of 1329 participants encompassing 44 instruments (see Table 1). For demonstration purposes, we utilized a subset of 31 licensed instruments with a total of 1458 item texts (see Table 2). These selected instruments and items cover a diverse array of concepts and constructs with varying syntactical structures. The inclusion of both German and English items allows us to investigate the potential for multilingual harmonization.

Figure 1. Workflow concept for detecting semantic similarities within multi-item questionnaires. The process involves: (1) inputting sentence data, including variable names and text; (2) converting text into vectors using models like SBERT or ADA; (3) building similarity pairs to identify semantic matches and generating scores; (4) selecting and downloading pairs for further analysis. Color coding indicates manual, automatic, and semi-automatic processes.

Table 1. Overview of IMAC-mind cohort studies

Table 2. Overview of questionnaires used in the studies

The process of semantic similarity search in the context of merging health-related datasets and the implementation of respective features into an online tool called “Semantic Search Helper”

The Maelstrom Institute has established protocols for harmonizing health-related datasets [Reference Fortier, Raina, Van den Heuvel, Griffith, Craig, Saliba, Stolk, Knoppers, Ferretti, Granda and Burton44]. Acknowledging that the guidelines do not extensively cover the semantic search process, our initial emphasis was on outlining the pertinent steps of this process from a researcher’s perspective. To aid researchers in manually identifying semantic harmonization opportunities in multi-item questionnaires and studies, we have delineated our approach into the following steps:

The manual search process

Firstly, we implemented a manual search process from the perspective of scientists defining the following activities as core features:

  1. 1. Collection of questionnaires and their items: Researchers need to be able to easily check relevant studies and their questionnaires for semantic similarities. Established sources such as LOINC (Logical Observation Identifiers Names and Codes) [Reference Forrey, CJ, DeMoor, Huff, Leavelle, Leland, Fiers, Charles, Griffin, Stalling, Tullis, Hutchins and Baenziger45] and MDM (Medical Data Models) [Reference Dugas, Neuhaus, Meidt, Doods, Storck and Bruland46] as well as own data sources are utilized to efficiently identify harmonization opportunities. Moreover, the aforementioned available catalogs also provide a valid source of information that can be integrated into the “Semantic Search Helper” toolbox.

  2. 2. Identification of similar questionnaire content: Questionnaires and their items are compared in terms of concepts and semantic similarity to form groups of similar questionnaires.

  3. 3. Filtering relevant semantic pairs: Semantically relevant pairs are selected based on specific research questions, allowing for efficient filtering through large volumes of data.

  4. 4. Preparation for syntactic harmonization: The selected item pairs are scrutinized for syntactic consolidatability to assess the granularity of the questions and ascertain their harmonization potential.

Identification of methods and techniques to support semantic similarity search

To improve the efficiency of the manual process, we explored applicable technological solutions. We focused on using advanced language-processing techniques, especially methods that help identify meaning in sentences (particularly attention-based embeddings). Sentence Embedding models such as SBERT and OpenAI embeddings were evaluated for their potential to facilitate semantic comparisons more effectively and efficiently than traditional nontransformer-based methods (similar to [Reference McElroy, Wood and Bond17]).

The platform Hugging Face offers many open-source models that provide a text/sentence-to-vector/embedding interface [34]. Our model selection was based on two characteristics: (a) the model was already tested for text similarity search tasks, and (b) the model was capable of handling multiple languages.

We selected the open-source sentence transformer model SBERT [Reference Reimers and Gurevych47] and the GPT-3 model ADA (“text-embedding-ada-002”) for our use case. The models have the following characteristics:

OpenAI embeddings

The OpenAI API provides an easy-to-use framework based on large datasets for converting texts into embeddings. We utilized the “textembedding-ada-002” model accessible via OpenAI’s REST API endpoint. We used the proposed configuration with the “cl100k base” embedding encoding. The questionnaire text was used as input, and we called the OpenAI Embedding API with the provided Python script (Python 3.8.) to create an embedding for each questionnaire.

SBERT sentence embedding

Sentence embeddings are a further option to make embeddings out of texts. This type of embedding is a variant of the renowned encoder BERT (Bidirectional Encoder Representations from Transformers), widely utilized in various applications [Reference Devlin, Chang, Lee and Toutanova32, Reference Reimers and Gurevych49]. For our use case, we employed SBERT, a model optimized for sentence embeddings [Reference Hao, Tan, Tang, Ni, Shao, Zhang, Rogers, Boyd-Graber and Okazaki48, Reference McInnes, Healy and Melville50]. We utilized the openly available model “sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2,” capable of handling multiple languages. We adhered to the guidelines for setting up text-to-embedding translation.

Embedding-related algorithms

Clustering is a well-known method in the realm of embeddings. K-nearest neighbors can be relatively easily calculated based on given vectors and distance metrics. The visualization of high-dimensional vectors can be achieved by dimension reduction techniques such as Uniform Manifold Approximation and Projection (UMAP) [Reference McInnes, Healy and Melville50], which simplifies complex structures into understandable formats, preserving the essential characteristics of the data. Discovering topics through embeddings represents another algorithm linked to embeddings, potentially valuable for data harmonization, especially with unfamiliar datasets. BERTopic employs a neural approach to topic modeling, utilizing a class-based TFIDF mechanism [Reference Grootendorst51]. Additionally, we selected the open-source library Faiss [Reference Douze, Guzhva, Deng, Johnson, Szilvasy, Mazaré, Lomeli, Hosseini and Jégou52] for efficient similarity search.

Data visualization methods

To enhance understanding of the underlying data, data visualization techniques are crucial. Advanced visualization tools such as Plotly [53], ECharts [Reference Li, Mei, Shen, Su, Zhang and Wang54], and Seaborn provide valuable means to illustrate relationships among elements and highlight critical details. These selected tools facilitate the creation of dynamic plots and charts that are interactive and user-friendly, making complex data more accessible. Another advantageous visualization technique we found useful in the realm of data harmonization is the utilization of interactive networks [55], which can be especially effective during the search and filtering phases. Interactive networks empower users to dynamically explore connections and patterns within the data, providing a potent means to visualize relationships and dependencies.

Pre-testing of transformer-based embeddings in the context of data harmonization with a group of domain experts

We conducted a pre-test with seven experts to evaluate the use of transformer-based embeddings in data harmonization, focusing on the feasibility of employing cosine similarity derived from these models. Domain experts from the authors’ network with prior experience in the field of psychology and cognitive neuroscience and in working with questionnaire data were contacted and invited to participate in the pretesting. In total, 7 experts rated 20 candidates for semantic similarity, assessing each for suitability for harmonization, similarity score, and providing qualitative feedback. We also provided an open comment opportunity to capture additional information on the evaluation procedures.

The selection of the questions was carried out as follows: For each of the 31 instruments, three groups were built based on the cosine score between the pairs of questions. These were the five most similar pairs, the five pairs closest to the overall cosine mean, and the five least similar pairs. For every instrument, one of the groups was randomly chosen, leading to 155 pre-selected pairs. We then randomly selected 20 pairs from the pre-selected ones to ensure that each instrument had the same chance to be included and that the three similarity groups were represented in the final selection.

We analyzed responses to categorize semantic pairings as positive or negative, aiding in the determination of their suitability for harmonization.

Prototype development based on the identified workflow

Based on the results from the pre-testing of embeddings with a small group of users, we developed a software prototype using Python 3.8 [Reference McKinney56] and the Streamlit library [Reference Streamlit57]. This prototype will incorporate algorithms and methods mentioned in the previous sections, such as UMAP for dimension reduction, Faiss for efficient similarity search, and SBERT along with ADA for embedding generation. The prototype demonstrates the feasibility of a supported harmonization workflow using the IMAC-Mind dataset as a use case.

Results

Workflow development for supported semantic search tasks

Based on the analysis of the manual process and the algorithms identified appropriately during the technical assessment, a concept for supported semantic harmonization and its implementation was developed (see Figure 1). The workflow incorporates the main activities of researchers in identifying harmonization candidates and includes the algorithms and methods described in the Methods chapter. The concept necessitates the sequential processing of the questionnaires, ultimately resulting in a potential list of harmonization pairs, which can be used for further syntactic verification (not covered by this concept). This concept resulted in four main steps as a workflow of a semantic search task:

  1. 1. Providing questionnaire data: The ability to upload questionnaires and their corresponding items is essential to initiate the automation process. Sources such as the Logical Observation Identifier Names and Codes (LOINC) database [Reference Forrey, CJ, DeMoor, Huff, Leavelle, Leland, Fiers, Charles, Griffin, Stalling, Tullis, Hutchins and Baenziger45], or the Portal of Medical Data Models (MDM) [Reference Dugas, Neuhaus, Meidt, Doods, Storck and Bruland46] can provide well-structured questionnaire metadata. Uploading Excel (.xlsx) or CSV (.csv) files is advantageous. For item matching, at minimum, the name of the questionnaire and the text of each question are required. Additional information such as variable names or study names may be helpful for the harmonization evaluation process.

  2. 2. Using models to represent sentences: Generate a vector for each sentence using advanced NLP models like ADA or a comparable model. These vectors represent the semantic meaning of each sentence.

  3. 3. Cosine similarity calculation: Calculate the cosine similarity between the vectors of all sentences from the questionnaires. Pair each sentence with every other sentence to determine similarities.

  4. 4. Filtering the list of pairs: Filter the list of sentence pairs based on relevant characteristics to identify the best matching pairs. These characteristics include the strength of similarity or specific instruments.

Figure 2. Mean score responses and model-based similarity scores. The correlation between average score responses (x-axis) and similarity scores generated by the embedding models (y-axis). Scatter plots: score responses versus model similarity scores. Each plot represents the correlation between mean score responses and similarity scores for SBERT (green) and ADA (red) models. The overall trend in the relationship between evaluative scores and model-derived similarity metrics is indicated by linear regression lines.

Evaluation within the scientific area: pre-testing results of expert ratings

The group of participants (N = 7) had a high level of experience in the field of data harmonization, averaging 7.14 years (standard deviation [SD] 6.440), working both in the clinical (43%) and research (57%) fields (psychological and neuroscientific research). The relevance of the topic, data harmonization, was rated with a mean of 8.14 (SD: 1.069) on a 10-point Likert scale, indicating high relevance. Table 3 shows the 20 examples that were used for the pre-testing with the corresponding similarity scores from the SBERT and ADA Models.

Table 3. Comparison of semantic similarity scores for questionnaire item pairs using SBERT and ADA algorithms

Similarity candidate ratings

Candidate pairs were rated by the experts according to the models: Within the group of positive candidate pairs, the pairs exhibited a mean agreement by the experts to be classified as positive of 87.0% (95% CI, 75.4–98.7), while the negative candidate pairs showed a mean agreement of 94.8% (95% CI, 88.3–101.2).

Similarity score rating

Similar findings were also observed for the 10-point similarity rating scale. Moreover, the sentence embedding (SE) method is strongly related to human ratings (0.837, p < 0.001), see Table 4 and Figure 2.

Figure 3. ‘Semantic Search Helper’ application interface. The user interface (UI) with different stages of the harmonization process: A bar chart overview of metadata distribution and a scatter plot for visualizing data points in semantic dimensions (Bottom Section). A filter tree for selecting specific data items and a filtered table displaying data based on applied filters (Right Section). Table with survey questions and semantic similarity scores, and a bar chart showing semantic coverage percentages (Top Section). A network graph visualizing semantic connections between questionnaire items (Middle Section). This interface facilitates the comparison of semantic similarities across survey questions, streamlining the data harmonization workflow for researchers.

Table 4. Spearman correlations and 95% confidence intervals

Analysis of the central tendency and variance shows that 11 pairs have a low variance (SD <1) highlighting pairs with a semantically clear rating. Especially for these pairs, the correlation between PSS and SES was 0.974 (p < 0.0001) and perfectly correlated, indicating that the SES highlights clear semantic similarity in the high score ranges (≥0.88) and clear semantic dissimilarities in the low score ranges (≤0.35). The remaining nine pairs had a greater variance for the PSS indicating that the pairs were harder to rate. The pairs (ID: 720) (LDH) “Alcohol-related problems in the past year” – (AEQ) “Drinking alcohol helps people deal with life’s problems” and the pair (ID: 385)” (STAI-T) “I am impatient” – (BSI) “Feeling tense or keyed up.” had the highest variability (SD of 2.936 and 2.760).

We interpreted the pre-testing results as positive and continued our work with a prototype implementation to support the search for harmonization opportunities with transformer-based embeddings.

Prototype development: Workflow implementation and application to the IMAC-Mind dataset

Our implementation of the workflow concept resulted in a prototype application called “Semantic Search Helper” (see Figure 3). We tested automation functionalities with the list of unlicensed instruments in Table 2.

Figure 4. Cosine similarity distributions for SBERT (light blue) and ADA (green). SBERT has a mean of 0.19 (blue line) and ADA has a mean of 0.76 (yellow line).

Embeddings and cosine similarity

The prototype application created 994.609 potential pairs in total and 15.085 with Faiss out of the IMAC-Mind dataset with 1458 questions and 31 instruments. The distribution of cosine similarity scores for the two models is shown in Figure 4. SBERT has a mean cosine similarity score of 0.195 (SD: 0.149), while ADA has a mean cosine similarity score of 0.765 (SD: 0.033), showing that the value scores are different between the models.

In the following, we describe only the results derived from the SBERT model, as there were no significant differences between SBERT and ADA.

Clustering and topic building

The clustering and BERTopic-based topic building revealed between 3 and 142 topics or clusters for the 1458 sentences. A clear clustering was observed with alcohol-related questions (Topic 1: N = 21) and smoking-related questions (Topic 2: N = 16). The remaining questions (Topic 0: N = 1421) were clustered together and presented a semantically unclear view (in Supplementary Table S1).

Relevant pair filtering

For our analysis, we selected all pairs with a similarity score higher than 0.751, which was the lower confidence interval (CI) bound obtained from the pre-testing. This selection resulted in 312 pairs (in Supplementary Table S2).

Based on these 312 pairs, the content coverage of the questionnaires in the application was displayed using a dependency wheel visualization and a table. The NEO (long version) with 62 items, which showed semantic similarities with other questionnaire items (cosine similarity >0.751), had the highest number of matches. With a total of 241 items, this accounted for a relative share of 25.73% of the questionnaire items covered. The NEO (short version) showed a relative coverage of 100%, covering all 30 questions, thus having the highest value among all 31 questionnaires, followed by STAIS-S at 85% and STAIS-T at 70% (in Supplementary Table S3). The “NEO-FFI (short version)” questionnaire has semantic similar content with the following eight questionnaires: (EBSK, SVF, TPF, STAI-S, STAI-T, NEO-FFI (long version), ADS, ADHS-LE). In general, two questionnaires (KINDL, BelVa) did not have content matches based on the chosen similarity score. This information is presented to the user both visually and in a summary of the interface.

Quantifying the automation potential

By considering the described findings and the strong correlations observed between the SBERT, ADA models, and human raters, we developed a theoretical hypothesis regarding the potential benefits of this methodology across the involved cohorts. Assuming that the confidence intervals for positive and negative candidate pairs can categorize the entire pool of harmonization candidates, the following results were obtained based on the estimated percentage of data that embedding models can automatically process without human intervention.

For our use case, 0.35% (N = 1,132) of the pairs are classified as positive candidate pairs based on a confidence interval (CI) of 0.676 to 0.912, while 73.39% (N = 235,826) are classified as negative candidate pairs based on a CI of 0.737–0.800. This indicates that 73.74% (N = 236,958) of the data can be automatically processed using embeddings, whereas 26.26% (N = 83,041) is unclear to the SBERT model and requires human review. The findings are comparable to the ADA model, with only 0.83% (N = 2,665) of the pairs being within the CI (0.862 to 0.945) for positive candidate pairs and 66.55% (N = 213,845) of the pairs being within the CI for negative candidate pairs. Consequently, 67.38% (N = 216,510) of the candidate pairs can be automatically processed with the ADA embeddings. Exactly 32.62% (N = 103,489) of the candidates remain uncertain and require human verification. This initial analysis suggests that, based on the confidence intervals applicable to the original dataset, a minority of possible candidates (26% and 32%) would require human verification for the use case considered.

Discussion

Our work shows that using NLP techniques to harmonize questionnaire data can be significantly beneficial. By employing semantic search methods, we compared 1458 items from 31 questionnaires. Similar studies [Reference Chen, Jutkowitz and Gross6,Reference Bandyopadhyay, Tingay, Borja, Griffiths, Akbari and Bedford9,Reference Dahal, Patten, Williamson, Patel, Premji and Tough12,Reference Chen, Jutkowitz, Iosepovici, Lin and Gross15] indicate that harmonization is time-consuming, resource-intensive, and rarely scalable. Our prototype streamlines the search process, assessing harmonization potential and providing preliminary insights, both considering standardized sources of information including terminologies, the language aspect, and providing researcher-friendly and usable presentation layers, before actual data integration. This approach quantifies insights that would otherwise only be roughly estimated.

This approach can enhance mental health research by integrating and comparing data from diverse populations, improving statistical power, and identifying subtle health trends and associations. For example, studying risk factors for alcohol use disorders (AUDs), can address the challenges posed by differing data collection methods and variable definitions across studies. Researchers can use the “Semantic Search Helper” to for example investigate the combined effects of genetic predispositions, early-life experiences, and socioeconomic status on AUD development across different demographic groups and also if information and data stem from a diverse set of instruments. This improves statistical power to detect associations between risk factors for AUD and outcomes, even between subgroups with smaller sample sizes, and can enhance generalizability and facilitate the identification of commonalities and differences in AUD risk factors. Harmonized data enables the exploration of interactions between various risk factors and the identification of potential mediators (e.g., mental health conditions) not only in AUD development but also in other mental health areas, where “Semantic Search Helper” can be applied in a similar fashion. It also supports meta-analyses and replication studies, thereby strengthening research findings. Our prototype can process 67.38 to 73.74% of potential candidates without human intervention, significantly reducing manual efforts and conserving resources. Expert testing during the conceptualization phase shows high agreement between semantically equivalent and expert-considered harmonizable items, proving the validity and usefulness of semantic embeddings for data harmonization of large health-related cohorts. It is important to note that, as the name suggests, the “Semantic Search Helper” serves as a tool to support the harmonization process by quantifying the comparability of questionnaires, thus facilitating their potential integration for specific research questions and analyses. This tool tests the semantic similarity of individual items while leveraging available standard terminologies, considering language-based aspects, and involving researchers and experts in the search process. It achieves this, in part, through various visualization and summary features that provide researchers with an intuitive sense of the data, which is valuable not only in cases where similarity is marginal but also to still consider deep clinical expertise and in the end transparency of the whole process. While our results are promising in identifying opportunities for semantic harmonization, it is further important to acknowledge the challenges that this emerging technology presents. Although NLP techniques can efficiently analyze semantic content, there is still a risk of over-reliance on automated methods at the expense of thorough documentation and description of the measurement instruments used. Such a development could contravene the principles of FAIR and impact the long-term traceability of collected data. It is therefore essential to employ NLP as an additional tool for the manual recording and documentation of meta-data, rather than as a replacement for it. Future studies should further compare semantic similarity and harmonization possibilities, consider multilingual questionnaires, and utilize NLP techniques to handle different informant-based information. These steps will enhance the robustness and accuracy of semantic embeddings in data harmonization. In this respect, it is also important to note the importance of developing clear and comprehensive data dictionaries [58] and/or making use of existing ones and integrating the respective information in the harmonization process. This is not only essential for a concrete overview of data sets from different cohorts but also for planning future studies in alignment with those cohorts that might be most important for potential later cross-cohort analyses. Our preliminary results encourage future research, including model fine-tuning and applications on larger texts or metadata elements, aiming to develop tools for semantic and syntactic mapping in metadata repositories and automate the harmonization process.

Limitations

While our investigation establishes a fundamental principle for utilizing embeddings in cohort study harmonizations for similarity searches and underscores and expands recent studies in this context (e.g., [Reference McElroy, Wood and Bond17]), this research is still in its early stages. The efficiency and potential of this approach in evaluating harmonization opportunities are significant strengths, which have been tested in a larger number of questionnaires and potential pairs (here: over 70,000) than in previous applications [Reference McElroy, Wood and Bond17] as well as now based on item texts from different languages. However, the user perspective, including input and additional evaluations from researchers and experts, should not be excluded from this process. For this reason, we have incorporated multiple layers into the “Semantic Search Helper,” drawing on information from standard terminologies in the field and allowing data adjustments at various levels to enable customized visualization. However, further examination and testing are necessary to fully appreciate its potential. Intercultural and historical effects on item formulation can influence item semantics, meaning identical formulations can differ across cultures or eras. These differences are not fully captured by semantic embeddings, posing a limitation for automated data harmonization algorithms. Despite its limitations, our research validates this approach’s efficacy and establishes a foundation for further exploration.

Our goal was to strengthen cross-cohort and -study based research in mental health by simplifying the time-intensive process of data harmonization through scalable, NLP-based methods. NLP provides an efficient way to identify harmonization potential and conserve resources, that should however not be treated independently from available standard sources and manual processes and the researchers’ expertise. This methodology can thus not completely replace human expertise and the nuanced semantic assessments, as our work demonstrates. In the long term, this research aims to establish pooling – combining data from different studies – as a common method to increase sample sizes and thereby strengthen the validity of scientific analyses.

Together, this research demonstrates that attention-based embeddings are effective for identifying semantic similarities and can thus assist researchers in this task. Using these techniques in questionnaire-based cohort data provides a viable approach for the initial phases of data harmonization and merits further investigation. Our code is available on GitHub under the MIT license upon request. The decision to release it upon request is driven by our goal to use user feedback, gathered through an evaluation questionnaire, to improve and adapt the prototype. Users can also submit suggestions for future iterations.

Abbreviations

ADA

Name of a specific model or method used in the context

ADHS-LE

ADHS-Langzeit-Erfassung

ADS

Alcohol Dependence Scale

AID

Adaptives Intelligenz Diagnostikum

APQ-2

Advanced Personality Questionnaire second version

AUDIT

Alcohol Use Disorders Identification Test

BDI

Beck Depression Inventory

BelVa

Belgian Violent behavior Assessment

BERT

Bidirectional Encoder Representations from Transformers

CBQ

Comprehensive Behavioral Questionnaire

EBSK

Eltern-Belastungs-Screening

EPDS

Edinburgh Postnatal Depression Scale

ESI

Environmental Stress Index

EtG

Ethyl glucuronide

FBB-ADHS

ADHD-specific Behavior Assessment Form

FBB-ANZ

Anxiety-specific Behavior Assessment Form

FBB-DES

Depression-specific behavior assessment form

FBB-SSV

Social Behavior Assessment Form

FRANCES

Franconian Cognition and Emotion Studies

FTND

Fagerström Test for Nicotine Dependence

IDS

Inventory of Depressive Symptomatology

IMAC-Mind

Improving mental health and reducing addiction in childhood and adolescence through mindfulness

KiGa

Kindergarten Adjustment Scale

KINDL

KINDL questionnaire for measuring health-related quality of life in children and adolescents

LES

Life Experiences Survey

LOINC

Logical Observation Identifiers Names and Codes

MARS

Mannheim Study of Children at Risk

MDM

Medical data models

NLP

Natural language processing

NEO-FFI

NEO Personality Inventory

POSEIDON

Pre- peri- and postnatal stress: epigenetic impact on depression

PSS

Perceived Stress Scale

ROLS

Rostocker Längsschnittstudie

SBERT

Sentence-BERT

SDQ

Strengths and Difficulties Questionnaire

STAI-S

State-Trait Anxiety Inventory – short form

STAI-T

State-Trait Anxiety Inventory – trait form

SVF

Stress Coping Inventory

TF-IDF

Frequency-inverse document frequency

TPF

Typical Personality Factors Questionnaire

UMAP

Uniform manifold approximation and projection

Supplementary material

The supplementary material for this article can be found at https://doi.org/10.1192/j.eurpsy.2024.1808.

Data availability statement

The data underlying this article cannot be shared publicly due to ethical reasons. The data will be shared on reasonable request to the corresponding author.

Financial support

The work was supported by the Federal Ministry of Education and Research Germany (BMBF 01GL1745F, subproject 1). With the public-funded research project IMAC-Mind: Improving Mental Health and Reducing Addiction in Childhood and Adolescence through Mindfulness: Mechanisms, Prevention and Treatment (2017–2021), the Federal Ministry of Education and Research contributes to improving the prevention and treatment of children and adolescents with substance use disorders and associated mental disorders. The project coordination was realized by the German Center of Addiction Research in Childhood and Adolescence at the University Medical Center Hamburg–Eppendorf. The work was also supported by the research project “CoviDrug: The role of pandemic and individual vulnerability in longitudinal cohorts across the life span: refined models of neurosociobehavioral pathways into substance (ab)use?”, which has been funded within the Call for proposals for interdisciplinary research on epidemics and pandemics in the context of the SARS-CoV-2 outbreak by the German Research Foundation (NE 1383/15-1, BA 2088/7-1).

Competing interest

Tobias Banaschewski served in an advisory or consultancy role for Lundbeck, Medice, Neurim Pharmaceuticals, Oberberg GmbH, Shire. He received conference support or speaker’s fees from Lilly, Medice, Novartis, and Shire. He has been involved in clinical trials conducted by Shire and Viforpharma. He received royalties from Hogrefe, Kohlhammer, CIP Medien, and Oxford University Press.

Ethics approval

The study has been performed according to the declaration of Helsinki and ethics approval has been obtained from each of the local ethics committees (Rostock A01/2004, Erlangen 4596, Mannheim: 2010-236N-MA, 2010-309N-MA).

Footnotes

K.G. and K.J. authors are contributed equally to this work.

References

Lydeard, S. The questionnaire as a research tool. Fam Pract. 1991;8(1):8491. doi:10.1093/fampra/8.1.84.CrossRefGoogle ScholarPubMed
Kelley, K, Clark, B, Brown, V, Sitzia, J. Good practice in the conduct and reporting of survey research. Int J Qual Health Care. 2003;15(3):261–6. doi:10.1093/intqhc/mzg031.CrossRefGoogle ScholarPubMed
Bauermeister, S, Orton, C, Thompson, S, Barker, RA, Bauermeister, JR, Ben-Shlomo, Y, et al. The dementias platform UK (DPUK) data portal. Eur J Epidemiol. 2020;35(6):601–11. doi:10.1007/s10654-020-00633-4.CrossRefGoogle ScholarPubMed
Salimi, Y, Domingo-Fernández, D, Bobis-Álvarez, C, Hofmann-Apitius, M, Birkenbihl, C. Family=Alzheimer’s Disease Neuroimaging Initiative suffix=for the ABV given=the Japanese Alzheimer’s Disease Neuroimaging Initiative. ADataViewer: Exploring semantically harmonized Alzheimer’s disease cohort datasets. Alzheimers Res Ther. 2022;14(1):69 doi:10.1186/s13195-022-01009-4.CrossRefGoogle Scholar
O’Connor, M, Moreno-Betancur, M, Goldfeld, S, Wake, M, Patton, G, Dwyer, T, et al. Investigators and the LC. Data resource profile: Melbourne children’s LifeCourse initiative (LifeCourse). Int J Epidemiol. 2022;e229–44. doi:10.1093/ije/dyac086.CrossRefGoogle Scholar
Chen, D, Jutkowitz, E, Gross, AL. Statistical harmonization of everyday functioning and dementia-related behavioral measures across nine surveys and trials. Alzheimers Dement Amst Neth. 2023;15(1):e12412 doi:10.1002/dad2.12412.CrossRefGoogle ScholarPubMed
Kezios, KL, Zimmerman, SC, Buto, PT, Rudolph, KE, Calonico, S, Zeki Al Hazzouri, A, et al. Overcoming data gaps in life course epidemiology by matching across cohorts. Epidemiology. 2024;35(5):610 doi:10.1097/EDE.0000000000001761.CrossRefGoogle ScholarPubMed
Aizpurua, E, Fitzgerald, R, de Barros, JF, Giacomin, G, Lomazzi, V, Luijkx, R, et al. Exploring the feasibility of ex-post harmonisation of religiosity items from the European social survey and the European values study. Meas Instrum Soc Sci. 2022;4(1):12 doi:10.1186/s42409-022-00038-x.CrossRefGoogle Scholar
Bandyopadhyay, A, Tingay, K, Borja, MC, Griffiths, L, Akbari, A, Bedford, H, et al. Harmonising data from different sources to conduct research using linked survey and routine datasets. Int J Popul Data Sci. 2018;3(4):157 doi:10.23889/ijpds.v3i4.750.Google Scholar
Bauermeister, S, Phatak, M, Sparks, K, Sargent, L, Griswold, M, McHugh, C, et al. Evaluating the harmonisation potential of diverse cohort datasets. Eur J Epidemiol. 2023;38(6):605–15. doi:10.1007/s10654-023-00997-3.CrossRefGoogle ScholarPubMed
Chow, S-M, Nahum-Shani, I, Baker, JT, Spruijt-Metz, D, Allen, NB, Auerbach, RP, et al. The ILHBN: Challenges, opportunities, and solutions from harmonizing data under heterogeneous study designs, target populations, and measurement protocols. Transl Behav Med. 2022;13(1):716.CrossRefGoogle Scholar
Dahal, KA, Patten, S, Williamson, T, Patel, A, Premji, S, Tough, S, et al. Harmonization of data from cohort studies– Potential challenges and opportunities. Int J Popul Data Sci. 2018;3(4): doi:10.23889/ijpds.v3i4.868.Google Scholar
Nguyen, T, Khadka, R, Phan, N, Yazidi, A, Halvorsen, P, Riegler, MA. Combining datasets to increase the number of samples and improve model fitting. arXiv. 2023. doi: 10.48550/arXiv.2210.05165CrossRefGoogle Scholar
Almeida, JR, Silva, LB, Bos, I, Visser, PJ, Oliveira, JL. A methodology for cohort harmonisation in multicentre clinical research. Inform Med Unlocked. 2021;27:100760 doi:10.1016/j.imu.2021.100760.CrossRefGoogle Scholar
Chen, D, Jutkowitz, E, Iosepovici, SL, Lin, JC, Gross, AL. Pre-statistical harmonization of behavioral instruments across eight surveys and trials. BMC Med Res Methodol. 2021;21(1):227 doi:10.1186/s12874-021-01431-6.CrossRefGoogle ScholarPubMed
Tomescu-Dubrow, I, Slomczynski, KM. Harmonization of cross-National Survey Projects on political behavior: Developing the analytic framework of survey data recycling. Int J Sociol Routledge. 2016;46(1):5872. doi:10.1080/00207659.2016.1130424.CrossRefGoogle Scholar
McElroy, E, Wood, T, Bond, R, et al. Using natural language processing to facilitate the harmonisation of mental health questionnaires: A validation study using real-world data. BMC Psychiatry. 2024;24:530 doi:10.1186/s12888-024-05954-2.CrossRefGoogle ScholarPubMed
Wilkinson, MD, Dumontier, M, IjJ, A, et al. The FAIR guiding principles for scientific data management and stewardship. Sci Data. 2016;3:160018 doi:10.1038/sdata.2016.18.CrossRefGoogle ScholarPubMed
Bietenbeck, A, Boeker, M, Schulz S, NPU. LOINC, and SNOMED CT: A comparison of terminologies for laboratory results reveals individual advantages and a lack of possibilities to encode interpretive comments. J Lab Med. 2018;42:267–75. doi:10.1515/labmed-2018-0103.Google Scholar
Drenkhahn, C, Ingenerf, J. The LOINC content model and its limitations of usage in the laboratory domain. Stud Health Technol Conform. 2020;437442. 10.3233/SHTI200198Google ScholarPubMed
Lee, D, Cornet, R, Lau, F, de Keizer, N. A survey of SNOMED CT implementations. J Biomed Inform. 2013;46:8796. doi:10.1016/j.jbi.2012.09.006.CrossRefGoogle ScholarPubMed
Sathappan, SMK, Jeon, YS, Dang, TK, Lim, SC, Shao, Y-M, Tai, ES, et al. Transformation of electronic health records and questionnaire data to OMOP CDM: A feasibility study using SG_T2DM dataset. Appl Clin Inform. 2021;12:757–67. doi:10.1055/s-0041-1732301.Google ScholarPubMed
O’Connor, M, Spry, E, Patton, G, Moreno-Betancur, M, Arnup, S, Downes, M, et al. Better together: Advancing life course research through multi-cohort analytic approaches. Ad Life Course Res. 2022;53: doi:10.1016/j.alcr.2022.100499.Google ScholarPubMed
O’Connor, M, Olsson, C, Lange, K, Downes, M, Moreno-Betancur, M, Mundy, L, et al. Progressing “positive epidemiology”: A cross-national analysis of adolescents’ positive mental health and outcomes during the COVID-19 pandemic. Epidemiology. 2025;36(1):2839. doi:10.1097/EDE.0000000000001798.CrossRefGoogle ScholarPubMed
Vaswani, A, Shazeer, N, Parmar, N, Uszkoreit, J, Jones, L, Gomez, AN, Kaiser, Ł, Polosukhin, I. Attention is All you Need. Advances in Neural Information Processing Systems Curran Associates, Inc.; 2017. Available from: https://proceedings.neurips.cc/paper_files/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html (accessed Nov 1, 2023)Google Scholar
Bi, Q, Goodman, KE, Kaminsky, J, Lessler, J. What is machine learning? A primer for the epidemiologist. Am J Epidemiol. 2019;188(12):2222–39. doi:10.1093/aje/kwz189.Google ScholarPubMed
Cer, D, Diab, M, Agirre, E, Lopez-Gazpio, I, Specia, L. SemEval-2017 task 1: Semantic textual similarity multilingual and Crosslingual focused evaluation. In: Bethard, S, Carpuat, M, Apidianaki, M, Mohammad, SM, Cer, D, Jurgens, D, editors. Proc 11th Int workshop semantic Eval SemEval-2017. Vancouver, Canada: Association for Computational Linguistics; 2017, p. 114. 10.18653/v1/S17-2001.Google Scholar
De Boom, C, Van Canneyt, S, Bohez, S, Demeester, T, Dhoedt, B. Learning semantic similarity for very short texts. 2015 IEEE Int Conf Data Min Workshop ICDMW; 2015, p. 12291234. doi: 10.1109/ICDMW.2015.86Google Scholar
Kang, B, Yoon, J, Kim, HY, Jo, SJ, Lee, Y, Kam, HJ. Deep-learning-based automated terminology mapping in OMOP-CDM. J Am Med Inform Assoc. 2021;28(7):1489–96. doi:10.1093/jamia/ocab030.CrossRefGoogle ScholarPubMed
Patil, A, Han, K, Jadon, A. A Comparative Study of Text Embedding Models for Semantic Text Similarity in Bug Reports. arXiv. 2023. doi: 10.48550/arXiv.2308.09193CrossRefGoogle Scholar
Reimers, N, Beyer, P, Gurevych, I. Task-Oriented Intrinsic Evaluation of Semantic Textual Similarity. Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, Japan: The COLING 2016 Organizing Committee, pp. 87–96. 2016. Available from: https://aclanthology.org/C16-1009 (accessed Oct 26, 2023)Google Scholar
Devlin, J, Chang, M-W, Lee, K, Toutanova, K.. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 Long and Short Papers, Minnesota: Association for Computational Linguistics, pp. 4171–4186; 2019. doi: 10.18653/v1/N19-1423CrossRefGoogle Scholar
Bergman, E, Sherwood, K, Forslund, M, Arlett, P, Westman, G. A natural language processing approach towards harmonisation of European medicinal product information. PLoS OnePublic Library of Science. 2022;17(10):e0275386 doi:10.1371/journal.pone.0275386.CrossRefGoogle ScholarPubMed
Hugging Face. The AI community building the future. 2024. Available from: https://huggingface.co/ (accessed May 27, 2024)Google Scholar
Honnibal, M, spaCy, Montani I. Industrial-strength Natural Language Processing in Python. Available from: https://spacy.io/ (accessed Feb 18, 2024)Google Scholar
OpenAI. ChatGPT. Available from: https://openai.com/blog/chatgpt (accessed Feb 18, 2024)Google Scholar
OpenAI. OpenAI Platform. Explore Dev Resour Tutor API Docs Dyn Ex Get Most OpenAIs Platf. 2024. Available from: https://platform.openai.com (accessed Feb 18, 2024)Google Scholar
Arnaud, N, Banaschewski, T, Nees, F, Bucholz, VN, Klein, M, Reis, O, et al. Achtsamkeit in der entwicklungsorientierten Suchtprävention und -therapie: Rational, design und Ziele des Forschungsverbundes IMAC-mind. Prax Kinderpsychol Kinderpsychiatr. 2020;69(4):353–74. doi:10.13109/prkk.2020.69.4.353.CrossRefGoogle ScholarPubMed
Meyer-Probst, B, Reis, O. Von der Geburt bis 25. Kindh Entwickl. 1999;8(1):5968. doi:10.1026//0942-5403.8.1.59.CrossRefGoogle Scholar
Holz, NE, Zabihi, M, Kia, SM, Monninger, M, Aggensteiner, P-M, Siehl, S, et al. A stable and replicable neural signature of lifespan adversity in the adult brain. Nat Neurosci. 2023;26(9):1603–12. doi:10.1038/s41593-023-01410-8.CrossRefGoogle ScholarPubMed
Laucht, M, Esser, G, Baving, L, Gerhold, M, Hoesch, I, Ihle, W, et al. Behavioral sequelae of perinatal insults and early family adversity at 8 years of age. J Am Acad Child Adolesc Psychiatry. 2000;39(10):1229–37. doi:10.1097/00004583-200010000-00009.CrossRefGoogle ScholarPubMed
Laucht, M, Esser, G, Schmidt, MH. Developmental outcome of infants born with biological and psychosocial risks. J Child Psychol Psychiatry. 1997;38(7):843–53. doi:10.1111/j.1469-7610.1997.tb01602.x.CrossRefGoogle ScholarPubMed
Eichler, A, Grunitz, J, Grimm, J, Walz, L, Raabe, E, Goecke, TW, et al. Did you drink alcohol during pregnancy? Inaccuracy and discontinuity of women’s self-reports: On the way to establish meconium ethyl glucuronide (EtG) as a biomarker for alcohol consumption during pregnancy. Alcohol. 2016;54:3944. doi:10.1016/j.alcohol.2016.07.002.CrossRefGoogle ScholarPubMed
Fortier, I, Raina, P, Van den Heuvel, ER, Griffith, LE, Craig, C, Saliba, M, Doiron D, Stolk, RP, Knoppers, BM, Ferretti, V, Granda, P, Burton, P. Maelstrom research guidelines for rigorous retrospective data harmonization. Int J Epidemiol 2017;46(1):103105. doi: 10.1093/ije/dyw075Google ScholarPubMed
W Forrey, A, CJ, McDonald, DeMoor, G, Huff, SM, Leavelle, D, Leland, D, Fiers, T, Charles, L, Griffin, B, Stalling, F, Tullis, A, Hutchins, K, Baenziger, J. Logical observation identifier names and codes (LOINC) database: A public use set of codes and names for electronic reporting of clinical laboratory test results. Clin Chem 1996;42:8190.CrossRefGoogle ScholarPubMed
Dugas, M, Neuhaus, P, Meidt, A, Doods, J, Storck, M, Bruland, P, et al. Portal of medical data models: Information infrastructure for medical research and healthcare. Database J Biol Databases Curation. 2016;2016:bav121 doi:10.1093/database/bav121.Google ScholarPubMed
Reimers, N, Gurevych, I. Sentence-BERT: sentence embeddings using siamese BERT-networks. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China: Association for Computational Linguistics, pp. 3982–3992; 2019. doi: 10.18653/v1/D19-1410CrossRefGoogle Scholar
Hao, S, Tan, B, Tang, K, Ni, B, Shao, X, Zhang, H, et al. BertNet: Harvesting knowledge graphs with arbitrary relations from pretrained language models. In: Rogers, A, Boyd-Graber, J, Okazaki, N, editors. Findings of the association for computational linguistics: ACL 2023. Toronto, Canada: Association for Computational Linguistics; 2023, p. 5000–15. 10.18653/v1/2023.findings-acl.309.CrossRefGoogle Scholar
Reimers, N, Gurevych, I. Making monolingual sentence embeddings multilingual using knowledge distillation. arXiv. 2020. doi: 10.48550/arXiv.2004.09813CrossRefGoogle Scholar
McInnes, L, Healy, J, Melville, J. UMAP: uniform manifold approximation and projection for dimension reduction. arXiv. 2020. doi: 10.48550/arXiv.1802.03426CrossRefGoogle Scholar
Grootendorst, M. BERTopic: Neural topic modeling with a class-based TF-IDF procedure. arXiv. 2022. doi: 10.48550/arXiv.2203.05794CrossRefGoogle Scholar
Douze, M, Guzhva, A, Deng, C, Johnson, J, Szilvasy, G, Mazaré, P-E, Lomeli, M, Hosseini, L, Jégou, H. The faiss library. 2024.Google Scholar
Inc. PT. Collaborative data science. Montreal, QC: Plotly Technologies Inc; 2015. Available from: https://plot.lyGoogle Scholar
Li, D, Mei, H, Shen, Y, Su, S, Zhang, W, Wang, J, et al. ECharts: A declarative framework for rapid construction of web-based visualization. Vis Inform. 2018;2(2):136–46. doi:10.1016/j.visinf.2018.04.011.CrossRefGoogle Scholar
Institute WH. Interactive network visualizations—pyvis 0.1.3.1 documentation. Available from: https://pyvis.readthedocs.io/en/latest/ (accessed May 10, 2024)Google Scholar
McKinney, W. Data structures for statistical computing in python. Proceedings of the 9th python in science conference. 2010;5661. doi: 10.25080/Majora-92bf1922-00aCrossRefGoogle Scholar
Streamlit, Kekre S.. GitHub; Available from: https://github.com/streamlit (accessed May 27, 2024)Google Scholar
Figure 0

Figure 1. Workflow concept for detecting semantic similarities within multi-item questionnaires. The process involves: (1) inputting sentence data, including variable names and text; (2) converting text into vectors using models like SBERT or ADA; (3) building similarity pairs to identify semantic matches and generating scores; (4) selecting and downloading pairs for further analysis. Color coding indicates manual, automatic, and semi-automatic processes.

Figure 1

Table 1. Overview of IMAC-mind cohort studies

Figure 2

Table 2. Overview of questionnaires used in the studies

Figure 3

Figure 2. Mean score responses and model-based similarity scores. The correlation between average score responses (x-axis) and similarity scores generated by the embedding models (y-axis). Scatter plots: score responses versus model similarity scores. Each plot represents the correlation between mean score responses and similarity scores for SBERT (green) and ADA (red) models. The overall trend in the relationship between evaluative scores and model-derived similarity metrics is indicated by linear regression lines.

Figure 4

Table 3. Comparison of semantic similarity scores for questionnaire item pairs using SBERT and ADA algorithms

Figure 5

Figure 3. ‘Semantic Search Helper’ application interface. The user interface (UI) with different stages of the harmonization process: A bar chart overview of metadata distribution and a scatter plot for visualizing data points in semantic dimensions (Bottom Section). A filter tree for selecting specific data items and a filtered table displaying data based on applied filters (Right Section). Table with survey questions and semantic similarity scores, and a bar chart showing semantic coverage percentages (Top Section). A network graph visualizing semantic connections between questionnaire items (Middle Section). This interface facilitates the comparison of semantic similarities across survey questions, streamlining the data harmonization workflow for researchers.

Figure 6

Table 4. Spearman correlations and 95% confidence intervals

Figure 7

Figure 4. Cosine similarity distributions for SBERT (light blue) and ADA (green). SBERT has a mean of 0.19 (blue line) and ADA has a mean of 0.76 (yellow line).

Supplementary material: File

Gottfried et al. supplementary material

Gottfried et al. supplementary material
Download Gottfried et al. supplementary material(File)
File 57.1 KB
Submit a response

Comments

No Comments have been published for this article.