bims-librar Biomed News
on Biomedical librarianship
Issue of 2025–06–01
35 papers selected by
Thomas Krichel, Open Library Society



  1. Med Ref Serv Q. 2025 May 26. 1-15
      This study sought to investigate the influence of a science educational background on librarian interactions with basic and life sciences programs. Through semi-structured interviews with 14 librarians, the study identified several key themes: understanding and enjoying science is beneficial; previous experience influences librarians' approaches; relationships are important; defining success is challenging; and collaborating with other campus units is unusual. Results suggest a science undergraduate or graduate degree and previous science-related work experience enhance librarians' comfort and communication with scientists and highlight a lack of professional continuing education related to researcher workflows.
    Keywords:  Basic science librarians; STEM librarians; educational background; librarian–researcher interactions; life science librarians
    DOI:  https://doi.org/10.1080/02763869.2025.2509084
  2. Comput Biol Med. 2025 May 28. pii: S0010-4825(25)00815-7. [Epub ahead of print]193 110464
       OBJECTIVE: Developing search strategies for synthesizing evidence on drug harms requires specialized expertise and knowledge. The aim of this study was to evaluate ChatGPT's ability to enhance search strategies for systematic reviews of drug harms by identifying missing and generating omitted keywords.
    MATERIALS AND METHODS: A literature search in PubMed identified systematic reviews of drug harms from 10 high-impact journals between 1-Nov-2013 to 27-Nov-2023. Sixteen search strategies used in these reviews were selected each with a single error of omission introduced. ChatGPT's (GPT-4) performance was evaluated based on error detection, similarity between the extracted and generated search strategies via strict and semantic keyword matching, and proportion of omitted keywords generated.
    RESULTS: ChatGPT identified the introduced errors in all search strategies. Under strict matching, the mean Jaccard's similarity measure was 0.17 (range: 0.00-0.52) and with semantic matching this increased to 0.23 (range: 0.00-0.53). Similarly, the mean proportion of keywords recreated by ChatGPT was 49 % using strict matching increasing to 71 % with semantic matching.
    DISCUSSION AND CONCLUSION: ChatGPT effectively detected errors and generated relevant keywords, showing potential as a tool for evidence retrieval on drug harms.
    Keywords:  Artificial intelligence; Drug-related side effects and adverse reactions; Evidence-based medicine; Patient safety; Systematic reviews as topic
    DOI:  https://doi.org/10.1016/j.compbiomed.2025.110464
  3. BMC Bioinformatics. 2025 May 28. 26(1): 140
       BACKGROUND: The exponential growth of scientific publications poses a formidable challenge for researchers seeking to validate emerging hypotheses or synthesize existing evidence. In this paper, we introduce Valsci, an open-source, self-hostable utility that automates large-batch scientific claim verification using any OpenAI-compatible large language model. Valsci unites retrieval-augmented generation with structured bibliometric scoring and chain-of-thought prompting, enabling users to efficiently search, evaluate, and summarize evidence from the Semantic Scholar database and other academic sources. Unlike conventional standalone LLMs, which often suffer from hallucinations and unreliable citations, Valsci grounds its analyses in verifiable published findings. A guided prompt-flow approach is employed to generate query expansions, retrieve relevant excerpts, and synthesize coherent, evidence-based reports.
    RESULTS: Preliminary evaluations across claims from the SciFact benchmark dataset reveal that Valsci significantly outperforms base GPT-4o outputs in citation hallucination rate while maintaining a low misclassification rate. The system is highly scalable, processing hundreds of claims per hour through asynchronous parallelization.
    CONCLUSIONS: By providing an open and transparent platform for large-batch literature verification, Valsci substantially lowers the barrier to comprehensive evidence-based reviews and fosters a more reproducible research ecosystem.
    Keywords:  Batch processing; Bibliometric scoring; Bioinformatics; Chain-of-Thought; Claim verification; Large language models; Open-source; Retrieval-augmented generation
    DOI:  https://doi.org/10.1186/s12859-025-06159-4
  4. AMIA Annu Symp Proc. 2024 ;2024 513-522
      The delivery of effective targeted therapies requires comprehensive analyses of the molecular profiling of tumors and matching with clinical phenotypes in the context of existing knowledge described in biomedical literature, registries, and knowledge bases. We evaluated the performance of natural language processing (NLP) approaches in supporting knowledge retrieval and synthesis from the biomedical literature. We tested PubTator 3.0, Bidirectional Encoder Representations from Transformers (BERT), and Large Language Models (LLMs) and evaluated their ability to support named entity recognition (NER) and relation extraction (RE) from biomedical texts. PubTator 3.0 and the BioBERT model performed best in the NER task (best F1-score 0.93 and 0.89, respectively), while BioBERT outperformed all other solutions in the RE task (best F1-score 0.79) and a specific use case it was applied to by recognizing nearly all entity mentions and most of the relations. Our findings support the use of AI-assisted approaches in facilitating precision oncology decision-making.
  5. J Am Acad Orthop Surg. 2025 May 20.
       BACKGROUND: Clubfoot is a congenital foot deformity treated primarily by the Ponseti method. Despite its effectiveness, challenges such as delayed treatment and nonadherence persist, leading parents to seek information online. With the rise of artificial intelligence (AI) chatbots as medical information sources, concerns have emerged regarding the accuracy, clarity, and safety of AI-generated content.
    OBJECTIVE: To evaluate the accuracy, clarity, and safety of AI-generated responses to common parental questions about clubfoot.
    METHODS: Nineteen frequently asked questions by parents of children with clubfoot were identified. Three AI language models-Google's Gemini, OpenAI's ChatGPT-4.0, and Meta AI-were each queried twice with these questions. Two pediatric orthopaedic surgeons independently assessed the AI responses for clarity, accuracy, and safety using a four-point scale (1 = lowest, 4 = highest). Statistical analyses were conducted using chi-square and t-tests with a significance level of P < 0.05.
    RESULTS: A total of 228 evaluations were obtained (76 per AI model). The overall average scores were 3.5 for clarity, 3.0 for accuracy, and 3.3 for safety. ChatGPT scored highest in clarity (mean = 3.5), Gemini in accuracy (mean = 3.2) and safety (mean = 3.4). No statistically significant differences were found among the models. Approximately 25% of responses were limited or inaccurate, and only 30% were deemed entirely safe.
    CONCLUSION: Although AI chatbots generally provide adequate information on clubfoot, notable concerns about accuracy and patient safety remain. It is recommended that patients consult healthcare professionals rather than rely solely on AI-generated advice. Developing AI models that access up-to-date, evidence-based medical databases could enhance the reliability and safety of health information provided to the public.
    DOI:  https://doi.org/10.5435/JAAOS-D-24-01274
  6. Am J Health Promot. 2025 May 29. 8901171251348208
      PurposeArtificial intelligence (AI) is increasingly integrated into healthcare, including psychiatric care. This study evaluates ChatGPT-4o's reliability in answering frequently asked antidepressant-related questions by comparing its performance with psychiatrists across four key dimensions: accuracy, conciseness, readability, and clarity.DesignA comparative study analyzing ChatGPT-4o-generated responses and those of psychiatrists with at least five years of clinical experience.SettingParticipants were recruited through institutional and professional networks and provided with standardized questions derived from authoritative treatment guidelines.SubjectsTwenty-six psychiatrists participated, and ChatGPT-4o responses were generated using a standardized prompt for each question.MeasuresTwo independent psychiatrists evaluated accuracy and conciseness using a blinded rating system. Readability was assessed with the Flesch-Kincaid Grade Level test, and clarity was measured with the Writing Clarity Index Calculator.AnalysisThe Shapiro-Wilk test assessed normality. Paired t-tests were used for normally distributed data, and the Wilcoxon signed-rank test for non-normally distributed data. Statistical significance was set at P < .05.ResultsChatGPT-4o showed comparable accuracy to psychiatrists (P = .0645) but was significantly more concise (P = .0019). Readability differences were not statistically significant (P = .0892), while psychiatrists provided clearer responses (P = .0059).ConclusionChatGPT-4o delivers accurate and concise responses, highlighting its potential as a patient education tool. However, psychiatrists offer greater clarity, underscoring the indispensable role of clinical expertise in psychiatric care.
    Keywords:  ChatGPT; accuracy; antidepressants; artificial intelligence in healthcare; psychiatrists
    DOI:  https://doi.org/10.1177/08901171251348208
  7. J Pediatr Soc North Am. 2025 Feb;10 100135
       Background: Large language models, including Chat Generative Pre-trained Transformer (ChatGPT) and Google Gemini have accelerated public accessibility to information, but their accuracy to medical questions remains unknown. In pediatric orthopaedics, no study has utilized board-certified expert opinion to evaluate the accuracy of artificial intelligence (AI) chatbots compared to evidence-based recommendations, including the American Academy of Orthopaedic Surgeons clinical practice guidelines (AAOS CPGs). The aims of this study were to compare responses by ChatGPT-4.0, ChatGPT-3.5, and Google Gemini with AAOS CPG recommendations on developmental dysplasia of the hip (DDH) regarding accuracy, supplementary and incomplete response patterns, and readability.
    Methods: ChatGPT-4.0, ChatGPT-3.5, and Google Gemini were prompted by questions created from 9 evidence-based recommendations from the 2022 AAOS CPG on DDH. The answers to these questions were obtained on July 1st, 2024. Responses were anonymized and independently evaluated by two pediatric orthopaedic attending surgeons. Supplementary responses were additionally evaluated on whether no, some, or many modifications were necessary. Readability metrics (response length, Flesch-Kincaid reading level, Flesch Reading Ease, Gunning Fog Index) were compared. Cohen's Kappa inter-rater reliability (κ) was calculated. Chi-square analyses and single-factor analysis of variance were utilized to compare categorical and continuous variables, respectively. Statistical significance was set with P ​< ​0.05.
    Results: ChatGPT-4.0, ChatGPT-3.5, and Google Gemini were accurate in 5/9, 6/9, 6/9, supplementary in 8/9, 7/9, 9/9, and incomplete in 7/9, 6/9, 7/9 recommendations, respectively. Of 24 supplementary responses, 5 (20.8%), 16 (66.7%), and 3 (12.5%) required no, some, and many modifications, respectively. There were no significant differences in accuracy (P ​= ​0.853), supplementary responses (P ​= ​0.325), necessary modifications (P ​= ​0.661), and incomplete responses (P ​= ​0.825). κ was highest for accuracy at 0.17. Google Gemini was significantly more readable in Flesch-Kincaid reading level, Flesch Reading Ease, and Gunning fog index (all, P ​< ​0.05).
    Conclusions: In the setting of DDH, AI chatbots demonstrated limited accuracy, high supplementary and incomplete response patterns, and complex readability. Pediatric orthopaedic surgeons can counsel patients and their families to set appropriate expectations on the utility of these novel tools.
    Key Concepts: (1)Responses by ChatGPT-4.0, ChatGPT-3.5, and Google Gemini were inadequately accurate, frequently provided supplementary information that required modifications ​and frequently lacked essential details from the AAOS CPGs on DDH.(2)Accurate, supplementary, and incomplete response patterns were not significantly different among the three chatbots.(3)Google Gemini provided responses that had the highest readability among the three chatbots.(4)Pediatric orthopaedic surgeons can play a role in counseling patients and their families on the limited utility of AI chatbots for patient education purposes.
    Level of Evidence: IV.
    Keywords:  American Academy of Orthopaedic Surgeons; ChatGPT; Clinical practice guideline; Developmental dysplasia of the hip; Google Gemini
    DOI:  https://doi.org/10.1016/j.jposna.2024.100135
  8. J Pediatr Ophthalmol Strabismus. 2025 May 27. 1-7
       PURPOSE: To evaluate the quality and readability of ChatGPT-4o-generated (ChatGPT) (OpenAI) patient education materials (PEMs) about pediatric ophthalmo-logic surgical procedures and compare these to PEMs from the American Association for Pediatric Ophthalmology and Strabismus (AAPOS) website.
    METHODS: The authors prompted ChatGPT-4o to provide PEMs on four procedures-strabismus surgery without adjustable sutures, strabismus surgery with adjustable sutures, pediatric cataract surgery, and nasolacrimal duct probing. The prompt requested responses at a 6th grade level in both Spanish and English. English ChatGPT responses were compared to AAPOS PEMs on quality (using the Quality of Generated Language Outputs for Patients [QGLOP] scale) and readability. English and Spanish ChatGPT responses were also compared on quality and readability.
    RESULTS: Based on average scores from the four procedures, AAPOS PEMs were superior to English ChatGPT responses on the accuracy, currency, and tone sub-scales of the QGLOP score (4.0 ± 0 vs 2.79 ± 0.79, P = .0021; 3.79 ± 0.26 vs 3.38 ± 0.71, P = .033; 4.0 ± 0 vs 3.42 ± 0.69, P = .042, respectively). There was no significant difference in readability between AAPOS PEMs and English ChatGPT responses. English and Spanish Chat- GPT responses did not significantly differ on quality or readability.
    CONCLUSIONS: ChatGPT-4o-generated PEMs on pediatric ophthalmologic surgical conditions are currently inferior in quality to PEMs on the AAPOS website. However, because ChatGPT is continually being updated and trained, this study should be repeated in the future to determine whether metrics improve over time. [J Pediatr Ophthalmol Strabismus. 20XX;X(X):XXX-XXX.].
    DOI:  https://doi.org/10.3928/01913913-20250404-01
  9. Int J Obstet Anesth. 2025 May 20. pii: S0959-289X(25)00280-8. [Epub ahead of print]63 104688
       BACKGROUND: Recent studies evaluating frequently asked questions (FAQs) on labor epidural analgesia (LEA) only used generic questions without incorporating detailed clinical information that reflects patient-specific inputs. We investigated the performance of ChatGPT in addressing these questions related to LEA with an emphasis on individual preferences and clinical conditions.
    METHODS: Twenty-nine questions for the AI chatbot were generated from the commonly asked questions relating to LEA based on clinical conditions. The generation of responses was performed in January 2025 with each question under individual sub-topics initiated as a "New chat" in ChatGPT-4o. Upon having the first questions answered, subsequent question(s) in the same sub-topic were continued in the same chat following the sequences as predefined. The readability of each response was graded using six readability indices, while the accuracy, Patient Education Materials Assessment Tool for Print (PEMAT) understandability and actionability was assessed by four obstetric anesthesiologists.
    RESULTS: The mean readability indices of the ChatGPT-4o responses to the questions were generally rated as fairly difficult to very difficult, which corresponded to a US grade level between 11th grade to college level entry. The mean (± standard deviation) accuracy of the responses was 97.7% ± 8.1%. The PEMAT understandability and actionability scores were 97.9% ± 0.9%) and 98.0% ± 1.4%), respectively.
    CONCLUSIONS: ChatGPT can provide accurate and readable information about LEA even under different clinical contexts. However, improvement is needed to refine the responses with suitable prompts to simplify the outputs and improve readability. These approaches will thereby meet the need for the effective delivery of reliable patient education information.
    Keywords:  Artificial intelligence; Chatbots; Context; Information; Labor epidural analgesia; Patient preferences; Prompts; Questions
    DOI:  https://doi.org/10.1016/j.ijoa.2025.104688
  10. J Pediatr Ophthalmol Strabismus. 2025 May 27. 1-7
       PURPOSE: To evaluate the success of Chat Generative Pre-trained Transformer (ChatGPT; OpenAl), Google Gemini (Alphabet, Inc), and Microsoft Copilot (Microsoft Corporation) artificial intelligence (AI) programs, which are offered free of charge by three different manufacturers, in answering questions related to pediatric ophthalmology correctly and to investigate whether they are superior to each other.
    METHODS: ChatGPT, Gemini, and Copilot were each asked 100 multiple-choice questions from the Ophtho-Questions online question bank, which is widely used for preparing for the high-stakes Ophthalmic Knowledge Evaluation Program examination. Their answers were compared to the official answer keys and categorized as correct or incorrect. The readability of the responses was assessed using the Flesch-Kincaid Grade Level, Flesch Reading Ease Score, and the Coleman-Liau Index.
    RESULTS: ChatGPT, Gemini, and Copilot chatbots answered 61 (61%), 60 (60%), and 74 (74%) questions correctly, respectively. The Copilot AI program had a significantly higher rate of correct answers to questions than ChatGPT and Gemini (P = .049 and .035). Three readability analyses revealed that Copilot had the highest average score, followed by ChatGPT and Gemini, which were more challenging than the recommended level.
    CONCLUSIONS: Although AI chatbots can serve as useful tools for acquiring information on pediatric ophthalmology, their responses should be interpreted with caution due to potential inaccuracies. [J Pediatr Ophthalmol Strabismus. 20XX;X(X):XXX-XXX.].
    DOI:  https://doi.org/10.3928/01913913-20250404-03
  11. Ann Otol Rhinol Laryngol. 2025 May 28. 34894251342969
       BACKGROUND: To compare the readability of patient education materials (PEMs) on rhinologic conditions and procedures from the American Rhinologic Society (ARS) with those generated by large language models (LLMs).
    METHODS: Forty-one PEMs from the ARS were retrieved. Readability was assessed through the Flesch Kincaid Reading Ease (FKRE) and Flesch Kincaid Grade Level (FKGL), in which higher FKRE and lower FKGL scores indicate better readability. Three LLMs-ChatGPT 4.o, Google Gemini, and Microsoft Copilot-were then used to translate each ARS PEM to the recommended sixth-grade reading level. Readability scores were calculated and compared for each translated PEM.
    RESULTS: A total of 164 PEMs were evaluated, including 123 generated by LLMs. The original ARS PEMs had a mean FKGL of 10.28, while AI-generated PEMs demonstrated significantly better readability, with a mean FKGL of 8.6 (P < .0001). Among the AI platforms, Gemini was the most easily readable, reaching a mean FKGL of 7.5 and FKRE of 65.5.
    CONCLUSION: LLMs improved the readability of PEMs, potentially enhancing accessibility to medical information for diverse populations. Despite these findings, healthcare providers and patients should cautiously appraise LLM-generated content, particularly for rhinology conditions and procedures.
    LEVEL OF EVIDENCE: N/A.
    Keywords:  artificial intelligence; clinical rhinology; health literacy; large language models; patient care; patient education materials; rhinology; surgical education
    DOI:  https://doi.org/10.1177/00034894251342969
  12. Clin Nurs Res. 2025 May 26. 10547738251343418
      Effective patient education is critical to an individual's treatment plan when living with diabetes, a debilitating disease requiring extensive knowledge and skills to effectively manage and prevent future morbidity. This descriptive, cross-sectional study assessed the appropriateness of commercially developed patient education materials for rural adults with diabetes. Using electronic health record data from patients (n = 132) with documented health literacy levels (HLL), their HLL scores were reviewed for suitability of educational materials provided by a single rural-border hospital. HLL was measured using the Newest Vital Sign, while educational materials were assessed by two independent reviewers using the Patient Education Materials Assessment Tool for Printable Materials to measure understandability and actionability, and the Simple Measure of Gobbledygook to measure the readability level of documents in both English and Spanish. The mean HLL for Spanish speakers (n = 77) was 2.01 (±1.56), while the mean HLL for English speakers (n = 55) was 2.09 (±1.46), indicating a high likelihood of low health literacy. The materials were evaluated with the English materials achieving an understandability score of 83.33% and an actionability score of 60.00%. By contrast, the Spanish materials scored lower, 61.54% and 20.00%, respectively. In addition, the readability of the materials measured a 9.0 grade level for English materials and a slightly higher 10.8 for Spanish materials. These findings highlight a significant discrepancy between the HLL and the appropriateness of educational materials for the sample population. Although commercially available educational products may provide a budget-concise solution to patient education, particularly in low-resourced organizations, they fail to address the health education needs of the individuals who receive them. Despite the project's limited sample size and single geographical location, it underscores the importance for healthcare organizations to address the health education needs of their communities.
    Keywords:  Newest Vital Sign; diabetes education; health education; health literacy; patient education; rural health
    DOI:  https://doi.org/10.1177/10547738251343418
  13. Urologie. 2025 May 19.
       BACKGROUND: The interest in digital information on pelvic floor dysfunction is constantly increasing. Various digital platforms offer an easy and anonymous way for patients to seek information about their condition. However, little is known about the quality of the information on the different platforms or about the how the quality of different sites compares.
    OBJECTIVE: The aim of this study was to investigate the completeness and quality of information on the search term "stress urinary incontinence" in comparison between different digital platforms.
    MATERIALS AND METHODS: A systematic analysis of the keyword search "stress urinary incontinence" was performed on Google and the social networks Facebook, YouTube, Instagram, and LinkedIn. The first 30 search results on each platform were evaluated. The results were categorized according to information content and readability. The Health On the Net Foundation (HON) seal was used to assess medical quality.
    RESULTS: The proportion of informative content was highest on YouTube (97%) and Google (93%). Content was predominantly provided by professional organizations on Google and YouTube. Information on conservative therapies dominated across all platforms. Surgical therapies were only discussed in up to 63% of results on Google and in up to 50% of results on YouTube. In most cases, there was also no comprehensive presentation of all surgical options. The readability of the texts was unsuitable for laypersons on all platforms, and HON certification was only present on Google (37%) and YouTube (3%).
    CONCLUSION: The results offer practical insights into the quality of digital information on stress urinary incontinence. However, they show deficits in readability and comprehensive presentation of surgical therapies. The physician-patient relationship remains indispensable for taking individual needs into account and avoiding misinformation.
    Keywords:  Digital media; Functional urology; Social media; Urinary incontinence; Urogenital surgical procedures
    DOI:  https://doi.org/10.1007/s00120-025-02593-7
  14. Musculoskeletal Care. 2025 Jun;23(2): e70130
       BACKGROUND: Patient education is a cornerstone of care for individuals with non-specific low back pain (LBP). However, little is known about whether accessible patient education materials (PEMs) meet people's information needs.
    METHODS: We conducted a scoping review following the JBI methodology and reported results according to PRISMA-ScR. We systematically reviewed three databases: Ovid MEDLINE, Scopus, and CINAHL. The search strategy was iteratively developed and peer-reviewed using the PRESS checklist. Eligible studies had to provide full access to the PEM designed for people with LBP. Study selection and data extraction were performed independently and in duplicate. Five reviewers conducted a consensus-based analysis by independently matching PEM content to eight categories of information needs derived from previous research.
    RESULTS: Of 9617 citations identified, 23 studies met inclusion criteria, yielding 41 unique PEMs. We excluded many citations (67.3%) because the PEM used in the study was missing. Most PEMs were in English (95%) and took the form of posters, booklets, or leaflets. Only eight PEMs (19.5%) reported readability assessment. Stakeholder involvement was reported in eight studies. Among PEMs with stakeholder input, characteristics from the PROGRESS + framework were rarely disclosed. Only one PEM addressed all eight identified information needs. The most frequently covered information needs were treatment options (65.9%) and imaging (61.0%), while information on prognosis and flare management was scarce (17.1%).
    CONCLUSION: Accessible PEMs for non-specific LBP rarely meet the full spectrum of patient information needs. Improving stakeholder involvement and readability assessment is essential to enhance the usefulness and equity of educational resources.
    Keywords:  health equity; information needs; low back pain; patient education; stakeholder involvement
    DOI:  https://doi.org/10.1002/msc.70130
  15. J Commun Healthc. 2025 May 27. 1-10
       BACKGROUND: Distressful clinician-to-patient dialogue such as a pre-cancer diagnosis of cervical dysplasia may interfere with information retention. Patient education material provided as an online resource offers a suitable option to review relevant health information outside the clinic. The aim of this study was to evaluate online resources (ORs) affiliated with healthcare institutions across Australia and the United Kingdom (UK) on their effectiveness to translate accessible and current knowledge to patients referred for loop electrosurgical excision procedure (LEEP) treatment.
    METHODS: A comprehensive directory of ORs related to LEEP was compiled from public hospital websites across Australia and the UK. Quantitative and qualitative methods were applied to evaluate resource reading-level (measured using three validated readability indices); actionability and understandability (measured using the Patient Education Material Assessment Tool [PEMAT]); and content (described using content analysis to assess disclosure practices associated with LEEP-related complications).
    RESULTS: All ORs (n = 39) exceeded the recommended reading level (Australia: x̄ = 10.07, σ = 1.01; UK: x̄ = 10.17, σ = 0.96). PEMAT results indicated higher percentages of ORs scored as understandable (Australia: 50.0%; UK: 69.7%) versus actionable (Australia: 33.3%; UK: 6.1%). Content analysis revealed widespread discordance in the disclosure of longer-term LEEP complications associated with pregnancy, fertility, and sexual function in both countries.
    CONCLUSIONS: Disclosures with significant health and wellness implications should be made with clear reference to peer reviewed science. Wider application of purpose-designed health literacy tools could improve measures of readability, actionability and understandability. International collaborations may provide opportunities to develop more comprehensive and patient-centred education materials to improve provider-to-patient knowledge translation.
    Keywords:  Cervical Cancer; Knowledge translation; LEEP; institutional communication; patient education; pre-cancer communication; provider patient communication; reproductive health
    DOI:  https://doi.org/10.1080/17538068.2025.2508346
  16. J Pediatr Soc North Am. 2025 May;11 100169
       Background: Patients newly diagnosed with Legg-Calve-Perthes disease (LCPD) often use the internet to learn about their disease. The American Medical Association (AMA) and the National Institutes of Health (NIH) recommend writing patient education resources at sixth- to eigth-grade reading levels. The purpose of this study was to determine the availability of online patient education resources for LCPD, assess the availability of translation to other languages, and analyze the readability of English-language resources.
    Methods: A cross-sectional analysis was conducted to determine the availability and readability of online, written patient education materials for LCPD. The top 50 pediatric orthopaedic hospitals, as reported by the U.S. News and World Report, major professional societies (the American Academy of Orthopaedic Surgeons [AAOS], the American Academy of Pediatrics [AAP], and the Pediatric Orthopaedic Society of North America [POSNA]), and international societies (POSNA "Alliance Societies") were included. The number of translations to other languages was determined among the institutions with English resources. Three different readability measures were used on resources written in English. Statistical analysis was performed using descriptive, univariate, and correlation analyses.
    Results: Seventy-four percent (37/50) of hospitals, 67% (2/3) of U.S.-based professional societies, and 16% (5/31) of international societies provided LCPD education resources. Of the hospitals and societies that provided English resources, 36% (14/39) provided translation to other languages. The most common language available was Spanish. Resources written in English had an average Flesch Reading Ease score of 60.5, Fry Graph grade level of 9.6, and Simple Measure of Gobbledygook (SMOG) grade level of 11.4. There was no significant relationship between hospital rank and availability (P > .05) or reading level (P > .05) of resources and no significant relationship between a state's percentage of Spanish speakers and availability of Spanish translations (P > .05).
    Conclusions: Most selected hospitals and U.S.-based professional societies provided information about LCPD; however, many did not provide translation to other languages. There is a discrepancy between the reading level of the resources queried and national reading level recommendations. There is a need for translated LCPD resources from hospitals and societies domestically and abroad to provide adequate patient education, particularly for individuals with lower reading levels or non-English speakers.
    Key Concepts: (1)Online patient educational materials across United States hospitals and major pediatric orthopaedic societies for Legg-Calve-Perthes disease (LCPD) were written above the recommended English reading levels.(2)There is a limited availability of patient educational material written in non-English languages despite a large percentage of non-English speakers in the United States.(3)While Spanish is the most common non-English language spoken in the United States, there was no increase in the availability of educational materials written in Spanish in states with a population >20% Spanish speakers.
    Level of Evidence: III.
    Keywords:  Health equity; Health literacy; Legg-Calve-Perthes disease; Limited English proficiency; Patient education materials
    DOI:  https://doi.org/10.1016/j.jposna.2025.100169
  17. Australas J Dermatol. 2025 May 30.
      Hidradenitis suppurativa (HS) or acne inversa is a chronic inflammatory skin condition characterised by painful nodules that significantly diminish quality of life. Patients diagnosed with HS commonly utilise online resources to gain insight into their condition. This study aimed to assess the readability, quality and relevance of 115 online HS resources found via Google. It was found that most of the websites exceeded the recommended 6th-grade reading level, with the average being around the 11th-grade level. JAMA and DISCERN criteria were used to assess the quality of the resources, which revealed limited reliability and timeliness. Our findings emphasise the need for accessible, high-quality, and up-to-date online resources to support informed decision-making for HS patients and promote prompt medical care.
    Keywords:  acne inversa; hidradenitis suppurativa; online resources; readability
    DOI:  https://doi.org/10.1111/ajd.14533
  18. Clin Exp Dermatol. 2025 May 28. pii: llaf236. [Epub ahead of print]
      Hidradenitis suppurativa (HS) is a chronic inflammatory skin disorder associated with significant physical, psychological, and social burdens. Despite increasing recognition, diagnostic delays remain common, often prompting patients to seek information online. This systematic review evaluated the quality and readability of HS-related information across AI-generated content, search engine derived resources, and social media platforms. A comprehensive search of PubMed, Embase, Cochrane Library, and Google Scholar identified 17 studies published between 2017 and 2024 that assessed HS-related online content using validated scoring tools and/or physician evaluation. More than 50% of studies rated online HS materials as variable in quality, with 36.4% rating them as moderate. Readability assessments revealed that most resources exceeded the recommended 6th to 8th grade reading level, limiting accessibility for patients. Social media platforms, particularly TikTok and YouTube, featured highly engaging but frequently inaccurate or anecdotal content, with physician-generated materials receiving lower engagement than non-medical sources. These findings highlight the critical need for simplified, evidence-based online resources to improve health literacy and support informed decision-making in HS. The prevalence of misinformation, particularly regarding alternative treatments and pharmaceutical scepticism, underscores the urgency of enhanced patient education strategies. Future efforts should focus on AI-driven readability improvements, clinician engagement in digital education, and collaboration with social media platforms to ensure the availability of accessible, high-quality HS information.
    DOI:  https://doi.org/10.1093/ced/llaf236
  19. Can Pharm J (Ott). 2025 May 22. 17151635251332612
       Background: The World Health Organization's right to health underscores the need for accessible, acceptable, and quality health services. Given that most Canadians use the Internet for health information, the readability of online pharmacy services information is crucial for accessibility.
    Methods: This study assessed the readability of online information about pharmacy services from Canadian provincial pharmacy regulatory authorities (PRAs) and community pharmacy banners. Public-facing website content was evaluated using various readability tests. Scores were compared to recommended reading grade levels by health organizations, and differences between PRA and community pharmacy banner websites were analyzed.
    Results: Website content from 9 PRAs and 10 community pharmacy banners was analyzed in June 2024. Average readability scores exceeded the recommended eighth-grade level, with summary scores ranging from 8.45 to 15.28. International English Language Testing System scores for all websites also surpassed reading benchmarks necessary for Canadian immigration. Mann-Whitney U tests indicated statistically significant differences between PRA and community pharmacy banner websites, with the latter being more readable.
    Discussion: The results suggest that both PRAs and community pharmacy banners provide information at an advanced reading level, hindering accessibility. This aligns with other research indicating that online health information is often too complex for the general public. Improving readability, particularly for new Canadians, is essential for better accessibility.
    Conclusion: Public health information on PRA and community pharmacy banner websites generally exceeds the recommended readability level, limiting accessibility. Implementing readability assessments and plain-language standards can enhance the accessibility and engagement of online health information.
    DOI:  https://doi.org/10.1177/17151635251332612
  20. Int J Environ Health Res. 2025 Mar 01. 1-10
      Air pollution is a significant public health problem affecting children's mortality, morbidity, and causing respiratory diseases and malignancy. Parents are increasingly using YouTube to obtain health information about their children. This study aims to analyze YouTube videos on the effects of air pollution on children's health. A total of 200 English-language videos about air pollution and child health were assessed. We used the Journal of American Medical Association (JAMA) benchmark criteria, Global Quality (GQ), and m-DISCERN scores to evaluate video quality, reliability, and content. Among the videos, 50 (79.4%) were categorized as "Informative" and 13 (20.6%) as "Misleading." Academic institution-produced videos had higher JAMA and m-DISCERN scores compared to other groups (p < 0.001 for both). The median duration of videos in the Informative group was 164.5 seconds, while it was 68 seconds in the Misleading group (p = 0.001). We found that videos longer than 4 minutes had fewer views. YouTube videos on air pollution and child health should provide more accurate information and better sources, and additional reference materials should be included for parents. Universities, academic institutes, and international organizations should play a more active role by creating educational videos on the impact of air pollution on child health.
    Keywords:  Air pollution; YouTube; child health; children; social media
    DOI:  https://doi.org/10.1080/09603123.2025.2474089
  21. BMC Public Health. 2025 May 24. 25(1): 1919
       BACKGROUND: Pancreatic cancer (PC) has become one of the leading causes of cancer-related deaths worldwide. Social media platforms are widely used for health information dissemination because of their visual appeal and entertainment value. This study evaluates the content, quality, and reliability of PC-related information on domestic short video platforms.
    METHODS: A total of 265 PC-related videos were retrieved from three short video-sharing platforms: TikTok, Bilibili, and Kwai. The Global Quality Scale (GQS) and the modified DISCERN score were employed to evaluate the quality and content of the videos, respectively. Correlation analysis was conducted to explore the relationships between different video variables.
    RESULTS: The overall quality of the video content was low, with median scores of 2 (IQR: 2-3) for both the GQS and the modified DISCERN score. Most of the videos related to PC were posted by healthcare professionals (219/265, 82.6%). Videos from specialists received more likes on short social video platforms than those from nonspecialists did (median: 678 vs. 270, P = 0.005). Educational videos scored highest in both the GQS (median: 3, IQR 2-3) and the modified DISCERN score (median: 2.5, IQR 2-3). There was a positive correlation between GQS and video duration (r = 0.31, P < 0.01) and the modified DISCERN score (r = 0.434, P < 0.001).
    CONCLUSION: The quality and reliability of the videos on these platforms were generally unsatisfactory in terms of source and content. Videos produced by healthcare professionals or institutions are more informative in terms of comprehensiveness, quality of information, and reliability than are those produced by non-healthcare professionals.
    Keywords:  Bilibili; DISCERN; GQS; Health information; Kwai; Pancreatic cancer; Quality; Reliability; Short video; TikTok
    DOI:  https://doi.org/10.1186/s12889-025-23130-3
  22. Clin Exp Ophthalmol. 2025 May 24.
      
    Keywords:   YouTube ; DISCERN score; global quality score; journal of the American medical association score; optic neuritis
    DOI:  https://doi.org/10.1111/ceo.14555
  23. Aust Endod J. 2025 May 24.
      This study aims to evaluate the content and quality of videos related to minimally invasive endodontics available on the YouTube platform. Twenty videos were analysed using the video information and quality index for content quality, the Global Quality Scale for educational value, the Journal of the American Medical Association benchmark for source reliability and the adapted DISCERN tool, known as the reliability score, for overall information accuracy. Data were assessed using Kolmogorov-Smirnov, Mann-Whitney U and Spearman's correlation tests. The findings indicate that low-quality videos had higher view counts, while videos of moderate quality achieved higher Global Quality Scale, reliability score and video information and quality index scores. Additionally, statistically significant positive correlations were identified among these scores. In conclusion, most online content regarding minimally invasive endodontics presents limited accuracy and reliability, and professional consultation is recommended.
    Keywords:  E‐health; YouTube; minimally invasive endodontics; social media
    DOI:  https://doi.org/10.1111/aej.12954
  24. JMIR Dermatol. 2025 May 28. 8 e65217
      
    Keywords:  MD; Reddit; Student Doctor Network; TikTok; cross-section; cross-sectional analysis; dermatology; dermatology match; digital health; health content; health information; health platform; information; media; medical residency; medical school; misinformation; online; online content; online data; online health; online platform; qualifications; residency; residency program; social media; social media content; social media posts; web platform
    DOI:  https://doi.org/10.2196/65217
  25. Strabismus. 2025 May 28. 1-9
      Purpose: To conduct a quality assessment of educational content on social media for patients with strabismus.
     Methods: TikTok and Instagram were searched for several hashtags. Educational videos meeting inclusion criteria under #strabismus and #lazyeye were analyzed using the Global Quality Scale (GQS), mDISCERN, and PEMAT-A/V tools. Statistical analysis was conducted to assess associations between video quality, metrics, and authorship.
     Results: Among the top 50 videos for #strabismus and #lazyeye on both platforms (n = 200), educational videos aimed at patients (n = 54) scored an average mDISCERN of 2.8/5, GQS of 3.3/5, PEMAT-A/V understandability of 78.5%, and actionability of 42.3%. Pearson's correlation analysis showed views and likes to be negatively correlated with mDISCERN and GQS scores (p < .05), but positively correlated with PEMAT-A/V actionability (p < .05). Content created by healthcare professionals, including optometrists and ophthalmologists, scored higher mDISCERN, GQS and PEMAT-A/V understandability scores than patient-created content (p < .05).
     Conclusion: This study highlights the significant role of social media in disseminating health information on strabismus. Educational content for strabismus has marked heterogeneity, and lower-quality videos yield more views and likes, which could lead to patients following misleading or incomplete medical advice. The positive correlation between actionability and lower-quality videos is particularly concerning, as these videos, while actionable, may encourage patients to take steps based on poorly explained information. To enhance social media as an educational tool, there is a clear need for more high-quality content created by qualified healthcare professionals, to ensure that patients receive accurate, safe guidance.
    Keywords:  Amblyopia; GQS; Instagram; Squint; TikTok; mDISCERN
    DOI:  https://doi.org/10.1080/09273972.2025.2509487
  26. Gastroenterol Nurs. 2025 May-Jun 01;48(3):48(3): 182-190
      TikTok has gained popularity among young adults. Individuals with inflammatory bowel disease (IBD), a chronic autoimmune disease commonly diagnosed between ages of 15 and 29, often use multiple social media platforms for medical information and support. This study aimed to assess the content and quality of the most viewed TikTok videos related to IBD. Using web scraping, 86 videos were collected from the 100 most viewed videos under three hashtags: #ibd, #crohns, #colitis. The analysis focused on themes, video metrics, characteristics, and quality using the modified DISCERN tool and Global Quality Scale. All videos were made by individuals with IBD, and 83% were personal narratives. Eight main themes emerged: ostomy, advocacy, disease management, interactions with the healthcare system, symptoms, social relationships, negative impact of IBD, and complications. The median DISCERN and global quality scores were 1.00 (interquartile range [IQR]: 1.00), and 2.00 (IQR: 2.00), respectively, indicating poor quality. The main themes may indicate a need for IBD-specific education regarding ostomies, advocacy, and disease management. TikTok offers a space for individuals with IBD to share information and experiences. Healthcare providers can utilize this platform to interact with and educate those with IBD; however, no healthcare provider videos were among the most popular.
    DOI:  https://doi.org/10.1097/SGA.0000000000000863
  27. J Cancer Educ. 2025 May 26.
      The information needs of cancer survivors are numerous and chronically difficult to meet, and the information sources currently available to them are far from ideal. The purpose of this study was to investigate the information-seeking behaviors and preferences of cancer survivors and explore the differences in information needs among cancer survivors of different types in Shanghai, China. A quota sampling method was used to select cancer survivors living in all districts of Shanghai as the study population. Data collection used the Unmet Information Needs Scale for Cancer Patients. ANOVA were used to assess differences in unmet need scores among subgroups. The mean age of 4195 cancer survivors was 63.2 ± 7.4 years, comprising 823 males (19.6%) and 3372 females (80.4%). Among the current sources of information, 36.2% of cancer survivors opted for consultations with tertiary-level doctors. In terms of desired sources, 43.9% preferred tertiary-level doctors. Regardless of the current or future situation, tertiary-level doctors, patient friends, and primary care doctor remain the top three preferred sources for health information. Kidney cancer, metastatic cancer, and nasopharyngeal cancer demonstrated the highest information requirement scores, scoring 39.4, 37.5, and 36.6, respectively. Cancer survivors of different cancer types showed differences in focus and information needs. Current and expected information sources for cancer survivors differ at different stages of survival. Tertiary-level doctors are the most popular providers of information. Unmet information needs for different cancer types are diverse and complex.
    Keywords:  Cancer survivors; Information seeking behaviors; Information seeking preferences
    DOI:  https://doi.org/10.1007/s13187-025-02628-9
  28. Glob Health Promot. 2025 May 26. 17579759251334395
       BACKGROUND: Many adolescents do not have access to sexual and reproductive health education due to the lack of structural support specific to cultural formatting and overall negative attitude toward this type of education. Therefore, adolescents often use the internet as the main source of health information. The purpose of this study was to investigate specific internet websites associated with online searching for information about sex and sexually transmitted infections (STIs).
    METHODS: A cross-sectional study was carried out on 702 high school students in Belgrade, the capital city of Serbia. Relevant data were obtained through an anonymous questionnaire. The questionnaire included socio-demographic characteristics, use of the internet and online platforms. Electronic health literacy was tested by the e-Health literacy scale.
    RESULTS: All students used the internet. The prevalence of seeking online information about sex was 31.9% and 12.4% about STIs. Being a boy, having lower grades, attending the humanities-languages program, being younger when using the internet for the first time, using forums and social media, but not websites run by physicians, and browsing online information about STIs were associated with seeking online information about sex. Girls, who had lower grades, had higher family income, were older when they first started using the internet, used social media and websites run by physicians and searched for online information about sex were more likely to seek online information about STIs.
    CONCLUSIONS: Adolescents in this study sought online information about sex more often than STIs. Social media were used to search both topics online, suggesting that participatory media which offer engaging content could be the preferred manner of online learning about sex and STIs. These online resources could be used for information delivery about reproductive health for high school students.
    Keywords:  adolescents; internet; online health information seeking; sex; sexually transmitted infections
    DOI:  https://doi.org/10.1177/17579759251334395
  29. AMIA Annu Symp Proc. 2024 ;2024 300-308
      Online health information sources (OHIS) offer potential for improving access to health information especially in areas with limited healthcare infrastructure. However, OHIS predominantly originates from Western societies potentially ignoring the specific needs and cultural contexts of diverse populations. There is limited research on the global suitability of OHIS content. This study explores the global relevance of OHIS for diverse populations through a case study examining user experiences of Nigerians living in multiple countries. Findings reveal OHIS usage patterns are influenced by the country of residence and local health services availability. The study highlights the need for culturally inclusive OHIS content to ensure equitable health information access globally. Ultimately, for OHIS to serve a global audience effectively, there needs to be reliable information sources that acknowledge and cater to different users' cultural backgrounds, including prevalent health issues, medical practices, beliefs, languages, and healthcare expectations.
  30. Aust N Z J Obstet Gynaecol. 2025 May 28.
       OBJECTIVE(S): Accessible and accurate information about long-acting reversible contraception (LARC) is important in addressing misperceptions and ensuring consumers can make informed decisions about their use, yet little is known about consumers' experiences of searching for and being provided information about LARC.
    METHODS: Individual semi-structured 'Zoom' interviews were conducted with people living in Australia aged 18-45, who were currently using a LARC or whose LARC was removed within the past 6 months. Data were analysed using reflexive thematic analysis.
    RESULTS: A total of 23 interviews were conducted between October 2022-January 2023. Participants were actively engaged in seeking information about LARC prior to initiation and/or during its use. Participants sought information from a variety of sources to validate, or support information previously acquired and made judgements on the accuracy of different sources. Hearing the experiences of other LARC users was highly valued and viewed as providing information not available from other sources, especially in regard to expected side effects. Although healthcare providers were seen as the best source of information, many participants reported dissatisfaction with clinical interactions.
    CONCLUSION: Consumers used a variety of information sources when making decisions about LARC and highly valued hearing other LARC users experiences. Healthcare providers and organisations should consider incorporating lived experiences of LARC into evidence-based information to meet the needs of consumers.
    Keywords:  contraceptive implant; health promotion; information seeking; intrauterine device; long‐acting reversible contraception; qualitative
    DOI:  https://doi.org/10.1111/ajo.70044
  31. Aust J Rural Health. 2025 Jun;33(3): e70060
       OBJECTIVE: The effects of childhood health, education and experiences can have long-term impacts on adult health and wellbeing. Access to health services and information can be complex especially in regional and rural areas of Australia. This research aimed to: (1) investigate how and where parents living in regional and rural Australia with young children search for health information and (2) explore how parents decide what is appropriate health information to enable them to meet the health needs of their families.
    SETTING: Regional and rural areas of Southern Queensland.
    PARTICIPANTS: Parents with a child under the age of 5 years.
    DESIGN: A convergent mixed methods design was utilised. Parents participated in an online survey and were invited to in-depth semi-structured telephone interviews about their health information search methods. Inductive content analysis was applied to the transcripts.
    RESULTS: The 11 interviewees searched for health information when their child was unwell, using the internet, family and friends and GPs and medical services. Websites were used for health information, whereas social media sites provided support and connection. The internet helped determine when to seek medical advice, and a preference was shown for Australian, hospital and government websites and websites recommended by GPs.
    CONCLUSION: The results may inform the development of targeted hospital and government websites to ensure all parents have easy access to evidence-based children's health information. GPs may also play a role in discussing internet-sourced health information with parents.
    Keywords:  children; general practitioners; health information; internet; parents; rural
    DOI:  https://doi.org/10.1111/ajr.70060
  32. Bioengineering (Basel). 2025 May 02. pii: 486. [Epub ahead of print]12(5):
      This work introduces TrialSieve, a novel framework for biomedical information extraction that enhances clinical meta-analysis and drug repurposing. By extending traditional PICO (Patient, Intervention, Comparison, Outcome) methodologies, TrialSieve incorporates hierarchical, treatment group-based graphs, enabling more comprehensive and quantitative comparisons of clinical outcomes. TrialSieve was used to annotate 1609 PubMed abstracts, 170,557 annotations, and 52,638 final spans, incorporating 20 unique annotation categories that capture a diverse range of biomedical entities relevant to systematic reviews and meta-analyses. The performance (accuracy, precision, recall, F1-score) of four natural-language processing (NLP) models (BioLinkBERT, BioBERT, KRISSBERT, PubMedBERT) and the large language model (LLM), GPT-4o, was evaluated using the human-annotated TrialSieve dataset. BioLinkBERT had the best accuracy (0.875) and recall (0.679) for biomedical entity labeling, whereas PubMedBERT had the best precision (0.614) and F1-score (0.639). Error analysis showed that NLP models trained on noisy, human-annotated data can match or, in most cases, surpass human performance. This finding highlights the feasibility of fully automating biomedical information extraction, even when relying on imperfectly annotated datasets. An annotator user study (n = 39) revealed significant (p < 0.05) gains in efficiency and human annotation accuracy with the unique TrialSieve tree-based annotation approach. In summary, TrialSieve provides a foundation to improve automated biomedical information extraction for frontend clinical research.
    Keywords:  artificial intelligence; biocuration; biomedical information extraction; biomedical literature annotation; biomedical literature schema; large language model; named entity recognition; natural-language processing; text mining
    DOI:  https://doi.org/10.3390/bioengineering12050486
  33. Front Bioinform. 2025 ;5 1523184
      Basic and clinical biomedical research relies heavily on modern large-scale datasets that include genomics, transcriptomics, epigenomics, metabolomics, and proteomics, among other "Omics". These research tools very often generate lists of candidate genes that are hypothesized or shown to be responsible for the biological effect in question. To aid the biological interpretation of experimentally-obtained gene lists, we developed pubCounteR, an R-package and web-based interface that screens publications by a user-defined set of keywords representing a specific biological context for experimentally-derived gene lists.
    Keywords:  gene list; information retrieval; literature search; publication activity; text mining
    DOI:  https://doi.org/10.3389/fbinf.2025.1523184