bims-librar Biomed News
on Biomedical librarianship
Issue of 2025–07–13
fifteen papers selected by
Thomas Krichel, Open Library Society



  1. Health Info Libr J. 2025 Jul 10.
      NHS knowledge and library specialists support researchers and undertake research. The aim of this pilot programme, delivered as part of the Knowledge for Healthcare Learning Academy offer, was to equip learners from the health and social care library and knowledge sector in England with: the knowledge, skills and confidence to undertake research which adds to the evidence base for the profession; and/or equips them to contribute as a member of a healthcare research team.
    Keywords:  continuing professional development (CPD); learning; librarians, health science; libraries, health care; library and information professionals; research capacity; research skills
    DOI:  https://doi.org/10.1111/hir.12579
  2. J Cannabis Res. 2025 Jul 07. 7(1): 39
      The Medical Cannabis Library (MCL) is a curated database designed to simplify the search for cannabis-related therapeutic research. Addressing the challenge of navigating vast and mixed-quality literature, the MCL consolidates over 11,000 relevant publications from Pubmed, focusing on cannabinoids as therapeutic agents. It features an advanced search interface allowing users to find information by diseases, conditions, symptoms, syndromes, and cannabinoids. The database also categorizes the relationship between cannabinoids and medical conditions as positive, negative, or neutral using natural language processing. This tool streamlines access to cannabis research, aiding professionals in understanding its therapeutic potential across various clinical indications. The MCL represents a significant advancement in organizing and accessing scientific knowledge in the field of medical cannabis.
    DOI:  https://doi.org/10.1186/s42238-025-00295-7
  3. Nature. 2025 Jul;643(8071): 329-331
      
    Keywords:  Biodiversity; Machine learning; Medical research; Research data
    DOI:  https://doi.org/10.1038/d41586-025-02069-w
  4. Global Spine J. 2025 Jul 11. 21925682251360655
      Study DesignCross-Sectional.ObjectivesAdult spinal deformity (ASD) affects 68% of the elderly, with surgical intervention carrying complication rates of up to 50%. Effective patient education is essential for managing expectations, yet high patient volumes can limit preoperative counseling. Language learning models (LLMs), such as ChatGPT, may supplement patient education. This study evaluates ChatGPT-3.5's accuracy and readability in answering common patient questions regarding ASD surgery.MethodsStructured interviews with ASD surgery patients identified 40 common preoperative questions, of which 19 were selected. Each question was posed to ChatGPT-3.5 in separate chat sessions to ensure independent responses. Three spine surgeons assessed response accuracy using a validated 4-point scale (1 = excellent, 4 = unsatisfactory). Readability was analyzed using the Flesch-Kincaid Grade Level formula.ResultsPatient inquiries fell into four themes: (1) Preoperative preparation, (2) Recovery (pain expectations, physical therapy), (3) Lifestyle modifications, and (4) Postoperative course. Accuracy scores varies: Preoperative responses averaged 1.67, Recovery and lifestyle responses 1.33, and postoperative responses 2.0. 59.7% of responses were excellent (no clarification needed), 26.3% were satisfactory (minimal clarification needed), 12.3% required moderate clarification, and 1.8% were unsatisfactory, with one response ("Will my pain return or worsen?") rated inaccurate by all reviewers. Readability analysis showed all 19 responses exceeded the eight-grade reading level by an average of 5.91 grade levels.ConclusionChatGPT-3.5 demonstrates potential as a supplemental patient education tool but provides varying accuracy and complex readability. While it may support patient understanding, the complexity of its responses may limit usefulness for individuals with lower health literacy.
    Keywords:  ChatGPT; adult; artificial intelligence; deformity; education; patient; spinal
    DOI:  https://doi.org/10.1177/21925682251360655
  5. J Asthma. 2025 Jul 11. 1-11
       OBJECTIVE: Asthma is the most common chronic disease of childhood, characterized by symptoms such as wheezing, shortness of breath, and coughing. With the advancement of technology, artificial intelligence applications are increasingly being used in various fields, among which ChatGPT is one of the most widely utilized. The aim of this study is to evaluate the reliability, quality, and readability of the answers provided by the ChatGPT-4o application to questions related to pediatric asthma.
    METHODS: The ChatGPT-4o application was used to record answers to 25 of the most frequently asked questions about asthma in children. To determine the reliability of the answers, we used the Global Quality Scale and modified DISCERN tool. We tested readability using seven indices: Automated Readability Index, Flesch Reading Ease Score, Flesch-Kincaid Grade Level, Gunning Fog Readability Index, Simple Measure of Gobbledygook, Coleman-Liau Readability Index, and Linsear Write Formula.
    RESULTS: The answers provided by the ChatGPT-4o application to questions about childhood asthma were found to have good reliability (88% by the first evaluator and 84% by the second evaluator) and high quality (88% by both evaluators). The application scored 10.77 ± 1.58 on the Flesch-Kincaid Grade Level scale, and in conjunction with the other indices, the results indicated that the answers required a high level of reading proficiency.
    CONCLUSIONS: Artificial intelligence can be a reliable tool for parents in providing information about pediatric asthma. However, these findings suggest that readability issues may hinder the clinical application of artificial intelligence-generated content in asthma diagnosis and treatment.
    Keywords:  ChatGPT; childhood; quality; readibility; reliability
    DOI:  https://doi.org/10.1080/02770903.2025.2531500
  6. J Am Acad Orthop Surg. 2025 Jul 03.
       INTRODUCTION: Ankle sprains are among the most common injuries treated by foot and ankle specialists. Despite increasing online health information-seeking behavior, existing information on ankle sprains has been criticized by physicians for poor quality and readability. ChatGPT-4 and Google Gemini are artificial intelligence chatbots known for their accessible interfaces and versatility, making them popular for patient self-education. This study evaluated the accuracy and readability of answers to frequently asked questions (FAQs) on ankle sprains from ChatGPT-4, Google Gemini, and Google Search.
    METHODS: ChatGPT-4, Gemini, and Google Search each generated 10 FAQs and answers about ankle sprains, with questions classified based on the Rothwell Classification System and answers graded using the Flesch-Kincaid Grade Level, Flesch-Kincaid Reading Ease Score, and Gunning Fog Score. ChatGPT-4 and Gemini were then asked to answer the 10 questions generated by Google Search. Two fellowship-trained foot and ankle surgeons and a current orthopaedic foot and ankle fellow independently graded responses for accuracy and comprehensiveness using a 4-point scale.
    RESULTS: Only 10% of FAQs generated by ChatGPT-4, Gemini, and Google Search overlapped. When answering the same 10 FAQs from Google Search, ChatGPT-4 and Gemini provided significantly longer responses than Google Search, with mean differences of 353.1 ± 42.8 words (P < 0.0001) and 87 ± 35 words (P < 0.0001), respectively. ChatGPT-4's answers were more difficult to read based on the Flesch-Kincaid Grade Level (3.2 ± 1.7, P = 0.0012) and Flesch-Kincaid Reading Ease Score (22.2 ± 9.3, P < 0.0001). Surgeons ranked ChatGPT-4's answers as the most comprehensive and accurate and Google Search's answers as the least.
    CONCLUSION: This study suggests that ChatGPT-4 and Google Gemini provide longer, more complex, yet more accurate and comprehensive answers than Google Search, highlighting their potential as patient education tools for orthopaedic conditions such as ankle sprains.
    DOI:  https://doi.org/10.5435/JAAOS-D-25-00260
  7. Int J Colorectal Dis. 2025 Jul 05. 40(1): 152
       PURPOSE: Robotic ventral rectopexy (RVR) has gained acceptance as a minimally invasive approach for treating rectal prolapse and rectocele. Although numerous surgical videos have been published, their educational quality remains underexplored. This study aimed to evaluate the overall quality, adherence to reporting guidelines, and educational value of the most-viewed RVR videos on YouTube, as rated by surgical trainees, fellows, and senior surgeons.
    METHODS: The 25 most-viewed YouTube videos on RVR were selected and assessed for adherence to LAP-VEGaS and consensus reporting guidelines, overall quality, and educational value. Surgeons' performance was evaluated using the Global Evaluative Assessment of Robotic Skills (GEARS) scale. A Bayesian ordinal regression model analyzed factors influencing video quality and utility ratings.
    RESULTS: Video quality and educational value varied significantly among viewer groups. Only 40% of videos underwent peer review before publication. Adherence to reporting guidelines was low (median conformity rate: 13.5%-16.7%). Inter-rater reliability differed across GEARS domains, with senior surgeons rating video quality and utility more critically than trainees and fellows. Videos with more likes and shorter online duration were more likely to be rated as high-quality.
    CONCLUSIONS: Online surgical videos on RVR offer easily accessible but potentially unreliable educational resources and exhibit significant variability in quality and guideline adherence. Structured, peer-reviewed video-based educational programs and standardized reporting practices are crucial for improving the educational impact of online surgical videos.
    Keywords:  Educational videos; Robotic rectopexy; Surgical videos; Ventral rectopexy; YouTube
    DOI:  https://doi.org/10.1007/s00384-025-04936-4
  8. JMIR Cardio. 2025 Jul 08. 9 e68817
       Background: Heart failure management involves comprehensive lifestyle modifications such as daily weights, fluid and sodium restriction, and blood pressure monitoring, placing additional responsibility on patients and caregivers, with successful adherence often requiring extensive counseling and understandable patient education materials (PEMs). Prior research has shown PEMs related to cardiovascular disease often exceed the American Medical Association's fifth- to sixth-grade recommended reading level. The large language model (LLM) ChatGPT may be a useful tool for improving PEM readability.
    Objective: We aim to assess the readability of heart failure-related PEMs from prominent cardiology institutions and evaluate GPT-4's ability to improve these metrics while maintaining accuracy and comprehensiveness.
    Methods: A total of 143 heart failure-related PEMs were collected from the websites of the top 10 institutions listed on the 2022-2023 US News & World Report for "Best Hospitals for Cardiology, Heart & Vascular Surgery." PEMs were individually entered into GPT-4 (version updated July 20, 2023), preceded by the prompt, "Please explain the following in simpler terms." Readability was assessed using the Flesch Reading Ease score, Flesch-Kincaid Grade Level (FKGL), Gunning Fog Index, Coleman-Liau Index, Simple Measure of Gobbledygook Index, and Automated Readability Index. The accuracy and comprehensiveness of revised GPT-4 PEMs were assessed by a board-certified cardiologist.
    Results: For 143 institutional heart failure-related PEMs analyzed, the median FKGL was 10.3 (IQR 7.9-13.1; high school sophomore) compared to 7.3 (IQR 6.1-8.5; seventh grade) for GPT-4's revised PEMs (P<.001). Of the 143 institutional PEMs, there were 13 (9.1%) below the sixth-grade reading level, which improved to 33 (23.1%) after revision by GPT-4 (P<.001). No revised GPT-4 PEMs were graded as less accurate or less comprehensive compared to institutional PEMs. A total of 33 (23.1%) GPT-4 PEMs were graded as more comprehensive.
    Conclusions: GPT-4 significantly improved the readability of institutional heart failure-related PEMs. The model may be a promising adjunct resource in addition to care provided by a licensed health care professional for patients living with heart failure. Further rigorous testing and validation is needed to investigate its safety, efficacy, and impact on patient health literacy.
    Keywords:  ChatGPT; GPT-4; artificial intelligence; health literacy; heart failure; large language models; patient education; readability
    DOI:  https://doi.org/10.2196/68817
  9. Cureus. 2025 Jun;17(6): e85595
      Background Haemorrhoids, anal fissures, and anal fistulae are common benign proctological conditions that heavily rely on self-management strategies to prevent morbidity. Online access to various treatment options has empowered patients in this regard. This study was conducted to assess the quality and readability of the available online information. Methods An online search using the Google search engine (Google, Inc., Mountain View, CA, USA) was carried out with the following terms: 'Treatment of Haemorrhoids', 'Treatment of Anal Fissure', and 'Treatment of Anal Fistula'. For each search term, the first 25 webpages developed for patient education were included. Thus, a total of 75 webpages were analysed for their quality using the DISCERN instrument and the Journal of the American Medical Association (JAMA) benchmarks. Their readability was assessed using the Flesch Reading Ease Score (FRES), Flesch-Kincaid Grade Level (FKGL), and the Simple Measure of Gobbledygook (SMOG). Results Across the three search terms, the average overall DISCERN score was 2.4 ± 0.8 (out of five), suggesting moderate to low-quality information. Only 16 (21%) webpages fulfilled all four JAMA benchmark criteria for quality. An average FRES score of 57.6 ± 9.0 indicated that the text was fairly difficult to read. FKGL and SMOG index levels of 8.0 ± 1.6 and 10.7 ± 1.0, respectively, correspond to the reading age range of 13- to 16-year-olds. Conclusion Available online information for patients with common proctological complaints is of suboptimal quality and is fairly difficult to read. It is essential to ensure that the information available meets high-quality standards and is readable by patients.
    Keywords:  anal fissure; anal fistula; haemorrhoids; proctology; readability analysis
    DOI:  https://doi.org/10.7759/cureus.85595
  10. Reprod Biomed Online. 2025 Mar 10. pii: S1472-6483(25)00114-2. [Epub ahead of print]51(2): 104907
       RESEARCH QUESTION: What is the tone of donor-conceived people (DCP) towards their donor-conceived status in YouTube videos assessed for sentiment analysis via ChatGPT, a natural language-processing tool?
    DESIGN: A cross-sectional analysis of YouTube videos using the keyword 'donor conceived' was performed by two independent reviewers. Videos were selected for inclusion if they described DCPs' opinions, stories or emotions about their life experiences. Reviewers categorized the sentiments of the videos as neutral, positive, negative or mixed. The auto-generated English-language YouTube transcript for each video was then inputted into ChatGPT (GPT-4o model) using the following prompt: 'Given this text, what is the sentiment conveyed? Is it positive or negative? Text: {transcript}'. Sentiment results generated via ChatGPT were categorized as neutral, positive, negative or mixed. Chi-squared goodness-of-fit and independence tests were performed to assess the observed proportions of sentiments compared with the expected distribution.
    RESULTS: Fifty YouTube videos were identified. The median video length was 6.91 min and the median number of views was 2217.5. Via ChatGPT, 44% of videos were determined to have a positive sentiment, 28% were negative, 24% were mixed and 4% were neutral, with a significant difference in the observed distribution of sentiments from expected (Χ2 [3, n = 50] = 16.240, P = 0.001). When compared with two human reviewers, ChatGPT sentiment results had 90% concordance with at least one reviewer.
    CONCLUSION: ChatGPT shows a strong performance in analysing YouTube video transcript sentiments when compared with human reviewers. The sentiments of DCP towards their donor-conceived status in YouTube videos have a positive leaning when analysed by ChatGPT.
    Keywords:  Artificial intelligence; Assisted reproductive technology; Donor conception; Sentiment analysis
    DOI:  https://doi.org/10.1016/j.rbmo.2025.104907
  11. Surg Endosc. 2025 Jul 07.
       BACKGROUND: The "10 Critical Steps of Hiatal Hernia Repair and Anti-Reflux Surgery" provides a framework for the standardization of benign foregut operations. This study aimed to determine if currently available educational videos on YouTube adequately detail all 10 of these critical steps.
    METHODS: YouTube videos that were published between January 1, 2019 and July 1, 2024 were identified using Google search methods. Only English language videos detailing elective, benign foregut surgery in adults were included. Two reviewer pairs independently assessed the videos for both the demonstration and description of the 10 critical steps. The publication date, number of video views, and shown operative video length were also captured. Data were analyzed using Chi-squared tests.
    RESULTS: A total of 52 videos were identified; 39 (75%) videos met inclusion criteria. Adherence to expert-recommended steps was inconsistent. Only 3 (8%) videos described and/or demonstrated all 10 critical steps. Steps 7 (bougie placement) and 9 (crural repair/use of mesh) were most often neither described nor shown (77% and 56%, respectively) while steps 3 (dissection of crura), 6 (intra-abdominal esophageal length), and 10 (antirefllux surgery/fundoplication) were the most frequently described and shown (62%, 62%, and 69%, respectively). Step 8 (gastroesophageal junction assessment/Collis gastroplasty) was often shown without description (64%). There was a weak positive correlation between operative video length and number of steps that were both described and shown (r = 0.24) and a moderate negative correlation between operative video length and videos that omitted at least one step (r = - 0.4).
    CONCLUSION: The results of our study highlight the need for more comprehensive hiatal hernia repair and anti-reflux surgery educational videos. We recommend updating currently available educational materials to be consistent with expert guidelines.
    Keywords:  Hiatal hernia repair; Minimally invasive surgery; Surgical education; Surgical techniques; Video-based learning; YouTube
    DOI:  https://doi.org/10.1007/s00464-025-11960-6
  12. Aesthetic Plast Surg. 2025 Jul 07.
       BACKGROUND: As the digital age progresses, individuals are increasingly seeking online websites for the dissemination of information. YouTube, the most popular free video streaming platform worldwide, garners over 2 billion users. Plastic surgery information is widely available on YouTube and acts as a common interface for the distribution of plastic surgery-related information. This investigation aimed to identify trends in YouTube videos describing common plastic surgery search terms based on narrator gender, channel type, and language patterns.
    METHODS: Top-ranking videos for six common plastic surgery search terms were queried. Video characteristics and transcripts for each video were extracted. Transcripts were subsequently analyzed using the Linguistic Inquiry and Word Count software. Comparisons were performed using various statistical analysis methods.
    RESULTS: Significant differences in audience engagement were noted by search topic (p < 0.05). Speakers in plastic surgery videos were predominantly male only (108 [60%]), with notable variation observed across channel type and search topic. Significant differences were found for videos on medical professional channels (p = 0.004) and videos regarding gynecomastia (p < 0.05). Linguistic pattern analysis revealed that female speakers used more authentic (p = 0.01) and emotional (p = 0.012) language than male speakers.
    CONCLUSION: With YouTube being the most visited online video platform website worldwide, plastic surgeons must understand trends to appropriately tailor their content and maximize their social media reach. Also, while plastic surgery is making strides in the right direction, continued work should be done in promoting adequate representation among online material.
    NO LEVEL ASSIGNED: This journal requires that authors assign a level of evidence to each submission to which Evidence-Based Medicine rankings are applicable. This excludes Review Articles, Book Reviews, and manuscripts that concern Basic Science, Animal Studies, Cadaver Studies, and Experimental Studies. For a full description of these Evidence-Based Medicine ratings, please refer to the Table of Contents or the online Instructions to Authors www.springer.com/00266 .
    Keywords:  Plastic surgery; Social media; Women in surgery; YouTube content
    DOI:  https://doi.org/10.1007/s00266-025-04851-7
  13. PeerJ. 2025 ;13 e19653
       Background: Gingival recession is a mucogingival problem that can cause esthetic concerns and sensitivity. While some patients remain unaware of it, others seek solutions through various means. The COVID-19 pandemic has increased the role of social media in accessing health information. The frequent use of YouTube has prompted researchers to evaluate its contents quality and reliability. This study aims to examine this issue by comparing two periods: before and after COVID-19.
    Methods: The first 100 videos for the keyword 'receding gums' shared in 2019 and 2024 were analyzed. A total of 23 and 35 videos were included, respectively. Quality was evaluated using Video Quality Index (VIQI) and Global Quality Scale (GQS), reliability with DISCERN and Modified DISCERN and content through an approach developed by the authors. Videos were also assessed based on their characteristics and sources. Data were analyzed using IBM SPSS 26.0. For two-group comparisons, independent sample t-tests were applied to normally distributed data, while the Mann-Whitney U test was used for non-normally distributed data. Categorical variables were analyzed with the Chi-square test. Correlations between variables were examined using Spearman's correlation analysis.
    Results: According to content analysis, useful videos were found to have higher scores of quality and reliability indices in both the 2019 and 2024 groups (p < 0.001). In 2024, the amount of videos uploaded by dental professionals was significantly more 'useful' compared to those from other sources (p = 0.45). There was a tendency of increase in sharing videos by healthcare professionals following the COVID-19 pandemic.
    Keywords:  COVID-19; Gingival recession; Gum recession; YouTube
    DOI:  https://doi.org/10.7717/peerj.19653
  14. Surg Endosc. 2025 Jul 09.
       BACKGROUND: Social media platforms have become a leading arena for disseminating health knowledge. Information on weight loss is no exception; health, fitness and nutrition are within the 5 most searched categories in 2023. Our study investigates the quality of information on TikTok and Instagram regarding the two most popular weight loss treatments: glucagon-like-peptide-1 (GLP-1) agonist and sleeve gastrectomy.
    METHODS: TikTok and Instagram were queried for contents about GLP-1 agonists or sleeve gastrectomy. The first 50 videos from each category were evaluated by two reviewers. The videos were divided into four categories: Instagram GLP-1, Instagram weight loss surgery (WLS), TikTok GLP-1, and TikTok WLS. The quality of the contents was assessed using the DISCERN score.
    RESULTS: 200 videos were collected for analysis. 31.5% were created by medical professionals and 68.5% were from non-medical professionals. More videos from Instagram regarding GLP-1 agonists followed by videos from TikTok regarding GLP-1 agonists were made by medical professionals compared to videos about surgery (p = 0.038). The average number of likes overall was 67,024 and saves or shares 6,305. The greatest number of likes were in the TikTok surgery group (p = 0.004) and the most saves or shares were in the Instagram surgery group (p < 0.001). DISCERN scores were low for all four categories, overall 28.0 out of 80. They were highest in the Instagram GLP-1 agonist group (p < 0.001).
    CONCLUSIONS: Despite the popularity of TikTok and Instagram, the quality of the health contents, as measured by DISCERN score, on these platforms remains poor. Most are created by non-medical professionals. Healthcare professionals should encourage patients to remain vigilant and educate them on misinformation on social media. This may also represent an opportunity for healthcare providers to generate their own social media content to ensure reliability of information as well as to direct patients to reliable social media content.
    Keywords:  Bariatric surgery; GLP-1; Obesity; Sleeve gastrectomy; Social media
    DOI:  https://doi.org/10.1007/s00464-025-11938-4