bims-librar Biomed News
on Biomedical librarianship
Issue of 2025–04–27
24 papers selected by
Thomas Krichel, Open Library Society



  1. F1000Res. 2024 ;13 1134
       Background: Health technology assessment (HTA) reports are based on comprehensive information retrieval. Current standards discourage the use of search restrictions, such as publication date and language. Given limited resources, it was unclear whether the effort invested in screening and translating studies published in languages other than English provided relevant additional information compared with the inclusion of English-language publications alone. We therefore analysed the impact of non-English publications on the conclusions of HTA reports produced by the German HTA agency, the Institute for Quality and Efficiency in Health Care (IQWiG).
    Methods: We determined whether non-English publications were included in all German HTA reports on non-drug interventions (published by IQWiG between 06/2007 to 08/2018) and on selected drug interventions. If at least one non-English publication was included, we assessed for each endpoint whether or not the exclusion of non-English publications changed the conclusion. If a non-English publication did not contain information relevant to the HTA report, we classified the publication as "not relevant".
    Results: Of 70 HTA reports, 38 (54%) included 126 non-English publications. In 4 reports (6%) with 50 endpoints investigated in 39 PICO questions, the exclusion of a total of 10 non-English publications led to a change in the conclusions for 13 endpoints (8 PICO questions). This was largely due to the fact that in many cases, non-English publications were the predominant or only literature available, resulting in a lack of analysable data after their exclusion.
    Conclusions: In general, studies only published in non-English languages have little influence on the conclusions of German HTA reports. For the vast majority of topics, a language restriction to English seems justified. Studies published in non-English languages may be useful in exceptional cases, for example when an intervention is only available in certain countries.
    Keywords:  Language; Publication Bias; Publications; Retrospective Studies
    DOI:  https://doi.org/10.12688/f1000research.151365.2
  2. Syst Rev. 2025 Apr 22. 14(1): 92
      The rise of powerful search engines (e.g., Google) make the searching for gray literature more feasible within the time and resources of a typical systematic review. However, there are no hypothesis-testing studies to guide us on how to conduct such a search. It is our belief that the "best practices" for incorporating Google searches might come from the collection of experiential evidence that users have had, from which can be drawn some tentative conclusions. It is our intention with this communication to relay our experience with Google searches for five projects and the lessons we think we have learned. We invite our systematic review colleagues to contribute their own experiences and thus to build up the experiential evidence about when and how to use Google as a search engine to supplement traditional computerized database searches.
    DOI:  https://doi.org/10.1186/s13643-025-02836-w
  3. Health Promot Int. 2025 Mar 05. pii: daaf044. [Epub ahead of print]40(2):
      People with chronic health conditions tend to look for information from a range of sources. Although information online is accessible, the quality varies and articles in scientific journals are not written for a general audience. Given this, plain language summaries (PLSs) may be a useful source of information, especially for people with chronic medical conditions. PLSs contain reliable information that is written in plain, easy-to-understand language. We investigated the health information-seeking behaviour of consumers with chronic health conditions and sought to understand what they see as important in a PLS with respect to content, design and structure, and the labels used. We conducted semi-structured focus groups and interviews with 19 participants from 6 countries, all with a chronic medical condition. Using reflexive thematic analysis, we developed four themes: (i) 'Accessing information in a competitive landscape of health information'; (ii) 'I really don't see the patient in this at all'; (iii) 'Co-design should be meaningful, not tokenistic'; and (iv) 'A way forward: Approach PLSs with creativity'. Participants highlighted the importance of PLS labels, the need for actionable content, and the recommendation to vary the reading level depending on audience needs. The results of this study suggest that PLSs produced with consumer input could elicit PLSs that better meet audience needs, focusing on information that is actionable, accessible, and written with useful content that is detailed but still respectful of the reader. Consequently, PLSs could become a more useful source of reliable information, particularly for people with chronic health conditions.
    Keywords:  consumer; health communication; health literacy; patients; plain language summaries; readability
    DOI:  https://doi.org/10.1093/heapro/daaf044
  4. Cureus. 2025 Mar;17(3): e81156
      Introduction Patient information guides are required, as they help to identify early diseases and complications and to prevent them. Artificial intelligence (AI) tools are being used to create patient education guides for easy accessibility to people. Methodology This was a cross-sectional study analyzing patient information guides regarding six anti-diabetic drugs created using ChatGPT (OpenAI, Inc., San Francisco, California, United States) and Gemini (Google LLC, Mountain View, California, United States), respectively. The patient information guides were on anti-diabetic drugs, including metformin, empagliflozin, liraglutide, glipizide, sitagliptin, and insulin glargine, and were created with the help of various prompts. Results There was no statistically significant difference found between any of the characteristics of the responses generated by the two AI tools, according to the P values obtained. Conclusion On comparing the two AI tools, there was not much difference noted in readability, reliability, and similarity.
    Keywords:  artificial intelligence; chatgpt; google gemini; insulin; metformin
    DOI:  https://doi.org/10.7759/cureus.81156
  5. Eur Geriatr Med. 2025 Apr 21.
       PURPOSE: ChatGPT, a comprehensive language processing model, provides the opportunity for supportive and professional interactions with patients. However, its use to address patients' frequently asked questions (FAQs) and the readability of the text generated by ChatGPT remain unexplored, particularly in geriatrics. We identified the FAQs about common geriatric syndromes and assessed the accuracy and readability of the responses provided by ChatGPT.
    METHODS: Two geriatricians with extensive knowledge and experience in geriatric syndromes independently reviewed the 28 responses provided by ChatGPT. The accuracy of the responses generated by ChatGPT was categorized on a rating scale from 0 (harmful) to 4 (excellent) based on current guidelines and approaches. The readability of the text generated by ChatGPT was assessed by administering two tests: the Flesch-Kincaid Reading Ease (FKRE) and the Flesch-Kincaid Grade Level (FKGL).
    RESULTS: ChatGPT-generated responses with an overall mean accuracy score of 88% (3.52/4). Responses generated for sarcopenia diagnosis and depression treatment in older adults had the lowest accuracy scores (2.0 and 2.5, respectively). The mean FKRE score of the texts was 25.2, while the mean FKGL score was 14.5.
    CONCLUSION: The accuracy scores of the responses generated by ChatGPT were high in most common geriatric syndromes except for sarcopenia diagnosis and depression treatment. Moreover, the text generated by ChatGPT was very difficult to read and best understood by college graduates. ChatGPT may reduce the uncertainty many patients face. Nevertheless, it remains advisable to consult with subject matter experts when undertaking consequential decision-making.
    Keywords:  Artificial intelligence; ChatGPT; Geriatric syndromes; Older adults
    DOI:  https://doi.org/10.1007/s41999-025-01202-2
  6. PEC Innov. 2025 Jun;6 100390
       Aim: This study evaluated the understandability, actionability, and readability of text on anemia generated by artificial intelligence (AI) chatbots.
    Methods: This cross-sectional study compared texts generated by ChatGPT-3.5, Microsoft Copilot, and Google Gemini at three levels: "normal," "6th grade," and "PEMAT-P version." Additionally, texts retrieved from the top eight Google Search results for relevant keywords were included for comparison. All texts were written in Japanese. The Japanese version of the PEMAT-P was used to assess understandability and actionability, while jReadability was used for readability. A systematic comparison was conducted to identify the strengths and weaknesses of each source.
    Results: Texts generated by Gemini at the 6th-grade level (n = 26, 86.7 %) and PEMAT-P version (n = 27, 90.0 %), as well as ChatGPT-3.5 at the normal level (n = 21, 80.8 %), achieved significantly higher scores (≥70 %) for understandability and actionability compared to Google Search results (n = 17, 25.4 %, p < 0.001). For readability, Copilot and Gemini texts demonstrated significantly higher percentages of "very readable" to "somewhat difficult" levels than texts retrieved from Google Search (p = 0.000-0.007).
    Innovation: This study is the first to objectively and quantitatively evaluate the understandability and actionability of educational materials on anemia prevention. By utilizing PEMAT-P and jReadability, the study demonstrated the superiority of Gemini in terms of understandability and readability through measurable data. This innovative approach highlights the potential of AI chatbots as a novel method for providing public health information and addressing health disparities.
    Conclusion: AI-generated texts on anemia were found to be more readable and easier to understand than traditional web-based texts, with Gemini demonstrating the highest level of understandability. Moving forward, improvements in prompts will be necessary to enhance the integration of visual elements that encourage actionable responses in AI chatbots.
    Keywords:  Anemia; ChatGPT; Copilot; Gemini; Patient education
    DOI:  https://doi.org/10.1016/j.pecinn.2025.100390
  7. Int J Impot Res. 2025 Apr 22.
      Failure to achieve spontaneous pregnancy within 12 months despite unprotected intercourse is called infertility. The rapid development of digital health data has led more people to search for healthcare-related topics on the Internet. Many infertile individuals and couples use the Internet as their primary source for information on infertility diagnosis and treatment. However, it is important to assess the readability, understandability, and actionability of the information provided by these sources for patients. There is a gap in the literature addressing this aspect. This study aims to compare the readability, understandability, and actionability of responses generated by Microsoft Copilot (MC), an AI chatbot, and Google Search (GS), an internet search engine, for infertility-related queries. Prospectively a Google Trends analysis was conducted to identify the top 20 queries related to infertility in February, 2024. Then these queries were entered into GS and MC in May 2024. Answers from both platforms were recorded for further analysis. Outputs were assessed using automated readability tools, and readability scores were calculated. Understandability and actionability of answers were evaluated using the Patient Education Materials Assessment Tool for Printable Materials (PEMAT-P) tool. GS was found to have significantly higher Automated Readability Index (ARI) and Flesch-Kincaid Grade Level (FKGL) scores than MC (p = 0.044), while no significant differences were observed in the Flesch Reading Ease, Gunning Fog Index, Simplified Measure of Gobbledygook (SMOG), and Coleman-Liau Index scores. Both GS and MC outputs had readability scores above the 8th-grade level, indicating advanced reading levels. According to PEMAT-P, MC outperformed GS in terms of understandability (68.65 ± 11.99 vs. 54.50 ± 15.09, p = 0.001) and actionability (29.85 ± 17.8 vs. 1 ± 4.47, p = 0.000). MC provides more understandable and actionable responses to infertility related queries, that it might have great potential for patient education.
    DOI:  https://doi.org/10.1038/s41443-025-01056-z
  8. Ann Plast Surg. 2025 May 01. 94(5): 500-506
       BACKGROUND: Patients access online resources for information about various treatment modalities, including gender-affirming surgery (GAS) patient populations. The present study provides a readability, content quality, technical quality, and qualitative characteristic assessment of online patient educational materials (OPEM) related to GAS.
    METHODS: An online search using search phrases related to GAS was performed. The first 20 unique search results for each search term were further assessed if identified as an OPEM. The readability, content quality, and technical quality were assessed. Characteristics including mention of World Professional Association for Transgender Health (WPATH) guidelines and discussion of insurance coverage were recorded. Statistical analyses included analysis of variance, chi-squared tests, and 2-sample t tests.
    RESULTS: Overall, 231 OPEM were identified (academic/hospital: 124, private practice: 47, online health reference: 43, other: 17). The average consensus grade level was 12.7, which was significantly higher than the National Institutes of Health/American Medical Association-recommended sixth-grade level (P < 0.001). With regard to content quality, academic/hospital websites had a significantly higher DISCERN score than private practice websites (P = 0.01). Notably, only 35% of OPEM mentioned WPATH guidelines. Academic/hospital websites mentioned WPATH guidelines at a significantly higher frequency than nonacademic/hospital websites (P < 0.001). Private practice websites had a significantly higher technical quality score than academic/hospital websites (P < 0.001).
    CONCLUSIONS: A significant discrepancy in readability was identified between analyzed websites and the National Institutes of Health/American Medical Association recommendation of a sixth-grade level. Further efforts are necessary to ensure accessible and accurate information is available for patients seeking OPEM related to GAS.
    Keywords:  gender-affirming surgery; health literacy; patient resources; readability
    DOI:  https://doi.org/10.1097/SAP.0000000000004238
  9. Cureus. 2025 Mar;17(3): e81163
      Social media has changed the digital health landscape, altering how the public consumes medical information and sparking concerns about the reliability of medical information. This study aims to assess the quality of online health information related to bullous pemphigoid by evaluating content type, information reliability, and user engagement across different social media platforms, using the Quality Evaluation Scoring Tool (QUEST) and Discerning the Quality of Written Consumer Health Information (DISCERN) questionnaires, quantitative tools designed to measure health information quality. Among the six social media platforms analyzed, educational content was most prevalent on YouTube (Google LLC, Mountain View, California, United States) (90.0%, n=10) and Threads (Meta Platforms, Inc., Menlo Park, California, United States) (100.0%, n=3), while TikTok (ByteDance Ltd., Beijing, China) (n=20) featured 60.0% educational and 30.0% patient experience content. X (X Corp., Bastrop, Texas, United States) (n=18) demonstrated the highest reliability, with a mean QUEST score of 16.8 (SD: 4.5) and a mean DISCERN score of 28.9 (SD: 6.7). In contrast, TikTok had the lowest reliability, with a mean QUEST score of 9.0 (SD: 3.0) and a mean DISCERN score of 20.1 (SD: 3.0). Dunn's test with Bonferroni correction revealed that TikTok scored significantly lower than X in both QUEST (Mean Difference: -7.83, p < 0.0001) and DISCERN scores (Mean Difference: -8.83, p <0.001). Instagram (n=9) outperformed TikTok in QUEST scores (Mean Difference: 6.66, p <0.001. These findings indicate that X offers the most reliable health information, whereas TikTok consistently ranks lower in reliability.
    Keywords:  bullous pemphigoid; facebook; instagram; online medical information; social media; thread; tiktok; twitter; x; youtube
    DOI:  https://doi.org/10.7759/cureus.81163
  10. World J Urol. 2025 Apr 21. 43(1): 237
       INTRODUCTION: The internet is a primary source of health information, our study addresses the need to assess the credibility, readability, and reliability of Spanish-language medical websites concerning nephrolithiasis. With the Spanish-speaking population in the United States rapidly approaching 20% and growing, our research endeavors to fill a critical gap by evaluating the quality of online resources in Spanish for this common medical condition.
    METHODS: In our study, we evaluated the quality of Spanish-language online resources on nephrolithiasis for the growing Spanish-speaking population in the United States. Using Google searches with terms like "cálculos renales" and "nefrolitiasis," we analyzed 80 unique websites using Fernandez-Huerta readability software, HON-code certification, JAMA benchmark criteria, source classification, and brief DISCERN. The evaluation was performed by an expert urologist and a medical student, with discrepancies resolved by a third experienced reviewer.
    RESULTS: Interrater reliability was strong for JAMA (ICC: 0.98) and DISCERN (ICC: 0.85). Websites mainly originated from Physician/Community Hospitals (43.75%) and University/Academic institutions (23.75%). Only 32.5% had HON-code certification, and 17.5% met all JAMA criteria. Readability typically matched an 8th to 9th grade level, but 58.75% scored below the good quality threshold on DISCERN, with University/Academic sources performing significantly better than Physician/Community sources (p = 0.0078). No significant differences were noted in credibility or readability between these groups.
    CONCLUSIONS: Our findings indicate that Spanish-language resources on nephrolithiasis often lack in quality, readability, and credibility, highlighting the need for enhanced online materials to support patient and provider management of kidney stones. University sources provide notably higher-quality information than other resource types.
    Keywords:  Healthcare Disparities; Patient-education; Urolithiasis
    DOI:  https://doi.org/10.1007/s00345-025-05588-3
  11. Cranio. 2025 Apr 21. 1-11
       OBJECTIVE: This systematic review aims to assess the quality, reliability, and content of TMD-related social media posts across multiple platforms.
    METHODS: A comprehensive search of three databases were conducted. TMD related content on social media sites were considered eligible. The extracted data was narratively synthesized.
    RESULTS: Out of 254 studies, seven were included. YouTube was the most commonly assessed social media platform (n=5). Information created by amateurs encourages greater interaction than that of experts or news outlets. A majority videos and threads cover treatment choices. However, there is a noticeable absence of discussion on complications and prevention methods. Health professionals' videos scored better in quality and reliability than -professionals' videos.
    CONCLUSION: There is a need for creating evidence-based, accessible, and engaging content on TMD on social media and consideration of implementing rigorous information verification methods to ensure that high-quality, expert-validated content is prioritized in search results.
    Keywords:  Temporomandibular joint disorders; YouTube; social media; systematic review
    DOI:  https://doi.org/10.1080/08869634.2025.2482271
  12. Sex Med. 2025 Apr;13(2): qfaf023
       Background: Patients dealing with sensitive issues like penile enlargement (PE) might benefit from YouTube videos. Therefore, it is essential that the textual content of these videos is clear, trustworthy, and of high quality.
    Aim: Are the AI-assisted acquired texts' qualities and comprehensibilities of YouTube videos about PE enough and suitable for the patients?
    Methods: On October 25, 2024, Google Trends analysis identified the 25 most searched phrases for "Penile enlargement." Non-related terms were excluded, and the relevant keywords were searched on YouTube. Only content about PE included; excluding duplicates, non-English videos, YouTube shorts, those under 30 seconds, silent, and music-only videos. Videos were transcribed using Whisper AI, and their quality was assessed by M.F.Ş, E.C.T., and Ç.D. using the GQS (global quality scale) and DISCERN, the readability was evaluated via Flesch-Kincaid (FKGL and FKRE) measures. High assessor agreement was noted (Pearson r = 0.912). Videos were categorized by uploader, and metrics such as views, likes, comments, and duration were recorded. The Chi-square test was used for categorical variable comparisons; the Kruskal-Wallis H-Test was applied when normality and homoscedasticity were not met, with Bonferroni post hoc correction for multiple comparisons.
    Outcomes: The mean DISCERN and GQS scores were 51.23 ± 13.1 and 3.32 ± 0.9, respectively. FKRE and FKGL scores were 73.12 ± 11.7 and 5.85 ± 2.1. Physicians (n = 67) produced the most videos, while academic institutions (n = 2) produced the least. No significant differences in text quality were found between groups (P = 0.067 and P = 0.051). Health-related websites exhibited lower FKRE compared to non-healthcare videos (P = 0.002), with a significant difference in FKGL as well (P = 0.019).
    Results: The video exhibited a high level of readability (indicating comprehensibility for almost a 6th-grade student). Text quality, view and like count of the videos uploaded by academic institutions was the highest.
    Clinical Implications: In PE, YouTube video's health information needs to be better quality and more trustworthy, according to our research. The language used in videos should be easier to understand.
    Strengths and Limitations: This study is the first scientific analysis of YouTube video transcripts on PE using AI, focusing specifically on English content, which limits its applicability to non-English speakers and other platforms. Exclusions of silent and shorter videos may result in the omission of valuable information.
    Conclusion: The need for better quality and trustworthiness in health-related YouTube information, especially for PE is essential. Content makers should stress clear, accessible language and minimize disinformation.
    Keywords:  YouTube; artificial intelligence; penile enlargement; transcription; whisperAI
    DOI:  https://doi.org/10.1093/sexmed/qfaf023
  13. Front Public Health. 2025 ;13 1542278
       Background: Short video social media platforms play a crucial role in public health by effectively disseminating health information. Despite this, many educational videos on dry eye care have not received sufficient attention. This study aimed to conduct a comprehensive analysis and evaluate the quality of educational short videos on dry eye care available on TikTok.
    Methods: On August 30, 2024, the top 200 videos related to dry eye were viewed from the Chinese version of TikTok using the platform's default ranking. The overall quality, reliability, comprehensibility, and applicability of the videos are systematically evaluated using the DISCERN and PEMAT-A/V assessment tools.
    Results: A total of 199 videos were included in the study and categorized based on account information: medical professional individual users, general professional individual users, for-profit organizations, non-profit organizations, and news organizations. Medical professionals were the predominant uploaders, contributing 81% of the videos. The overall misinformation rate was 2%. A majority of the videos (85.9%) addressed at least two aspects of dry eye, while only 14.1% covered three or more topics. The videos scored 22.4 ± 6.4 for reliability and 17.4 ± 6.2 for treatment options. Upon evaluation, the understandability and actionability of these videos were found to be 79.1% and 60.4%, respectively.
    Conclusion: TikTok holds significant potential for disseminating health information, primarily through content created by medical professionals. Currently, much of the content focuses on the symptoms and management of dry eye, with limited discussion on its definition, classification, and diagnosis. While most video content is reliable, there is a risk of incomplete or inaccurate information, these videos can serve as a reference. Therefore, the public should exercise caution when seeking information on dry eye through TikTok and individuals experiencing symptoms are advised to consult healthcare professionals for accurate diagnosis and treatment.
    Keywords:  care; dry eye; education; quality; short video; social media
    DOI:  https://doi.org/10.3389/fpubh.2025.1542278
  14. Oral Maxillofac Surg. 2025 Apr 22. 29(1): 91
       PURPOSE: Internet platforms, such as YouTube, play a significant role in disseminating health information. This study assessed the quality and reliability of YouTube videos related to ameloblastoma.
    METHODS: This cross-sectional study examined videos using the Global Quality Scale (GQS) to assess content quality and the DISCERN tool to determine reliability. The creators' profiles and target audiences were also evaluated. Data were analyzed descriptively and analytically.
    RESULTS: A total of 65 videos were analyzed. Most videos (84.6%) were educational, while 15.4% were case reports. Specialists produced 40% of the videos, general dentists 21.5%, and students 6.2%. Regarding content quality, 23.2% of the videos received the highest GQS score, while 27.6% were rated as low quality (scores 1-2). According to DISCERN, 1.5% of videos were classified as maximum quality, 21.6% as high quality, and 17% as low quality. A significant difference in GQS scores was observed between oral and maxillofacial surgeons and general dentists (p = 0.025). A strong positive correlation was found between GQS and DISCERN scores (r = 0.79).
    CONCLUSION: While most videos exhibited good production quality, their overall reliability was moderate, with notable deficiencies. Additional validation strategies are needed to enhance the accuracy and educational value of online health information.
    CLINICAL TRIAL: Not applicable.
    Keywords:  Ameloblastoma; Internet; Odontogenic tumor; Social media; Video-audio media
    DOI:  https://doi.org/10.1007/s10006-025-01381-z
  15. PeerJ. 2025 ;13 e19280
       Purpose: This study aimed to assess the quality, reliability, content, and usefulness of YouTube videos related to anesthesia in geriatric and elderly patients.
    Methods: Using Google Trends, the most popular search terms in the past five years, "geriatric anesthesia" and "anesthesia in the elderly," were identified. A search on YouTube using these terms initially yielded 200 of the most viewed videos for each term. After applying exclusion criteria, 87 videos were included for detailed analysis. Major themes and topics related to anesthesia in geriatric and elderly patients were identified using a pre-determined qualitative thematic analysis method. The usefulness of the videos was assessed using the specially developed Geriatric and Elderly Anesthesia Usefulness Score (GAEUS). The overall quality and reliability of the videos were evaluated using the Global Quality Scale (GQS) and the Modified DISCERN Scale (M-DISCERN), respectively. The average of the quality, reliability, and usefulness scores calculated by the researchers was used for consistency analysis.
    Results: In our study, 48.3% (42) of the videos on geriatric and elderly patients concerning anesthesia on YouTube were created by personal blogs. The quality of the videos was measured using the GQS, with a mean score of 3.34 and a median of 3 (range: 1-5), showing no significant difference according to the video source (p = 0.166). Reliability was assessed using the M-DISCERN scale, with a mean score of 3.37 and a median of 3.50 (range: 1-5), again showing no significant difference according to the video source (p = 0.097). Usefulness was measured using the GAEUS score, with a mean score of 15.30 and a median of 12.5 (range: 2-63), which showed a significant difference according to the video source (p = 0.000). The average duration of videos with low usefulness was 31.59 minutes (range: 5-44), while the average duration of moderately and highly useful videos was 59.37 minutes (range: 19.44-119.05). This duration difference was statistically significant (u = 2.569, p = 0.010).
    Conclusion: In our study, we examined YouTube videos covering anesthesia topics for geriatric and elderly patients. The highest usefulness scores were obtained from personal blogs; however, all sources generally showed low usefulness. The quality of the videos was assessed using the GQS, and their reliability was evaluated with the M-DISCERN. On both scales, the videos showed moderate performance across all sources. These findings indicate a need for more comprehensive and informative content on YouTube, especially for the education of healthcare professionals and patients. To better address the needs of elderly patients, the richness of content and educational value of these videos should be enhanced.
    Keywords:  Anesthesia; Content quality; Elderly; Geriatrics; Health education; Perioperative care; Reliability; Usefulness; Video analysis; YouTube
    DOI:  https://doi.org/10.7717/peerj.19280
  16. Clin Exp Ophthalmol. 2025 Apr 24.
      
    Keywords:  YouTube health information; medical video quality assessment; optic neuritis education
    DOI:  https://doi.org/10.1111/ceo.14548
  17. Clin Exp Optom. 2025 Apr 21. 1-7
       CLINICAL RELEVANCE: YouTube is a widely used medical information source, offering educational content on health conditions and treatments. However, its unregulated nature raises concerns about accuracy, reliability, and accessibility. Assessing online medical content quality is important to ensure that patients receive credible, evidence-based guidance.
    BACKGROUND: Keratoconus is a progressive corneal disorder that can lead to visual impairment. Corneal cross-linking (CXL) is an effective treatment that strengthens the corneal structure and halts disease progression. YouTube is frequently used by patients seeking information about diseases and treatments. This study aimed to evaluate the quality and reliability of YouTube videos on CXL.
    METHODS: A cross-sectional analysis was conducted on YouTube videos related to CXL. Five commonly searched keywords were identified using Google Trends, and a YouTube search was performed. Among the first 100 videos ranked by relevance, 69 met the inclusion criteria. Video characteristics, including source, content type, narrator gender, and accessibility features, were recorded. Quality and reliability were assessed using the DISCERN, Global Quality Score (GQS), and Journal of the American Medical Association (JAMA) scoring systems. Statistical analyses examined differences in video quality based on information source and gender.
    RESULTS: The overall quality of YouTube videos on CXL was average, with a mean DISCERN score of 42.5 ± 11.8, GQS of 3.28 ± 0.94, and JAMA score of 2.75 ± 0.68. Videos uploaded by clinicians (39.1%) and health channels (31.9%) had significantly higher quality scores than patient-uploaded videos (29%) (p < 0.05). No significant correlation was found between video popularity and quality metrics. Narrator gender had no significant effect on video quality or popularity metrics (p > 0.05).
    CONCLUSION: While clinician- and health channel-produced videos demonstrated higher quality, this did not translate into greater engagement, highlighting a disconnect between popularity and informational value.
    Keywords:  Corneal cross-linking; DISCERN score; Journal of the American Medical Association score; YouTube; global quality score
    DOI:  https://doi.org/10.1080/08164622.2025.2486701
  18. Emerg Med Int. 2025 ;2025 2582984
      Objective: The aim of our study was to evaluate the accuracy and reliability of videos available on YouTube and Google showing the use of automated external defibrillators. Methods: Videos available on YouTube and Google between 2020 and 2023 were searched using the search terms "Defibrillator," "Resuscitation," "Basic life support," "Cardiac arrest," "CPR," "Cardiac shock," "Chest trust," or "First aid." Data such as the year the video was uploaded, number of views, and video length were collected. The videos were watched and evaluated by two independent emergency physicians. According to the 6-stage evaluation criteria, 1 point was given if the information given in the video was correct and 0 point was given if no information was given. The maximum score was determined as 6 and the minimum score as 1. Results: Out of a total of 315 videos uploaded to the specified platforms, 29 met the inclusion criteria. After the evaluation, the average score given to the videos was 5.45 ± 1.02. When the videos were categorized as low and medium-high according to their fidelity levels, there was no statistically significant difference between these two groups in terms of the number of views, video length, and the score given (p=0.469, 0.078, and 0.110, respectively). Videos from institutions were shorter, with a median length of 180 s compared to 289 s for noninstitution uploads (p=0.047). Both groups received similar scores, with a median of 6 for each (p=0.257). Conclusion: The main findings of our study were that most of the videos were uploaded by health institutions and were shorter. Video scores did not differ according to the level of loyalty of the mannequins used and the uploading source.
    Keywords:  AED; basic life support; cardiac arrest; defibrillator; resuscitation
    DOI:  https://doi.org/10.1155/emmi/2582984
  19. J Hand Ther. 2025 Apr 24. pii: S0894-1130(25)00044-4. [Epub ahead of print]
       BACKGROUND: Resistance exercise is considered first-line management for lateral elbow tendinopathy (LET). TikTok enables sharing of short-form videos with algorithms distributing videos with higher engagement (e.g., views/likes).
    PURPOSE: To assess resistance exercise video quality for LET on TikTok between general users and self-identified health care professionals and explore associations between engagement and video quality.
    STUDY DESIGN: Social media review using a systematic review framework.
    METHODS: TikTok was searched using five hashtags. Included videos were in English and focused on resisted exercise. Creator and video engagement metrics were extracted, with viewing index and engagement rate calculated. The DISCERN and Tennis Elbow Exercise Education Score (TEEES) were used to assess video quality. A Kruskal-Wallis test examined differences between general users and self-identified health care professionals and Spearman's rho (rs) correlations explored associations between video quality and engagement metrics. Median (interquartile range) were reported.
    RESULTS: Ninety-eight videos were included with no significant differences in engagement between general users and self-identified health care professionals. Overall, resistance exercise video quality was poor with no significant differences between general users (DISCERN 26 [22-29] out of 75, TEEES 8 [5-10] out of 25) and self-identified health care professionals (DISCERN 25 [21-27], TEEES 7 [5-9]; p = 0.37 and p = 0.35, respectively). Video quality was weakly correlated with video likes, favourites, and viewing index (rs = 0.23, p = 0.03; rs = 0.23, p = 0.02; rs = 0.21, p = 0.04, respectively).
    CONCLUSIONS: The quality of resistance-based exercise videos for LET on TikTok was poor, regardless of user profile. While some engagement metrics were associated with video quality, viewers should not rely on these metrics to indicate video quality.
    Keywords:  Critical appraisal; Evidence-based; Rehabilitation; Social media; Tennis elbow
    DOI:  https://doi.org/10.1016/j.jht.2025.02.016
  20. J Cancer Educ. 2025 Apr 24.
      Esophageal cancer is a prevalent cancer worldwide, with incidence rates rising significantly in recent years, particularly in high-risk regions where case numbers continue to escalate. As one of the most visited social media platforms globally, TikTok has emerged as a key source for health information. This study aims to evaluate Chinese TikTok videos on esophageal cancer, focusing on content integrity, informational quality, source credibility, usefulness, and reliability.: We assessed 246 esophageal cancer-related videos using the Content Integrity Score based on Goobie's coding scheme for content evaluation and the DISCERN instrument for assessing information reliability and treatment guidance. Videos were categorized by duration, source, publication year, and user engagement metrics (likes, comments, favorites, shares). Statistical analysis was conducted using Shapiro-Wilk, Kruskal-Wallis and post hoc Bonferroni tests. Additionally, Spearman's correlation analysis was applied to explore relationships between key quantitative variables. Of the videos analyzed, 212(86.2%) were uploaded by doctors, and 168(68.3%) lacked content related to risk factors. and 217(88.2%) had a quality score rated as poor or below. User engagement metrics, including likes, comments, saves, and shares, exhibited strong intercorrelations. However, only shares showed a weak correlation with DISCERN scores (R = 0.134, p = 0.036). Additionally, DISCERN scores were positively correlated with video duration (R = 0.352, p < 0.001). Chinese TikTok videos on esophageal cancer generally offer low-quality information that lacks accuracy and is insufficient for guiding patients in making informed health decisions. Given the low quality and reliability of such content, TikTok is not a suitable source for patient education.
    Keywords:  E-health; Esophageal Cancer; Information Quality; Patient Education; Social Media; TikTok
    DOI:  https://doi.org/10.1007/s13187-025-02630-1
  21. Front Public Health. 2025 ;13 1501879
      The rapid development and widespread use of Internet technology have facilitated access to health information for the general public. However, the behavior of acquiring health information is influenced by multiple factors, resulting in differences and even inequalities. This paper aims to explore the influencing factors of user health information acquisition behavior, find feasible ways to optimize such behavior and maximize the utility of health information for users. By utilizing the Zhihu Q&A platform to obtain user health information and integrating social capital theory, the study identifies the influencing factors of user health information acquisition behavior. It utilizes fuzzy set qualitative comparative analysis to examine these factors as antecedent variables, with the dissemination heat of health information as the outcome variable. The study reveals that the structure, relationships, and cognitive dimensions of social capital significantly impact user health information acquisition behavior. The configuration path for user health information acquisition behavior generated by the fsQCA 3.0 software provides recommendations to maximize the utility of health information.
    Keywords:  fsQCA; health inequality; health information acquisition behavior; influencing factors; social capital
    DOI:  https://doi.org/10.3389/fpubh.2025.1501879
  22. J Adolesc Health. 2025 Apr 24. pii: S1054-139X(25)00098-9. [Epub ahead of print]
       PURPOSE: The purpose of this study was to understand the lived experience of adolescents as they seek, view, and evaluate health information using online and offline sources available to them.
    METHODS: Adolescents aged 13-18 years were recruited using purposeful sampling to participate in virtual focus groups. Focus group questions were semistructured and addressed online and offline experiences with health information across a variety of health topics. Focus groups were recorded and transcribed verbatim. Three investigators were involved in the inductive analysis of the transcripts, identifying themes, and leveraging the full range of codes toward a conceptual model.
    RESULTS: A total of 36 adolescents participated across 7 focus groups (Mage = 15.9, SD = 1.3). Adolescents were 69.4% female and 41.7% White. Thematic analysis revealed the following 2 major themes: (1) interconnectedness of online and offline sources and (2) challenges and strategies for navigating online health information. The conceptual model describes adolescents' processes in deciding how and whether to seek health information.
    DISCUSSION: Adolescents rely on online sources to supplement offline sources, to address sensitive health questions, and to seek community, but they encounter challenges when assessing the credibility of online sources.
    Keywords:  Adolescents; Digital health literacy; Focus groups; Health information; Information seeking; Mental health; Online; Qualitative; Sexual health; Social media
    DOI:  https://doi.org/10.1016/j.jadohealth.2025.02.016
  23. Oral Surg Oral Med Oral Pathol Oral Radiol. 2025 Jan 02. pii: S2212-4403(24)00975-1. [Epub ahead of print]
       OBJECTIVE: Dentofacial disharmony is commonly addressed with orthognathic surgery. High-quality and readable information is essential for patients to make informed decisions about their healthcare. This scoping review aimed to collate and synthesize the literature regarding website-based information on orthognathic surgery.
    STUDY DESIGN: Five databases and the grey literature were searched without location or publication year restrictions. Eligible studies were selected using the PCC (Population, Concept, Context) framework under Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews guidelines.
    RESULTS: Nine cross-sectional studies satisfied inclusion criteria. The quality of information was found to be low to moderate. Much of the website content was too complex for the average reader to comprehend. Key information regarding risks and complications was commonly omitted. Higher ranking websites were identified as having lower Ensuring Quality Information for Patients scores. The search terms "orthognathic surgery" and "jaw operation" yielded more accurate information compared to alternative terms.
    CONCLUSIONS: Websites contain low to moderate quality information regarding orthognathic surgery, often providing content too complex to read and understand. Improving the deficiencies identified in the present review is essential to aid informed decision-making by patients contemplating or undergoing an orthognathic surgery pathway.
    DOI:  https://doi.org/10.1016/j.oooo.2024.12.022
  24. Soc Sci Med. 2025 Apr 10. pii: S0277-9536(25)00389-2. [Epub ahead of print]375 118059
       RATIONALES: Using AI for health information seeking is a novel behavior, and as such, developing effective communication strategies to optimize AI adoption in this area presents challenges. To lay the groundwork, research is needed to map out users' behavioral underpinnings regarding AI use, as understanding users' needs, concerns and perspectives could inform the design of targeted and effective communication strategies in this context.
    OBJECTIVE: Guided by the planned risk information seeking model and the comprehensive model of information seeking, our study examines how socio-psychological factors (i.e., attitudes, perceived descriptive and injunctive norms, self-efficacy, technological anxiety) and factors related to information carriers (i.e., trust in and perceived accuracy of AI), shape users' latent profiles. In addition, we explore how individual differences in demographic attributes and anthropocentrism predict membership in these user profiles.
    METHODS: We conducted a quota-sampled survey with 1051 AI-experienced users in Hong Kong. Latent profile analysis was used to examine users' profile patterns. The hierarchical multiple logistic regression was employed to examine how individual differences predict membership in these user profiles.
    RESULTS: The latent profile analysis revealed five heterogeneous profiles, which we labeled "Discreet Approachers," "Casual Investigators," "Apprehensive Moderates," "Apathetic Bystanders," and "Anxious Explorers." Each profile was associated with specific predictors related to individual differences in demographic attributes and/or aspects of anthropocentrism.
    CONCLUSION: The findings advance theoretical understandings of using AI for health information seeking, provide theory-driven strategies to empower users to make well-informed decisions, and offer insights to optimize the adoption of AI technology.
    Keywords:  Artificial intelligence; Audience segmentation; Comprehensive information seeking model; Health information seeking; Planned risk information seeking model
    DOI:  https://doi.org/10.1016/j.socscimed.2025.118059