bims-librar Biomed News
on Biomedical librarianship
Issue of 2025–03–09
twenty-two papers selected by
Thomas Krichel, Open Library Society



  1. Med Ref Serv Q. 2025 Mar 07. 1-14
      As part of the All of Us Data Training and Engagement for Academic Libraries Program, a toolkit was created to provide resources for academic librarians to increase library visibility and research capacity. The toolkit, created from program documents, is intended to simplify library worker's task of bringing the All of Us Research Program to academic institutions to enhance the academic library's research capacity for biomedical and public health data and is intended for wide use. The impact of the toolkit was seen during the 2024 Medical Library Association Annual Conference.
    Keywords:  All of Us Research Program; Library training program; data engagement; digital toolkit; library supporting faculty and student research
    DOI:  https://doi.org/10.1080/02763869.2025.2469915
  2. Epilepsy Behav. 2025 Mar 04. pii: S1525-5050(25)00100-3. [Epub ahead of print]166 110361
       OBJECTIVE: To evaluate and compare the types of information available online about epilepsy genetics to people with epilepsy and their family members on websites and social media platforms.
    METHODS: To identify websites, we used the search terms "epilepsy," "genetic epilepsy," and "epilepsy genetics" in Google, using Incognito mode. We selected the first ten English, non-sponsored websites for analysis. Websites were then classified according to their target audience (healthcare professionals vs. lay persons), and as epilepsy-specific or not and genetic-focused or not. On YouTube, X, Facebook, and TikTok, we searched using the terms "epilepsy" and "genetic epilepsy." Posts were assessed for genetic-related themes, and the frequency of discussion was recorded.
    RESULTS: Among 31 websites assessed, only one did not discuss genetics in any capacity. However, many sources did not expand upon the potential role of genetics in epilepsy, and genetic testing was not frequently discussed. On social media, epilepsy genetics was more frequently discussed using a more specific search term, "genetic epilepsy" than using "epilepsy." In posts that addressed genetics, the importance of genetic testing was the most frequent topic discussed. Other complexities, including offspring epilepsy risk, mode of inheritance, genetic heterogeneity, and variable expressivity, were very seldom discussed. Genetics was more frequently addressed in posts on YouTube and X than in those on Facebook or TikTok.
    SIGNIFICANCE: Online information about epilepsy genetics is difficult to encounter by chance; users must actively seek out this information. While information is available, accessibility and depth of the discussion is limited.
    Keywords:  Behavior; Ethical, legal, and social implications; Genetic; Information seeking; Media exposure; Online information
    DOI:  https://doi.org/10.1016/j.yebeh.2025.110361
  3. J Can Health Libr Assoc. 2024 Dec;45(3): 161-175
       Introduction: The University of Ottawa Heart Institute's Berkman Library space is outdated. Budget constraints and tiny square footage leave little room for error. A needs assessment using user experience (UX) research methods was conducted from 2022 to 2023 to inform strategic decisions on updating and reorganizing furnishings to better support library patrons and their needs.
    Methods: Data was collected via an electronic survey, "guerilla" interviews, observations of library patrons, and a physical survey of communal spaces in the building. Resulting qualitative data were compiled and examined for common themes. Low fidelity mockups of furnishings and space arrangements were prototyped and presented to patrons for feedback.
    Results: Quiet was one of the most valued attributes of the library space and showed itself to be a unique quality of the library when compared to communal spaces within the hospital. Survey and interview responses consistently cited soft, comfortable furnishings as desirable additions. Observed behaviours support the continued need for desks with a deep surface area to accommodate multiple devices used in tandem. Flexible use of computer hardware, better access to power outlets, and adjustable lighting were identified as additional gaps.
    Discussion: Methods showcase light-weight space assessment strategies that are of particular interest to solo librarians or small library teams working in a hospital environment. Results identify library qualities that address institutional gaps and provide insight into the motivators, needs, and behaviours of hospital staff. Centering patron behaviours and preferences in the project's methodology provides data to support decision-making for near term upgrades and long-term library policy.
    DOI:  https://doi.org/10.29173/jchla29774
  4. J Can Health Libr Assoc. 2024 Dec;45(3): 147-160
       Introduction: The objectives of this comparative study were 1) to compare the effectiveness of in-person classroom instruction versus an online video for teaching literature searching skills to graduate students, and 2) to evaluate their perceptions of the instruction methods received and their instruction preferences for learning literature searching skills.
    Methods: Students enrolled in a translational medicine graduate course in Fall 2022 were invited to participate in the study. Participants were randomly assigned to a control group (receiving in-person instruction) or to an intervention group (receiving a link to a narrated PowerPoint video). Using Qualtrics online survey tool, participants completed one pre-test and two post-tests to assess learning and retention, and a survey to evaluate their perceptions and attitudes.
    Results: 12 out of 17 students participated. Both instruction methods were effective for delivering this information literacy content. The in-person group improved more than the online video group from pre- to post-test 1 and from pre- to post-test 2, but the difference was not statistically significant. However, the online video group rated the pace, perceived effectiveness, and clarity of library instruction, and their confidence to perform specific search tasks, more favourably than the in-person group, although the difference was not significant. Participants valued being able to access library training on their own schedule most of all.
    Discussion: The sample size for this study was small, making it difficult for differences to attain statistical significance. Creating an online video to deliver this content ended up being very time-intensive compared to providing synchronous instruction.
    DOI:  https://doi.org/10.29173/jchla29792
  5. Med Ref Serv Q. 2025 Mar 03. 1-12
      Covidence is a tool that facilitates systematic review screening. Librarians obtained Covidence review data from a large research institution to identify the effect that the total number of citations, team members, and librarian collaboration has on the possibility of completion and time needed for screening. The fewer the citations and the fewer the citations per collaborator, the more likely the team is to complete title and abstract and full text screening, and the faster they will finish the screening process. This relationship was stronger for the number of citations per collaborator than the number of citations alone.
    Keywords:  Covidence; screening; systematic reviews; time
    DOI:  https://doi.org/10.1080/02763869.2025.2471879
  6. Med Ref Serv Q. 2025 Mar 03. 1-13
      Librarians build health literacy, beginning with a focus on literacy and information literacy and moving to health literacy for different age and ethnic groups, and levels of literacy groups. Today with media, librarians have concern for the method of how to build literacy. They apply a readability formula to assess health information in multi-languages and different formats to assure each constituency can build their literacy level. This article presents definitions of the many types of literacy, readability formulas, literacy assessment tools and readability apps. The generative AI tool perplexity designed original properly targeted health information.
    Keywords:  Digital literacy; ehealth literacy; health literacy; librarians; literacy; literacy assessment tools; mhealth literacy; perplexity.ai; readability
    DOI:  https://doi.org/10.1080/02763869.2025.2470521
  7. J Bone Joint Surg Am. 2025 Feb 28.
       BACKGROUND: The readability of most online patient educational materials (OPEMs) in orthopaedic surgery is above the American Medical Association/National Institutes of Health recommended reading level of sixth grade for both English- and Spanish-language content. The current project evaluates ChatGPT's performance across English- and Spanish-language orthopaedic OPEMs when prompted to rewrite the material at a sixth-grade reading level.
    METHODS: We performed a cross-sectional study evaluating the readability of 57 English- and 56 Spanish-language publicly available OPEMs found by querying online in both English and Spanish for 6 common orthopaedic procedures. Five distinct, validated readability tests were used to score the OPEMs before and after ChatGPT 4.0 was prompted to rewrite the OPEMs at a sixth-grade reading level. We compared the averages of each readability test, the cumulative average reading grade level, average total word count, average number of complex words (defined as ≥3 syllables), and average number of long sentences (defined as >22 words) between original content and ChatGPT-rewritten content for both languages using paired t tests.
    RESULTS: The cumulative average reading grade level of original English- and Spanish-language OPEMs was 9.6 ± 2.6 and 9.5 ± 1.5, respectively. ChatGPT significantly lowered the reading grade level (improved comprehension) to 7.7 ± 1.9 (95% CI of difference, 1.68 to 2.15; p < 0.05) for English-language content and 8.3 ± 1.3 (95% CI, 1.17 to 1.45; p < 0.05) for Spanish-language content. English-language OPEMs saw a reduction of 2.0 ± 1.8 grade levels, whereas Spanish-language OPEMs saw a reduction of 1.5 ± 1.2 grade levels. Word count, use of complex words, and long sentences were also reduced significantly in both languages while still maintaining high accuracy and similarity compared with original content.
    CONCLUSIONS: Our study supports the potential of artificial intelligence as a low-cost, accessible tool to assist health professionals in improving the readability of orthopaedic OPEMs in both English and Spanish.
    CLINICAL RELEVANCE: TK.
    DOI:  https://doi.org/10.2106/JBJS.24.01078
  8. Updates Surg. 2025 Mar 05.
      This study aimed to evaluate the quality and accuracy of responses provided by two user-interactive AI chatbots, namely ChatGPT and ChatSonic, in response to patient queries regarding laparoscopic repair of inguinal hernias, and additionally determine the suitability of these chatbots in addressing patient queries related to inguinal hernia repair. Ten questions regarding laparoscopic repair of inguinal hernias were developed and presented to ChatGPT 4.0 and ChatSonic. Responses were evaluated by two experienced surgeons blinded to the source, using the Global Quality Score (GQS) and modified DISCERN Score to gauge response quality and reliability. ChatGPT demonstrated high-quality responses (GQS = 4 & 5) for all ten questions according to one evaluator, and for seven out of ten questions according to the other. Similarly, ChatGPT showed high reliability (DISCERN = 4 & 5) for nine responses according to one evaluator, and for three responses according to the other, with only slight agreement between evaluators for both GQS (kappa = 0.20) and modified DISCERN scores (kappa = 0.08). ChatSonic also provided high-quality and reliable responses for a majority of questions, albeit to a lesser extent than ChatGPT, and both demonstrating limited concordance in responses (p > 0.05). Overall, Both ChatGPT and ChatSonic demonstrated potential utility in providing responses to patient queries about hernia surgery. However, due to inconsistencies in reliability and quality, ongoing refinement and validation of AI generated medical information remain necessary before widespread clinical adoption.
    Keywords:  AI Chatbots; ChatGPT; ChatSonic; Hernia repair; Inguinal hernias; Laparoscopic repair
    DOI:  https://doi.org/10.1007/s13304-025-02158-5
  9. Cureus. 2025 Jan;17(1): e78313
       INTRODUCTION: The rise of artificial intelligence (AI), including generative chatbots like ChatGPT (OpenAI, San Francisco, CA, USA), has revolutionized many fields, including healthcare. Patients have gained the ability to prompt chatbots to generate purportedly accurate and individualized healthcare content. This study analyzed the readability and quality of answers to Achilles tendon rupture questions from six generative AI chatbots to evaluate and distinguish their potential as patient education resources.
    METHODS: The six AI models used were ChatGPT 3.5, ChatGPT 4, Gemini 1.0 (previously Bard; Google, Mountain View, CA, USA), Gemini 1.5 Pro, Claude (Anthropic, San Francisco, CA, USA) and Grok (xAI, Palo Alto, CA, USA) without prior prompting. Each was asked 10 common patient questions about Achilles tendon rupture, determined by five orthopaedic surgeons. The readability of generative responses was measured using Flesch-Kincaid Reading Grade Level, Gunning Fog, and SMOG (Simple Measure of Gobbledygook). The response quality was subsequently graded using the DISCERN criteria by five blinded orthopaedic surgeons.
    RESULTS: Gemini 1.0 generated statistically significant differences in ease of readability (closest to average American reading level) than responses from ChatGPT 3.5, ChatGPT 4, and Claude. Additionally, mean DISCERN scores demonstrated significantly higher quality of responses from Gemini 1.0 (63.0±5.1) and ChatGPT 4 (63.8±6.2) than ChatGPT 3.5 (53.8±3.8), Claude (55.0±3.8), and Grok (54.2±4.8). However, the overall quality (question 16, DISCERN) of each model was averaged and graded at an above-average level (range, 3.4-4.4).
    DISCUSSION AND CONCLUSION: Our results indicate that generative chatbots can potentially serve as patient education resources alongside physicians. Although some models lacked sufficient content, each performed above average in overall quality. With the lowest readability and highest DISCERN scores, Gemini 1.0 outperformed ChatGPT, Claude, and Grok and potentially emerged as the simplest and most reliable generative chatbot regarding management of Achilles tendon rupture.
    Keywords:  achilles tear; artificial intelligence; deep learning; orthopedics; sports medicine
    DOI:  https://doi.org/10.7759/cureus.78313
  10. Int Forum Allergy Rhinol. 2025 Mar 04. e23560
      
    Keywords:  American Medical Association; otolaryngology; reading level
    DOI:  https://doi.org/10.1002/alr.23560
  11. J Esthet Restor Dent. 2025 Mar 02.
       OBJECTIVE: This study aimed to evaluate the reliability, consistency, and readability of responses provided by various artificial intelligence (AI) programs to questions related to Restorative Dentistry.
    MATERIALS AND METHODS: Forty-five knowledge-based information and 20 questions (10 patient-related and 10 dentistry-specific) were posed to ChatGPT-3.5, ChatGPT-4, ChatGPT-4o, Chatsonic, Copilot, and Gemini Advanced chatbots. The DISCERN questionnaire was used to assess the reliability; Flesch Reading Ease and Flesch-Kincaid Grade Level scores were utilized to evaluate readability. Accuracy and consistency were determined based on the chatbots' responses to the knowledge-based questions.
    RESULTS: ChatGPT-4, ChatGPT-4o, Chatsonic, and Copilot demonstrated "good" reliability, while ChatGPT-3.5 and Gemini Advanced showed "fair" reliability. Chatsonic exhibited the highest "DISCERN total score" for patient-related questions, while ChatGPT-4o performed best for dentistry-specific questions. No significant differences were found in readability among the chatbots (p > 0.05). ChatGPT-4o showed the highest accuracy (93.3%) for knowledge-based questions, while Copilot had the lowest (68.9%). ChatGPT-4 demonstrated the highest consistency between repetitions.
    CONCLUSION: Performance of AIs varied in terms of accuracy, reliability, consistency, and readability when responding to Restorative Dentistry questions. ChatGPT-4o and Chatsonic showed promising results for academic and patient education applications. However, the readability of responses was generally above recommended levels for patient education materials.
    CLINICAL SIGNIFICANCE: The utilization of AI has an increasing impact on various aspects of dentistry. Moreover, if the responses to patient-related and dentistry-specific questions in restorative dentistry prove to be reliable and comprehensible, this may yield promising outcomes for the future.
    Keywords:  accuracy; artificial intelligence; consistency; readability; reliability
    DOI:  https://doi.org/10.1111/jerd.13447
  12. ANZ J Surg. 2025 Mar 05.
       BACKGROUNDS: Abdominal aortic aneurysms (AAA) carry significant risks, yet patient understanding is often limited, with online resources typically low quality. ChatGPT, an artificial intelligence (AI) chatbot, presents a new frontier in patient education, but concerns remain about misinformation. This study evaluates the quality of ChatGPT-generated patient information on AAA.
    METHODS: Eight patient questions on AAA were sourced from a reputable online resource for patient information funded by the Australian Government's Healthdirect Australia (HDA) website and input into ChatGPT's free (ChatGPT-4o mini) and paid (ChatGPT-4) models. A vascular surgeon evaluated response appropriateness. Readability was assessed using the Flesch-Kincaid test. The Patient Education Materials Assessment Tool (PEMAT) measured understandability and actionability, with responses scoring ≥75% for both considered high-quality.
    RESULTS: All responses were deemed clinically appropriate. Mean response length was longer for ChatGPT than HDA. Readability was at a college level for ChatGPT, while HDA was at a 10th to 12th-grade level. One response was high-quality (generated by paid ChatGPT) with a PEMAT actionability score of ≥75%. Actionability scores were otherwise low across all sources with ChatGPT responses more likely to contain identifiable actions, although these were often not clearly presented. ChatGPT responses were marginally more understandable than HDA.
    CONCLUSIONS: ChatGPT-generated information on AAA was appropriate and understandable, outperforming HDA in both aspects. However, AI responses are at a more advanced reading level and lack actionable instructions. AI chatbots show promise as supplemental tools for AAA patient education, but further refinement is needed to enhance their effectiveness in supporting informed decision-making.
    Keywords:  AAA; AI; ChatGPT; aortic abdominal aneurysm; artificial intelligence; patient information; vascular surgery
    DOI:  https://doi.org/10.1111/ans.70053
  13. Oral Surg Oral Med Oral Pathol Oral Radiol. 2025 Jan 11. pii: S2212-4403(25)00003-3. [Epub ahead of print]
       OBJECTIVE: This study aimed to compare the quality and readability of the responses generated by 3 publicly available artificial intelligence (AI) chatbots in answering frequently asked questions (FAQs) related to Oral and Maxillofacial Radiology (OMR) to assess their suitability for patient education.
    STUDY DESIGN: Fifteen OMR-related questions were selected from professional patient information websites. These questions were posed to ChatGPT-3.5 by OpenAI, Gemini 1.5 Pro by Google, and Copilot by Microsoft to generate responses. Three board-certified OMR specialists evaluated the responses regarding scientific adequacy, ease of understanding, and overall reader satisfaction. Readability was assessed using the Flesch-Kincaid Grade Level (FKGL) and Flesch Reading Ease (FRE) scores. The Wilcoxon signed-rank test was conducted to compare the scores assigned by the evaluators to the responses from the chatbots and professional websites. Interevaluator agreement was examined by calculating the Fleiss kappa coefficient.
    RESULTS: There were no significant differences between groups in terms of scientific adequacy. In terms of readability, chatbots had overall mean FKGL and FRE scores of 12.97 and 34.11, respectively. Interevaluator agreement level was generally high.
    CONCLUSIONS: Although chatbots are relatively good at responding to FAQs, validating AI-generated information using input from healthcare professionals can enhance patient care and safety. Readability of the text content in the chatbots and websites requires high reading levels.
    DOI:  https://doi.org/10.1016/j.oooo.2025.01.001
  14. Arthrosc Sports Med Rehabil. 2025 Feb;7(1): 101027
       Purpose: To investigate whether ChatGPT-3.5 and -4.0 can serve as a viable tool to create readable patient education materials for patients with common orthopaedic upper- and lower-extremity conditions.
    Methods: Using ChatGPT versions 3.5 and 4.0, we asked the artificial intelligence program a series of 2 questions pertaining to patient education for 50 common orthopaedic upper-extremity pathologies and 50 common orthopaedic lower-extremity pathologies. Two templated questions were created and used for all conditions. Readability scores were calculated using the Python library Textstat. Multiple readability test scores were generated, and a consensus reading level was created taking into account the results of 8 reading tests.
    Results: ChatGPT-3.5 produced only 2% and 4% of responses at the appropriate reading level for upper- and lower-extremity conditions, respectively, compared with 54% produced by ChatGPT-4.0 for both upper- and lower-extremity conditions (both P < .0001). After a priming phase, ChatGPT-3.5 did not produce any viable responses for either the upper- or lower-extremity conditions, compared with 64% for both upper- and lower-extremity conditions by ChatGPT-4.0 (both P < .0001). Additionally, ChatGPT-4.0 was more successful than ChatGPT-3.5 in producing viable responses both before and after a priming phase based on all available metrics for reading level (all P < .001), including the Automated Readability index, Coleman-Liau index, Dale-Chall formula, Flesch-Kincaid grade, Flesch Reading Ease score, Gunning Fog score, Linsear Write Formula score, and Simple Measure of Gobbledygook index.
    Conclusions: Our results indicate that ChatGPT-3.5 and -4.0 unreliably created readable patient education materials for common orthopaedic upper- and lower-extremity conditions at the time of the study.
    Clinical Relevance: The findings of this study suggest that ChatGPT, while constantly improving as evidenced by the advances from version 3.5 to version 4.0, should not be substituted for traditional methods of patient education at this time and, in its current state, may be used as a supplemental resource at the discretion of providers.
    DOI:  https://doi.org/10.1016/j.asmr.2024.101027
  15. J Hum Hypertens. 2025 Mar 07.
      Patient education is an important aspect in the management of hypertension and YouTube has become a popular platform for sharing hypertension-related information, but concerns persist regarding the accuracy and reliability of such content in the context of patient education. To address this issue, this cross-sectional type of observational study aims to assess YouTube's information on hypertension using the Global Quality Score (GQS) and Reliability Score (DISCERN SCORE), which evaluate the quality and trustworthiness of the YouTube content, and the Video Power Index (VPI). Majority of the videos are uploaded by physicians (26, 33.3 percent) and most give information about causes/etiology, treatment, complications (n = 51, 65.38%) each. The median GQS for all the videos is 4. The median reliability score is 4. The statistical test used is Kruskal - Wallis Test. The p value of VPI, GQS, and Reliability score among various types of uploader are 0.264, 0.347 and 0.452. Hence, there is no statistical significance (p value > 0.05). The analysis therefore suggests that the quality of these videos is moderate. Although they usually provide accurate information, they lack comprehensiveness and may have a biased perspective.
    DOI:  https://doi.org/10.1038/s41371-025-00999-6
  16. Arthrosc Sports Med Rehabil. 2025 Feb;7(1): 101002
       Purpose: To evaluate the reliability and quality of the educational content of YouTube videos about the anterolateral ligament (ALL).
    Methods: A standard search of the YouTube database was performed. All English-language videos were included for analysis. Video reliability was assessed with the use of the DISCERN tool, Journal of the American Medical Association (JAMA) benchmark criteria, and Global Quality Score (GQS). Data regarding the duration of the videos, date of publication, and number of likes and views were collected. Furthermore, videos were categorized based on video source (health professional, company, or private user), type of information (surgical technique, overview, or anatomy, radiology, and patient experience), and video content (education or patient experience/testimony).
    Results: A total of 94 videos were included in the analysis. Of these videos, 88 (93.6%) were published by health professionals, whereas 4 (4.3%) were published by companies and 2 (2.1%) were published by private users. Most of the videos were about surgical technique (57.4%), and almost all the videos (98.9%) had an educational aim, with the exception of 1 video that reported a patient experience (1.1%). The mean length of the videos was 648.4 ± 973.5 seconds, and the mean online period was 34.5 ± 27.0 months. The mean DISCERN score, JAMA score, and GQS were 32.9 ± 15.9, 1.5 ± 0.9, and 2.3 ± 1.0, respectively. Videos that provided an overview of the ALL were the best in terms of all quality scores and were significantly higher quality than videos about surgical technique and anatomy, radiology, and patient experience for all scores (P < .001). No difference was found between surgical technique and anatomy, radiology, and patient experience (DISCERN score, P > .99; JAMA score, P = .839; and GQS, P > .99).
    Conclusions: The educational content of YouTube videos about the ALL of the knee showed low quality and validity based on the DISCERN score, JAMA score, and GQS.
    Clinical Relevance: With the growing use of social media by patients to gather information about their medical conditions, it is crucial for orthopaedic health care providers to recognize the limitations of social media videos discussing the ALL as potential sources of knowledge for their patients.
    DOI:  https://doi.org/10.1016/j.asmr.2024.101002
  17. J ECT. 2025 Feb 24.
       OBJECTIVES: This study was carried out with a descriptive design to evaluate the performance, comprehensiveness, reliability, and quality of Turkish YouTube videos addressing ECT.
    METHODS: The Turkish keywords "Elektrokonvülsif Terapi, Elektrokonvülsif Tedavi" and EKT [Electroconvulsive Therapy, Electroconvulsive Treatment and ECT] were searched on September 15, 2023, and the first 30 for each keyword and total 90 YouTube videos were examined. A form developed by the researchers was used for comprehensiveness, DISCERN for reliability, and Global Quality Scale for quality.
    RESULTS: The videos' (N = 27) average "number of views" was 546881.3 ± 1754392.2, the average "comprehensiveness score" was 4.8 ± 3.0, the average "reliability score" was 3.1 ± 2.2, and the average "quality score" was 2.6 ± 1.5. When the videos were examined for reliability, 55.6% contained good and useful information for the audience, whereas only 33.3% were of high quality. There was a significant positive correlation between comprehensiveness, reliability, and quality (P = 0.000).
    CONCLUSIONS: People with psychiatric problems or their relatives resort to the Internet and platforms such as YouTube on topics for some information they are unfamiliar with, it is recommended that the content prepared in these areas be comprehensive, reliable, and of high quality.
    DOI:  https://doi.org/10.1097/YCT.0000000000001127
  18. Digit Health. 2025 Jan-Dec;11:11 20552076251324439
       Background: Amyotrophic lateral sclerosis (ALS) is a rare neurodegenerative disease that leads to progressive motor weakness and eventual death. Recent years have seen an increase in online information on ALS, with the popular video platform YouTube becoming a prominent source. We aimed to evaluate the quality, reliability, actionability, and understandability of ALS videos on YouTube.
    Methods: A search was performed using the keyword "Amyotrophic Lateral Sclerosis" on YouTube. A total of 240 videos were viewed and assessed by two independent raters. Video characteristics such as type of uploader, views, likes, comments, and Video Power Index were also collected.
    Results: Videos had moderate to low quality and reliability (Global Quality Scale [GQS] and modified DISCERN [mDISCERN] median 2.5), and poor understandability and actionability (PEMAT total median 8.5). Among the general video characteristics, only length of video showed a significant positive correlation across the tools (with mDISCERN [p < 0.001]; with GQS [p < 0.001]; with PEMAT [p < 0.001]). Videos from physicians (p = 0.024, sig <0.05), other healthcare professionals (p = 0.017, sig <0.05), and educational channels (p = 0.001, sig <0.05) had better quality when compared to others.
    Conclusion: YouTube videos are a poor source of information for ALS as videos tend to have moderate to low quality and reliability and are poorly understandable and actionable. Longer videos, and videos uploaded by those in the healthcare and educational fields, were found to perform relatively better. Thus, when using YouTube, caution and careful attention to the video characteristics are recommended.
    Keywords:  YouTube; amyotrophic lateral sclerosis; digital information; patient education; video quality
    DOI:  https://doi.org/10.1177/20552076251324439
  19. Health Info Libr J. 2025 Mar 06.
       BACKGROUND: Good maternal health is essential (UN Sustainable Development Goal 3). Pregnant women need to effectively evaluate and utilize health information for proper health decision making.
    PURPOSE: To examine the ability of pregnant women to evaluate and utilize maternal health information in the Coastal region of Tanzania (a region with high levels of maternal morbidity).
    METHODOLOGY: Mixed research approach and descriptive cross-sectional design were used to collect data from 132 pregnant women and 8 nurses/midwives using questionnaires and focus group discussions as data collection methods. IBM SPSS version 21 was used to analyse quantitative data, while thematic analysis was used to analyse qualitative data.
    FINDINGS: Many of the pregnant women surveyed (64/128) had low or very low perceived abilities to evaluate maternal health information, but most women, according to health care staff, made appropriate decisions to seek help to avoid major risks. The higher the level of education of a pregnant woman, the higher, generally, her perceived evaluation skills.
    CONCLUSION: Low ability to evaluate maternal health information affects the effective utilization of maternal health services. Collaboration between libraries and health facilities is recommended for the repackaging of information in a user-friendly format.
    Keywords:  consumer health information; health literacy; maternal health services; patient education
    DOI:  https://doi.org/10.1111/hir.12568
  20. JMIR Cancer. 2025 03 03. 11 e56098
       Background: The lack of information and awareness about clinical trials, as well as misconceptions about them, are major barriers to cancer clinical trial participation. Digital and social media are dominant sources of health information and offer optimal opportunities to improve public medical awareness and education by providing accurate and trustworthy health information from reliable sources. Infotainment, material intended to both entertain and inform, is an effective strategy for engaging and educating audiences that can be easily disseminated using social media and may be a novel way to improve awareness of and recruitment in clinical trials.
    Objective: The purpose of this study was to evaluate whether an infotainment video promoting a clinical trial, disseminated using social media, could drive health information seeking behaviors.
    Methods: As part of a video series, we created an infotainment video focused on the promotion of a specific cancer clinical trial. We instituted a dissemination and marketing process on Facebook to measure video engagement and health information seeking behaviors among targeted audiences who expressed interest in breast cancer research and organizations. To evaluate video engagement, we measured reach, retention, outbound clicks, and outbound click-through rate. Frequencies and descriptive statistics were used to summarize each measure.
    Results: The video substantially increased health information seeking behavior by increasing viewership from 1 visitor one month prior to launch to 414 outbound clicks from the video to the clinical trial web page during the 21-day social media campaign period.
    Conclusions: Our study shows that digital and social media tools can be tailored for specific target audiences, are scalable, and can be disseminated at low cost, making it an accessible educational, recruitment, and retention strategy focused on improving the awareness of clinical trials.
    Keywords:  cancer; cancer clinical trials; digital media; education and awareness; edutainment; health information; infotainment; lack of information; low cost; medical awareness; public audience; public engagement; recruitment; research participants; social media; video series
    DOI:  https://doi.org/10.2196/56098