bims-librar Biomed News
on Biomedical librarianship
Issue of 2024‒03‒10
24 papers selected by
Thomas Krichel, Open Library Society



  1. Ir J Med Sci. 2024 Mar 06.
      BACKGROUND & AIMS: Liver cirrhosis affects 4.5 million adults in the United States (US). As more patients educate themselves online, we evaluated the accessibility, quality, understandability, accuracy, readability, and comprehensiveness (AQUA-RC) of online patient education materials for cirrhosis.METHODS: Cross-sectional analysis using Google® and YouTube® on a cleared internet browser from 12 cities across the US. The authors independently reviewed the top 25 search results from each platform using search terms "liver cirrhosis" and "cirrhosis". Accessibility was evaluated from twelve cities in six regions across the US. We evaluated resource quality using the DISCERN score, understandability using the PEMAT score, readability using the Flesch-Kinkaid score, and accuracy/comprehensiveness using a list of author-generated criteria. AQUA-RC was compared between 1) academic websites (AW) vs. non-academic websites (NAW), and 2) websites vs. YouTube® videos.
    RESULTS: 28 websites and 25 videos were included. Accessibility was equal across all regions. Websites had higher average quality scores than videos, although this was not statistically significant (p = 0.84). Websites were more understandable than videos (p < 0.00001). Both websites and videos were 100% accurate. Readability for websites was higher than recommended standards. Websites were more comprehensive than videos (p = 0.02).
    CONCLUSION: Online patient education materials for cirrhosis in the US are equally accessible, but readability and understandability are too complex. Websites are of greater quality, accuracy, and comprehensiveness than YouTube videos, which are often narrowly focused and targeted at the medical community rather than patients. Further efforts should be made to improve online patient education and expand content across platforms.
    Keywords:  Internet; Liver Cirrhosis; Patient Education; Quality; Readability; Understandability
    DOI:  https://doi.org/10.1007/s11845-024-03645-1
  2. PLoS One. 2024 ;19(3): e0300180
      BACKGROUND: The development of short popular science video platforms helps people obtain health information, but no research has evaluated the information characteristics and quality of short videos related to cervical cancer. The purpose of this study was to evaluate the quality and reliability of short cervical cancer-related videos on TikTok and Kwai.METHODS: The Chinese keyword "cervical cancer" was used to search for related videos on TikTok and Kwai, and a total of 163 videos were ultimately included. The overall quality of these videos was evaluated by the Global Quality Score (GQS) and the modified DISCERN tool.
    RESULTS: A total of 163 videos were included in this study, TikTok and Kwai contributed 82 and 81 videos, respectively. Overall, these videos received much attention; the median number of likes received was 1360 (403-6867), the median number of comments was 147 (40-601), and the median number of collections was 282 (71-1296). In terms of video content, the etiology of cervical cancer was the most frequently discussed topic. Short videos posted on TikTok received more attention than did those posted on Kwai, and the GQS and DISCERN score of videos posted on TikTok were significantly better than those of videos posted on Kwai. In addition, the videos posted by specialists were of the highest quality, with a GQS and DISCERN score of 3 (2-3) and 2 (2-3), respectively. Correlation analysis showed that GQS was significantly correlated with the modified DISCERN scores (p<0.001).
    CONCLUSION: In conclusion, the quality and reliability of cervical cancer-related health information provided by short videos were unsatisfactory, and the quality of the videos posted on TikTok was better than that of videos posted on Kwai. Compared with those posted by individual users, short videos posted by specialists provided higher-quality health information.
    DOI:  https://doi.org/10.1371/journal.pone.0300180
  3. World Neurosurg. 2024 Mar 06. pii: S1878-8750(24)00369-3. [Epub ahead of print]
      AIM: The aim of this study is to analyze the quality, reliability, comprehensibility, and the mentioned medical concepts of traumatic brain injury (TBI) related videos on YouTube. MATERIAL - METHODS: "Traumatic brain injury" and related Google Trends queries were used and 328 videos included. Categorization, quality assessment via a 5-point Global Quality Scale, reliability evaluation using a modified DISCERN scale, comprehensibility assessment using Flesch-Kincaid Reading Ease and Flesch-Kincaid Grade Level, and consensus-based classification for usefulness were performed. Target audience categorization was content-based. Text mining techniques were used to identify biomedical terms in video transcripts.RESULTS: Most of the videos were intended for non-healthcare professionals and deemed useful. The videos had intermediate quality and moderate level of reliability. The comprehensibility of the videos exceeded the recommended levels. Videos predominantly covered TBI symptoms, the severity of the condition, its impact on individuals, and possibly strategies related to diagnosis and treatment.
    CONCLUSIONS: Results suggest the requirement for higher quality and reliability in YouTube content about TBI. Emphasis should be placed on clear and accurate language to promote comprehensibility. Continued research, guidelines, education, and platform oversight can enhance the spread of reliable health information on social media, benefiting creators and consumers in this field.
    Keywords:  Comprehensibility; Quality; Reliability; Text Mining; Traumatic Brain Injury; YouTube
    DOI:  https://doi.org/10.1016/j.wneu.2024.02.150
  4. Int Wound J. 2024 Mar;21(3): e14729
      We aimed to assess the quality and reliability of pressure injury-related videos uploaded on YouTube, analyse the sources and contents, and examine the correlation between video parameters. We searched YouTube using two keywords, "pressure ulcer" and "pressure sore", on August 20, 2022. We sorted the videos according to their number of views and included the top 100 videos for each keyword. The quality of videos was assessed using the Global Quality Scale (GQS), while their reliability was evaluated by the modified DISCERN (mDISCERN) tool. In addition, we evaluated the videos in which content was included, analysed the correlations and differences between GQS, mDISCERN, and video parameters. We initially found a total of 100 videos for each keyword and finally included and analysed 77 videos. The mean scores for the mDISCERN and GQS were 2.35 ± 0.98 and 3.09 ± 0.99, respectively. Both GQS and mDISCERN showed statistically significant correlations with each other (rho = 0.54, p < 0.0001*) and with the length of the videos, respectively (rho = 0.36, p = 0.001*), (rho = 0.29, p = 0.01*). Of the videos created by physicians, 8 (57.1%) included content related to treatment, while of the videos created by nonphysician health personnel, 22 (57.9%) included content related to prevention. Analysing whether there were differences in video parameters based on the sources, we observed significant differences between sources in GQS (p < 0.0001*), mDISCERN (p < 0.0001*), and video length (p = 0.001*). In the post-hoc analysis, videos uploaded by physicians or nonphysician health personnel showed higher quality and reliability than videos uploaded by other sources. Therefore, the results of this study could be useful for healthcare providers, as well as patients and caregivers, to search for high-quality and reliable YouTube videos related to pressure injury.
    Keywords:  YouTube; pressure injury; pressure sore; pressure ulcer; video analysis
    DOI:  https://doi.org/10.1111/iwj.14729
  5. Restor Dent Endod. 2024 Feb;49(1): e10
      Objectives: This study aimed to evaluate the content, quality and demographics of YouTube videos about rubber dam as an information source for clinicians and dental students.Materials and Methods: "Rubber dam," "rubber dam application," "dental isolation," "rubber dam isolation," and "dental dam" were determined as keywords for the detection of YouTube videos related to rubber dam. Seventy 3 videos were evaluated and a total of 34 videos met the inclusion criteria. All selected videos were evaluated according to 8 parameters. The videos were scored 1 if the videos contained information about the selected parameter, but if the videos did not contain enough information, they were scored 0. The data were statistically analyzed with the analysis of variance and post hoc Tukey test (p < 0.05).
    Results: We found that 41% of the videos have poor, 47% have moderate, and 12% have good information. There is a statistically significant difference in time between poor and good information content (p < 0.05). There is a statistically significant difference between the poor and good information in the video information and quality index 1.
    Conclusions: Rubber dam-related videos available on YouTube are generally moderately informed and insufficient. YouTube is currently not sufficient as a source of information for patients and clinicians at the moment. The YouTube platform should be developed and enriched with quality information on current and dental issues.
    Keywords:  Dental isolation; Education; Rubber dam; YouTube
    DOI:  https://doi.org/10.5395/rde.2024.49.e10
  6. PLoS One. 2024 ;19(3): e0298597
      Gum bleeding is a common dental problem, and numerous patients seek health-related information on this topic online. The YouTube website is a popular resource for people searching for medical information. To our knowledge, no recent study has evaluated content related to bleeding gums on YouTube™. Therefore, this study aimed to conduct a quantitative and qualitative analysis of YouTube videos related to bleeding gums. A search was performed on YouTube using the keyword "bleeding gums" from Google Trends. Of the first 200 results, 107 videos met the inclusion criteria. The descriptive statistics for the videos included the time since upload, the video length, and the number of likes, views, comments, subscribers, and viewing rates. The global quality score (GQS), usefulness score, and DISCERN were used to evaluate the video quality. Statistical analysis was performed using the Kruskal-Wallis test, Mann-Whitney test, and Spearman correlation analysis. The majority (n = 69, 64.48%) of the videos observed were uploaded by hospitals/clinics and dentists/specialists. The highest coverage was for symptoms (95.33%). Only 14.02% of the videos were classified as "good". The average video length of the videos rated as "good" was significantly longer than the other groups (p <0.05), and the average viewing rate of the videos rated as "poor" (63,943.68%) was substantially higher than the other groups (p <0.05). YouTube videos on bleeding gums were of moderate quality, but their content was incomplete and unreliable. Incorrect and inadequate content can significantly influence patients' attitudes and medical decisions. Effort needs to be expended by dental professionals, organizations, and the YouTube platform to ensure that YouTube can serve as a reliable source of information on bleeding gums.
    DOI:  https://doi.org/10.1371/journal.pone.0298597
  7. Surg Endosc. 2024 Mar 05.
      INTRODUCTION: Generative artificial intelligence (AI) chatbots have recently been posited as potential sources of online medical information for patients making medical decisions. Existing online patient-oriented medical information has repeatedly been shown to be of variable quality and difficult readability. Therefore, we sought to evaluate the content and quality of AI-generated medical information on acute appendicitis.METHODS: A modified DISCERN assessment tool, comprising 16 distinct criteria each scored on a 5-point Likert scale (score range 16-80), was used to assess AI-generated content. Readability was determined using the Flesch Reading Ease (FRE) and Flesch-Kincaid Grade Level (FKGL) scores. Four popular chatbots, ChatGPT-3.5 and ChatGPT-4, Bard, and Claude-2, were prompted to generate medical information about appendicitis. Three investigators independently scored the generated texts blinded to the identity of the AI platforms.
    RESULTS: ChatGPT-3.5, ChatGPT-4, Bard, and Claude-2 had overall mean (SD) quality scores of 60.7 (1.2), 62.0 (1.0), 62.3 (1.2), and 51.3 (2.3), respectively, on a scale of 16-80. Inter-rater reliability was 0.81, 0.75, 0.81, and 0.72, respectively, indicating substantial agreement. Claude-2 demonstrated a significantly lower mean quality score compared to ChatGPT-4 (p = 0.001), ChatGPT-3.5 (p = 0.005), and Bard (p = 0.001). Bard was the only AI platform that listed verifiable sources, while Claude-2 provided fabricated sources. All chatbots except for Claude-2 advised readers to consult a physician if experiencing symptoms. Regarding readability, FKGL and FRE scores of ChatGPT-3.5, ChatGPT-4, Bard, and Claude-2 were 14.6 and 23.8, 11.9 and 33.9, 8.6 and 52.8, 11.0 and 36.6, respectively, indicating difficulty readability at a college reading skill level.
    CONCLUSION: AI-generated medical information on appendicitis scored favorably upon quality assessment, but most either fabricated sources or did not provide any altogether. Additionally, overall readability far exceeded recommended levels for the public. Generative AI platforms demonstrate measured potential for patient education and engagement about appendicitis.
    Keywords:  Appendicitis; Artificial intelligence; ChatGPT; Health literacy; Large language models; Online medical information
    DOI:  https://doi.org/10.1007/s00464-024-10739-5
  8. Indian J Otolaryngol Head Neck Surg. 2024 Feb;76(1): 987-991
      Hearing aids and implants are used to treat hearing loss, with cochlear implants being the most successful option for severe sensorineural hearing loss. Patients frequently use the internet as a trusted source of clinical information before committing to any therapeutic procedure, including receiving a cochlear implant. A health resource's readability and dependability influence its value to patients. Readability refers to how easily language can be understood, whereas reliability refers to the correctness and consistency of the information presented. JAMA standards and the DISCERN tool were used to assess the reliability of the websites listed. For readability analysis, the FRE, FKG and GFI were chosen. The acceptable readability level was set to < 7 for the FKG, GF score over 17 as the equivalent of college-level education and ≥ 80.0 for the FRE. The readability scores vary across the sources, suggesting a range of comprehension levels required for understanding the cochlear implant patient information found on Google. There was a statistical difference detected in Discern score between the groups (p = 0.008). The mean discern score was significantly higher in hospital generated sources when compared to industry (3.13 ± 0.69 vs. 2.11 ± 0.78, p = 0.03).
    Keywords:  Cochlear implant; Health literacy; Patient information; Readability and reliability
    DOI:  https://doi.org/10.1007/s12070-023-04341-9
  9. Korean J Fam Med. 2024 Mar 07.
      Background: The Web is an important source of health information, but the quality of such online information is highly variable. This study evaluates the quality of health articles published on Naver News, Korea's most popular portal, using the Health Information Quality Assessment Tool (HIQUAL).Methods: We collected 712 health-related articles published on Naver News from May 1 to 7, 2023. After applying exclusion criteria, we selected 116 articles for analysis. Two clinicians independently assessed the quality of these articles using the HIQUAL, which scores articles based on five domains: "reliability," "usefulness," "understandability," "sufficiency," and "transparency."
    Results: Overall article quality was generally considered recommendable (mean±standard deviation: 7.52±2.00). "Usefulness," one item of "reliability," and "understandability" were the three items with the highest levels of satisfaction. "Sufficiency" criteria for costs, risks, and benefits received low scores. Quality scores for articles focused on health risk factors and intervention showed a statistically significant difference.
    Conclusion: While the overall quality of health information in Korean online newspaper articles is acceptable, room for improvement remains in some areas, particularly with regard to the fair presentation of costs, risks, and benefits. The study highlights the need for ongoing quality improvement and evaluation initiatives for online health information.
    Keywords:  Consumer Health Information; Health Information Quality Assessment Tool; Health Literacy; Newspaper Article
    DOI:  https://doi.org/10.4082/kjfm.23.0197
  10. Liver Int. 2024 Mar 05.
      BACKGROUND & AIMS: Short videos, crucial for disseminating health information on metabolic dysfunction-associated steatotic liver disease (MASLD), lack a clear evaluation of quality and reliability. This study aimed to assess the quality and reliability of MASLD-related videos on Chinese platforms.METHODS: Video samples were collected from three platforms (TikTok, Kwai and Bilibili) during the period from November 2019 to July 2023. Two independent reviewers evaluated the integrity of the information contained therein by scoring six key aspects of its content: definition, epidemiology, risk factors, outcomes, diagnosis and treatment. The quality and reliability of the videos were assessed using the Journal of the American Medical Association (JAMA) criteria, the Global Quality Score (GQS) and the modified DISCERN score.
    RESULTS: A total of 198 videos were included. The video content exhibited an overall unsatisfactory quality, with a primary emphasis on risk factors and treatment, while diagnosis and epidemiology were seldom addressed. Regarding the sources of the videos, the GQS and modified DISCERN scores varied significantly between the platforms (p = .003), although they had generally similar JAMA scores (p = .251). Videos created by medical professionals differed significantly in terms of JAMA scores (p = .046) compared to those created by nonmedical professionals, but there were no statistically significant differences in GQS (p = .923) or modified DISCERN scores (p = .317).
    CONCLUSIONS: The overall quality and reliability of the videos were poor and varied between platforms and uploaders. Platforms and healthcare professionals should strive to provide more reliable health-related information regarding MASLD.
    Keywords:  metabolic dysfunction-associated steatotic liver disease; patient education; public health; quality; reliability
    DOI:  https://doi.org/10.1111/liv.15871
  11. Front Public Health. 2024 ;12 1328544
      Background: This study examined access to technology, internet usage, and online health information-seeking behaviors, in a racially diverse, lower-income population.Methods: Data were obtained via a cross-sectional survey of low-income communities in Houston, Los Angeles, and New York between April and August 2023. Binary responses to the following online health information-seeking behaviors, internet and technology access, were examined: using the internet to (i) understand a medical diagnosis, (ii) fill a prescription, (iii) schedule a healthcare appointment, (iv) email communication with a healthcare provider, and (v) access electronic health records and medical notes.
    Results: 41% of survey respondents identified as non-Hispanic Black individuals, 33% as non-Hispanic White individuals, and 22% as Hispanic individuals. 69% reported a pre-tax annual household income of less than $35,000. 97% reported ownership/access to a smart device; 97% reported access to reliable internet. In the past year, only 59% reported using the internet to better understand their medical diagnosis, 36% reported filling a prescription online, 47% scheduled a medical appointment online, 47% viewed electronic health records online, and 56% emailed healthcare providers. Female sex, higher incomes, and having at least a bachelor's degree were significantly associated with all five online health information-seeking attributes.
    Conclusion: Despite high technology adoption rates, we observed suboptimal online health information-seeking behaviors. This underutilization has potential adverse implications for healthcare access and use given the documented advantage of HIT. Efforts to increase health information-seeking behaviors should explore the identification of HIT barriers, and patient education to increase familiarity and usage in this population.
    Keywords:  health disparities; internet use; perception; technology; trust
    DOI:  https://doi.org/10.3389/fpubh.2024.1328544
  12. BMC Bioinformatics. 2024 Mar 06. 25(1): 101
      PURPOSE: The expansion of research across various disciplines has led to a substantial increase in published papers and journals, highlighting the necessity for reliable text mining platforms for database construction and knowledge acquisition. This abstract introduces GPDMiner(Gene, Protein, and Disease Miner), a platform designed for the biomedical domain, addressing the challenges posed by the growing volume of academic papers.METHODS: GPDMiner is a text mining platform that utilizes advanced information retrieval techniques. It operates by searching PubMed for specific queries, extracting and analyzing information relevant to the biomedical field. This system is designed to discern and illustrate relationships between biomedical entities obtained from automated information extraction.
    RESULTS: The implementation of GPDMiner demonstrates its efficacy in navigating the extensive corpus of biomedical literature. It efficiently retrieves, extracts, and analyzes information, highlighting significant connections between genes, proteins, and diseases. The platform also allows users to save their analytical outcomes in various formats, including Excel and images.
    CONCLUSION: GPDMiner offers a notable additional functionality among the array of text mining tools available for the biomedical field. This tool presents an effective solution for researchers to navigate and extract relevant information from the vast unstructured texts found in biomedical literature, thereby providing distinctive capabilities that set it apart from existing methodologies. Its application is expected to greatly benefit researchers in this domain, enhancing their capacity for knowledge discovery and data management.
    Keywords:  Biomedical domain; Database curation; Mining tool; Named-entity recognition; Natural language process; Relation extraction; Text mining
    DOI:  https://doi.org/10.1186/s12859-024-05710-z
  13. J Turk Ger Gynecol Assoc. 2024 Mar 06. 25(1): 24-29
      Objective: This study aimed to evaluate the quality of laparoscopic myomectomy videos on YouTube and WebSurg.Material and Methods: We searched using the keyword "laparoscopic myomectomy" on WebSurg and selected surgical interventions in the gynecology section. Eleven videos on WebSurg were enrolled. We selected the 22 most-relevant videos on YouTube to create a comparison group, with a ratio of 1:2. Sound in videos, number of subscribers, views, likes, and comments, number of days since videos were uploaded and durations of videos were recorded. View/day, like/view, like/subscriber, and view/subscriber ratios were calculated. The videos were evaluated with usefulness score (US), global quality scoring (GQS), modified discern score (mDS) and laparoscopic surgery video educational guidelines (LAP-VEGaS).
    Results: The view/day ratio was lower in WebSurg compared to YouTube [1.3 (1.9) vs. 7.5 (30.6), respectively; p=0.039]. No difference was found between WebSurg and YouTube in terms of US, GQS and mDS. On LAP-VEGaS assessment, WebSurg was found to be superior to YouTube in terms of intraoperative findings [2 (1-2) vs. 1 (0-2), p=0.001], additional materials [1 (0-2) vs. 1 (0-1), p=0.041], audio/written commentary [2 (2-2) vs. 2 (0-2), p=0.037], image quality [2 (2-2) vs. 2 (0-2), p=0.023], questions and total score [12 (11-13) vs. 10.5 (4-13), p=0.006]. The proportion of high-quality video was higher in WebSurg compared to YouTube, when the cut-off value of total score of 11 or 12 was used as 10 (100%) vs. 10 (50%), p=0.011 and 9 (90%) vs. 5 (25%), p=0.001, respectively.
    Conclusion: WebSurg was better compared to YouTube in terms of quality of laparoscopic myomectomy videos.
    Keywords:  LAP-VEGaS; Laparoscopic myomectomy; WebSurg; YouTube; global quality scoring
    DOI:  https://doi.org/10.4274/jtgga.galenos.2023.2023-5-7
  14. Otolaryngol Head Neck Surg. 2024 Mar 03.
      OBJECTIVE: Social media may inform health care decisions among younger patient populations. TikTok is a social media platform that allows users to post short-form videos. This study aimed to assess the quality of sinusitis-related videos on TikTok.STUDY DESIGN: We searched TikTok on January 29, 2023, for sinusitis-related hashtags: #sinusitis, #sinus, #sinusinfection.
    SETTING: Internet.
    METHODS: The number of views/shares per day, uploader type (nonmedical influencer, lay individual, and medical professional) content categories (medical advice, marketing, comedy, and lifestyle/acceptability), and content type (educational vs factual) were collected. The Patient Education Materials Assessment Tool for Audiovisual Material and Journal of the American Medical Association criteria score was used to measure understandability, actionability, and reliability. The Global Quality Scale (GQS) was used to evaluate the quality of videos; the harm/benefit score was used to evaluate causative effects. Analyses were performed using analysis of variance (α = .05).
    RESULTS: There were 221 videos identified, which garnered over 300 million views and 1 million shares. Almost half of the videos were published by nonmedical influencers. When controlling for covariates, nonmedical influencers and lay uploaders were more likely to have harmful harm/benefit scores, less understandable videos, and lower GQS scores compared to medical professionals. Less than half of videos posted by nonmedical influencers categorized as educational were factual (46.7%); lay individuals and medical professionals had higher rates of factual educational content (79.9% and 83.7%, respectively).
    CONCLUSION: Most nonmedical influencer-posted TikTok videos about sinusitis are inaccurate, despite being portrayed as medical advice/educational. Rhinologists must find modern ways to disseminate true disease-related content via social media to combat medical misinformation.
    Keywords:  TikTok; health care education; public health; sinusitis; social media
    DOI:  https://doi.org/10.1002/ohn.688
  15. Eur Arch Otorhinolaryngol. 2024 Mar 04.
      PURPOSE: ChatGPT (Chat-Generative Pre-trained Transformer) has proven to be a powerful information tool on various topics, including healthcare. This system is based on information obtained on the Internet, but this information is not always reliable. Currently, few studies analyze the validity of these responses in rhinology. Our work aims to assess the quality and reliability of the information provided by AI regarding the main rhinological pathologies.METHODS: We asked to the default ChatGPT version (GPT-3.5) 65 questions about the most prevalent pathologies in rhinology. The focus was learning about the causes, risk factors, treatments, prognosis, and outcomes. We use the Discern questionnaire and a hexagonal radar schema to evaluate the quality of the information. We use Fleiss's kappa statistical analysis to determine the consistency of agreement between different observers.
    RESULTS: The overall evaluation of the Discern questionnaire resulted in a score of 4.05 (± 0.6). The results in the Reliability section are worse, with an average score of 3.18. (± 1.77). This score is affected by the responses to questions about the source of the information provided. The average score for the Quality section was 3.59 (± 1.18). Fleiss's Kappa shows substantial agreement, with a K of 0.69 (p < 0.001).
    CONCLUSION: The ChatGPT answers are accurate and reliable. It generates a simple and understandable description of the pathology for the patient's benefit. Our team considers that ChatGPT could be a useful tool to provide information under prior supervision by a health professional.
    Keywords:  Artificial intelligence; ChatGPT; Chatbot; Healthcare; Natural language processing; Rhinology
    DOI:  https://doi.org/10.1007/s00405-024-08581-5
  16. J Endourol. 2024 Mar 05.
      Introduction Artificial intelligence (AI) platforms such as ChatGPT and Bard are increasingly utilized to answer patient healthcare questions. We present the first study to blindly evaluate AI-generated responses to common endourology patient questions against official patient education materials. Methods 32 questions and answers spanning kidney stones, ureteral stents, BPH, and UTUC were extracted from official Urology Care Foundation (UCF) patient education documents. The same questions were input into ChatGPT 4.0 and Bard, limiting responses to within  10% of the word count of the corresponding UCF response to ensure fair comparison. Six endourologists blindly evaluated responses from each platform using Likert scales for accuracy, clarity, comprehensiveness, and patient utility. Reviewers identified which response they believed was not AI-generated. Lastly, Flesch-Kincaid Reading Grade Level formulas assessed the readability of each platform response. Ratings were compared using ANOVA and Chi-Square tests. Results ChatGPT responses were rated the highest across all categories including accuracy, comprehensiveness, clarity, and patient utility while UCF answers were consistently scored the lowest, all p<0.01. Sub-analysis revealed that this trend was consistent across question categories (i.e., kidney stones, BPH, etc.). However, AI-generated responses were more likely to be classified at an advanced reading level while UCF responses showed improved readability (college or higher reading level: ChatGPT = 100%, Bard = 66%, UCF = 19%), p<0.001. When asked to identify which answer was not AI-generated, 54.2% of responses indicated ChatGPT, 26.6% indicated Bard, and only 19.3% correctly identified it as the UCF response. Conclusions In a blind evaluation, AI-generated responses from ChatGPT and Bard surpassed the quality of official patient education materials in endourology, suggesting that current AI platforms are already a reliable resource for basic urologic care information. AI-generated responses do, however, tend to require a higher reading level, which may limit their applicability to a broader audience.
    DOI:  https://doi.org/10.1089/end.2023.0696
  17. JMIR Dermatol. 2024 Mar 06. 7 e50163
      
    Keywords:  AI; ChatGPT; LLM; LLMs; NLP; artificial intelligence; comprehensibility; comprehensible; dermatology; disparities; disparity; generative; health disparities; health education; health information; health literacy; language model; language models; large language model; natural language processing; online information; patient education; public education; readability; understandability
    DOI:  https://doi.org/10.2196/50163
  18. JMIR Med Educ. 2024 Mar 04. 10 e48393
      BACKGROUND: Access to reliable and accurate digital health web-based resources is crucial. However, the lack of dedicated search engines for non-English languages, such as French, is a significant obstacle in this field. Thus, we developed and implemented a multilingual, multiterminology semantic search engine called Catalog and Index of Digital Health Teaching Resources (CIDHR). CIDHR is freely accessible to everyone, with a focus on French-speaking resources. CIDHR has been initiated to provide validated, high-quality content tailored to the specific needs of each user profile, be it students or professionals.OBJECTIVE: This study's primary aim in developing and implementing the CIDHR is to improve knowledge sharing and spreading in digital health and health informatics and expand the health-related educational community, primarily French speaking but also in other languages. We intend to support the continuous development of initial (ie, bachelor level), advanced (ie, master and doctoral levels), and continuing training (ie, professionals and postgraduate levels) in digital health for health and social work fields. The main objective is to describe the development and implementation of CIDHR. The hypothesis guiding this research is that controlled vocabularies dedicated to medical informatics and digital health, such as the Medical Informatics Multilingual Ontology (MIMO) and the concepts structuring the French National Referential on Digital Health (FNRDH), to index digital health teaching and learning resources, are effectively increasing the availability and accessibility of these resources to medical students and other health care professionals.
    METHODS: First, resource identification is processed by medical librarians from websites and scientific sources preselected and validated by domain experts and surveyed every week. Then, based on MIMO and FNRDH, the educational resources are indexed for each related knowledge domain. The same resources are also tagged with relevant academic and professional experience levels. Afterward, the indexed resources are shared with the digital health teaching and learning community. The last step consists of assessing CIDHR by obtaining informal feedback from users.
    RESULTS: Resource identification and evaluation processes were executed by a dedicated team of medical librarians, aiming to collect and curate an extensive collection of digital health teaching and learning resources. The resources that successfully passed the evaluation process were promptly included in CIDHR. These resources were diligently indexed (with MIMO and FNRDH) and tagged for the study field and degree level. By October 2023, a total of 371 indexed resources were available on a dedicated portal.
    CONCLUSIONS: CIDHR is a multilingual digital health education semantic search engine and platform that aims to increase the accessibility of educational resources to the broader health care-related community. It focuses on making resources "findable," "accessible," "interoperable," and "reusable" by using a one-stop shop portal approach. CIDHR has and will have an essential role in increasing digital health literacy.
    Keywords:  French; controlled; curriculum; digital health; educational personnel; knowledge management; language; medical education; medical informatics; search engine; semantic web; students; teaching; vocabulary
    DOI:  https://doi.org/10.2196/48393
  19. J Drugs Dermatol. 2024 Mar 01. 23(3): e93-e96
      BACKGROUND: Hidradenitis suppurativa (HS) is a painful, chronic inflammatory skin disease that negatively affects patient quality of life, and conventional treatments are variably effective. As a result, patients often turn to complementary and alternative medicine (CAM) for pain relief. Social media enables HS patients to share treatment recommendations. TikTok is a popular social media platform, but little is known about the HS treatments discussed in TikTok videos.&nbsp; Objective: To evaluate the content and quality of information on TikTok regarding CAM HS therapies.&nbsp; Methods: A cross-sectional analysis was conducted by performing a search in TikTok using the terms #hidradenitissuppurativa, #hswarrior, #naturalremedy, #complementarymedicine, #alternativemedicine, and #HStreatment. Two independent reviewers evaluated video quality using the DISCERN and AVA instruments. Linear regressions compared the engagement, DISCERN, and AVA scores among different uploader types.RESULTS: In total, 91 TikTok videos were analyzed. Videos were uploaded by non-physicians (82.4), dermatologists (6.6%), and private companies (11.0%). The average DISCERN and AVA scores were 36.2 and 1.6, respectively (poor quality). Common CAM therapies were natural salves, turmeric, Epsom salts, elimination diets, and zinc supplements. Physician-uploaded videos were of significantly higher quality than videos by other uploader types, with an average DISCERN and AVA score of 44.3 (P&lt;0.009) and 2.6 (P&lt;0.001), respectively (fair quality).
    CONCLUSION: TikTok videos were poor quality (low DISCERN and AVA scores); physician-uploaded videos were fair quality. Dermatologists can improve video quality by adequately discussing the supporting evidence, mechanisms of action, and remaining questions for HS treatments.&nbsp; J Drugs Dermatol. 2024;23(3):e93-96. doi:10.36849/JDD.7738e.
    DOI:  https://doi.org/10.36849/jdd.7738
  20. Gen Hosp Psychiatry. 2024 Feb 29. pii: S0163-8343(24)00045-8. [Epub ahead of print]
      
    Keywords:  Grey literature; Publication bias; Scoping review; Search strategy
    DOI:  https://doi.org/10.1016/j.genhosppsych.2024.02.015
  21. Cont Lens Anterior Eye. 2024 Mar 04. pii: S1367-0484(24)00013-4. [Epub ahead of print] 102130
      INTRODUCTION: Artificial Intelligence (AI) chatbots are able to explain complex concepts using plain language. The aim of this study was to assess the accuracy of three AI chatbots answering common questions related to contact lens (CL) wear.METHODS: Three open access AI chatbots were compared: Perplexity, Open Assistant and ChatGPT 3.5. Ten general CL questions were asked to all AI chatbots on the same day in two different countries, with the questions asked in Spanish from Spain and in English from the U.K. Two independent optometrists with experience working in each country assessed the accuracy of the answers provided. Also, the AI chatbots' responses were assessed if their outputs showed any bias towards (or against) any eye care professional (ECP).
    RESULTS: The answers obtained by the same AI chatbots were different in Spain and the U.K. Also, statistically significant differences were found between the AI chatbots for accuracy. In the U.K., ChatGPT 3.5 was the most and Open Assistant least accurate (p < 0.01). In Spain, Perplexity and ChatGPT were statistically more accurate than Open Assistant (p < 0.01). All the AI chatbots presented bias, except ChatGPT 3.5 in Spain.
    CONCLUSIONS: AI chatbots do not always consider local CL legislation, and their accuracy seems to be dependent on the language used to interact with them. Hence, at this time, although some AI chatbots might be a good source of information for general CL related questions, they cannot replace an ECP.
    Keywords:  Artificial intelligence chatbots; Contact lens opticians; Contact lenses; Eye care professionals; Ophthalmologists; Optometrists
    DOI:  https://doi.org/10.1016/j.clae.2024.102130
  22. Nature. 2024 Mar 04.
      
    Keywords:  Information technology; Publishing; Scientific community
    DOI:  https://doi.org/10.1038/d41586-024-00616-5
  23. J Cell Physiol. 2024 Mar 08.
      A popular preprint server, bioRxiv, is important as a tool for increased visibility for life science research. If used properly, however, bioRxiv can also be an important tool for training, as it may expose trainees (degree-seeking students undertaking research or internships directly related to their field of study) to the peer review process. Here, we offer a comprehensive guide to using bioRxiv as a training tool, as well as offer suggestions for improvements in bioRxiv, including confusion that may be caused by bioRxiv articles appearing on PubMed.
    Keywords:  PubMed; STEM; bioRxiv; mentoring; preprint; training
    DOI:  https://doi.org/10.1002/jcp.31234