bims-librar Biomed News
on Biomedical librarianship
Issue of 2024–12–29
twenty-two papers selected by
Thomas Krichel, Open Library Society



  1. Front Psychol. 2024 ;15 1514348
      Boredom and curiosity are common everyday states that drive individuals to seek information. Due to their functional relatedness, it is not trivial to distinguish whether an action, for instance in the context of a behavioral experiment, is driven by boredom or curiosity. Are the two constructs opposite poles of the same cognitive mechanism, or distinct states? How do they interact? Can they co-exist and complement each other? Here, we systematically review similarities and dissimilarities of boredom and curiosity with respect to their subjective experience, functional role, and neurocognitive implementation. We highlight the usefulness of Information Theory for formalizing information-seeking in the context of both states and provide guidelines for their experimental investigation. Our emerging view is that despite their distinction on an experiential level, boredom and curiosity are closely related on a functional level, providing complementary drives on information-seeking: boredom, similar to hunger, arises from a lack of information and drives individuals to avoid contexts with low information yield, whereas curiosity constitutes a mechanism similar to appetite, pulling individuals toward specific sources of information. We discuss predictions arising from this perspective, concluding that boredom and curiosity are independent, but coalesce to optimize behavior in environments providing varying levels of information.
    Keywords:  boredom; curiosity; exploration; exploration and exploitation; information; information-seeking
    DOI:  https://doi.org/10.3389/fpsyg.2024.1514348
  2. Int Urol Nephrol. 2024 Dec 23.
      This paper evaluated the bibliometric study by Li et al. (Int Urol Nephrol, 2024) on machine learning in renal medicine. Although the study claims to summarize the forefront trends and hotspots in this field, several key issues require further clarification to effectively guide future research. Firstly, while the authors used the "*" wildcard to broaden the search scope, they screened articles only by document type and language, without specific filtering based on titles, abstracts, or full texts. This approach may have led to the inclusion of irrelevant studies, potentially compromising analytical accuracy. Secondly, the authors conducted the search using the Topic (TS) field, which may include articles not closely related to the intended topic. We recommend using Title (TI), Abstract (AB), and Author Keywords (AK) as filtering criteria in future studies to improve search precision. Finally, in the keyword co-occurrence analysis, the authors did not merge synonyms, leading to distortions in keyword frequency rankings; for example, "machine learning" and "machine learning (ML)" were treated as separate terms. We believe that synonym merging would enhance the accuracy of keyword analysis. Overall, the search strategy by Li et al. demonstrates issues such as imprecise scope and lack of synonym integration. To ensure the comprehensiveness and accuracy of future research, we suggest refining the search strategy, employing precise screening steps, and integrating synonyms to improve the quality of bibliometric studies.
    DOI:  https://doi.org/10.1007/s11255-024-04335-8
  3. Front Vet Sci. 2024 ;11 1463332
      It is said that dogs are human's best friend. On occasion, dogs can be raised and trained to provide additional specific benefits to humans suffering from a range of physical or mental conditions, working as assistance dogs. In this article, we employed innovative techniques to review the vast and constantly expanding literature on the subject, which covers a multitude of aspects. The 450 articles obtained through keyword search on Scopus were initially described in terms of year of publication, geographical context and publication destination, and were subsequently analysed through automated text mining to detect the most important words contained within them. Lastly, a generative model of topic analysis (Latent Dirichlet Allocation-LDA) described the content of the collection of documents, dividing it into the appropriate number of topics. The results yielded interesting insights across all domains, demonstrating the potential of automated text mining and topic analysis as a useful tool to support the researchers in dealing with complex and time-consuming subjects' reviews, integrating the work done with traditional reviewing methods.
    Keywords:  Latent Dirichlet Allocation; assistance dogs; generative model; service dogs; text mining; topic analysis
    DOI:  https://doi.org/10.3389/fvets.2024.1463332
  4. J Med Internet Res. 2024 Dec 24. 26 e59843
       BACKGROUND: Adequate health literacy has been shown to be important for the general health of a population. To address this, it is recommended that patient-targeted medical information is written at a sixth-grade reading level. To make well-informed decisions about their health, patients may want to interact directly with peer-reviewed open access scientific articles. However, studies have shown that such text is often written with highly complex language above the levels that can be comprehended by the general population. Previously, we have published on the use of large language models (LLMs) in easing the readability of patient-targeted health information on the internet. In this study, we continue to explore the advantages of LLMs in patient education.
    OBJECTIVE: This study aimed to explore the use of LLMs, specifically ChatGPT (OpenAI), to enhance the readability of peer-reviewed scientific articles in the field of ophthalmology.
    METHODS: A total of 12 open access, peer-reviewed papers published by the senior authors of this study (ET and RA) were selected. Readability was assessed using the Flesch-Kincaid Grade Level and Simple Measure of Gobbledygook tests. ChatGPT 4.0 was asked "I will give you the text of a peer-reviewed scientific paper. Considering that the recommended readability of the text is 6th grade, can you simplify the following text so that a layperson reading this text can fully comprehend it? - Insert Manuscript Text -". Appropriateness was evaluated by the 2 uveitis-trained ophthalmologists. Statistical analysis was performed in Microsoft Excel.
    RESULTS: ChatGPT significantly lowered the readability and length of the selected papers from 15th to 7th grade (P<.001) while generating responses that were deemed appropriate by expert ophthalmologists.
    CONCLUSIONS: LLMs show promise in improving health literacy by enhancing the accessibility of peer-reviewed scientific articles and allowing the general population to interact directly with medical literature.
    Keywords:  ChatGPT; LLMs; artificial intelligence; health literacy; large language models; medical information; ophthalmology; patient education; peer review; readability; uveitis
    DOI:  https://doi.org/10.2196/59843
  5. BMJ Open. 2024 Dec 23. 14(12): e079756
       INTRODUCTION: Healthcare organisations should create and disseminate readable information to citizens and patients to reduce the level of health literacy required for individuals. Several systematic reviews have reported overviews of the readability of written health information for each subject area. This systematic review of systematic reviews will synthesise evidence from existing systematic reviews to provide an overview of studies that assessed readability in diverse areas and identify gaps to be filled in future research and practice to improve the readability of health information.
    METHODS AND ANALYSIS: The following databases will be searched: PubMed, MEDLINE, CINAHL, PsycINFO, Web of Science Core Collection, Scopus, Cochrane Database of Systematic Reviews, and Google Scholar. The reference lists of the eligible studies will be searched to identify additional studies. The eligible literature will be systematic reviews of published empirical studies that quantitatively assessed the readability of health information for patients and/or citizens. Two independent reviewers will screen the titles, abstracts and full texts of all identified literature. A lead researcher will extract the data, and a second reviewer will check the results. Discrepancies will be resolved through discussion, and a third reviewer will be involved if necessary. Data, such as the characteristics of the systematic review, characteristics of the reviewed literature in the systematic review, subject area, material characteristics, definition(s) of readability, readability assessment methods, main results and quality assessment, will be extracted. The findings will be synthesised in a numerical summary and a descriptive and narrative review.
    ETHICS AND DISSEMINATION: This review does not involve any ethical concerns. The results and implications of this systematic review will be disseminated at conferences and in peer-reviewed journals.
    Keywords:  Health Equity; Health Literacy; PUBLIC HEALTH; Patient-Centered Care
    DOI:  https://doi.org/10.1136/bmjopen-2023-079756
  6. J Fr Ophtalmol. 2024 Dec 20. pii: S0181-5512(24)00345-0. [Epub ahead of print]48(3): 104400
       PURPOSE: To evaluate the compatibility and readability of ChatGPT-4 in providing responses to common inquiries about strabismus and amblyopia.
    MATERIALS AND METHODS: A series of commonly asked questions were compiled, covering topics such as the definition, prevalence, diagnostic approaches, surgical and non-surgical treatment alternatives, postoperative guidelines, surgery-related risks, and visual prognosis associated with strabismus and amblyopia. Each question was asked three times on the online ChatGPT-4 platform both in English and French, with data collection conducted on February 18, 2024. The responses generated by ChatGPT-4 were evaluated by two independent pediatric ophthalmologists, who classified them as "acceptable," "unacceptable," or "incomplete." Additionally, an online readability assessment tool called "readable" was utilized for readability analysis.
    RESULTS: The majority of responses, totaling 97% of the questions regarding strabismus and amblyopia, consistently met the criteria for acceptability. Only 3% of responses were classified as incomplete, with no instances of unacceptable responses observed. The average Flesch-Kincaid Grade Level and Flesch Reading Ease Score were calculated as 14.53±1.8 and 23.63±8.2, respectively. Furthermore, the means for all readability indices, including the Coleman-Liau index, the Gunning Fog index, and the SMOG index, were found to be 15.75±1.4, 16.96±2.4, and 16.05±1.6, respectively.
    CONCLUSIONS: ChatGPT-4 consistently produced acceptable responses to the majority of the questions asked (97%). Nevertheless, the readability of these responses proved challenging for the average layperson, requiring a college-level education for comprehension. Further improvements, particularly in terms of readability, are necessary to enhance the advisory capacity of this AI software in providing eye and health-related guidance for patients, physicians, and the general public.
    Keywords:  Amblyopia; Amblyopie; Artificial intelligence; ChatGPT-4; Intelligence artificielle; Strabisme; Strabismus
    DOI:  https://doi.org/10.1016/j.jfo.2024.104400
  7. Cureus. 2024 Nov;16(11): e74398
       OBJECTIVE: The study aims to compare ChatGPT and Google Gemini-generated patient education guides regarding claustrophobia during MRI, mammography screening, and MR safe and unsafe items and the importance of knowing what items can be carried into an MR room.
    METHODS: The study utilized ChatGPT 3.5 and Google Gemini to create patient education guides concerning claustrophobia during MRI, mammography screening, and MR safe and unsafe items. A Flesch-Kincaid calculator was used to evaluate readability and ease of understanding. QuillBot (QuillBot, Inc., Chicago, USA) was used to generate a similarity score to evaluate possible plagiarism. In order to assess the scientific reliability of the AI-generated responses, we utilized a modified DISCERN score. R Studio 4.3.2 (The R Foundation for Statistical Computing, Vienna, Austria) was used for statistical analyses, with unpaired t-tests used to determine statistical significance between variables.
    RESULTS: The average number of words in ChatGPT and Google Gemini were 468.7±132.07 and 328.7±163.65, respectively. The mean number of sentences was 35.67±18.15 for ChatGPT and 30.33±12.22 for Google Gemini. Ease of readability for ChatGPT responses was 36.30±7.01 and for Google Gemini 46.77±4.96. The similarity scores for the ChatGPT responses were 0.50±0.62 and for Google Gemini 9.43±6.20. The reliability score was evaluated at 2.67±0.25 for ChatGPT and 2.67± 0.58 for Google Gemini.
    CONCLUSION: The AI generated by ChatGPT and Google Gemini had no statistically significant difference in regard to word count, average word per sentence, average syllables per word, grade level comprehension score, or scientific reliability. However, the ease score was significantly greater for the ChatGPT response compared to Google Gemini. In addition, the similarity score was much higher in Google Gemini than in ChatGPT responses.
    Keywords:  artificial intelligence; chatgpt; claustrophobia; google gemini; magnetic resonance; mammography; mr; mr safety; mri; patient education brochure
    DOI:  https://doi.org/10.7759/cureus.74398
  8. Eur Arch Otorhinolaryngol. 2024 Dec 26.
       PURPOSE: The artificial intelligence (AI) chatbot ChatGPT has become a major tool for generating responses in healthcare. This study assessed ChatGPT's ability to generate French preoperative patient-facing medical information (PFI) in rhinology at a comparable level to material provided by an academic source, the French Society of Otorhinolaryngology (Société Française d'Otorhinolaryngologie et Chirurgie Cervico-Faciale, SFORL).
    METHODS: ChatGPT and SFORL French preoperative PFI in rhinology were compared by analyzing responses to 16 questions regarding common rhinology procedures: ethmoidectomy, sphenoidotomy, septoplasty, and endonasal dacryocystorhinostomy. Twenty rhinologists assessed the clarity, comprehensiveness, accuracy, and overall quality of the information, while 24 nonmedical individuals analyzed the clarity and overall quality. Six readability formulas were used to compare readability scores.
    RESULTS: Among rhinologists, no significant difference was found between ChatGPT and SFORL regarding clarity (7.61 ± 0.36 vs. 7.53 ± 0.28; p = 0.485), comprehensiveness (7.32 ± 0.77 vs. 7.58 ± 0.50; p = 0.872), and accuracy (inaccuracies: 60% vs. 40%; p = 0.228), respectively. Non-medical individuals scored the clarity of ChatGPT significantly higher than that of the SFORL (8.16 ± 1.16 vs. 6.32 ± 1.33; p < 0.0001). The non-medical individuals chose ChatGPT as the most informative source significantly more often than rhinologists (62.8% vs. 39.7%, p < 0.001).
    CONCLUSION: ChatGPT-generated French preoperative PFI in rhinology was comparable to SFORL-provided PFI regarding clarity, comprehensiveness, accuracy, readability, and overall quality. This study highlights ChatGPT's potential to increase accessibility to high quality PFI and suggests its use by physicians as a complement to academic resources written by learned societies such as the SFORL.
    Keywords:  Accuracy; Artificial intelligence; Preoperative information; Readability; Rhinology
    DOI:  https://doi.org/10.1007/s00405-024-09180-0
  9. Eur J Ophthalmol. 2024 Dec 24. 11206721241306142
       INTRODUCTION: Keratoconus patients turn to the internet for answers to their disease expectations. Webpages are not filtered or submitted to evaluation before getting published. We aim to evaluate the quality and readability of the online information regarding keratoconus in Portugal and Brazil.
    METHODS: Two independent ophthalmologists and one ophthalmologist supervisor evaluated 30 Portuguese and 30 Brazilian websites by order of appearance in Google with the word "Queratocone" and "Ceratocone", respectively. Two quality scores were used: a quality index of consumer health information (DISCERN) and the Journal of the American Medical Association (JAMA) benchmark. Readability was evaluated with 3 scores: FleschKincaid Reading Ease (FRE), FleschKincaid Grade (FKG) and Automated Readability Index (ARI).
    RESULTS: Sites from private hospitals or clinics were the most prevalent in both countries, followed by health platform sites. Final JAMA benchmark was 1.13 ± 1.18 in Portugal and 1.07 ± 1.00 in Brazil. Final DISCERN was 34.07 ± 11.71 in Portugal and 38.17 ± 10.51 in Brazil. FRE and FKG scores denoted "difficult to read" and "college school level" in both countries; ARI denoted "professor" level in Portugal and "college student" level in Brazil needed to understand the text, a statistically significant difference. There was no correlation between Google ranking and quality and readability scores.
    DISCUSSION AND CONCLUSIONS: The information on keratoconus available online to Portuguese-speaking patients is of poor quality and difficult to interpret. Ophthalmologists have a shared responsibility to tackle this challenge through multifold efforts. Educating our patients on how to find reputable websites can help them navigate their life with keratoconus.
    Keywords:  Keratoconus; online information; quality scores; readability
    DOI:  https://doi.org/10.1177/11206721241306142
  10. J Am Acad Orthop Surg. 2024 Dec 18.
       INTRODUCTION: The importance of web resources for educating orthopaedic trauma patients is well recognized. Yet these resources often exceed the sixth-grade reading level and are too complex for most patients to understand. This study examines the readability of the American Academy of Orthopaedic Surgeons' (AAOS) pediatric trauma-related educational content compared with similar materials from the Pediatric Orthopaedic Society of North America (POSNA).
    METHODS: Eleven AAOS (https://www.orthoinfo.org/) and nine POSNA (https://orthokids.org) pediatric trauma education articles were included. Articles' readability was assessed by two independent examiners using (1) the Flesch-Kincaid Grade Level (FKGL) and the Flesch Reading Ease (FRE) algorithms. The FKGL mean was compared with the sixth-grade level and the average American reading level using a one-sample t-test. A two-sample t-test evaluated the differences in readability between the AAOS and POSNA materials.
    RESULTS: A total of 15 unique articles were included. Excellent agreement (>90%) was noted between reviewers for FKGL and FRE. The average (SD) FKGL for AAOS and POSNA articles were 8.5 (0.8) and 9.0 (1.5), respectively, and the FRE scores were 61.2 (3.8) for AAOS and 61.7 (7.7) for POSNA. Only one AAOS article met the sixth-grade reading level. The average readability levels of the AAOS and POSNA articles were markedly higher than the recommended sixth-grade level (P < 0.001). No notable difference was found in FKGL (P = 0.47) or FRE (P = 0.89) when comparing AAOS and POSNA articles.
    DISCUSSION: This study indicates that the readability of the AAOS pediatric trauma and POSNA trauma-related articles is well above that which is recommended for the general public, and this has remained so for more than a decade. Both POSNA and AAOS trauma education materials have high readability levels, potentially hindering patient understanding. Improving the readability of these widely used trauma education resources is overdue and will likely improve patient comprehension.
    DOI:  https://doi.org/10.5435/JAAOS-D-24-00617
  11. Int Orthod. 2024 Dec 19. pii: S1761-7227(24)00115-3. [Epub ahead of print]23(1): 100959
       INTRODUCTION: Orthognathic surgery combined with orthodontic treatment is commonly undertaken to address facial and dental disharmony. Content uploaded onto social media platforms can be an important source of healthcare information. This scoping review aimed to collate and synthesize data from studies that investigated the quality of information communicated through these platforms.
    METHODS: Five databases and grey literature were searched, with no restrictions on location and year published. The PCC (Population, Concept, Context) framework was adopted to select eligible studies. Inclusion criteria included literature that assessed the quality of information on orthognathic surgery in relation to orthodontic treatment in the context of social media.
    RESULTS: A total of 12 cross-sectional studies satisfied the selection criteria. The most investigated social media platform was YouTube. The quality of information on all platforms examined was found to be poor to moderate. The most popular content featured patients sharing their treatment experiences, but these often lacked clinically relevant information, focusing mainly on satisfaction with treatment, outcomes and their supervising healthcare professionals. In contrast, healthcare professionals provided content that contained better quality information but had lower views and engagement. The least frequently mentioned topics were found to be complications and their duration, costs, and postoperative expectations. Certain search terms such as "#jawsurgery", "#surgeryfirst" and "#genioplasty" appeared to resonate more with the public. There was heterogeneity across the methodologies and quality of information instruments used.
    CONCLUSIONS: Information communicated through social media platforms regarding orthognathic surgery was of poor to moderate quality. Orthodontic professionals should be aware of the limitations of health information on these online platforms and guide individuals to more accurate sources.
    Keywords:  Jaw surgery; Orthodontics; Orthognathic surgery; Patient information; Social media; YouTube
    DOI:  https://doi.org/10.1016/j.ortho.2024.100959
  12. Med Acupunct. 2024 Dec;36(6): 350-358
       Background: In recent years, there has been a notable increase in the number of patients seeking information from online health websites. As the information available on these websites can significantly impact the overall health of individuals in a society, it is vital for online health information to be presented in a manner that is readable and credible to the general public. To address this concern, the objective of the study was to examine and assess the credibility and readability of websites about acupuncture as a pain management approach.
    Methods: The research population was represented by websites containing information about pain management using acupuncture found through Google, Yahoo, and Bing search engines and with the keywords acupuncture and pain. The websites were evaluated in terms of confirming HON and JAMA criteria. FRE, GFI, FKGL, CLI, SMOG, and ARI indexes were used to assess their readability, and Mann-Whitney and Kruskal-Wallis tests were used to compare the data.
    Results: This study evaluated 52 websites that appeared on the first three pages of search engine results. Out of the 52 websites analyzed, 40 websites (76.9%) met the validity criteria set by JAMA, while only 21 websites (40.4%) were deemed valid based on the HONcode standards. According to the average readability index (12.38), the contents presented on the websites were at the Difficult to Read level. The mean readability scores of websites across various domains exhibit no statistically significant difference (p > 0.05).
    Conclusion: The majority of websites that provide acupuncture-related information do meet the verification set by HONcode. Additionally, the scientific and educational content on these websites tends to be more complex than what is recommended by the AMA and NIH. This is concerning given the increasing popularity of online content and the impact it has on public health. Providers of this information must prioritize making it readable to a wide audience.
    Keywords:  Acupuncture; Public health informatics; health website; internet; patient education; readability
    DOI:  https://doi.org/10.1089/acu.2023.0136
  13. Front Psychiatry. 2024 ;15 1408384
       Objectives: We aimed to assess the quality of information regarding depression on Chinese websites and popular video platforms.
    Methods: We conducted searches on website platforms (Baidu, Bing) and video platforms (Bilibili, Douyin) using search terms "depression", "depressive disorder", "depression treatment", "depressive anxiety", "depressed patient", and "depressive symptoms". We collected the first 50 results with each search term in each platform. Each website and video included in this study was assessed using The DISCERN instrument (DISCERN), Journal of American Medical Association benchmark criteria (JAMA), Hexagonal Radar Schema (HRS), and Global Quality Scores (GQS).
    Results: A total of 177 websites, 216 Bilibili videos, and 244 Douyin videos were included. Among all the platforms, websites had the highest median scores on DISCERN and HRS, at 33 and 2 respectively, but were still classified as "poor" and "very poor" according to the classification. Bilibili, Douyin, and websites had median scores of 3, 2, and 2 respectively in JAMA, indicating a moderate level of quality. Bilibili, Douyin, and websites all had a median score of 2 for GQS, and were of poor quality. Only the percentage score for JAMA was more than half of the weighted score, while none of the other scales reached half of the score. The median percentage scores of the websites in DISCERN, HRS, and GQS were higher than those of Bilibili and Douyin (P < 0.001). Compared to other sources, news media on websites and medical organizations on videos demonstrated higher quality (all P values < 0.05).
    Conclusions: The findings of the study indicated an overall low quality of online depression information. Collaborative efforts between platforms and professionals are necessary to improve the comprehensiveness and quality of depression-related information available online. In addition, the platform needs to prioritize optimizing the algorithm of recommendations and present real high-quality health information to the audience.
    Keywords:  Chinese videos; Chinese websites; depression; information quality; internet
    DOI:  https://doi.org/10.3389/fpsyt.2024.1408384
  14. J Pediatr Urol. 2024 Dec 17. pii: S1477-5131(24)00671-5. [Epub ahead of print]
       INTRODUCTION: Posterior urethral valve (PUV) is a congenital condition marked by obstructing persistent urogenital membrane, leading to urinary tract infections, bladder dysfunction, and kidney damage. It affects males only, mostly suspected antenatally and confirmed in early infancy. It requires early diagnosis and intervention to prevent long-term complications. Effective communication of follow-up and treatment protocols with families is crucial, as informed patients play an active role in decision-making and experience less anxiety.
    METHODS: A cross-sectional study was conducted on June 20, 2024, by searching YouTube with the keyword "posterior urethral valve." From the first 120 videos, 60 were excluded for being off-topic, duplicates, non-English, or lacking audio. The remaining 60 videos were analyzed for views, duration, comments, likes, and upload source. Video content was categorized into general information, ultrasound imaging, case discussion, and surgical teaching. Two independent urology specialists evaluated the video quality using DISCERN and JAMA Benchmark Criteria.
    RESULTS: Of the 60 analyzed videos, 71.6 % were uploaded by physicians. The most common content was general information on symptoms and treatment. The mean JAMA score was 2.45, and the mean DISCERN score was 45.8. According to DISCERN, 8.3 % of videos were "very poor," 25 % were "poor," 30 % were "fair," 11.6 % were "good," and 23 % were "excellent." Videos uploaded by physicians had significantly higher DISCERN reliability and total scores, indicating better quality. No significant correlation was found between DISCERN scores and likes, comments, or views, but video duration was positively correlated with higher scores.
    DISCUSSION: Among the analyzed videos, those uploaded by physicians had significantly higher DISCERN total scores, emphasizing their superior quality compared to other sources. Video duration positively correlated with DISCERN and JAMA scores, while metrics such as likes, comments, and views did not reflect content quality.
    CONCLUSION: Creating evidence-based, clear, and engaging content that follows DISCERN and JAMA criteria is essential. Future studies should explore animated formats, diverse keywords, and track content trends over time.
    Keywords:  DISCERN score; Posterior urethral valve; Youtube
    DOI:  https://doi.org/10.1016/j.jpurol.2024.12.011
  15. Int J Med Inform. 2024 Dec 22. pii: S1386-5056(24)00439-8. [Epub ahead of print]195 105776
       BACKGROUND: Usefulness of hysteroscopic metroplasty to improve reproductive outcomes is controversial and debated among reproductive specialists and, consequently, patients.
    METHODS: We performed a cross-sectional analysis to assess the quality, reliability, and level of misinformation in YouTube, Instagram, and TikTok videos about hysteroscopic metroplasty. Videos on each social network retrieved using "hysteroscopy" and "septate uterus" or "uterine septum" as keywords were assessed using Patient Education Materials Assessment Tool for audio-visual (PEMAT A/V) content, the modified DISCERN (mDISCERN), Global Quality Scale (GQS), Video Information and Quality Index (VIQI) and Misinformation assessment.
    RESULTS: One hundred fifteen videos (47 from YouTube, 35 from TikTok, and 33 from Instagram) were included. Median PEMAT A/V Understandability and Actionability scores were 50.0 % (IQR: 0-63.3) and 20 % (IQR: 0.0-50.0). Actionability was higher YouTube relative to TikTok and Instagram (50.0 % [IQR: 0-75.0] vs. 0 % [IQR: 0-66.0]; p < 0.001 and 0 % [IQR: 0-33.0], p < 0.001), with no significant differences for Understandability. Median mDISCERN was 3 (IQR: 0-3), with higher score for YouTube compared to TikTok and Instagram (3 [IQR: 0-3] vs. 0 [IQR: 0-23], p < 0.001 and 1 [IQR 0-3.0], p < 0.001 respectively). VIQI showed low-level content (median 9; IQR: 4-19) and no differences among social media platforms. GQS was low (median 2 [IQR 1-4.5]), with no differences in subgroup analysis. Videos on YouTube (97.9 %) and TikTok (62.9 %) favored hysteroscopic metroplasty, while Instagram content was against in 63.6 % of cases (p < 0.001). While most videos were found to be either completely or partially accurate, significant differences in misinformation levels were observed across platforms, with YouTube exhibiting the lowest rate of inaccuracies compared to TikTok and Instagram. (8.5 % vs. 42.9 % and 36.4 %; p < 0.001, respectively).
    CONCLUSION: Quality of social networks' videos on hysteroscopic metroplasty seems unsatisfactory, but with low levels of misinformation. YouTube and TikTok were more favorable than Instagram regarding the usefulness of metroplasty. In contrast, a favorable opinion regarding its efficacy in fertility improvement was retrieved in each social.
    Keywords:  Hysteroscopy; Infertility; Internet; Misinformation; Mullerian anomalies; Social media; Uterine septum
    DOI:  https://doi.org/10.1016/j.ijmedinf.2024.105776
  16. ANZ J Surg. 2024 Dec 26.
       BACKGROUND: Laparoscopic cholecystectomy is one of the most commonly performed surgeries worldwide. With the rise of online platforms like YouTube and WebSurg, surgical videos have become popular educational tools. However, the quality of these videos varies, raising concerns about their reliability. This study evaluates the changes in the quality of laparoscopic cholecystectomy videos over two decades and compares the educational value of YouTube and WebSurg videos.
    METHODS: Twenty-four videos were analyzed and divided into four groups: older YouTube, older WebSurg, newer YouTube, and newer WebSurg. Six videos from each group were selected based on relevance and popularity. Three independent surgeons assessed them using standardized scoring systems, including the JAMA Benchmark Score, Global Quality Score (GQS), LAP-VEGaS, and a new Laparoscopic Cholecystectomy-Specific Score (LAP Chole-SS). Video characteristics such as duration, uploader type, and viewer engagement were also considered. Statistical comparisons were made between the groups.
    RESULTS: The analysis showed that although newer videos were longer, the difference was not statistically significant (P = 0.74). Newer videos had significantly better image quality (P < 0.001). WebSurg videos consistently outperformed YouTube videos across all evaluation criteria. WebSurg scored higher in the JAMA Benchmark (2.62 versus 1.66), GQS (3.58 versus 2.63), LAP-VEGaS (13.7 versus 6.58), and LAP Chole-SS (4.08 versus 2.33).
    CONCLUSION: In conclusion, WebSurg provides higher-quality, peer-reviewed content, making it more reliable for surgical education, while YouTube offers accessibility but often lacks educational rigor.
    Keywords:  cholecystectomy; education; laparoscopic surgery; video; youTube
    DOI:  https://doi.org/10.1111/ans.19359
  17. Cureus. 2024 Dec;16(12): e76102
      Objective Proper preparation is necessary before performing certain procedures on a patient. Recently, videos created using social media content have been used as a preparation method. This has become particularly important in medical education since the COVID-19 pandemic. Central venous catheter placement is an important procedure performed in critically ill patients. This study assessed the quality of videos about central venous catheter placement available on YouTube. Methods A search was conducted on YouTube with the keyword "central venous catheter placement" on June 1, 2024. The review process involved two independent medical experts who comprehensively assessed the relevance and content of each video. To assess the quality of the videos, the DISCERN score, mDISCERN score, Global Quality Score (GQS), and American Medical Association Journal (JAMA) rating scale were used as evaluation tools. Results Top 41 English-language videos deemed most relevant were evaluated, each with over 25,000 views. Analysis of the video sources revealed that 36.6% of the videos were uploaded by doctors. Moreover, 87.8% of the videos were audio narrations accompanied by real patients or mannequins, and the remaining videos comprised audio narration with animated images. The mean JAMA score of the videos was 1.73, the mean mDISCERN score was 2.66, the mean DISCERN score was 40.46, and the mean GQS score was 3.8. A statistically significant difference was observed between the mDISCERN scores of videos containing real visual content and those containing animated visual content (p = 0.038). Moreover, the same two groups also exhibited higher quality scores in the GQS (p = 0.029). Conclusion The quality of information related to central venous catheter placement on YouTube is highly variable. Specifically, videos that provide valuable information and those that have the potential to mislead viewers do not have a noticeable difference in terms of views and popularity. For a medical practitioner or physician seeking reliable information, a useful and safe approach is preferring videos uploaded by medical professionals. It is important to prioritize the professional identity of the content creator rather than the video's popularity or the number of comments it has received.
    Keywords:  central line placement; medicine education; quality; video analysis; youtube
    DOI:  https://doi.org/10.7759/cureus.76102
  18. Ann Plast Surg. 2024 Dec 18.
       INTRODUCTION: YouTube has become a popular source of health information, including plastic surgery. Given the platform's wide reach and potential influence on patient decisions, this study aimed to assess the quality of information available on YouTube for African audiences seeking plastic surgery procedures.
    METHODS: This cross-sectional study extracted data from YouTube videos on plastic surgery relevant to Africa. A search strategy identified videos in English using keywords. The first 50 results for each term were included, with duplicates removed. Next, the metadata of videos published from inception to June 9, 2024, were extracted. Two reviewers independently assessed videos using standardized tools to evaluate reliability (modified DISCERN and JAMA criteria) and engagement (likes-to-views ratio [LVR] and comments-to-views ratio [CVR]). The Mann-Whitney U test was used for unadjusted bivariable comparisons. Then ordinal logistic and beta regression analyses were used to evaluate the primary (modified DISCERN and JAMA scores) and secondary (LVR and CVR) outcomes, with a statistical significance level set at 0.05.
    RESULTS: Eight hundred ninety-seven plastic surgery videos were analyzed, and 3.9% were published by African entities. Large subscriber count (coefficient = -6.9e-8, 95% confidence interval [CI] [-1.13e-7, -2.9e-8], P = 0.001), African-authored (coefficient = -0.85, 95% CI [-1.44, -0.25], P = 0.005), and advertising (coefficient = -1.01, 95% CI [-1.63, -0.57], P < 0.001) videos had lower modified DISCERN scores. Advertising videos equally had lower JAMA scores (coefficient = -1.29, 95% CI [-1.83, -0.74], P < 0.001). Academic videos had lower LVR (coefficient = -0.48, 95% CI [-0.66, -0.30], P < 0.001), whereas independent videos had higher LVR (coefficient = 0.40, 95% CI [0.26, 0.54], P < 0.001). Academic videos had lower CVR (coefficient = -0.40, 95% CI [-0.67, -0.13], P = 0.003), whereas videos with other purposes had higher CVR (coefficient = 0.37, 95% CI [0.10, 0.64], P = 0.007).
    CONCLUSIONS: This study underscores a potential disparity in the quality of online plastic surgery information based on video sources and purposes.
    DOI:  https://doi.org/10.1097/SAP.0000000000004186
  19. Cleft Palate Craniofac J. 2024 Dec 22. 10556656241297557
      To analyze the content and sources of TikTok posts regarding cleft conditions; to examine the quality and reliability of educational content related to cleft conditions.
    Cross-sectional, observational study of cleft-related TikTok content.
    N/A.
    N/A.
    N/A .
    Frequencies of TikTok post likes, views, comments, shares, and saves; engagement rate; post themes; and post educational content accuracy, quality, and reliability (assessed with the Patient Education Materials Assessment Tool and modified DISCERN [mDISCERN] rating scales).
    A total of 252 cleft-related TikTok posts were reviewed. Most posts were created by family members of patients (n = 193, 76.59%), while posts from plastic surgeons were underrepresented (n = 11, 4.37%). Positive Messages were the most common video theme (n = 80; 31.75%), and 24 (9.52%) of the TikTok posts were classified as providing educational content. Plastic surgeons had the highest mDISCERN score (2.75) among all content creator categories, indicating more accurate and reliable educational content. When content creator categories were stratified by physician versus nonphysician creators, physician posts had significantly higher mDISCERN scores (P = .001). Despite objectively better educational content, plastic surgeons had the lowest engagement rate (1.82%), while family members had the highest engagement rate (7.04%).
    Cleft teams should anticipate that patients and their caregivers will most likely access social media to learn more about cleft conditions and seek support. Patients and caregivers should be urged to stay cautious of the information they find online and ideally be offered reputable sources of information and support.
    Keywords:  TikTok; cleft lip and palate; patient education; social media
    DOI:  https://doi.org/10.1177/10556656241297557
  20. JMIR Cancer. 2024 Dec 24. 10 e55300
       Background: Following medical cannabis legalization in Thailand in 2019, more people are seeking medical cannabis-related information, including women living with breast cancer. The extent to which they access cannabis-related information from internet sources and social media platforms and the quality of such content are relatively unknown and need further evaluation.
    Objective: This study aims to analyze the factors determining cannabis-related content quality for breast cancer care from internet sources and on social media platforms and examine the characteristics of such content accessed and consumed by Thai breast cancer survivors.
    Methods: A mixed methods study was conducted between January 2021 and May 2022, involving a breast cancer survivor support group. The group identified medical cannabis-related content from frequently accessed internet sources and social media platforms. The contents were categorized based on content creators, platforms, content category, and upload dates. Four researchers used the Quality Evaluation Scoring Tool (QUEST) to assess content quality, with scores ranging from 0 to 28. Contents were expert-rated as either high or poor. The QUEST interobserver reliability was analyzed. Receiver-operating characteristic curve analysis with the Youden index was used to determine the QUEST score cut-off point. Statistical significance was set at P<.05. Fairclough Critical Discourse Analysis was undertaken to examine the underlying discourses around poor-quality content.
    Results: Sixty-two Thai-language cannabis-related items were evaluated. The content sources were categorized as follows: news channels (21/62, 34%), government sources (16/62, 26%), health care providers (12/62, 19%), and alternative medicine providers (12/62, 19%). Most of the contents (30/62, 48%) were uploaded to YouTube, whereas 31% (19/62) appeared on websites and Facebook. Forty of 62 content items (64%) were news-related and generic cannabis advertisements while 8 of 62 (13%) content items had no identifiable date. The interobserver QUEST score correlation was 0.86 (P<.001). The mean QUEST score was 12.1 (SD 7.6). Contents were considered "high" when the expert rating was >3. With a QUEST score of 15 as the threshold, the sensitivity and specificity for differentiating between high and poor content quality were 81% and 98%, respectively. Content creation was the only significant factor between high- and poor-quality content. Poor-quality contents were primarily created by alternative medicine providers and news channels. Two discourses were identified: advocacy for cannabis use normalization and cannabis romanticization as a panacea. These discourses overly normalize and romanticize the use of cannabis, focusing on indications and instructions for cannabis use, and medical cannabis promotion, while neglecting discussions on cannabis contraindications and potential side effects.
    Conclusions: The varying quality of medical cannabis-related information on internet sources and social media platforms accessed and shared by Thai breast cancer survivors is an issue of concern. Given that content creators are the sole predictive factors of high content quality, future studies should examine a wider range of cannabis-related sources accessible to both the public and patients to gain a more comprehensive understanding of the issue.
    Keywords:  Thailand; breast cancer; cannabis; critical discourse analysis; digital literacy; health literacy; legislation; medical cannabis; mixed method study
    DOI:  https://doi.org/10.2196/55300
  21. Sci Rep. 2024 Dec 28. 14(1): 31080
       BACKGROUND: Understanding the impact of digital health literacy and health information-seeking behavior on the self-perceived health and depression symptoms of older adults is crucial, particularly as the number of older internet users is increasing.
    METHODS: This study utilized data from the Health Information National Trends Survey to examine the relationship between these factors and the health outcomes of adults aged 50 and above.
    RESULTS: The study found that digital health literacy has a positive but non-significant relationship with self-perceived health when other factors are considered. However, education level and body mass index consistently predicted self-perceived health. Moreover, higher digital health literacy was associated with a reduced likelihood of perceived depression symptoms, even after adjusting for demographic and health-related factors.
    CONCLUSIONS: These findings highlight the importance of digital health literacy in the mental well-being of older adults and provide insights for shaping future health policies and interventions.
    Keywords:  Depression symptoms; Digital health literacy; Health information seeking behavior; Older adults; Self-perceived health
    DOI:  https://doi.org/10.1038/s41598-024-82187-z