bims-librar Biomed News
on Biomedical librarianship
Issue of 2024‒03‒03
twenty-two papers selected by
Thomas Krichel, Open Library Society



  1. ArXiv. 2024 Jan 19. pii: arXiv:2401.11048v1. [Epub ahead of print]
      PubTator 3.0 (https://www.ncbi.nlm.nih.gov/research/pubtator3/) is a biomedical literature resource using state-of-the-art AI techniques to offer semantic and relation searches for key concepts like proteins, genetic variants, diseases, and chemicals. It currently provides over one billion entity and relation annotations across approximately 36 million PubMed abstracts and 6 million full-text articles from the PMC open access subset, updated weekly. PubTator 3.0's online interface and API utilize these precomputed entity relations and synonyms to provide advanced search capabilities and enable large-scale analyses, streamlining many complex information needs. We showcase the retrieval quality of PubTator 3.0 using a series of entity pair queries, demonstrating that PubTator 3.0 retrieves a greater number of articles than either PubMed or Google Scholar, with higher precision in the top 20 results. We further show that integrating ChatGPT (GPT-4) with PubTator APIs dramatically improves the factuality and verifiability of its responses. In summary, PubTator 3.0 offers a comprehensive set of features and tools that allow researchers to navigate the ever-expanding wealth of biomedical literature, expediting research and unlocking valuable insights for scientific discovery.
  2. Behav Anal Pract. 2024 Mar;17(1): 347-358
      Incorporating literature into practice can help behavior analysts provide better services and achieve better outcomes. In addition, behavior analysts have an ethical obligation to remain current with the scholarly literature and to use it to inform services. Despite the merits of maintaining regular contact with the published literature, barriers exist to doing so. In this tutorial, we present a system that was created for a human service agency to increase practitioner access to the scholarly literature. The system consisted of an electronic search request form, a literature team, and a liaison. We present 7 years of data including the frequency of use, topics of interest, and other noteworthy patterns of submitter responding. We discuss the value of this type of system, limitations of its design, and considerations for practitioners who may wish to implement a similar system in their agency. We discuss modifications that could be made to fit organizations of diverse sizes and with different resources, while presenting ideas for improvement and expansion of the system.
    Keywords:  Evidence-based practice; Information literacy; Literature access; Research-to-practice gap; Scientist-practitioner model
    DOI:  https://doi.org/10.1007/s40617-023-00887-w
  3. Health Info Libr J. 2024 Mar;41(1): 1-3
      In this editorial, Anthea Sutton and Veronica Parisi reflect on ChatGPT, how it may contribute to systematic searching, and provide their overview of some recent training they attended on ChatGPT, AI and systematic literature reviews.
    Keywords:  artificial intelligence; automation; database searching; literature searching
    DOI:  https://doi.org/10.1111/hir.12517
  4. Front Public Health. 2024 ;12 1310437
      Artificial intelligence (AI) chatbots have the potential to revolutionize online health information-seeking behavior by delivering up-to-date information on a wide range of health topics. They generate personalized responses to user queries through their ability to process extensive amounts of text, analyze trends, and generate natural language responses. Chatbots can manage infodemic by debunking online health misinformation on a large scale. Nevertheless, system accuracy remains technically challenging. Chatbots require training on diverse and representative datasets, security to protect against malicious actors, and updates to keep up-to-date on scientific progress. Therefore, although AI chatbots hold significant potential in assisting infodemic management, it is essential to approach their outputs with caution due to their current limitations.
    Keywords:  artificial intelligence; eHealth; health information management; infodemic; misinformation
    DOI:  https://doi.org/10.3389/fpubh.2024.1310437
  5. Eur Arch Otorhinolaryngol. 2024 Feb 28.
      PURPOSE: Chat generative pretrained transformer (ChatGPT) has the potential to significantly impact how patients acquire medical information online. Here, we characterize the readability and appropriateness of ChatGPT responses to a range of patient questions compared to results from traditional web searches.METHODS: Patient questions related to the published Clinical Practice Guidelines by the American Academy of Otolaryngology-Head and Neck Surgery were sourced from existing online posts. Questions were categorized using a modified Rothwell classification system into (1) fact, (2) policy, and (3) diagnosis and recommendations. These were queried using ChatGPT and traditional web search. All results were evaluated on readability (Flesch Reading Ease and Flesch-Kinkaid Grade Level) and understandability (Patient Education Materials Assessment Tool). Accuracy was assessed by two blinded clinical evaluators using a three-point ordinal scale.
    RESULTS: 54 questions were organized into fact (37.0%), policy (37.0%), and diagnosis (25.8%). The average readability for ChatGPT responses was lower than traditional web search (FRE: 42.3 ± 13.1 vs. 55.6 ± 10.5, p < 0.001), while the PEMAT understandability was equivalent (93.8% vs. 93.5%, p = 0.17). ChatGPT scored higher than web search for questions the 'Diagnosis' category (p < 0.01); there was no difference in questions categorized as 'Fact' (p = 0.15) or 'Policy' (p = 0.22). Additional prompting improved ChatGPT response readability (FRE 55.6 ± 13.6, p < 0.01).
    CONCLUSIONS: ChatGPT outperforms web search in answering patient questions related to symptom-based diagnoses and is equivalent in providing medical facts and established policy. Appropriate prompting can further improve readability while maintaining accuracy. Further patient education is needed to relay the benefits and limitations of this technology as a source of medial information.
    Keywords:  Accessibility; Accuracy; ChatGPT; Large language model; Patient education; Patient questions; Readability
    DOI:  https://doi.org/10.1007/s00405-024-08524-0
  6. Clin Spine Surg. 2024 Feb 20.
      STUDY DESIGN: Retrospective Observational Study.OBJECTIVE: The objective of this study was to assess the utility of ChatGPT, an artificial intelligence chatbot, in providing patient information for lumbar spinal fusion and lumbar laminectomy in comparison with the Google search engine.
    SUMMARY OF BACKGROUND DATA: ChatGPT, an artificial intelligence chatbot with seemingly unlimited functionality, may present an alternative to a Google web search for patients seeking information about medical questions. With widespread misinformation and suboptimal quality of online health information, it is imperative to assess ChatGPT as a resource for this purpose.
    METHODS: The first 10 frequently asked questions (FAQs) related to the search terms "lumbar spinal fusion" and "lumbar laminectomy" were extracted from Google and ChatGPT. Responses to shared questions were compared regarding length and readability, using the Flesch Reading Ease score and Flesch-Kincaid Grade Level. Numerical FAQs from Google were replicated in ChatGPT.
    RESULTS: Two of 10 (20%) questions for both lumbar spinal fusion and lumbar laminectomy were asked similarly between ChatGPT and Google. Compared with Google, ChatGPT's responses were lengthier (340.0 vs. 159.3 words) and of lower readability (Flesch Reading Ease score: 34.0 vs. 58.2; Flesch-Kincaid grade level: 11.6 vs. 8.8). Subjectively, we evaluated these responses to be accurate and adequately nonspecific. Each response concluded with a recommendation to discuss further with a health care provider. Over half of the numerical questions from Google produced a varying or nonnumerical response in ChatGPT.
    CONCLUSIONS: FAQs and responses regarding lumbar spinal fusion and lumbar laminectomy were highly variable between Google and ChatGPT. While ChatGPT may be able to produce relatively accurate responses in select questions, its role remains as a supplement or starting point to a consultation with a physician, not as a replacement, and should be taken with caution until its functionality can be validated.
    DOI:  https://doi.org/10.1097/BSD.0000000000001582
  7. J Am Med Inform Assoc. 2024 Feb 29. pii: ocae037. [Epub ahead of print]
      OBJECTIVE: Most existing fine-tuned biomedical large language models (LLMs) focus on enhancing performance in monolingual biomedical question answering and conversation tasks. To investigate the effectiveness of the fine-tuned LLMs on diverse biomedical natural language processing (NLP) tasks in different languages, we present Taiyi, a bilingual fine-tuned LLM for diverse biomedical NLP tasks.MATERIALS AND METHODS: We first curated a comprehensive collection of 140 existing biomedical text mining datasets (102 English and 38 Chinese datasets) across over 10 task types. Subsequently, these corpora were converted to the instruction data used to fine-tune the general LLM. During the supervised fine-tuning phase, a 2-stage strategy is proposed to optimize the model performance across various tasks.
    RESULTS: Experimental results on 13 test sets, which include named entity recognition, relation extraction, text classification, and question answering tasks, demonstrate that Taiyi achieves superior performance compared to general LLMs. The case study involving additional biomedical NLP tasks further shows Taiyi's considerable potential for bilingual biomedical multitasking.
    CONCLUSION: Leveraging rich high-quality biomedical corpora and developing effective fine-tuning strategies can significantly improve the performance of LLMs within the biomedical domain. Taiyi shows the bilingual multitasking capability through supervised fine-tuning. However, those tasks such as information extraction that are not generation tasks in nature remain challenging for LLM-based generative approaches, and they still underperform the conventional discriminative approaches using smaller language models.
    Keywords:  biomedical multitasking; large language model; natural language processing; supervised fine-tuning
    DOI:  https://doi.org/10.1093/jamia/ocae037
  8. Hand (N Y). 2024 Feb 27. 15589447241232095
      BACKGROUND: The National Institutes of Health (NIH) and the American Medical Association (AMA) recommend a sixth-grade reading level for patient-directed content. This study aims to quantitatively evaluate the readability of online information sources related to carpal tunnel surgery using established readability indices.METHODS: Web searches for "carpal tunnel release" and "carpal tunnel decompression surgery" queries were performed using Google, and the first 20 websites were identified per query. WebFX online software tools were utilized to determine readability. Indices included Flesch Kincaid Reading Ease, Flesch Kincaid Grade Level, Coleman Liau Index, Automated Readability Index, Gunning Fog Score, and the Simple Measure of Gobbledygook Index. Health-specific clickthrough rate (CTR) data were used in order to select the first 20 search engine results page from each query.
    RESULTS: "Carpal tunnel release" had a mean readability of 8.46, and "carpal tunnel decompression surgery" had a mean readability of 8.70. The range of mean readability scores among the indices used for both search queries was 6.17 to 14.0. The total mean readability for carpal tunnel surgery information was found to be 8.58. This corresponds to approximately a ninth-grade reading level in the United States.
    CONCLUSION: The average readability of carpal tunnel surgery online content is three grade levels above the recommended sixth-grade level for patient-directed materials. This discrepancy indicates that existing online materials related to carpal tunnel surgery are more difficult to understand than the standards set by NIH and AMA.
    Keywords:  carpal tunnel surgery; carpal tunnel syndrome; diagnosis; health literacy; nerve; patient education; readability; research and health outcomes; specialty; surgery
    DOI:  https://doi.org/10.1177/15589447241232095
  9. J Nerv Ment Dis. 2024 Jan 25.
      ABSTRACT: To our knowledge, this is the first study that specifically aims to assess the readability and quality of online information about schizophrenia. The analysis is performed on 93 of 195 websites that appeared in an advanced Google search of the term "schizophrenia" performed on a single day. The websites were categorized as commercial, nonprofit, professional, and government. The websites were evaluated using the Health on the Net Foundation (HON) code certificate, DISCERN tool, and Journal of the American Medical Association (JAMA) benchmark criteria for quality and the Flesch Reading Ease Formula, Flesch-Kincaid Grade Level (FKGL) Formula, Simple Measure of Gobbledygook, and Gunning Fog indices for readability. A total of 21.5% of all websites had a HON code certificate, 50.5% were accepted as high quality (JAMA score ≥3), and 25.8% reached the recommended readability level (FKGL ≤8). Only three websites scored at a fifth- to sixth-grade reading level. Commercial and government websites had significantly lower DISCERN scores. Commercial websites had significantly lower Flesch Reading Ease Score and FKGL score than nonprofit websites. In conclusion, the current findings indicate that the quality of online information on schizophrenia is generally acceptable, but the readability is insufficient. Website creators, physicians, and health authorities should be more sensitive to the readability of online information about schizophrenia, considering the poor cognitive capacity of the patients and the unique nature of the disease.
    DOI:  https://doi.org/10.1097/NMD.0000000000001763
  10. Dermatol Ther (Heidelb). 2024 Feb 24.
      INTRODUCTION: Patients search on the Internet for information about various medical procedures and conditions. The main aim of this study was to evaluate the readability of online health information related to atopic dermatitis (AD). Online resources are becoming a standard in facilitating shared decision-making processes. With a pipeline of new therapeutic options like immunomodulators, understanding of the complexity of AD by the patients is crucial.METHODS: The term "atopic dermatitis" translated into 23 official European Union languages was searched using the Google search engine. The first 50 records in each language were evaluated for suitability. Included materials were barrier-free, focused on patient education, and were not categorized as advertisements. Article sources were classified into four categories: non-profit, online shops, pharmaceutical companies, and dermatology clinic. Readability was assessed with Lix score.
    RESULTS: A total of 615 articles in Swedish, Spanish, Slovenian, Slovak, Romanian, Portuguese, Polish, Lithuanian, Latvian, Irish, Italian, Hungarian, Greek, German, French, Finnish, Estonian, English, Dutch, Danish, Czech, Croatian, and Bulgarian were evaluated. The overall mean Lix score was 56 ± 8, which classified articles as very hard to comprehend. Significant differences in mean Lix scores were observed across all included languages (all P < 0.001). Articles released by non-profit organizations and pharmaceutical companies had the highest readability (P < 0.001). Low readability level was correlated with high article prevalence (R2 = 0.189, P = 0.031).
    CONCLUSIONS: Although there was an abundance of online articles related to AD, the readability of the available information was low. As online health information has become essential in making shared decisions between patients and physicians, an improvement in AD-related materials is needed.
    Keywords:  Atopic dermatitis; Health information; Internet content; Patient education
    DOI:  https://doi.org/10.1007/s13555-024-01115-1
  11. J Clin Epidemiol. 2024 Feb 23. pii: S0895-4356(24)00055-6. [Epub ahead of print] 111300
      OBJECTIVE: To determine whether clinical trial register (CTR) searches can accurately identify a greater number of completed randomized clinical trials (RCTs) than electronic bibliographic database (EBD) searches for systematic reviews of interventions, and to quantify the number of eligible ongoing trials.STUDY DESIGN AND SETTING: We performed an evaluation study and based our search for RCTs on the eligibility criteria of a systematic review that focused on the underrepresentation of people with chronic kidney disease in cardiovascular RCTs. We conducted a combined search of ClinicalTrials.gov and the WHO International Clinical Trials Registry Platform (ICTRP) through the Cochrane Central Register of Controlled Trials (CENTRAL) to identify eligible RCTs registered up to June 1st, 2023. We searched Cochrane CENTRAL, EMBASE, and MEDLINE for publications of eligible RCTs published up to June 5th, 2023. Finally, we compared the search results to determine the extent to which the two sources identified the same RCTs.
    RESULTS: We included 92 completed RCTs. Of these, 81 had results available. 66 completed RCTs with available results were identified by both sources (81% agreement [95% CI 71 - 88]). We identified seven completed RCTs with results exclusively by CTR search (9% [95% CI 4 - 17]) and eight exclusively by EBD search (10% [95% CI 5 - 18]). 11 RCTs were completed but lacked results (four identified by both sources (36% [95% CI 15 - 65]), one exclusively by EBD search (9% [95% CI 1 - 38]), and six exclusively by CTR search (55% [95% CI 28 - 79]). Also, we identified 42 eligible ongoing RCTs: 16 by both sources (38% [95% CI 25 - 53]) and 26 exclusively by CTR search (62% [95% CI 47 - 75]). Lastly, we identified four RCTs of unknown status by both sources.
    CONCLUSION: CTR searches identify a greater number of completed RCTs than EBD searches. Both searches missed some included RCTs. Based on our case study, researchers (e.g. information specialists, systematic reviewers) aiming to identify all available RCTs should continue to search both sources. Once the barriers to performing CTR searches alone are targeted, CTR searches may be a suitable alternative.
    Keywords:  Clinical trial register search; clinical trial registry; information retrieval; randomized clinical trials; randomized controlled trials; systematic review
    DOI:  https://doi.org/10.1016/j.jclinepi.2024.111300
  12. EJVES Vasc Forum. 2024 ;61 62-76
      Objective: Due to its video based approach, YouTube has become a widely accessed educational resource for patients and trainees. This systematic review characterised and evaluated the peer reviewed literature investigating YouTube as a source of patient or trainee education in vascular surgery.Data sources: A comprehensive literature search was conducted using EMBASE, MEDLINE, and Ovid HealthStar from inception until 19 January 2023. All primary studies and conference abstracts evaluating YouTube as a source of vascular surgery education were included.
    Review methods: Video educational quality was analysed across several factors, including pathology, video audience, and length.
    Results: Overall, 24 studies were identified examining 3 221 videos with 123.1 hours of content and 37.1 million views. Studies primarily examined YouTube videos on diabetic foot care (7/24, 29%), peripheral arterial disease (3/24, 13%), carotid artery stenosis (3/24, 13%), varicose veins (3/24, 13%), and abdominal aortic aneurysm (2/24, 8%). Video educational quality was analysed using standardised assessment tools, author generated scoring systems, or global author reported assessment of quality. Six studies assessed videos for trainee education, while 18 studies evaluated videos for patient education. Among the 20 studies which reported on the overall quality of educational content, 10/20 studies deemed it poor, and 10/20 studies considered it fair, with 53% of studies noting poor educational quality for videos intended for patients and 40% of studies noting poor educational quality in videos intended for trainees. Poor quality videos had more views than fair quality videos (mean 27 348, 95% CI 15 154-39 543 views vs. 11 372, 95% CI 3 115-19 629 views, p = .030).
    Conclusion: The overall educational quality of YouTube videos for vascular surgery patient and trainee education is suboptimal. There is significant heterogeneity in the quality assessment tools used in their evaluation. A standardised approach to online education with a consistent quality assessment tool is required to better support online patient and trainee education in vascular surgery.
    Keywords:  Education; Systematic Review; Vascular Surgery; YouTube
    DOI:  https://doi.org/10.1016/j.ejvsvf.2024.01.054
  13. Plast Reconstr Surg Glob Open. 2024 Feb;12(2): e5604
      Background: The internet serves as a vital health information resource, yet the quality of data on specific health conditions, especially in Arabic, is often overlooked. This research assesses the quality of Arabic online information about cleft lip and palate (CLP) and proposes avenues for enhancement.Methods: From July to August 2022, a systematic evaluation of Arabic articles on CLP was performed using the DISCERN tool for quality assessment. Searches on Google and Bing resulted in 119 articles that met the study's criteria.
    Results: The quality of available Arabic information on CLP displayed substantial gaps. Commercial sources dominated (49.6%), followed by private (32.8%) and nonprofit entities (17.6%). The average DISCERN score was 2.26 of 5 (SD = 1.06), indicating the need for enhanced content, particularly concerning treatment risks.
    Conclusions: The study underscores the subpar quality of Arabic CLP information online, which might mislead patients and impede access to accurate advice. Nonprofit organizations should bolster their online footprint, offering refined health content. A deep dive into DISCERN scores reveals pinpointed improvement areas. Clinicians should direct patients and their families to reliable information sources. Addressing these gaps promises improved CLP knowledge in Arabic, fostering superior patient education and outcomes for those with this condition.
    DOI:  https://doi.org/10.1097/GOX.0000000000005604
  14. J Korean Med Sci. 2024 Feb 26. 39(7): e61
      BACKGROUND: Public health ethics (PHE) is a dynamic area within bioethics that addresses the complex moral implications of public health measures in the face of growing health threats. YouTube is a powerful and widely used platform for disseminating health-related information. The primary objective of this study is to assess videos related to PHE on YouTube. The aim is to gauge the extent of misinformation in collecting PHE videos on the platform.METHODS: On October 25, 2023, a thorough investigation on YouTube was undertaken, employing pre-determined search phrases involving 'public health,' 'healthcare,' 'health services administration,' and 'health policy and ethics.' The research encompassed a total of 137 videos that were selected according to strict inclusion and exclusion criteria. The videos were evaluated using the Global Quality Scale to measure quality and the modified DISCERN tool to evaluate reliability. The researchers identified video sources and compared several video attributes across different quality groups.
    RESULTS: A total of 137 videos were analyzed, and 65 (47.45%) were classified as high quality, 52 (37.23%) as moderate quality, and 21 (15.32%) as low quality. In high-quality videos, academic, government, physician, and university-hospital sources predominated, whereas Internet users and news sources were connected with low-quality videos. Significant differences in DISCERN score, per day views, likes, and comments were seen across the quality groups (P = 0.001 for views per day and P = 0.001 for other characteristics). According to the findings, low-quality videos had higher median values for daily views, likes, and comments.
    CONCLUSION: Although nearly half of the videos were high-quality, low-quality videos attracted greater attention. Critical contributors to high-quality videos included academic, government, physician, and university-hospital sources. The findings highlight the importance of quality control methods on social media platforms and strategies to direct users to trustworthy health information. Authors should prioritize appropriate citations and evaluate YouTube and other comparable platforms for potential promotional low-quality information.
    Keywords:  Ethics; Information Science; Internet; Public Health; Social Media
    DOI:  https://doi.org/10.3346/jkms.2024.39.e61
  15. Front Public Health. 2024 ;12 1309632
      Background: For high-quality colonoscopies, adequate bowel preparation is a prerequisite, closely associated with the diagnostic accuracy and therapeutic safety of colonoscopy. Although popular-science short videos can help people quickly access health information, the overall quality of such short videos as a source of health information regarding bowel preparation before colonoscopy is unclear. Therefore, we intend to conduct a cross-sectional study to investigate the quality of bowel preparation information before colonoscopy through short videos taken on TikTok and Bilibili.Methods: The Chinese phrases "colonoscopy" and "bowel preparation" were used as keywords to search for and screen the top 100 videos in the comprehensive rankings on TikTok and Bilibili. The Global Quality Score (GQS) and the modified DISCERN score were used to assess the quality of the information provided in these short videos.
    Results: A total of 186 short videos were included in this study; 56.5% of them were posted by health professionals, whereas 43.5% of them were posted by nonhealth professionals. The overall quality of these videos was unsatisfactory, with a median DISCERN score of 3 (2-4) and a median GQS of 3 (3-4). The radar maps showed that videos posted by gastroenterologists had higher completeness scores regarding outcomes, management, and risk factors, while nongastroenterologists had higher completeness scores concerning adverse effects, symptoms, and definitions of bowel preparation. Additionally, the median DISCERN score and GQS of the videos posted by gastroenterologists were 3 (3-4) and 3 (3-4), respectively, whereas the quality of the videos posted by patients was the worst, with a median DISCERN score of 2 (1-2) and a median GQS of 2 (1.25-3).
    Conclusion: In conclusion, the overall quality of health information-related videos on bowel preparation before colonoscopy posted on specified short video platforms was not satisfactory. Gastroenterologists provide more information on the outcomes, management, and risk factors for bowel preparation before colonoscopy, while nongastroenterologists focus on adverse effects, symptoms, and definitions of bowel preparation.
    Keywords:  bowel preparation; colonoscopy; health information; overall quality; short videos
    DOI:  https://doi.org/10.3389/fpubh.2024.1309632
  16. Saudi Pharm J. 2024 Apr;32(4): 101997
      Background: The goal of this study was to identify and evaluate the use of Arabic YouTube videos on BD as a resource for patient education.Methods: A cross-sectional evaluation of YouTube videos as a source of information for patients with BD in Arabic was performed. The study was observational and, because it did not involve human subjects, it followed the STROBE guidelines whenever possible. The quality of the videos was assessed using the validated DISCERN instrument. The search strategy involved entering the term "bipolar disorder" in the YouTube search bar, and only YouTube videos in Arabic were included.
    Results: A total of 58 videos were included in this study after removing duplicates and videos unrelated to BD (Figure 1). The most common source of videos was others (38%), followed by physician (33%), educational (26%), and hospital (3%). Resources covering symptoms and prognosis were mostly in the "others" category (41%). The resources or videos that covered treatment options were mainly created by physicians (41%). However, resources or videos that included a personal story mainly belonged to the "others" category (67%).
    Conclusion: Visual health-related instructional resources still have a significant shortage. This study highlights the poor quality of videos about serious illnesses like BD. Evaluation and promotion of the creation of visual health-related educational resources should be the primary goal of future study.
    Keywords:  Bipolar disorder; DISCERN instrument; Educational material; Educational videos; YouTube
    DOI:  https://doi.org/10.1016/j.jsps.2024.101997
  17. Ecancermedicalscience. 2023 ;17 1646
      [This corrects the article DOI: 10.3332/ecancer.2023.1573.].
    Keywords:  Arabic; YouTube; misinformation; prostate cancer; social media; videos
    DOI:  https://doi.org/10.3332/ecancer.2023.1646
  18. Orthop J Sports Med. 2024 Feb;12(2): 23259671241228543
      Background: Video material is one of the most common types of digital information on social media platforms.Purpose: To assess the validity and informational value of TikTok videos with regard to the anterior cruciate ligament (ACL).
    Study Design: Cross-sectional study.
    Methods: Using the terms "anterior cruciate ligament" AND/OR "ACL," 420 TikTok videos were retrieved, of which 100 videos were included in the analysis. The video duration and the number of likes, shares, and views were recorded. Videos were categorized based on the source (private user, physical therapist, surgeon, and researcher), subject type (patient experience, physical therapy, injury mechanism, anatomy, clinical test, and surgical technique), content (patient experience/testimony, education, and rehabilitation), and any background audio (music or voice). Video quality assessments were conducted using the DISCERN instrument, The Journal of the American Medical Association (JAMA) benchmark criteria, and the Global Quality Score (GQS). Associations between quantitative variables were tested using the Spearman rank correlation. One-way analysis of variance or Mann-Whitney test was performed to assess whether video quality differed by video characteristics.
    Results: A total of 41 videos were published by private users (41%), 34 by physical therapists (34%), 23 by surgeons (23%), and 2 by researchers (2%). Most of the information regarded patient experience (34%), followed by physical therapy (20%), injury mechanism (19%), anatomy (11%), clinical test (9%), and surgical technique (7%). The mean video length was 40.55 ± 41.58 seconds. The mean number of views was 151,084.39 ± 487,150.02, while the mean numbers of comments, likes, and shares were 72.80 ± 249.68, 6781.49 ± 29,163.96, and 98.71 ± 307.76, respectively. The mean DISCERN, JAMA, and GQS scores were 15.73 ± 1.44, 0.19 ± 0.39, and 1.16 ± 0.37, respectively, indicating the poorest quality on all 3 indices. The number of views, likes, shares, comments, and video lengths were all positively correlated with DISCERN (except the number of shares), JAMA, and GQS scores (P < .05). A significant difference was found in DISCERN scores between videos by private users and those by surgeons (15.37 ± 0.98 vs 16.22 ± 1.57; P = .007).
    Conclusion: Our analysis has demonstrated that the educational value of these videos on the ACL on TikTok is poor. Given the rapid growth of TikTok, further research is needed.
    Keywords:  TikTok; anterior cruciate ligament; arthroscopy; knee; reels; rehabilitation; return to sport; social media
    DOI:  https://doi.org/10.1177/23259671241228543
  19. Strategies Trauma Limb Reconstr. 2023 Sep-Dec;18(3):18(3): 163-168
      Background: Osseointegration (OI) limb has been performed for over 30 years and is an example of an advance in technology and surgical technique which has led to improvements in patient mobility and quality of life. An increasing number of patients seek information about osseointegration. The aim of this study was to categorise the most frequently asked questions by patients using the Google search engine and the most frequently accessed websites with the highest return on answers. The secondary aims of this study were to assess the quality of the information provided on those websites and to stratify, by category, which websites provide the best quality information.Materials and methods: Ten permutations and conjugations of the word 'osseointegration' were entered into Google. The first fifty 'People also ask' and associated websites by Google's machine learning and natural language processing engine were collected for each search term. The Rothwell classification system of questions by topic (Fact, Value, Policy) and websites by category was used (Commercial, Academic, Medical Practice, Single Surgeon Personal, Government, Social Media). Website quality was assessed using the Journal of the American Medical Association (JAMA) benchmark criteria (Likert-style rating 0-4). Pearson's Chi-squared and Student's t-tests were performed for statistical analysis as appropriate (significance, p < 0.05).
    Results: The 10 search terms generated 454 questions and referenced 408 websites. Of the 454 questions generated, the most common question categories were fact (70.8%), value (19.2%), and policy (10%). The most common website type was social media (37.4%). The most common question types were technical details (30.4%), specific activity (20.6%), and cost (14.1%). Only 1.6% of questions related to risks and complications. Generally, website quality was poor with 64.1% having a JAMA score of 0 or 1. Websites that were categorised as 'Government' had the highest overall JAMA scores: 71.4% had a score of 4.
    Conclusion: Based on Google search engine's results, the most commonly asked questions about osteointegration related to technical details, specific activities and cost; only 1.6% related to risks and complications. Interestingly, social media websites represented the highest volume of search result referrals. Overall, the quality of websites was poor with the most factual information coming from governmental websites.
    How to cite this article: Murphy EP, Sheridan GA, Page BJ, et al. Modern Internet Search Analytics and Osseointegration: What Are Patients Asking and Reading Online? Strategies Trauma Limb Reconstr 2023;18(3):163-168.
    Keywords:  Cost; Google; Online health information; Osseointegration; Prosthesis; Search analytics
    DOI:  https://doi.org/10.5005/jp-journals-10080-1603
  20. Indian J Psychiatry. 2024 Jan;66(1): 111-115
      The United Nations Organizations observe various global public health days (GPHDs) (also known as world days) throughout the year. We aimed to assess the impact of GPSDs on online health information-seeking behavior related to substance use in the Indian context. We used the Google Trends data for this study using standard guidelines. We conducted a topic search for the query "Substance abuse" to capture online information-seeking behavior (OHISB) for substance use disorder-related searches. The data were analyzed using the Joinpoint Regression software. Joinpoint regression model analysis was conducted to determine the statistically significant daily percent change (DPC) in the RSV trend. We found that there were significant changes in OHISB for substance use-related disorders on the International Day Against Drug Abuse and Illicit Trafficking over the last five years. The increase in the pattern of online search for substance use-related information was not observed around the World No Tobacco Day, World Mental Health Day, and World Suicide Prevention Day. Since most of these world health days are likely to continue to be observed in the coming years, it is important to regularly assess their impact as well as make the necessary modifications to make them more effective in achieving the desired objectives.
    Keywords:  Health days; internet; online health behavior; trends
    DOI:  https://doi.org/10.4103/indianjpsychiatry.indianjpsychiatry_473_23