Despite high promises, several limitations concerning AI technologies and our specific use case are worth noting. First, large language models (LLMs) like BERT, BART, T5, and PEGASUS were trained on vast amounts of human-written texts and may struggle to ground their generated text on factual information (Petroni et al.,
2020). Due to the inherent biases present in the training data, these models might inadvertently perpetuate or exacerbate existing biases in generated headlines. Moreover, LLMs may incorporate factual inaccuracies, outdated information, or misleading content, leading to inconsistencies or contradictions in generated headlines (Tang et al.,
2023), especially when dealing with complex or ambiguous input texts. Consequently, ensuring factual grounding, accurate representation of source content, and addressing biases in the models remain significant challenges for LLMs in producing reliable, bias-free headlines. Second, most NLP foundation models can only handle relatively short passages (i.e., 512-1,024 words), which should suffice title/abstract comprehension but fall short of a full-length peer-reviewed journal article. More recent models, such as Big Bird and Longformer, accept longer text, up to 4,096 words. But in general, today’s mainstream NLP model architectures, such as transformers and recurrent neural networks, are not optimized to consume or produce long passages (e.g., text with over 4,000 words). Third, the substantial growth in the number of parameters in LLMs has led to increased computational costs for fine-tuning such large-scale models. This drawback could potentially render them impractical for real-world applications. Fourth, our application focuses solely on obesity-related scientific discoveries owing to our team’s domain expertise. To capitalize on large language models’ potential for generating news reports concerning distinct scientific domains, those models need to be fine-tuned on large-scale, diverse text corpora, with examples pairing news reports with scientific publications. Constructing such datasets demands substantial human and financial resources. Fifth, this study built AI models to autogenerate news headlines, an initial step for creating full-length news articles reporting scientific findings. Substantial future endeavors and technological advancement may be required to fulfill that goal.