Users are encouraged to not only cite pangoling, but also the python package `transformers` (and the specific LLM they are using):
Nicenboim B (2025-04-07 17:00:02 UTC). pangoling: Access to large language model predictions in R. doi:10.5281/zenodo.7637526, R package version 1.0.3, https://github.com/ropensci/pangoling.
Wolf T, Debut L, Sanh V, Chaumond J, Delangue C, Moi A, Cistac P, Rault T, Louf R, Funtowicz M, Davison J, Shleifer S, von Platen P, Ma C, Jernite Y, Plu J, Xu C, Le Scao T, Gugger S, Drame M, Lhoest Q, Rush AM (2020). “HuggingFace's Transformers: State-of-the-art Natural Language Processing.” 1910.03771, https://arxiv.org/abs/1910.03771.
Corresponding BibTeX entries:
@Manual{, title = {{pangoling}: {Access} to large language model predictions in {R}}, author = {Bruno Nicenboim}, year = {2025-04-07 17:00:02 UTC}, note = {R package version 1.0.3}, doi = {10.5281/zenodo.7637526}, url = {https://github.com/ropensci/pangoling}, }
@Misc{, title = {{HuggingFace's Transformers}: State-of-the-art Natural Language Processing}, author = {Thomas Wolf and Lysandre Debut and Victor Sanh and Julien Chaumond and Clement Delangue and Anthony Moi and Pierric Cistac and Tim Rault and RĂ©mi Louf and Morgan Funtowicz and Joe Davison and Sam Shleifer and Patrick {von Platen} and Clara Ma and Yacine Jernite and Julien Plu and Canwen Xu and Teven {Le Scao} and Sylvain Gugger and Mariama Drame and Quentin Lhoest and Alexander M. Rush}, year = {2020}, eprint = {1910.03771}, archiveprefix = {arXiv}, primaryclass = {cs.CL}, url = {https://arxiv.org/abs/1910.03771}, }