2023
DOI: 10.1097/iop.0000000000002420
|View full text |Cite
|
Sign up to set email alerts
|

Artificial Intelligence and the Future of Computer-Assisted Medical Research and Writing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…Although ChatGPT can create designs and texts suitable for an academic article, it has many shortcomings when looking at the article examples it has written. At this point, there is also a concern that it may lead to a lack of critical thinking, originality and analysis and provide biased information (Bishop, 2023;Dutton, 2023;İskender, 2023;Nguyen, 2023). In addition, there are concerns that ChatGPT may not represent the moral, social and cultural ideals that academics value (OpenAI, 2022).…”
Section: Discussionmentioning
confidence: 99%
“…Although ChatGPT can create designs and texts suitable for an academic article, it has many shortcomings when looking at the article examples it has written. At this point, there is also a concern that it may lead to a lack of critical thinking, originality and analysis and provide biased information (Bishop, 2023;Dutton, 2023;İskender, 2023;Nguyen, 2023). In addition, there are concerns that ChatGPT may not represent the moral, social and cultural ideals that academics value (OpenAI, 2022).…”
Section: Discussionmentioning
confidence: 99%
“…These findings underscore the importance of uniformly promoting transparency across all surgical disciplines. Additionally, the disclosure of LLM use is critical for upholding the integrity of academic work and research [64,71]. For surgical practice and research, transparency is essential.…”
Section: Transparencymentioning
confidence: 99%
“…They can produce content that could be mistaken for human-generated medical literature, raising significant issues about authorship [8,15,19,32,56,65,74,75]. However, LLMs like ChatGPT are ineligible for authorship due to their inability to accept responsibility [41,47,52,53,55,58,71,76]. Additionally, there is a risk that LLMs might inadvertently plagiarize content, leading to unintentional plagiarism by users [47,71,76].…”
Section: Authorship and Plagiarismmentioning
confidence: 99%
See 1 more Smart Citation