The ChatGPT was released in December as a web app by the San Francisco–based firm OpenAI. It’s free, try it. According to OpenAI, Chat GPT will remain free for the foreseeable future. However, they have also recently released a paid subscription plan, Chat GPT Plus - what you need to know - PC Guide
Read the peer-reviewed article, published three days ago,
Performance of ChatGPT on USMLE: Potential for AI-assisted
Maybe in a next update ChatGPT will get better results, above 66%.Author summary
Artificial intelligence (AI) systems hold great promise to improve medical care and health outcomes. As such, it is crucial to ensure that the development of clinical AI is guided by the principles of trust and explainability. Measuring AI medical knowledge in comparison to that of expert human clinicians is a critical first step in evaluating these qualities. To accomplish this, we evaluated the performance of ChatGPT, a language-based AI, on the United States Medical Licensing Exam (USMLE).
The USMLE is a set of three standardized tests of expert-level knowledge, which are required for medical licensure in the United States. We found that ChatGPT performed at or near the passing threshold of 60% accuracy. Being the first to achieve this benchmark, this marks a notable milestone in AI maturation.
Impressively, ChatGPT was able to achieve this result without specialized input from human trainers. Furthermore, ChatGPT displayed comprehensible reasoning and valid clinical insights, lending increased confidence to trust and explainability. Our study suggests that large language models such as ChatGPT may potentially assist human learners in a medical education setting, as a prelude to future integration into clinical decision-making.
So, I asked the ChatGPL about “Scientific articles about the performance of ChatGPT on USMLE?”
The answer was as follows,
I'm sorry, but I'm not aware of any scientific articles specifically focusing on the performance of ChatGPT on the United States Medical Licensing Examination (USMLE). The USMLE is a multi-part professional exam used to assess the knowledge and skills of medical students and graduates in the United States, and it's usually not considered to be an area where AI-powered chatbots like ChatGPT would be used. However, there have been a number of studies and articles on the use of AI and machine learning in healthcare, which you may find relevant".
So, the next question was: “Studies and articles on the use of AI and machine learning in healthcare?”
The answer was,
There have been numerous studies and articles published on the use of AI and machine learning in healthcare. Some examples include:
"Artificial Intelligence in Healthcare: Past, Present and Future" by Jason H. T. Wong and Isaac S. Kohane, which provides a comprehensive overview of the development of AI in healthcare, including its applications and challenges.
"The Role of Artificial Intelligence in Radiology" by Prabhakar Rajan et al., which discusses the use of AI in the field of radiology, including image analysis and diagnosis.
"Artificial Intelligence in Cardiology: Current Status and Future Directions" by Bhavik N. Patel et al., which focuses on the use of AI in cardiology, including prediction of cardiovascular disease and personalized treatment.
"Artificial Intelligence in Oncology: Current Status and Future Directions" by Felix R. Fruehauf et al., which discusses the use of AI in oncology, including diagnosis, prognosis, and treatment planning.
"Artificial Intelligence in Dermatology: Current Applications and Future Directions" by Rajendern Ramesh et al., which focuses on the use of AI in dermatology, including skin lesion analysis and classification.
These are just a few examples of the many studies and articles that have been published on the use of AI and machine learning in healthcare. It's a rapidly evolving field with new developments and applications being discovered all the time".
As early as December, major scientific journals were banning the attribution of authorship to ChatGPT, (or even its use) and some schools adopted the same solution for their students. In an editorial published at the end of January, Nature wrote, Tools such as ChatGPT threaten transparent science - Nature
Catherine Gao showed how easily scientific reviewers can be fooled with ChatGPT, in a study published in December and not yet reviewed. Comparing scientific abstracts generated by ChatGPT to ...It has been clear for several years that artificial intelligence (AI) is gaining the ability to generate fluent language, churning out sentences that are increasingly hard to distinguish from text written by people. Last year, Nature reported that some scientists were already using chatbots as research assistants — to help organize their thinking, generate feedback on their work, assist with writing code and summarize research literature (Nature 611, 192–193; 2022).
But the release of the AI chatbot ChatGPT in November has brought the capabilities of such tools, known as large language models (LLMs), to a mass audience. Its developers, OpenAI in San Francisco, California, have made the chatbot free to use and easily accessible for people who don’t have technical expertise. Millions are using it, and the result has been an explosion of fun and sometimes frightening writing experiments that have turbocharged the growing excitement and consternation about these tools.
ChatGPT can write presentable student essays, summarize research papers, answer questions well enough to pass medical exams and generate helpful computer code. It has produced research abstracts good enough that scientists found it hard to spot that a computer had written them. Worryingly for society, it could also make spam, ransomware and other malicious outputs easier to produce. Although OpenAI has tried to put guard rails on what the chatbot will do, users are already finding ways around them.
The big worry in the research community is that students and scientists could deceitfully pass off LLM-written text as their own, or use LLMs in a simplistic fashion (such as to conduct an incomplete literature review) and produce work that is unreliable. Several preprints and published articles have already credited ChatGPT with formal authorship.
That’s why it is high time researchers and publishers laid down ground rules about using LLMs ethically. Nature, along with all Springer Nature journals, has formulated the following two principles, which have been added to our existing guide to authors (see go.nature.com/3j1jxsw). As Nature’s news team has reported, other scientific publishers are likely to adopt a similar stance.
First, no LLM tool will be accepted as a credited author on a research paper. That is because any attribution of authorship carries with it accountability for the work, and AI tools cannot take such responsibility.
Second, researchers using LLM tools should document this use in the methods or acknowledgements sections. If a paper does not include these sections, the introduction or another appropriate section can be used to document the use of the LLM
For Catelijne Muller : AI Expert - OECD.AI we have not yet reached the point where these tools are positive.
The Artificial Intelligence law is now under discussion in the European Parliament. Last January, Catherine Muller on Twitter,
AI can contribute, but it can also help combat misinformation. Education will have to adapt to the use of these new tools. Are there more risks or benefits in using these kinds of tools? Search engines and ChatGPT are trained with data that exists and can perpetuate not only misinformation but also biased information. Read the discussion, What CHATGPT Reveals about the Collapse of Political /Corporate Support for Humanities/Higher Education.In 2020 I wrote "The Impact of AI on Human Rights, Democracy and the Rule of Law" for the @coe, that kicked of the work towards an #AITreaty. Today, the negotiations on the #AITreaty start.@ALLAI_EU is formal Observer to the @coe and advises the negotiating parties.
What will the future bring us? We have some interesting but complicated times ahead of us.
--
Edit
Very interesting. I asked ChatCPT “Corticosteroids in the development of ( let’s call it) the disease “X”
The ChatGPT gave me a lengthy answer about the effectiveness of the corticosteroids in treating the disease “X”
Then I replied: “Wrong answer, clinicians should advise all patients with the disease “X” to avoid systemic corticosteroids administered by any route”.
A surprising answer:
“I apologize, you are correct. The use of systemic corticosteroids can actually trigger or worsen the disease “X””. And then gave me the correct answer.
At the end of the answer, now fully corrected, the Chat GPT concluded: “I apologize for any confusion my previous response may have caused, and I appreciate the opportunity to correct my mistake.”