
OpenAI has updated its ChatGPT usage policy, prohibiting the use of the AI system to provide medical, legal, or any other advice that requires professional licensing.
The changes are detailed in the company’s official Usage Policies and took effect from October 29.
Under the new rules, users are forbidden from using ChatGPT for: consultations that require professional certification (including medical or legal advice); facial or personal recognition without a person’s consent; making critical decisions in areas such as finance, education, housing, migration, or employment without human oversight; academic misconduct or manipulation of evaluation results.
OpenAI states that the updated policy aims to enhance user safety and prevent potential harm that could result from using the system beyond its intended capabilities.
As reported by NEXTA, the bot will no longer give specific medical, legal, or financial advice.
ChatGPT is now officially an “educational tool”, not a “consultant.”
The reason for this change has been chalked up to “regulations and liability fears” to avoid lawsuits.
Now, instead of providing direct advice, ChatGPT will “only explain principles, outline general mechanisms and tell you to talk to a doctor, lawyer or financial professional.”
Based on the new explicit rules, there will be “no more naming medications or giving dosages… no lawsuit templates… no investment tips or buy/sell suggestions.”
This clampdown directly addresses the fears that have long surrounded the technology.
