WeChat ID :
Whoever says
that laws or certain fields of knowledge
don’t need to be learned —
because “if something happens, we can just ask ChatGPT” —
.
Do you know that
after news broke about a foreign woman
who used ChatGPT to self-diagnose and treat herself
by following the medication instructions it gave —
she died!!!!
.
Now, ChatGPT (OpenAI)
has issued new regulations
prohibiting consultations related to:
drug trading, illegal activities, medicine, law,
and investment solicitation.
.
However, this doesn’t mean
that ChatGPT is completely forbidden
from answering questions about law or medicine.
.
What’s prohibited is giving personalized advice
that could be considered professional service without a license.
.
The permitted scope is
providing general information for education and reference purposes only.
.
Users are responsible for verifying, interpreting,
and applying such information
in a way that complies with the law.
.
These rules were announced
to protect OpenAI
from lawsuits resulting from user misuse or misinterpretation.
Everything else remains the same.
.
Therefore,
if you plan to fight a legal case by consulting ChatGPT —
β οΈ be careful, your case might backfire.
Because ChatGPT doesn’t know all the facts,
and real courts judge based on evidence, documents, and intent —
not on the prompts that make you look good.
.
Join the discussion:
https://www.facebook.com/Ex.MatchingProperty/posts/pfbid0XCCQUjtNykEmzRi7ng5rvnrnJe9imzF9GTkxoVERj2ciTFCPR1PMkMYZxWjzMfSDl