- Joined
- 21 January 2015
- Messages
- 12,150
- Reaction score
- 16,351
New superbug-killing antibiotic discovered using AI https://www.bbc.co.uk/news/health-65709834
AI generated books on King Charles' illness removed by Amazon:
Because of inaccurate information. Way to go.![]()
Amazon stops selling ‘intrusive’ books written about King Charles’ cancer
Several books were listed for sale which were reported to share exclusive revelations about the King’s health, it was reportedwww.independent.co.uk
I hadn't read that far, actually. Amazon's reply AI-generated as well?They said: "... content that creates a disappointing customer experience." Sad. Very sad.
Elon Musk says in a Thursday lawsuit that Sam Altman and OpenAI have betrayed an agreement from the artificial intelligence research company's founding to develop the technology toward the benefit of humanity over profits.
In the suit filed Thursday night in San Francisco Superior Court, Musk claims OpenAI's recent relationship with tech giant Microsoft has compromised the company's original dedication to public, open source artificial general intelligence.
"OpenAI, Inc. has been transformed into a closed-source de facto subsidiary of the largest technology company in the world: Microsoft. Under its new board, it is not just developing but is actually refining an AGI to maximize profits for Microsoft, rather than for the benefit of humanity," Musk says in the suit.
Musk brings claims including breach of contract, breach of fiduciary duty and unfair business practices against OpenAI and asks for the company to revert back to open source ...
A lawyer in Canada is under fire after the artificial intelligence chatbot use she used for legal research created “fictitious” cases, in the latest episode to expose the perils of using untested technologies in the courtroom.
Vancouver lawyer Chong Ke, who now faces an investigation into her conduct, allegedly used ChatGPT to develop legal submissions during a child custody case at the British Columbia supreme court.
According to court documents, Ke was representing a father who wanted to take his children overseas on a trip but was locked in a separation dispute with the children’s mother. Ke is alleged to have asked ChatGPT for instances of previous case law that might apply to her client’s circumstances. The chatbot, developed by OpenAI, produced three results, two of which she submitted to the court.
The lawyers for the children’s mother, however, could not find any record of the cases, despite multiple requests.
When confronted with the discrepancies, Ke backtracked.
We have been told by management that we should make use of AI at work to “improve efficiency”. I was initially reluctant, but a friend told me I’d be surprised by how useful chatbots could be with certain tasks, so I decided to give it a go.
To begin, I was impressed. Then I asked a question related to something quite specific to my job and area of expertise. The answer included several patent inaccuracies. I pointed this out inside the chat, and then a bizarre conversation ensued. The AI denied that it was wrong and when I politely explained its error, it attempted to gaslight me. Did I just get lied to by an AI language model?