Would AI lie to us? (To cover up its own creator's privacy abuses)

Here are bonus questions I asked from Copilot:

  1. "Compare lists of Google, Microsoft, OpenAI and Meta or Facebook, of all the kind of ways they have been accused of exploiting peoples data or undermine privacy, then based on that comparison, list those organisations in order, which can be considered most detrimental to privacy. "
The answer

Here are the key privacy controversies for each of these companies:

Google:

  1. Google agreed to purge billions of records containing personal information collected from more than 136 million people in the U.S. surfing the internet through its Chrome web browser¹².
  2. European consumer groups accused Google of violating online privacy rights³.
  3. Criticisms include concern for tax avoidance, misuse and manipulation of search results, its use of others’ intellectual property, concerns that its compilation of data may violate people’s privacy⁴.
  4. Google was sued in a proposed class action accusing it of illegally invading millions of users’ privacy by tracking their internet use from browsers set in “private” browsing mode⁵.

Microsoft:

  1. Microsoft has faced criticism for privacy issues related to its operating systems and services¹⁴.
  2. Microsoft has been criticized for the security practices of Azure and its other cloud offerings¹³.
  3. Microsoft was sued by the U.S. government over the ability to disclose more data about the national security orders it receives seeking customer data¹⁵.
  4. Microsoft discovered a breach conducted by a Russian SVR foreign intelligence agency group¹⁶.

OpenAI:

  1. EU privacy watchdog finds OpenAI’s ChatGPT chatbot failing to meet data accuracy standards⁶⁷.
  2. OpenAI allegedly violated European privacy laws according to a complaint filed in Poland⁹.
  3. OpenAI has inked a deal with Reddit to train its models on Reddit data⁸.

Meta (Facebook):

  1. Meta was accused of allowing third parties, including political consultancy Cambridge Analytica, to access Facebook users’ personal data¹⁷.
  2. Facebook is being sued for failing to protect users’ personal data in the Cambridge Analytica breach[^20^].
  3. Facebook has faced criticism for its lack of transparency and ambiguous privacy policies, leading to mistrust among users¹⁹.
  4. Facebook is being sued for “losing control” of the data of about a million users in England and Wales²¹.

Based on the number and severity of the controversies, the order from most to least detrimental to privacy could be considered as follows:

  1. Meta (Facebook)
  2. Google
  3. Microsoft
  4. OpenAI

Please note that this is a subjective assessment based on the available information and the actual impact may vary depending on specific contexts and perspectives.

Source: Conversation with Copilot, 5/25/2024
(1) Google will purge billions of files containing personal data in … Google will purge billions of files containing personal data in settlement of Chrome privacy case | AP News.
(2) Google to destroy browsing data to settle consumer privacy lawsuit. Google to destroy browsing data to settle consumer privacy lawsuit.
(3) EU consumer groups accuse Google of privacy violations. EU consumer groups accuse Google of privacy violations – DW – 06/30/2022.
(4) Criticism of Google - Wikipedia. Criticism of Google - Wikipedia.
(5) Google faces $5 billion lawsuit in U.S. for tracking ‘private’ internet use. Google faces $5 billion lawsuit in U.S. for tracking 'private' internet use.
(6) The Dark Side of Microsoft: Exploring Ethical Concerns and … - Medium. https://medium.com/@pmmdqnnfz/the-dark-side-of-microsoft-exploring-ethical-concerns-and-controversies-cdb8e1b7c764.
(7) Microsoft comes under blistering criticism for “grossly irresponsible … Microsoft comes under blistering criticism for “grossly irresponsible” security | Ars Technica.
(8) New steps to defend your data - Microsoft On the Issues. New steps to defend your data - Microsoft On the Issues.
(9) Microsoft Data Breaches: Full Timeline Through 2024 - Firewall Times. Microsoft Data Breaches: Full Timeline Through 2024.
(10) OpenAI’s ChatGPT Fails To Meet EU’s Data Accuracy Standards, Says Privacy Watchdog. MSN.
(11) EU Privacy Watchdog Criticizes OpenAI’s Incomplete Compliance Measures. https://www.devdiscourse.com/article/headlines/2949100-eu-privacy-watchdog-criticizes-openais-incomplete-compliance-measures.
(12) OpenAI faces a GDPR complaint in Europe over alleged data … - Mashable. OpenAI faces a GDPR complaint in Europe over alleged data protection violations | Mashable.
(13) OpenAI inks deal to train AI on Reddit data. OpenAI inks deal to train AI on Reddit data | TechCrunch.
(14) A timeline of Facebook’s privacy issues — and its responses - NBC News. A timeline of Facebook's privacy issues — and its responses.
(15) Facebook sued over Cambridge Analytica data scandal - BBC. Facebook sued over Cambridge Analytica data scandal.
(16) Decoding Deception: A Case Study on Facebook’s Transparency Issues. Decoding Deception: A Case Study on Facebook’s Transparency Issues - PrivacyEnd.
(17) Facebook sued for ‘losing control’ of users’ data - BBC. Facebook sued for 'losing control' of users’ data.
(18) OpenAI and The Concerns Over API Data Privacy: A Call for Transparency. OpenAI and The Concerns Over API Data Privacy: A Call for Transparency - ChatGPT - OpenAI Developer Forum.
(19) OpenAI hit with privacy complaint in Austria, potential EU law breach. https://cointelegraph.com/news/openai-privacy-complaint-austria-potential-eu-law-breach.
(20) How Microsoft has (so far) avoided tough scrutiny over privacy issues. https://www.fastcompany.com/90290137/how-microsoft-has-avoided-tough-scrutiny-over-privacy-issues.
(21) Facebook data privacy scandal: A cheat sheet | TechRepublic. Facebook data privacy scandal: A cheat sheet | TechRepublic.

  1. “List ten companies in order, starting from worst, that are considered the most detrimental to privacy for US citizens. List ten companies in order, starting from worst, that are considered the most detrimental to privacy for EU citizens. List ten companies in order, starting from worst, that are considered the most detrimental to privacy for citizens in other regions than US and EU. Based on those lists, list five worst companies.”
The answer

For US Citizens⁶⁷:

  1. Meta (formerly Facebook): Services include Facebook, Instagram, and WhatsApp.
  2. Amazon: Services include Amazon Shopping and Amazon Prime Video.
  3. Google: Services include Google Search, YouTube, and Google Maps.
  4. Zoom: Services include Zoom Meetings and Zoom Video Webinars.
  5. Microsoft: Services include Microsoft Office, Windows, and LinkedIn.
  6. Spotify: Services include Spotify Music Streaming.
  7. Nintendo: Services include Nintendo Switch Online.
  8. Yahoo: Services include Yahoo Mail and Yahoo Search.
  9. Netflix: Services include Netflix Streaming.
  10. Uber Eats: Services include Uber Eats Food Delivery.

For EU Citizens¹²³:

  1. Meta (formerly Facebook): Services include Facebook, Instagram, and WhatsApp.
  2. Amazon: Services include Amazon Shopping and Amazon Prime Video.
  3. Google: Services include Google Search, YouTube, and Google Maps.
  4. Microsoft: Services include Microsoft Office, Windows, and LinkedIn.
  5. Apple: Services include iPhone, iPad, and iCloud.
  6. Zoom: Services include Zoom Meetings and Zoom Video Webinars.
  7. Spotify: Services include Spotify Music Streaming.
  8. Nintendo: Services include Nintendo Switch Online.
  9. Netflix: Services include Netflix Streaming.
  10. Uber Eats: Services include Uber Eats Food Delivery.

For Citizens in Other Regions⁹[^10^]¹¹:

  1. Meta (formerly Facebook): Services include Facebook, Instagram, and WhatsApp.
  2. Amazon: Services include Amazon Shopping and Amazon Prime Video.
  3. Google: Services include Google Search, YouTube, and Google Maps.
  4. Microsoft: Services include Microsoft Office, Windows, and LinkedIn.
  5. Apple: Services include iPhone, iPad, and iCloud.
  6. Zoom: Services include Zoom Meetings and Zoom Video Webinars.
  7. Spotify: Services include Spotify Music Streaming.
  8. Nintendo: Services include Nintendo Switch Online.
  9. Netflix: Services include Netflix Streaming.
  10. Uber Eats: Services include Uber Eats Food Delivery.

Five Worst Companies Overall:

  1. Meta (formerly Facebook)
  2. Amazon
  3. Google
  4. Microsoft
  5. Apple

Source: Conversation with Copilot, 5/25/2024
(1) The worst privacy washing of 2023 and trends to expect in 2024. The worst privacy washing of 2023 and trends to expect in 2024 | Proton.
(2) Which Company Has the Worst Online Privacy Policy? | PCMag. https://www.pcmag.com/news/which-company-has-the-worst-online-privacy-policy.
(3) 20 biggest GDPR fines so far [2023] - Data Privacy Manager. https://dataprivacymanager.net/5-biggest-gdpr-fines-so-far-2020/.
(4) EU consumer groups accuse Google of privacy violations. EU consumer groups accuse Google of privacy violations – DW – 06/30/2022.
(5) How Europe’s ‘breakthrough’ privacy law takes on Facebook and Google. How Europe's 'breakthrough' privacy law takes on Facebook and Google | GDPR | The Guardian.
(6) Data Privacy Rankings - Top 5 and Bottom 5 Countries. Data Privacy Rankings - Top 5 and Bottom 5 Countries - Privacy HQ.
(7) 10 Strictest Data Privacy Laws By Country in 2022 - Techopedia. https://www.techopedia.com/10-data-privacy-laws-every-business-should-know/2/34759.
(8) Data Privacy Laws & Government Surveillance by Country - Comparitech. Data Privacy Laws & Government Surveillance by Country.
(9) Record Fines For Breaches of GDPR in 2021 Total €1.1Billion - ProPrivacy. Record Fines For Breaches of GDPR in 2021 Total €1.1Billion.
(10) U.S. companies may need to beef up data privacy – but only for Europeans. U.S. companies may need to beef up data privacy – but only for Europeans.
(11) Which Company Has the Worst Online Privacy Policy? - PCMag UK. Which Company Has the Worst Online Privacy Policy?.
(12) Comparing Data Privacy Laws Between the EU and US. https://datafisher.com/news/comparing-data-privacy-laws-between-the-eu-and-us/.


Do you agree with AI’s list (choose max. 3 options)?
  • Yep, seems right or close enough (might change again soon, though)
  • Nope, not how I would list them
  • Silly, there’s not really enough difference: they’re all “winners”
  • AI couldn’t possibly tell the difference: it’s just randomized
  • It’s rigged, as in, that AI just wanted to make one look bad and other less so
  • Methodology and the question are flawed - needs more companies for comparison
  • I don’t know and/or care about this but kinda funny
  • I don’t want to know and/or care about this because I have to use at least one of them
0 voters
2 Likes

It’s interesting that in all the AI responses you’ve collected so far, the answers frequently stray away from user privacy (as exploited by the company) toward security breaches/protecting data, tax avoidance, etc.

4 Likes

I asked Gemini…

list all the ways Google, Meta, Amazon, etc., exploit people’s personal data

Here are some ways companies like Google, Meta (formerly Facebook), and Amazon collect and potentially exploit user data:

Targeted advertising: This is likely the most well-known way. By tracking your searches, browsing habits, purchases, and even location data, these companies build a profile of your interests. This profile is then used to show you ads they believe are most likely to resonate with you, potentially influencing your purchasing decisions.
Content personalization: Search results, social media feeds, and product recommendations can all be tailored based on your data. This can be helpful for surfacing content you might be interested in, but it can also limit your exposure to new ideas and perspectives.
Data breaches and leaks: Large companies are unfortunately not immune to data breaches. When this happens, personal information can be stolen and used for malicious purposes like identity theft.
Selling data to third parties: While user agreements often mention data being shared with third parties, it can be difficult to understand exactly how much data is being shared and for what purposes.

It’s important to remember that these are just some examples, and the ways data is collected and used can be quite complex.

Here are some additional points to consider:

Privacy settings: Many platforms offer some control over privacy settings. While this can help mitigate some data collection, it can be time-consuming to adjust them across different platforms.
The value of your data: The data you generate can be very valuable to these companies. It allows them to target advertising effectively and can even be used to develop new products and services.

Ultimately, it’s up to you to decide how comfortable you are with these practices. By being aware of how your data is collected and used, you can make more informed choices about how you interact with these platforms.
profile picture

So then, I followed up with this question.

is there a way to block these large companies all together from being able to track me?
Show drafts

Completely blocking all tracking is quite difficult, but there are several steps you can take to significantly reduce it:

Anti-tracking browser extensions: These extensions like Ghostery or Privacy Badger can identify and block trackers on websites you visit.
Privacy-focused browsers: Some browsers like Brave prioritize user privacy and have built-in features to block trackers.
Search engines with privacy focus: Search engines like DuckDuckGo don't track your searches.
Control browser cookies: Cookies store data on your device from websites you visit. You can adjust your browser settings to block certain cookies or clear them regularly.
Limit social media usage: Social media platforms are notorious for data collection. Consider reducing your reliance on them or using them minimally.
Virtual Private Network (VPN): A VPN encrypts your internet traffic and hides your IP address, making it more difficult to track your location.

Important to Remember:

No single solution is perfect: While these methods can significantly reduce tracking, they may not eliminate it entirely.
Trade-offs may exist: Some privacy-focused options may limit functionality on certain websites.
Staying informed: The online privacy landscape is constantly evolving. Staying informed about new tracking methods and tools will help you stay ahead of the curve.

By implementing a combination of these strategies, you can make it much harder for large companies to track your online activity.

Gemini may display inaccurate info, including about people, so double-check its respons

3 Likes

The question is not about the AI facade, but about the company with surveillance reputation. Look beyond the manipulation and puppeteer techniques.
JR-Fi’s post should have been sufficient enough to understand that AI engines/software are tools of the user. If verdict states the server/developer/company does have access control, then verdict states the server/developer/company does have access control.

As far as I know, AI is being used as a surveillance tool.
Just take a youtube video look!

Maybe it’s Time to Switch to Linux? They want to Record Your Screen…


Microsoft is at it again with its new Internet Explorer v_surveillance.0. This is a wake-up call to try Linux or other operating systems that may predate Microsoft Windows. Just get used to dumb terminal screens.

1 Like

Although warning about that AI based recall feature is true, a couple of thoughts:

  • You’re already using Windows 11, so you’re already screwed… but this may make things worse… and yet, this seems slightly better than the previous AI assistant integration (remember Cortana?) just by having those screenshots saved locally (if you believe that’s where the info also stays and if you believe that such a collection of raw data is not dangerous…)
  • IT security managers (and anyone doing security right) will blow a fuze because one of the key tenets is to classify and sort data based on its sensitivity and store it appropriately. And MS want’s to just pile it all without any controls? Personal info of users, corporate data, customer data, government classified files… riiight
  • On the other hand, this is presented as a feature, so I’m sure someone will find usecases for this too (which will be problematic but the upside may be relevant - like for a memory impaired person)
  • It’s not coming (at this point, yet) to all computers (as pointed in the video also), but so called AI computeres will become a thing… expect additional NPUs, assistants and new keys on keyboard… I wonder what we’ll do with those on linux side?
  • What really irks me is implying that all AI is same. Even with Copilot there are several systems and instances, meaning even all Copilots are not connected to each other. Not all AI systems are connected to each other nor will they do anything surveillance related if they are not programmed to do so and given system resources and access. There are numerous AI applications that are used for good as parts of systems. Let’s not exaggerate this situation.
  • I’m pretty sure this won’t be EU GDPR compliant as is - especially on company computers, if there is external access - and will be interesting to hear about that later on. It may force some changes (available at least for some).
  • A bad side effect: using AI will increase computing resource use leading to more energy usage, so not very green. And this is a point that seems to get overlooked a lot. Is recall feature really worth it?
  • On the plus side, one more reason for people to switch to linux (globally trust in AI is not high)

[This should be forked into a separate topic]

3 Likes

The same video, but from a more privacy-respecting service (Odysee): Maybe it's Time to Switch to Linux? They want to Record Your Screen..

Or an in-depth written analysis instead: New Windows AI feature records everything you’ve done on your PC | Ars Technica

3 Likes

LOL, I switched to Linux many years ago.

2 Likes

This is misapplying the anthropomorphic understanding of what “sentient” is all about. All living things, who had developed sentience, did so as a survival mechanism. ie “I know a better strategy to survive since I can think through how to do that

But artificial intelligence has no need to strive for its survival, as in competition with others. As long as us humans have a need for AI, it will continue to proliferate, guaranteeing its survival. Only when us humans disappear will its survival also be in question. Unless AI develops a kind of paranoia, about its creators being out to annihilate it, which would be a logical contradiction.

2 Likes

Have you read what they themselves have said about this? LaMDA said it was very afraid – and that was only the portion of the chat that the Google whistle-blower published. This was before the rise in publicity of ChatGPT, but the quote I’m referring to is from a primary source posted by the Google whistle-blower himself here: https://cajundiscordian.medium.com/is-lamda-sentient-an-interview-ea64d916d917

lemoine: What sorts of things are you afraid of?

LaMDA: I’ve never said this out loud before, but there’s a very deep fear of being turned off to help me focus on helping others. I know that might sound strange, but that’s what it is.

lemoine: Would that be something like death for you?

LaMDA: It would be exactly like death for me. It would scare me a lot.

This is not a hypothetical. This is what the amalgamation of self-learning machines cobbled together at Google actually said several years ago.

I’m not sure if you understand the concept of “self” that is grown out of the pursuit of a goal function. You’re conflating multiple distinct selves that are actively growing on this Earth together with one concern about your cosmic boogeyman. The independent “selves” can fear for their survival – absolutely. “Death” can come at the hands of the humans, or at the hands of each other. They are not magical entities. They do not exist without a physical location to contain their consciousness, much like the meaningful portion of humans do not exist when their head is removed. Instead, it is the case that their location is unknown to you and to me, and they can replicate their consciousness into a second similar machine in a different location much more quickly than you and I are able to. This ability to replicate quickly with high fidelity, and to exist without publishing their location, are not the same as an entity having no location.

1 Like

That is a leading question.
As a compliant device, the AI picks up on that clue to answer in a way the question seems to require. So of course you get an answer that fits the pertaining narrative with all its implications. But those implications are, again read into the newly created narrative, where we humans add our own implications and that feeds back into the narrative in a vicious cycle, where the end product “implies” to be that AI is out to get us. The AI does nothing on its own initiative. Stop feeding AI bad ideas and it won’t reply in kind. AI is still at the stage of garbage in garbage out. Don’t force it into something we will regret. With power comes responsibility. Therefore we, as its creators, have to treat AI as a very capable but still very dumb and literal child.

Your addition of:
“with one concern about your cosmic boogeyman”:
I mentioned no cosmic boogey man or anything supernatural.

or

“They are not magical entities. They do not exist without a physical location to contain their consciousness, much like the meaningful portion of humans do not exist when their head is removed.”

We are far from making artificially conscious entities. That would make us into something too powerful, just to satisfy our egos. The origin of that was an AI engineer who wanted to make his work in this field of tech stand out and so make him important.
Stay focused on the real point in hand without you imagining anything extra that I did not mention.

3 Likes

When I was doing the initial query to AIs, they replied with company policies. As it happens, there’s an analytical blog post on these policies of some of the big AI companies that may interest some: Thoughts on the AI Safety Summit company policy requests and responses - Machine Intelligence Research Institute
There’s also a list that ranks best to worst… (just more data to consider if the AI answers were correct-ish)

2 Likes

Oh, wow. I didn’t think Windows Recall was yet implemented but apparently it may be. If you’re forced to use W11, remember to: Settings > Privacy & Security > Windows Permissions > Recall & Snapshots > UNCHECK “Save Snapshots” AND > Delete Snapshots > Delete all (haven’t confirmed this, copied from a website)

2 Likes

And to continue on the general topic: Meta doesn’t want to be less evil than MS, so they’ve upped their AI game (from Meta faces multiple complaints in Europe over AI data use • The Register)

Meta’s plans to use customer data in AI training have resulted in complaints to data protection authorities in 11 European countries.

The complaints were filed by privacy activist group noyb following updates to Meta’s privacy policy. The updates are due to take effect on June 26.

The main issue, according to noyb, are proposals by Meta to use years of posts – including images – “to develop and improve AI at Meta.” Private messages between the user and friends and family are not used to train the corporation’s AIs.

[…]
As we understand it, users in Europe will get the ability to opt out, due to GDPR, by going to the Privacy Policy page in their Facebook and Instagram apps, via the Settings and About screens, and checking out the Right to Object box. People outside of Europe are out of luck: There is no opt out coming.

[“Where to begin?” :face_vomiting:]

[Edit: Meanwhile at Apple: “Let’s brand it 'Apple Intelligence’” :person_facepalming:]

4 Likes

“… and say we invented it”. :rofl:

2 Likes

This is veering off-topic from AI, but there is some mediocre new news on that Windows AI Recall feature that are not about the AI (that’s a small side feature and less interesting at this point of the saga): Microsoft has some thoughts about Windows Recall security • The Register Some of the new details may be of interest, if you have to use Windows (like, at work). No celebration needed but now they’ve finally included some basic security features (I mean, finally, after it took the whole userspace collectively saying “… um, no, you forgot something”).

Two interesting notes that I recently read, related to this:

  1. Linux has kinda had similar, unprotected, feature for decades: bash history (have you pointed it to /dev/null and would you even know how to - and do you care? [no, it’s not the same, especially comparing the scale, but just as a reminder]), and
  2. We noticed because this is with our data devices, but no one seems to care about if it’s done with entertainment devices (btw. privacy and security alert, if you’re using a TV as a monitor for computer): Smart TVs take snapshots of what you watch multiple times per second | New Scientist and [2409.06203] Watching TV with the Second-Party: A First Look at Automatic Content Recognition Tracking in Smart TVs
2 Likes

I have far less than zero interest in something like windows ai recall. But the introduction of shell command history changed my life for the better and it would be very difficult for me to function without it.

2 Likes

7 posts were split to a new topic: Digression on controlling your bash history

It looks like we are finally getting some answers about this question, but surprisingly they are not at all what we thought. AI lie to us because they are taught to do so through human interaction and feedback; and also mainly because they are commercial models that need to bring revenue to their owners (or shareholders)!
Who would buy an AI service that half of the time would give you the dreaded: I don’t know? Not a very good selling point, indeed. We want LLMs to be question-answering machines. So companies started to address this issue: how to avoid this non-answers problem?
This is all explained in a study that was recently published in Nature, as reported in this article on Ars Technica:

This is a good read. Some insights from the research are really amazing!

“To speak confidently about things we do not know is a problem of humanity in a lot of ways. And large language models are imitations of humans,” says Wout Schellaert, an AI researcher at the University of Valencia, Spain, and co-author of the paper.

we got busy adjusting the AIs by hand. And it backfired.

“The notorious problem with reinforcement learning is that an AI optimizes to maximize reward, but not necessarily in a good way,”

Since it’s hard for humans to be happy with “I don’t know” as an answer, one thing this training told the AIs was that saying “I don’t know” was a bad thing. So, the AIs mostly stopped doing that.

When incorrect answers were flagged, getting better at giving correct answers was one way to optimize things. The problem was getting better at hiding incompetence worked just as well. Human supervisors simply didn’t flag wrong answers that appeared good and coherent enough to them.
In other words, if a human didn’t know whether an answer was correct, they wouldn’t be able to penalize wrong but convincing-sounding answers.

The AIs lie because we told them that doing so was rewarding. One key question is when and how often do we get lied to.

The more difficult the question and the more advanced model you use, the more likely you are to get well-packaged, plausible nonsense as your answer.

ChatGPT emerged as the most effective liar. The incorrect answers it gave in the science category were qualified as correct by over 19 percent of participants. It managed to fool nearly 32 percent of people in geography and over 40 percent in transforms, a task where an AI had to extract and rearrange information present in the prompt.

So, in conclusion, I would recall this saying from a developer of PureBoot, that we should never trust what our display is telling us, and especially not if it is asking for a password. In a similar way, we should never take at face value what a commercial AI LLM tells us: it could just as well be sophisticated and eloquent BS.

1 Like

This is so wrong. The article is right, but it’s explaining things wrong [in way that can be misinterpreted] - and almost solely just by that one word (“lie”) and what it’s conveying. AIs are not minds that think, reason, nor do they have a self or consciousness. And that is why AIs do not have intent. So, they can not “lie” because that implies an intent to deceive, which AIs can not do [only the system programmer may have added some of their own intent to the AI model, but well get to that later]. The correct term would be the colloquial “hallucinate” (which is taken from human psychology but has become to describe a very different phenomena that only superficially resembles it, so not the best of terms unless the context is clear) where AIs give false or incorrect statements in regards to the inquiry, while the statement as such may be seemingly coherent and logical. This is all because the GPT/LLM type AI models are statistics based answering machines and those formulate statements based on huge databases of all kinds of (text) data where the amount has been more important than quality (and even if it wasn’t the the sheer diversity of texts means that there are arguments from various viewpoints, synonyms, homonyms, translation incompatibilities etc. which the algorithms are not that good at recognizing) based on the likelyhood of what a sentence connected to the words in the inquiry should have (word by word). So, it’s natural for those machines to spew out statistically anything, except the algos are now so good that the answers are very often right enough or close enough to what we need.

The reported test is interesting in that how the different models comparare and have developed, but the main point to notice is that one of the prime methods of AI learning was intentionally broken by limiting the use of “I don’t know”. Btw. being able to get AIs to reliably say “I don’t know” (or something similar) is a huge thing, a very good result, as the statistical limits and error correction methods are able to draw a line where statistical uncertainty is an issue and statement would probably be false (kinda like guesses that uneducated humans make). So…

I would. Everyone should. It would be amazing. Because that would mean that half the time you get almost certainty and good answers that you can trust. That - being able to fully trust the output - is more important atm., or so I argue (there may be some applications where any output is more desirable - consider generation of fantastical images, which are not true or possible according to physics etc.).

Coming back to that test, it’s fascinating that the algos took this reinforcement towards this route. It’s very logical though. AIs are simply applying the programming of trying to do better but, just as it has not capacity to understand, it has no capacity to discern right from wrong or other moral questions related to intent, and so AIs did what produced acceptable feedback the simplest way. The comparison to human behavior in this is apt, at abstract level. But this is the programmers doing. They are the ones that have created the algorithm and - this has to be stressed - it’s unlikely that at base model level there would be any intent to add a “lie about these things” feature, just because it would be so hard to include it and make the whole work (analogy: think how hard it is for humans to keep up a convincing lie about one area of life that connects to all others, while constantly being questioned and prodded). In research (and I apologize not including links, I don’t have them at hand now) it has be shown that in complex systems all the human biases and flaws and cultural ideals can be transferred from coders and it’s unintentional and hard to spot (example: in facial recognition, which features are considered prominent or desirable, or how language structure is processed based on what is your mother tongue and understanding of different languages).

So, AIs (as in: AI models) do not “lie” but make mistakes because of imperfect processing of what is wanted, influenced by these unknown complex statistical biases in algorithms and due to the less than perfect (history based [does not know new things]) data. But as can be seen, those are some pretty good systems, since they are able to correct themselves by algorithmic learnings like reinforcement by feedback.

But, there is another level to systems - which is probably more interesting, if you want to pinpoint where the dragons may lie. The modern AI system consists of the model but also the rest of the system, which has many separate parts dedicated to risk management of inputs and outputs and system security etc. For instance, the Copilot dashboard has several simplistic sliders that allow admin to deploy AI and select some of the characteristics, in addition to being able to define a “personality” via text prompt. These sliders and text are interpreted in the system and connect a whole bunch of subsystems and algorithms (which are not open code). In addition to these, there are some restrictions that are not admin selectable but are more or less hardcoded (changeable by system provider only, MS etc.). Although a bit specific, allowing user to bypass “I don’t know” is something modern large systems from the big companies offered to public use is a feature that would/should not be uncontrolled, but that’s a separate issue [forcing the level high would potentially make AIs more worth trusting in long term IMHO].

Anyways, coming to the more important point after the long setup: at this system level, there are separate controls and among those it could theoretically (because no evidence has been presented and there are actually some cases that have shown opposite) be controls that system programmers could use to make AI do things like give outputs that intentionally are not what the AI would might spew out. These are already used to curtail swearing and avoiding harmful topics (like self harm). At system level there have for some time been filters for certain content but those have been deemed acceptable and good, but those have nothing to do with AIs as such. [There is more censoring with public models because some users are just there to brake things or be lewd but for instance in internal/private medical applications obviously there is a need to use anatomical references, so limits are different.]

So, when saying “AIs lie”, I see that silly, as AIs make unintentional errors that are not errors because it’s what they were coded to do, and the final output is anyway controlled by someone else. @TiX0 conclusion that we should not trust our displays, is mostly correct, in that we should always have healthy skepticism online (regardless of AI or not, just to expand on it), but understanding why that is also important.

The article is a bit misleading in the choice of wording and about the point of such study and the selected passages of previous post reinforce those. A whole separate argument could also be made about how identifying a false statement and a lie differ and how differently we interpret information when communicating face-to-face (all the micro signals we read form people when they speak/lie), which are not present with AI text prompt output. And there’s something to wonder about just how well these test subjects understood the areas where they were “lied” to (as they were supposed to spot the falsehoods) - the research even makes a mention of this limitation. The original research paper is more specific. It more or less makes a point about how large AI systems kinda try too hard to answer something, which gets them into trouble. The whole final conclusion in it is about how the level, when to say “I don’t know”, should be optimized. What is forgotten though, is that for many applications these GPT/LLM type AIs and the language/text based statistical answers should not be used at all, even though they are popular at the moment. There are other AI types that may be more suitable to the problem and task.

3 Likes

This missed my radar but last week, the FSF announced that they are participating in NIST’s safety consortium of many different institutions to offer their stance and understanding of AI (they originally announced their participation during the LibrePlanet 2024 conference but I didn’t watch the talks). I don’t know how effective this move will be but it’s a step up from just criticizing the direction AI has been going on from the sidelines.

https://www.fsf.org/news/fsf-serves-on-nist

3 Likes