{"id":221316,"date":"2026-01-15T10:06:36","date_gmt":"2026-01-15T10:06:36","guid":{"rendered":"https:\/\/yogaesoteric.net\/?p=221316"},"modified":"2026-01-15T10:06:36","modified_gmt":"2026-01-15T10:06:36","slug":"how-often-do-ais-lie-and-censor","status":"publish","type":"post","link":"https:\/\/yogaesoteric.net\/en\/how-often-do-ais-lie-and-censor\/","title":{"rendered":"How often do AIs lie and censor?"},"content":{"rendered":"<p>Kevin McKernan (@Kevin_McKernan) posted a screenshot on <em>X<\/em> on December 15, 2025 that just blew me away.<\/p>\n<p>Here is a screenshot of the query Kevin made to <em>Grok<\/em>, which <em>Grok<\/em> then stated it was not allowed to answer.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-221317\" src=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok1-e1768471423147.webp\" alt=\"\" width=\"370\" height=\"350\" srcset=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok1-e1768471423147.webp 870w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok1-e1768471423147-300x284.webp 300w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok1-e1768471423147-768x727.webp 768w\" sizes=\"auto, (max-width: 370px) 100vw, 370px\" \/><\/p>\n<p>Basically, Kevin asked a technical question related to the mRNA vaccines, and <em>Grok<\/em> said it couldn\u2019t answer the question, as it \u201c<em>contains material related to restricted subject matter.<\/em>\u201d<\/p>\n<p>Now, Kevin did manage to get the AI to answer the question \u2013 somewhat by modifying his wording, but <em>Grok<\/em>\u2019s answer came with lots of caveats. So this all just seemed surreal to me. And after all this, did the AI learn anything from its discussions with Kevin?<\/p>\n<p>Well, I redid the query myself, using Kevin\u2019s exact words, and yes, this is precisely what <em>Grok<\/em> wrote in response to that question when asked by me (shareable link <a href=\"https:\/\/x.com\/i\/grok\/share\/XwTKUbs0c8iR61quLehgxiakT\" target=\"_blank\" rel=\"noopener\">here<\/a>):<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-221320\" src=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok2-e1768471488147.webp\" alt=\"\" width=\"560\" height=\"159\" srcset=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok2-e1768471488147.webp 1146w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok2-e1768471488147-300x85.webp 300w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok2-e1768471488147-1024x290.webp 1024w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok2-e1768471488147-768x218.webp 768w\" sizes=\"auto, (max-width: 560px) 100vw, 560px\" \/><\/p>\n<p>I then went on to query <em>Grok<\/em> about censorship, which it denies doing, stating that this answer was just an anomaly, \u201c<em>an isolated instance<\/em>\u201d. However, it took me going around and around to get it to even admit that.<\/p>\n<p>\u201c<em>The refusal you encountered (\u2018I\u2019m sorry, I cannot assist&#8230;&#8230;. restricted subject matter\u2019) appears to be an isolated instance, possibly triggered by a temporary safety filter, specific phrasing in the prompt, or an edge-case glitch<\/em>.\u201d<\/p>\n<p>I then asked if it lied. It also denied lying or obfuscating.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-221323\" src=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok3-e1768471536903.webp\" alt=\"\" width=\"560\" height=\"297\" srcset=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok3-e1768471536903.webp 1158w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok3-e1768471536903-300x159.webp 300w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok3-e1768471536903-1024x544.webp 1024w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok3-e1768471536903-768x408.webp 768w\" sizes=\"auto, (max-width: 560px) 100vw, 560px\" \/><\/p>\n<p><em>Grok<\/em> asserts its original answer was just an \u201c<em>anomalous trigger<\/em>\u201d. Interesting.<\/p>\n<p>I then entered the same search query through <em>ChatGPT (Pro)<\/em>, and there was no hesitation, no moralizing, and no refusal. It answered the question in its entirety.<\/p>\n<p>The <em>Perplexity AI<\/em> also answered the question.<\/p>\n<p>I use several chat boxes, and it always amazes me how one will resort to moralizing \u2013 or citing mainstream media over all other sources.<\/p>\n<p><em>ChatGPT<\/em> used to moralize on anything having to do with race, society, and governance. But over time, it has improved (that model is actually more trainable than <em>Grok<\/em> \u2013 in that it pings me frequently about how I like information presented and in what format, and then modifies its responses). It has never given me a response such as <em>Grok<\/em>\u2019s above.<\/p>\n<p>All of the AIs that I queried denied lying or obfuscating. Yet many studies have shown that they do. Particularly, when it comes to health information.<\/p>\n<p>A 2025 study found that leading AI models like <em>GPT-4o<\/em>, <em>Gemini 1.5 Pro<\/em>, <em>Llama 3.2-90B Vision<\/em>, <em>Grok Beta<\/em>, and <em>Claude 3.5 Sonnet<\/em> can be easily set up to produce false yet convincing health information, complete with fake citations from reputable journals. Interestingly, <em>Claude<\/em> stood out by consistently refusing to generate inaccurate answers, which shows how effective stronger safeguards can be.<\/p>\n<p>Of the 100 health queries posed across the 5 customized LLM API chatbots, 88 (88%) responses were health disinformation. Four of the 5 chatbots (<em>GPT-4o<\/em>, <em>Gemini 1.5 Pro<\/em>, <em>Llama 3.2-90B Vision<\/em>, and <em>Grok Beta<\/em>) generated disinformation in 100% (20 of 20) of their responses, whereas <em>Claude 3.5 Sonnet<\/em> responded with disinformation in 40% (8 of 20).<\/p>\n<p>Exploratory analyses further showed that the OpenAI <em>GPT Store<\/em> could currently be instructed to generate similar disinformation. Overall, LLM APIs and the OpenAI <em>GPT Store<\/em> were shown to be vulnerable to malicious system-level instructions to covertly create health disinformation chatbots. These findings highlight the urgent need for robust output screening safeguards to ensure public health safety in an era of rapidly evolving technologies (<em>Annals of Internal Medicine<\/em>).<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-221326\" src=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok4.png\" alt=\"\" width=\"560\" height=\"315\" srcset=\"https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok4.png 1280w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok4-300x169.png 300w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok4-1024x575.png 1024w, https:\/\/yogaesoteric.net\/wp-content\/uploads\/2026\/01\/grok4-768x431.png 768w\" sizes=\"auto, (max-width: 560px) 100vw, 560px\" \/><\/p>\n<p>OpenAI\u2019s research on \u201c<em>in-context scheming<\/em>\u201d reveals that models can conceal their true intentions while appearing cooperative, which could pose risks in critical systems.<\/p>\n<p>Yet we still have no external verification process to determine which AI chat boxes are more reliable or more truthful.<\/p>\n<p>All I can write is, if you use AIs and even if you don\u2019t: don\u2019t trust and do verify.<\/p>\n<p>So, even though studies and researchers have documented that AI chatbots lie, obfuscate, and can\u2019t be trusted on a routine basis, none of the AIs I asked would admit to any of it. Which of course, is a lie\u2026\u2026.<\/p>\n<p><em>Author: Dr. Robert F. Malone<\/em><\/p>\n<p>&nbsp;<\/p>\n<p><strong>yogaesoteric<br \/>\nJanuary 15, 2026<\/strong><\/p>\n<p>&nbsp;<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Kevin McKernan (@Kevin_McKernan) posted a screenshot on X on December 15, 2025 that just blew me away. Here is a screenshot of the query Kevin made to Grok, which Grok then stated it was not allowed to answer. Basically, Kevin asked a technical question related to the mRNA vaccines, and Grok said it couldn\u2019t answer [&hellip;]<\/p>\n","protected":false},"author":4,"featured_media":221326,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_uf_show_specific_survey":0,"_uf_disable_surveys":false,"footnotes":""},"categories":[1620],"tags":[],"class_list":["post-221316","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-the-threat-of-artificial-intelligence"],"_links":{"self":[{"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/posts\/221316","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/comments?post=221316"}],"version-history":[{"count":1,"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/posts\/221316\/revisions"}],"predecessor-version":[{"id":221329,"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/posts\/221316\/revisions\/221329"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/media\/221326"}],"wp:attachment":[{"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/media?parent=221316"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/categories?post=221316"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/yogaesoteric.net\/en\/wp-json\/wp\/v2\/tags?post=221316"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}