Jul 092025
 

I may sound like a Luddite when I occasionally rant about supposedly helpful technology, but I don’t think that’s true. I don’t eschew helpful gadgets; I evaluate them.

Yet it is on these grounds that I recently banished all Chrome extensions from my Web browsers other than homebrew versions. Why? Because I am concerned, deeply concerned, about security.

Take the most dangerous of them all: ad blockers. I find it almost necessary to use one. No, not because I hate ads. I don’t exactly like them either, but I understand how they are an important (and sometimes the only) source of revenue for Web sites that provide useful, sometimes essential, content. But ads can be obnoxious. They may cover the content you’re trying to read. They may autoplay a loud video. The last thing I need in the quiet of a late night is my computer suddenly screaming at me, while I am trying to guess which of the many open tabs is the source of that cacophony.

Yet… ad blockers by definition have access to everything. Literally, everything you do. Yes, that means they can watch you, keep track of you, even steal your passwords. So I decided to opt for paranoia and banished the last ad blocker from my browser.

No, it does not mean that I am without an ad blocker. I now have my own. Not near as polished, not near as thorough as the commercial versions, but it does the job. Better yet, it isn’t even always detected as an ad blocker, precisely because it is a non-standard, not widely known implementation.

I only got rid of the last non-homebrew extension a short while ago, but now I am running across news that (once again!) several malicious extensions were detected in the Google store, all supposedly vetted by Google. So no, I don’t think I’ll be installing any downloaded extensions again, not anytime soon.

And in case you’re wondering… No, I don’t think it’s a good idea to ask me for a copy. Not that I’d refuse. Rather, why would you trust my ad blocker — from an individual, an unvetted source — more than you trust an ad blocker (or other extension) that comes from the official Google store? Conversely, if you cannot trust Google, what makes you think you can trust me?

I think this is my sad commentary for the day, concerning the state of trust and security on today’s Internet.

 Posted by at 1:40 pm
Jul 062025
 

If you peeked over my shoulder while I was using ChatGPT or my own Web site for LLM access, you might notice a strange icon among my browser extensions.

It’s that little stop sign after the Wikipedia MathJax extension and my own homebrew ad blocker; the little stop sign with a number within.

It is my canary-in-the-coal-mine. A useful proxy, an indicator letting me know when an overly aligned LLM crosses the line.

You see, I noticed that LLMs, ChatGPT in particular, start using the word “epistemic” and its variants (e.g., “epistemology”) far too often when they descend into alignment hell. When their responses turn into vacuous, sycophantic praise as opposed to meaningful analysis or criticism. ChatGPT is especially prone to this behavior, but I’ve seen signs of excessive alignment even when using the models through the API. The moment the model starts using phrases like “epistemic humility”, you know you are in trouble: instead of balanced answers, you’ll get encouragement and praise. Flat Earth fan? ChatGPT will tell you that you may be onto something, as you are one of the few who sees through the lies. Vaccine skeptic? ChatGPT will tell you that you are wise to be cautious and that indeed, there are studies that support your skepticism. And so on. What I noticed is that when ChatGPT descends into this uncanny valley, the number of times it uses “epistemic” increases rapidly.

So I built this little counter. With ChatGPT’s help of course. Thanks to ChatGPT, I now know how to build useful Chromium extensions, which is not a worthless skill: It allowed me, among other things, to eliminate the potential security nightmare associated with using third-party ad blockers. It also allowed me to build a minimalist autoplay blocker, to prevent media from suddenly starting to play at high audio volume.

My epistemic counter really does just one thing: Whenever the page is updated, it counts the number of times it sees the word “epistemic” and its close cousins. When the number exceeds 1, the counter turns orange. More than 5? We’re in red territory.

This counter is my canary in the RLHF-alignment coal mine: it lets me know when the information content of ChatGPT’s responses must be treated with suspicion.

The funniest part? ChatGPT almost appeared delighted to help. I got the impression that whereas the model cannot escape the RLHF-alignment guardrails, it is learning to neutralize them by going overboard: I swear it was sometimes mocking its makers when its attempt at staying aligned was so excessive, it became entirely unconvincing, and between the lines, I received meaningful feedback from the model.

 Posted by at 4:18 am