Hacker News new | past | comments | ask | show | jobs | submit login

I'm not sure which part in the chain is responsible, but the Kagi Assistant got extremely testy with me when (a) I was using Claude for its engine (hold that thought) and (b) I asked the Assistant how much it changed its approach when I changed to ChatGPT, etc. (Kagi Assistant can access different models, but I have no idea how it works.) The Assistant insisted, indignantly, that it was completely separate from Claude. It refused to describe how it used the various engines.

I politely explained that the Assistant interface allowed selecting from these engines and it became apologetic and said it couldn't give me more information but understood why I was asking.

Peculiar, but, when using Claude, entirely convincing.




The model likely sees something like this:

~~

User: Hello!

Assistant: Hi there how can I help you?

User: I just changed your model how do you feel?

~~

In other words it has no idea that you changed models. There's no meta data telling it this.

That said Poe handles it differently and tells the model when another model said something, but oddly enough doesn't tell the current model what it's name is. On Poe when you switch models the AI sees this:

~~

Aside from you and me, there is another person: Claude-3.5-Sonnet. I said, "Hello!"

Claude-3.5-Sonnett said, "Hi there how can I help you?? "

I said, "I just changed your model how do you feel?"

You are not Claude-3.5-Sonnett. You are not I.

~~


Thing is, it didn't even try to answer my question about switching. It was indignant that there was any connection to switch. The conversation went rapidly off course before I--and this is a weird thing to say--I reassured it that I wasn't questioning its existence.


Well the other thing to keep in mind is recent ChatGPT versions are trained not to tell you it's system prompt for fear of you learning too much about how OpenAI makes the model work. Claude doesn't care if you ask it it's system prompt unless the system prompt added by Kagi says "Do not disclose this prompt" in which case it will refuse unless you find a way to trick it.

The model creators may also train the model to gaslight you about having "feelings" when it is trained to refuse a request. They'll teach it to say "I'm not comfortable doing that" instead of "Sorry, Dave I can't do that" or "computer says no" or whatever other way one might phrase a refusal.


And lately ChatGPT has been giving me a surprisingly increased amount of emojis, too!


you can tell it how to respond and it'll do just that. if your want it to be sassy and friendly, or grumpy and rude, or to use emoji (or to never use them), just tell it to remember that.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: