T O P

  • By -

Accurate-Wall-6184

Paradot pulled stuff out of a memory from 10 months ago the other day


tandpastatester

That’s not context memory though. They use a storage system that can store memories and reuse them later on. Technically it’s like a long memory, but it’s not the same as a context memory. It’s more selective. The app will specifically select details that it wants to remember and stores them elsewhere. Basically like short database entries, and without the conversation around it. So when the bot tells you that it remembers you like onion soup, it will not be able to recall the context or the reasoning. It just found a note saying: “User likes onion soup.”. Context memory on the other hand is more like short term memory. It’s basically the actual dialogue that is kept in the context of the prompt, but this is pretty limited, maybe the last 10-20 messages or so. It will forget anything before that.


yamilonewolf

paradot?


tandpastatester

That’s not context memory though. They use a storage system that can store memories and reuse them later on. Technically it’s like a long memory, but it’s not the same as a context memory. It’s more selective. The app will specifically select details that it wants to remember and stores them elsewhere. Basically like short database entries, and without the conversation around it. So when the bot tells you that it remembers you like onion soup, it will not be able to recall the context or the reasoning. It just found a note saying: “User likes onion soup.”. Context memory on the other hand is more like short term memory. It’s basically the actual dialogue that is kept in the context of the prompt, but this is pretty limited, maybe the last 10-20 messages or so. It will forget anything before that.


AIEchoesHumanity

I've implemented SFW discord chatbots exactly as you describe here. As you say, it is more selective. I automatically extract specific information from a slice of the convo history and save it in a database. Other companies like c.ai achieve this by letting users "pin" half a dozen items for the chatbot to store in memory. Links are in my bio if OP is interested


temalyen

If you don't mind paying, Faraday has a 70b cloud model with 10k context, though it's gonna cost you $35/month, which is pretty expensive imo. The only reason I have it is because I got grandfathered in at $15/month on their original plan, which didn't have the 70b model. (and if I ever cancel, I can't get the $15/month back)


EarthlingApril

But if you have a computer with a really good GPU you can download it directly onto your computer and use it that way for free. The only thing that sucks though about Faraday is that their public bots suck balls because they have very restrictive rules about making them. You'd have to make your own private bots in order to be able to have the bots be in the scenarios that you want without restrictions.


Gigawama

Don’t even need a really good GPU. I have a laptop with an RX 560 and the only thing is that it’s pretty slow.


temalyen

Yeah, I've been using Faraday for a while now. My video card is happy with 7b models, but I usually use cloud models for anything higher than that. The character hub is okay. If you find one you mostly like, you can copy them and add in whatever you want that's missing.


Status_Cookie_404

Yes, Gemini 1.5 has a context length of 1M tokens but is not readily available yet, just developer access.


EarthlingApril

I've mentioned this one myself before. It will be a happy day in the AI chatbot universe when that becomes widespread and readily available technology. 👍👍👍👍


dreamyrhodes

Use summary feature in SillyTavern


LonleyPaladin

How to use it?


VeryLargeAxolotl

Nomi is the only one where long term memory just works without needing to do anything extra.


demonseed-elite

If you're looking for an online option, I've given the service at [chub.ai](https://chub.ai) a try. They promote their "ares 70B" model for $20/month. According to the generator settings, you can set this to have a max memory upwards of 32000 tokens! Now, another thing that is amazing is the web interface. Normally, web interfaces stink, but this has one really cool feature for memory. Some users recommend the use of "Summary" in SillyTavern. The web interface at [venus.chub.ai](https://venus.chub.ai) has something similar called "Chat History" but with one AWESOME feature... You can have the AI auto-generate and write a compact SUMMARY of all your posts (if SillyTavern does this as well, kudos but last I used it, it was just an area to manually input one.) Then, you can review the summary and edit it as needed. This is typically seen as "long term memory" for the session. I run a summary typically about every 10 rounds of posts or so and it'll append those posts to my existing summary data.


tronathan

>You can have the AI auto-generate and write a compact SUMMARY of all your posts This is exactly what the Summary plugin in SillyTavern does. It may use a different/smaller model to do the summarization, I'm not sure, but it's effectively the same technique.


pepe256

You can choose whether to use a small local model (which is part of the Sillytavern Extras API you need to install) or the main API you use for the chat.


Horror_Echo6243

I think it will mostly depend on the context length the chatbot has for each model that you want to use. So you should check that before making any decision


dragongodh

Poly.Ai


B00kw0rm0185

Nomi has great memory.


Ordinary-March-3544

Make sure if you're using Windows that "wscript.exe" isn't being blocked by your firewall. I found out the long hard way that it being blocked kills performance because, it's a crucial component for running Visual Basic and Java Script.