bnaur

joined 1 year ago
[–] bnaur@lemmy.world 1 points 1 year ago* (last edited 1 year ago)

For years now I have only read ebooks on my phone, so one evening I decided to get back to the habit of reading real books.

So I take my time and carefully pick just the right book, gather some pillows, turn off the lights and lay comfortable on the couch. And after a few confused moments of flipping through pages I remembered that these fucking things didn't work in the dark. And I really don't like to read under a bright light anymore so back to reddit it was for that evening.

That said, I think I'll skip this one, doesn't sound too comfortable.

[–] bnaur@lemmy.world 1 points 1 year ago* (last edited 1 year ago) (1 children)

Speaking as just a hobbyist, a more developer oriented community focused on the topic would be nice, if someone is up to the task.

It's currently hard to find any good information about how to actually use LLMs as part of a software project as most of the related subreddits etc. are more focused on shitposting and you don't currently really want to talk about these in general tech/programming forums without a huge Don't shoot I'm not one of them! disclaimer.

[–] bnaur@lemmy.world 3 points 1 year ago (1 children)

Regarding little Bobby, is there any known guaranteed way to harden the current systems against prompt injections?

This is something that I'm personally more worried about than Skynet or mass unemployment now that everyone and their dog is rushing to integrate LLMs into to their systems (ok worried maybe a wrong word, but let's just say I have the popcorns ready for the moment the first mass breaches happen with something like the Windows Copilot).

[–] bnaur@lemmy.world 2 points 1 year ago (3 children)

At least I'm interested but more technical discussion about this would probably fit better in some comp sci or programming community? Though most of those are a bit hostile to the LLM related topics these days because of all the hype and low effort spam.

[–] bnaur@lemmy.world 2 points 1 year ago (1 children)

Is the whole "You are an LLM by OpenAI, system date is etc." prompt part of the system message?

A few days ago when I was talking about controlled natural languages with it and asked it to give a summary of the chat so far in Gellish it spit that out.

[–] bnaur@lemmy.world 2 points 1 year ago (3 children)

If these commands were in a system message it would generally refuse to help you.

Doesn't it usually fairly easily give its system message to the user? I have had that happen purely by accident.

[–] bnaur@lemmy.world 4 points 1 year ago (5 children)

I'm not sure if I'd call that reverse engineering any more than using a web browsers View Source feature.

But it's interesting how it works behind the scenes and that only way to get these models to interface with the external world is by using the natural language interface and hoping for the best.