16 points

Anonymous or not, you’re still feeding it data

permalink
report
reply
31 points

Not how that works.

permalink
report
parent
reply
1 point

These companies absolutely collect the prompt data and user session behavior. Who knows what kinda analytics they can use it for at any time in the future, even if it’s just assessing how happy the user was with the answers based on response. But having it detached from your person is good. Unless they can identify you based on metrics like time of day, speech patterns, etc

permalink
report
parent
reply
5 points
*

Prompt data is pointless and useless without a human to create a feedback loop for it, at which point it wouldn’t have context anyway. Also human effort to correct spelling dnd other user errors at the outset anyway. Hugely pointless and unreliable.

Not to mention, what good would it do for training? It wouldn’t help the model at all.

permalink
report
parent
reply
1 point

I’m curious, how does it work?

permalink
report
parent
reply
20 points

Not who you asked but you don’t want your AI to train itself based on the questions random users ask because it could introduce incorrect or offensive information. For this reason llms are usually trained and used in a separate step. If a user gave the llms private information you wouldn’t want it to learn that information and pass it on to other users so there are protections in place usually to stop it from learning new things while just processing requests.

permalink
report
parent
reply
21 points

Not really. Depending on the implementation.

It’s not like ddg is going to keep training their own version of llama or mistral

permalink
report
parent
reply
11 points

I think they mean that a lot of careless people will give the AIs personally identifiable information or other sensitive information. Privacy and security are often breached due to human error, one way or another.

permalink
report
parent
reply
15 points

But these open models don’t really take new input into their models at any point. They don’t normally do that type of inference training.

permalink
report
parent
reply
3 points

But that’s a human error as you said, the only way to fix it is by using it correctly as an user. AI is a tool and it should be handled correctly like any other tool, be it a knife, a car, a password manager, a video recording program, a bank app or whatever.

I think a bigger issue here is that many people don’t care about their personal information as much as their lives.

permalink
report
parent
reply
16 points
*

https://duckduckgo.com/duckduckgo-help-pages/aichat/ai-chat-privacy/

your conversations are not used to train chat models by DuckDuckGo or the underlying model providers

permalink
report
parent
reply
-16 points

And this is why I stopped using DDG. I swear, I’m just going to have to throw away my computer in the future if this fucking AI bullshit isn’t thrown away like the thieving, energy-sucking, lying pile of garbage that it is.

permalink
report
reply
15 points

If it’s using different AI models and allowing anonymity, I am not sure what’s the issue? Do you also object to using a calculator?

permalink
report
parent
reply
9 points

Calculator?! Those thieving, energy-sucking piles of garbage! Abacus till I die!

But seriously, AI is insidious in how it data mines us to give us answers, and data mines our questions to build profiles of users. I distrust assurances of anonymity by big data corpos.

permalink
report
parent
reply
3 points
*

I am not sure what method DDG is using for their model updates, I think it’s only fair if journalists follow up with them for clarification. Local LLMs, ones you can download to your machine for use, would circumvent privacy concerns if you’re not updating the weights in some way

Edit to clarify I meant updating the weights using online learning, but it’s still possible to update weights using pre trained weights you can download

permalink
report
parent
reply
6 points

You can disable ads and AI in the settings.

permalink
report
parent
reply
-5 points
*

I don’t see how we can prove this. Paying them to also spy on us is bad but allowing them replace our software c/localllama with their service is even worse. My funds are better spent on local AI development or device upgrade.

permalink
report
reply
11 points

Honest question. How does their service “replace” an open source LLM? If I’ve got locallama on my machine, how does using their service replace my local install?

permalink
report
parent
reply
-1 points
*

Yes, it does the same with less control, privacy.

permalink
report
parent
reply
2 points

So it isnt replacing it’s offering an alternative tradeoff with more convenience/less control. I dont see how thats a bad thing?

permalink
report
parent
reply
1 point

How anonymous is that thing ?
Ai needs data training & correction from us as user

permalink
report
reply
0 points

“Keep in mind that, as a model running through DuckDuckGo’s privacy layer, I cannot access personal data, browsing history, or user information. My responses are generated on-the-fly based on the input you provide, and I do not have the ability to track or identify users.”

permalink
report
parent
reply
5 points

Let’s be honest, regardless of whether or not this is true, it’s been instructed to say that.

permalink
report
parent
reply
2 points

You can train models of all kinds without disclosing anything personal about a user. Also see differential privacy

permalink
report
parent
reply
11 points

Training and fine tuning happens offline for LLMs, it’s not like they continuously learn by interacting with users. Sure, the company behind it might record conversations and use them to further tune the model, but it’s not like these models inherently need that

permalink
report
parent
reply
11 points
*

I could use that!

Update: it works fantastic and lets you switch easily to different AI models

permalink
report
reply

Technology

!technology@lemmy.world

Create post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


Community stats

  • 18K

    Monthly active users

  • 11K

    Posts

  • 519K

    Comments