bolha.us is one of the many independent Mastodon servers you can use to participate in the fediverse.
We're a Brazilian IT Community. We love IT/DevOps/Cloud, but we also love to talk about life, the universe, and more. | Nós somos uma comunidade de TI Brasileira, gostamos de Dev/DevOps/Cloud e mais!

Server stats:

249
active users

#slms

0 posts0 participants0 posts today

"To prevent AI models from memorizing their input, we know exactly one robust method: differential privacy (DP). But crucially, DP requires you to precisely define what you want to protect. For example, to protect individual people, you must know which piece of data comes from which person in your dataset. If you have a dataset with identifiers, that's easy. If you want to use a humongous pile of data crawled from the open Web, that's not just hard: that's fundamentally impossible.

In practice, this means that for massive AI models, you can't really protect the massive pile of training data. This probably doesn't matter to you: chances are, you can't afford to train one from scratch anyway. But you may want to use sensitive data to fine-tune them, so they can perform better on some task. There, you may be able to use DP to mitigate the memorization risks on your sensitive data.

This still requires you to be OK with the inherent risk of the off-the-shelf LLMs, whose privacy and compliance story boils down to "everyone else is doing it, so it's probably fine?".

To avoid this last problem, and get robust protection, and probably get better results… Why not train a reasonably-sized model entirely on data that you fully understand instead?"

desfontain.es/blog/privacy-in-

desfontain.esFive things privacy experts know about AI - Ted is writing things… and that AI salespeople don't want you to know!

"Rather than building massive, complex large language models, many organizations choose smaller language models that focus on niche applications such as supply chain management or inventory control. This is the “lean AI” concept, and it entails purpose-built models able to deliver value without the high costs and complexity associated with larger systems, according to Linthicum.

“We have Agentic AI and certainly using things like small language models where we’re leveraging generative AI and AI in general for more tactical implementation,” he said. “It’s dealing with supply chain integration, dealing with inventory control. We’re not building LLMs, and I don’t think the businesses out there are going to get the value from building huge LLMs that they think they’re going to get.”"

siliconangle.com/2024/12/31/ge

SiliconANGLE · The contrarian CIO approach to gen AI hype, explored - SiliconANGLECIOs push back on the gen AI hype, opting for lean AI and small wins over costly, large-scale implementations to deliver real business value.