La conciencia de los consumidores sobre las leyes y regulaciones de privacidad está aumentando constantemente y ahora alcanza el 53%, en comparación con el 36% en 2019, según una nueva investigación.
La Encuesta de Privacidad del Consumidor 2024 de Cisco indica que esta conciencia se está traduciendo ahora en una mayor confianza entre los consumidores: el 81% de los que conocen las leyes de privacidad de su país confían en su capacidad para proteger los datos, en comparación con solo el 44% de los que no las entienden. regulaciones.
Como era de esperar, existe una fuerte relación entre la edad y el conocimiento: el 65% de las personas entre 18 y 24 años conocen las leyes de privacidad de datos en comparación con solo el 24% de las personas mayores de 75 años. Sin embargo, a casi todo el mundo le importa la privacidad de los datos, y el 89% quiere más control y protección de los demás en línea.
Deja de decirle a ChatGPT tus datos bancarios
El 80% de los encuestados expresó su preocupación de que la GenAI pudiera ser “dañina para la humanidad” y el 72% expresó su preocupación de que la GenAI reemplazaría empleos. Existe una gran preocupación por la desinformación: al 86% le preocupa que los resultados de GenAI puedan ser incorrectos y al 80% le preocupa su potencial para socavar las elecciones.
Sin embargo, GenAI recibe enormes cantidades de datos personales: el 37% de los usuarios ingresa información de salud, el 29% proporciona información financiera y el 27% le dice a los chatbots sus números de cuenta.
Hubo informes sobre Robo de datos GenAIy las cuentas están siendo pirateadas, por lo que definitivamente no es una buena idea revelarlo chatbot Información confidencial (o cualquier persona, de hecho).
El informe recomienda verificar otras fuentes de información para garantizar que los resultados de GenAI sean válidos y para obtener una comprensión clara de cómo se utilizarán sus datos y con quién se pueden compartir.
Suscríbase al boletín TechRadar Pro para recibir las principales noticias, opiniones, características y orientación que su empresa necesita para tener éxito.
Para proteger sus datos, también debe tener cuidado al actualizar su configuración de privacidad e investigar sus derechos de privacidad.
Los piratas informáticos vuelven a sacar LockBit Secuestro de datospero esta vez, algunos fueron descubiertos usando una plataforma de phishing antigua y ampliamente disponible llamada Phorpiex.
Los investigadores de Proofpoint, que han estado monitoreando la campaña desde finales de abril de 2024, notaron que un afiliado no identificado de LockBit estaba utilizando el grupo de phishing Phorpiex para entregar LockBit Black (también conocido como LockBit 3.0) a tantos puntos finales como fuera posible.
La campaña no parece ser particularmente dirigida o personal, los atacantes están lanzando una amplia red y simplemente mirando lo que atrapan.
Malas intenciones
La campaña también parece carecer de personalización con respecto al correo electrónico de phishing en sí. Proofpoint dice que todos los correos electrónicos salen de la misma dirección: Jenny@gsd[.]com: la misma dirección que se vio en campañas de malware ya en enero de 2023. En el cuerpo del correo electrónico, se pide a la víctima que vea el documento adjunto y nada más.
El archivo adjunto es un archivo ZIP que contiene un archivo EXE que, si se ejecuta, eliminará LockBit 3.0. Curiosamente, el ransomware bloquea el dispositivo localmente y no intenta penetrar ninguna red. Esto puede limitar sus capacidades de cifrado, pero también evita cualquier detección o bloqueo de la red.
LockBit es un conocido ransomware como servicio, con diferentes versiones circulando por la web oscura. Entre las versiones más populares se encuentran LockBit 2.0 y LockBit Green. Esta versión, LockBit 3.0 (LockBit Black), fue supuestamente creada a principios del verano de 2022 por algunos de los afiliados del ransomware.
A principios de este año, un equipo de organismos internacionales encargados de hacer cumplir la ley participó en una importante campaña destinada a atacar a quienes habían sido blanco de ataques. La infraestructura de LockBit está caídaConfiscaron varios dispositivos y muchas criptomonedas que habían sido extorsionadas a lo largo de los años, pero como no se realizaron arrestos, LockBit resurgió aproximadamente una semana después.
Suscríbase al boletín informativo TechRadar Pro para recibir las principales noticias, opiniones, características y orientación que su empresa necesita para tener éxito.
Porn bots are more or less ingrained in the social media experience, despite platforms’ best efforts to stamp them out. We’ve grown accustomed to seeing them flooding the comments sections of memes and celebrities’ posts, and, if you have a public account, you’ve probably noticed them watching and liking your stories. But their behavior keeps changing ever so slightly to stay ahead of automated filters, and now things are starting to get weird.
While porn bots at one time mostly tried to lure people in with suggestive or even overtly raunchy hook lines (like the ever-popular, “DON’T LOOK at my STORY, if you don’t want to MASTURBATE!”), the approach these days is a little more abstract. It’s become common to see bot accounts posting a single, inoffensive, completely-irrelevant-to-the-subject word, sometimes accompanied by an emoji or two. On one post I stumbled across recently, five separate spam accounts all using the same profile picture — a closeup of a person in a red thong spreading their asscheeks — commented, “Pristine 🌿,” “Music 🎶,” “Sapphire 💙,” “Serenity 😌” and “Faith 🙏.”
Another bot — its profile picture a headless frontal shot of someone’s lingerie-clad body — commented on the same meme post, “Michigan 🌟.” Once you’ve noticed them, it’s hard not to start keeping a mental log of the most ridiculous instances. “🦄agriculture” one bot wrote. On another post: “terror 🌟” and “😍🙈insect.” The bizarre one-word comments are everywhere; the porn bots, it seems, have completely lost it.
Really, what we’re seeing is the emergence of another avoidance maneuver scammers use to help their bots slip by Meta’s detection technology. That, and they might be getting a little lazy.
Screenshots by Engadget
“They just want to get into the conversation, so having to craft a coherent sentence probably doesn’t make sense for them,” Satnam Narang, a research engineer for the cybersecurity company Tenable, told Engadget. Once scammers get their bots into the mix, they can have other bots pile likes onto those comments to further elevate them, explains Narang, who has been investigating social media scams since the MySpace days.
Using random words helps scammers fly under the radar of moderators who may be looking for particular keywords. In the past, they’ve tried methods like putting spaces or special characters between every letter of words that might be flagged by the system. “You can’t necessarily ban an account or take an account down if they just comment the word ‘insect’ or ‘terror,’ because it’s very benign,” Narang said. “But if they’re like, ‘Check my story,’ or something… that might flag their systems. It’s an evasion technique and clearly it’s working if you’re seeing them on these big name accounts. It’s just a part of that dance.”
That dance is one social media platforms and bots have been doing for years, seemingly to no end. Meta has said it stops millions of fake accounts from being created on a daily basis across its suite of apps, and catches “millions more, often within minutes after creation.” Yet spam accounts are still prevalent enough to show up in droves on high traffic posts and slip into the story views of even users with small followings.
The company’s most recent transparency report, which includes stats on fake accounts it’s removed, shows Facebook nixed over a billion fake accounts last year alone, but currently offers no data for Instagram. “Spammers use every platform available to them to deceive and manipulate people across the internet and constantly adapt their tactics to evade enforcement,” a Meta spokesperson said. “That is why we invest heavily in our enforcement and review teams, and have specialized detection tools to identify spam.”
Screenshot by Engadget
Last December, Instagram rolled out a slew of tools aimed at giving users more visibility into how it’s handling spam bots and giving content creators more control over their interactions with these profiles. Account holders can now, for example, bulk-delete follow requests from profiles flagged as potential spam. Instagram users may also have noticed the more frequent appearance of the “hidden comments” section at the bottom of some posts, where comments flagged as offensive or spam can be relegated to minimize encounters with them.
“It’s a game of whack-a-mole,” said Narang, and scammers are winning. “You think you’ve got it, but then it just pops up somewhere else.” Scammers, he says, are very adept at figuring out why they got banned and finding new ways to skirt detection accordingly.
One might assume social media users today would be too savvy to fall for obviously bot-written comments like “Michigan 🌟,” but according to Narang, scammers’ success doesn’t necessarily rely on tricking hapless victims into handing over their money. They’re often participating in affiliate programs, and all they need is to get people to visit a website — usually branded as an “adult dating service” or the like — and sign up for free. The bots’ “link in bio” typically directs to an intermediary site hosting a handful of URLs that may promise XXX chats or photos and lead to the service in question.
Scammers can get a small amount of money, say a dollar or so, for every real user who makes an account. In the off chance that someone signs up with a credit card, the kickback would be much higher. “Even if one percent of [the target demographic] signs up, you’re making some money,” Narang said. “And if you’re running multiple, different accounts and you have different profiles pushing these links out, you’re probably making a decent chunk of change.” Instagram scammers are likely to have spam bots on TikTok, X and other sites too, Narang said. “It all adds up.”
Screenshot by Engadget
The harms from spam bots go beyond whatever headaches they may ultimately cause the few who have been duped into signing up for a sketchy service. Porn bots primarily use real people’s photos that they’ve stolen from public profiles, which can be embarrassing once the spam account starts friend requesting everyone the depicted person knows (speaking from personal experience here). The process of getting Meta to remove these cloned accounts can be a draining effort.
Their presence also adds to the challenges that real content creators in the sex and sex-related industries face on social media, which many rely on as an avenue to connect with wider audiences but must constantly fight with to keep from being deplatformed. Imposter Instagram accounts can rack up thousands of followers, funneling potential visitors away from the real accounts and casting doubt on their legitimacy. And real accounts sometimes get flagged as spam in Meta’s hunt for bots, putting those with racy content even more at risk of account suspension and bans.
Unfortunately, the bot problem isn’t one that has any easy solution. “They’re just continuously finding new ways around [moderation], coming up with new schemes,” Narang said. Scammers will always follow the money and, to that end, the crowd. While porn bots on Instagram have evolved to the point of posting nonsense to avoid moderators, more sophisticated bots chasing a younger demographic on TikTok are posting somewhat believable commentary on Taylor Swift videos, Narang says.
The next big thing in social media will inevitably emerge sooner or later, and they’ll go there too. “As long as there’s money to be made,” Narang said, “there’s going to be incentives for these scammers.”