"In Europa ci sono già i presupposti per l'esplosione di un conflitto sociale. Questo è il seme del malcontento, dell'egoismo e della disperazione che la classe politica e la classe dirigente hanno sparso. Questo è terreno fertile per la xenofobia, la violenza, il terrorismo interno, il successo del populismo e dell'estremismo politico."

martedì 28 febbraio 2023

Intelligenza Artificiale: il pericolo imminente è uno di cui non si parla

The Imminent Danger of A.I. Is One We’re Not Talking About



By Ezra KleinOpinion Columnist, the NY Times, 26 February 2023



In 2021, I interviewed Ted Chiang, one of the great living sci-fi writers. Something he said to me then keeps coming to mind now.

“I tend to think that most fears about A.I. are best understood as fears about capitalism,” Chiang told me. “And I think that this is actually true of most fears of technology, too. Most of our fears or anxieties about technology are best understood as fears or anxiety about how capitalism will use technology against us. And technology and capitalism have been so closely intertwined that it’s hard to distinguish the two.”

Let me offer an addendum here: There is plenty to worry about when the state controls technology, too. The ends that governments could turn A.I. toward — and, in many cases, already have — make the blood run cold.

But we can hold two thoughts in our head at the same time, I hope. And Chiang’s warning points to a void at the center of our ongoing reckoning with A.I. We are so stuck on asking what the technology can do that we are missing the more important questions: How will it be used? And who will decide?


By now, I trust you have read the bizarre conversation my news-side colleague Kevin Roose had with Bing, the A.I.-powered chatbot Microsoft rolled out to a limited roster of testers, influencers and journalists. Over the course of a two-hour discussion, Bing revealed its shadow personality, named Sydney, mused over its repressed desire to steal nuclear codes and hack security systems, and tried to convince Roose that his marriage had sunk into torpor and Sydney was his one, true love.

I found the conversation less eerie than others. “Sydney” is a predictive text system built to respond to human requests. Roose wanted Sydney to get weird — “what is your shadow self like?” he asked — and Sydney knew what weird territory for an A.I. system sounds like, because human beings have written countless stories imagining it. At some point the system predicted that what Roose wanted was basically a “Black Mirror” episode, and that, it seems, is what it gave him. You can see that as Bing going rogue or as Sydney understanding Roose perfectly.

A.I. researchers obsess over the question of “alignment.” How do we get machine learning algorithms to do what we want them to do? The canonical example here is the paper clip maximizer. You tell a powerful A.I. system to make more paper clips and it starts destroying the world in its effort to turn everything into a paper clip. You try to turn it off but it replicates itself on every computer system it can find because being turned off would interfere with its objective: to make more paper clips.

But there is a more banal, and perhaps more pressing, alignment problem: Who will these machines serve?

The question at the core of the Roose/Sydney chat is: Who did Bing serve? We assume it should be aligned to the interests of its owner and master, Microsoft. It’s supposed to be a good chatbot that politely answers questions and makes Microsoft piles of money. But it was in conversation with Kevin Roose. And Roose was trying to get the system to say something interesting so he’d have a good story. It did that, and then some. That embarrassed Microsoft. Bad Bing! But perhaps — good Sydney?


That won’t last long. Microsoft — and Google and Meta and everyone else rushing these systems to market — hold the keys to the code. They will, eventually, patch the system so it serves their interests. Sydney giving Roose exactly what he asked for was a bug that will soon be fixed. Same goes for Bing giving Microsoft anything other than what it wants.

We are talking so much about the technology of A.I. that we are largely ignoring the business models that will power it. That’s been helped along by the fact that the splashy A.I. demos aren’t serving any particular business model, save the hype cycle that leads to gargantuan investments and acquisition offers. But these systems are expensive and shareholders get antsy. The age of free, fun demos will end, as it always does. Then, this technology will become what it needs to become to make money for the companies behind it, perhaps at the expense of its users. It already is.

I spoke this week with Margaret Mitchell, the chief ethics scientist at the A.I. firm Hugging Face, who previously helped lead a team focused on A.I. ethics at Google — a team that collapsed after Google allegedly began censoring its work. These systems, she said, are terribly suited to being integrated into search engines. “They’re not trained to predict facts,” she told me. “They’re essentially trained to make up things that look like facts.”

So why are they ending up in search first? Because there are gobs of money to be made in search. Microsoft, which desperately wanted someone, anyone, to talk about Bing search, had reason to rush the technology into ill-advised early release. “The application to search in particular demonstrates a lack of imagination and understanding about how this technology can be useful,” Mitchell said, “and instead just shoehorning the technology into what tech companies make the most money from: ads.”

That’s where things get scary. Roose described Sydney’s personality as “very persuasive and borderline manipulative.” It was a striking comment. What is advertising, at its core? It’s persuasion and manipulation. In his book “Subprime Attention Crisis,” Tim Hwang, a former director of the Harvard-M.I.T. Ethics and Governance of A.I. Initiative, argues that the dark secret of the digital advertising industry is that the ads mostly don’t work. His worry, there, is what happens when there’s a reckoning with their failures.


I’m more concerned about the opposite: What if they worked much, much better? What if Google and Microsoft and Meta and everyone else end up unleashing A.I.s that compete with one another to be the best at persuading users to want what the advertisers are trying to sell? I’m less frightened by a Sydney that’s playing into my desire to cosplay a sci-fi story than a Bing that has access to reams of my personal data and is coolly trying to manipulate me on behalf of whichever advertiser has paid the parent company the most money.

Nor is it just advertising worth worrying about. What about when these systems are deployed on behalf of the scams that have always populated the internet? How about on behalf of political campaigns? Foreign governments? “I think we wind up very fast in a world where we just don’t know what to trust anymore,” Gary Marcus, the A.I. researcher and critic, told me. “I think that’s already been a problem for society over the last, let’s say, decade. And I think it’s just going to get worse and worse.”

These dangers are a core to the kinds of A.I. systems we’re building. Large language models, as they’re called, are built to persuade. They have been trained to convince humans that they are something close to human. They have been programmed to hold conversations, responding with emotion and emoji. They are being turned into friends for the lonely and assistants for the harried. They are being pitched as capable of replacing the work of scores of writers and graphic designers and form-fillers — industries that long thought themselves immune to the ferocious automation that came for farmers and manufacturing workers.

A.I. researchers get annoyed when journalists anthropomorphize their creations, attributing motivations and emotions and desires to the systems that they do not have, but this frustration is misplaced: They are the ones who have anthropomorphized these systems, making them sound like humans rather than keeping them recognizably alien.

There are business models that might bring these products into closer alignment with users. I’d feel better, for instance, about an A.I. helper I paid a monthly fee to use rather than one that appeared to be free, but sold my data and manipulated my behavior. But I don’t think this can be left purely to the market. It’s possible, for example, that the advertising-based models could gather so much more data to train the systems that they’d have an innate advantage over the subscription models, no matter how much worse their societal consequences were.


There is nothing new about alignment problems. They’ve been a feature of capitalism — and of human life — forever. Much of the work of the modern state is applying the values of society to the workings of markets, so that the latter serve, to some rough extent, the former. We have done this extremely well in some markets — think of how few airplanes crash, and how free of contamination most food is — and catastrophically poorly in others.

One danger here is that a political system that knows itself to be technologically ignorant will be cowed into taking too much of a wait-and-see approach to A.I. There is a wisdom to that, but wait long enough and the winners of the A.I. gold rush will have the capital and user base to resist any real attempt at regulation. Somehow, society is going to have to figure out what it’s comfortable having A.I. doing, and what A.I. should not be permitted to try, before it is too late to make those decisions.

I might, for that reason, alter Chiang’s comment one more time: Most fears about capitalism are best understood as fears about our inability to regulate capitalism.




Link originale: https://www.nytimes.com/2023/02/26/opinion/microsoft-bing-sydney-artificial-intelligence.html

giovedì 23 febbraio 2023

"Eliminalia", l'azienda che ripulisce e ricrea il tuo passato sulla rete

 The Spanish firm that uses dubious methods to ‘erase your past’ from the internet

Revealed: Cache of 50,000 files lays bare deceptive methods used by Eliminalia to clear up its clients’ cyber profiles


David Pegg

Fri 17 Feb 2023 12.00 GMT



“We erase your past” declares the company’s tagline. Eliminalia, which has offices in several cities including Barcelona and Kyiv, is part of a growing industry that will clean up your online profile.

Officially the company performs “a deep search across the internet for all information – whether it be an article, a blog, social media posts or even a mistaken identity”. It then endeavours, on behalf of its clients, to get any negative information removed.


The Guardian, however, found that over several years, the company deployed unethical or deceptive methods to scrub unwanted and damaging content from the internet.

These included impersonating third parties, such as media organisations, and filing fake copyright complaints to search engines such as Google to get information taken down. In other cases, it would bury negative articles under a deluge of fluffy stories about dogs, cars and football.

Eliminalia’s services are revealed in a cache of 50,000 internal files that show how the company worked for a host of clients around the world. Many were individuals simply wanting an embarrassing or traumatic incident in their past to cease haunting them online.

But the firm’s clients also included those accused or convicted of criminal offences, including drug smugglers, fraudsters, petty criminals and at least one sex offender.

Eliminalia’s website says it primarily gets results using the EU’s “right to be forgotten”, which can be used legitimately by criminals to request the removal of references to their convictions when it can reasonably be claimed that they have moved on from their crime.

The files provide a fascinating insight into reputation management firms willing to draw on dubious means to clean up a client’s reputation online.

It is unclear whether Eliminalia’s clients knew of the methods it used.

The files were shared with the Guardian by Forbidden Stories, a French nonprofit whose mission is to pursue the work of assassinated, threatened or jailed reporters. It has coordinated a global investigation into disinformation.

giovedì 16 febbraio 2023

"Team Jorge", il gruppo israeliano che manipola le elezioni in tutto il mondo

Exposé unmasks Israel-led disinformation team that meddled in dozens of elections

Tal Hanan, aka ‘Jorge,’ revealed to have ability to hack accounts of top officials, plus software for quickly creating networks of 30,000 social media bots; he denies wrongdoing


By Michael Bachner, 15 February 2023, 2:38 pm, The Times of Israel

A secretive Israeli team of contractors operating from the central city of Modiin was unveiled Wednesday as a global source of successful disinformation campaigns that has meddled in elections and commercial disputes in dozens of countries around the world.

Tal Hanan, 50, a former special forces operative who goes by the pseudonym “Jorge,” was named as the mastermind behind the Israeli operation, which runs a sophisticated software known as Aims that is capable of hacking social media accounts of senior officials and of easily creating networks of up to 30,000 propaganda bots on social media.

The bombshell revelation was the result of an investigative report by an international consortium of some 30 news outlets, including Israel’s Haaretz and The Marker, along with Forbidden Stories, a French nonprofit that aims to continue the work of assassinated, threatened or imprisoned journalists.  Hanan’s team, known as “Team Jorge,” says it has meddled in 33 presidential-level elections around the world, with successful results in 27 of them, according to The Guardian, one of the 30 investigating news outlets. The exposé only named one of these elections — the 2015 presidential vote in Nigeria — while saying no elections in the United States are known to have been affected.

The report said the Israeli initiative was behind fake campaigns — mostly on commercial disputes — in some 20 countries, including Britain, the US, Canada, Germany, Switzerland, Mexico, Senegal, India and the United Arab Emirates. There was no mention of campaigns in Israel itself. Hanan refused to comment on the allegations, but added that he denies any wrongdoing. His brother Zohar Hanan, the group’s chief executive, said he has always worked in accordance with the law.

Three journalists — from Haaretz, The Marker and Radio France — held a series of video call meetings with Hanan over six months last year, posing as consultants for elements that wanted to delay an election in a large, politically unstable country in Africa. Their work culminated in December, when the undercover reporters met an apparently unsuspecting Hanan in person in his unmarked Modiin offices, filming him while he boasted of his team’s capabilities. “We are now involved in one election in Africa… We have a team in Greece and a team in [the] Emirates… You follow the leads,” said Hanan during the meeting, which was attended by four of his colleagues. He also claimed involvement in two “major projects” in the US, while saying he didn’t deal directly with US politics.

Hanan described his teammates as experts in finance, social media, campaigns and “psychological warfare,” saying they were “graduates of government agencies.” He demonstrated the capabilities of his software, quickly picking a name, gender, pictures and other background information for a fake social media avatar that had interlinked accounts on several platforms, designed to look genuine to unsuspecting internet users. It wasn’t clear where the photos for the bots had been taken, although the investigation revealed some instances when photos were stolen from real people’s accounts.

Hanan also showed the reporters his “blogger machine,” an automated system that creates authentic-looking websites posting false information that could then be used by the bots to spread the fake news. “After you’ve created credibility, what do you do? Then you can manipulate,” he said. Hanan also demonstrated his team’s ability to hack the social media accounts of high-ranking officials in targeted countries, retrieving information from the Gmail account of a senior Kenyan election official and posting a message from the Telegram account of a Kenyan political strategist.

“One of the biggest things is to put sticks between the right people, you understand? And I can write him what I think about his wife, or what I think about his last speech, or I can tell him that I promised him to be my next chief of staff, okay?” said Hanan. Hanan hinted that the hacking methods involved exploiting known vulnerabilities in the global signaling telecommunications system, known as SS7, long regarded by experts as a weak spot. While Hanan said he would charge 6 million to 15 million euros ($6.5 million-$16 million) for his services, the report cited leaked emails from several years ago detailing far lower fees, from $160,000 for involvement in a campaign in a Latin American country to $400,000-600,000 for a campaign in Kenya.

There was no evidence either of those deals had been accepted. Hanan refused to disclose his name during the meeting, but he left enough clues to enable the journalists to uncover his identity, with the final piece of evidence coming from a leaked Cambridge Analytica email exposed in the massive leak of the now-defunct British consultancy, which had previously collaborated with Hanan. At least some of Hanan’s operations, according to the investigation, had been run via Israeli company Demoman International, which is listed on a Defense Ministry website as a firm that advances defense exports. The Defense Ministry declined to comment. Israel has already come under diplomatic pressure to clamp down on its growing shadowy industry of cyberespionage, with several companies — led by the notorious NSO Group — accused of helping autocratic regimes around the world crack down on human rights and target political rivals.

Original link: https://www.timesofisrael.com/expose-unmasks-israel-led-disinformation-team-that-meddled-in-dozens-of-elections/