It had only been a few years since the company that put up that sign suffered its own humiliating breach. Claiming that you can stop data breaches is a lot like claiming you can stop employees from stealing office supplies. You can minimize, you can reduce, you can mitigate. But you can’t actually stop someone from walking out the front door with a box of paperclips. Even the National Security Agency, the nation’s largest employer of mathematicians, can’t solve this problem.

Advertisement

Simply put, the best vulnerability scanner in the world can’t detect stupidity or the malicious intent of an employee. Just ask the CIA.

Zero Doomsday

In the first few years of business, Netragard’s team built up their reputation by hacking major Las Vegas casinos, pharmaceutical companies, and nuclear research—with permission, of course. The penetration testing side of the business was booming, but the company was also pulling in millions from the sale of zero-day exploits—a practice that remains controversial to this day, particularly among privacy advocates concerned about government spies stockpiling “cyber weapons.”

Advertisement

“We were making money doing offensive penetration testing,” Desautels says, “but we were also known throughout the zero-day world as being the go-to guys if you needed to break something.”

Who was it buying up all these previously undiscovered exploits? Desautels just smiles. “People always ask me that and I always give them the same answer,” he says. “Public and private sector.” So in other words, defense contractors, private security firms, and three-letter agencies. To Desautels, the idea that the government is stockpiling zero days for the purpose of mass surveillance is preposterous. “These exploits cost millions of dollars a piece,” he says. “And they have a shelf life of basically three months.” After that, the odds of getting detected rapidly increases; so do the odds that the vulnerability will be fixed.

Advertisement

For more than five years after the company was founded, the majority of Netragard’s zero-day customers were all based in America. But between 2013 and 2014, that changed. The company had received a special recommendation from a trusted buyer—which Desautels refers to only as “a very well known entity here in the US”—and so Netragard began dealing to a Milan-based firm known as Hacking Team, which had done business with the FBI, among other US agencies.

The decision to work with these foreign hackers would ultimately threaten not only Netragard’s lucrative zero-day business but its entire reputation.

Advertisement

Hacking Team, a major supplier of surveillance and intrusion software for governments worldwide, got hacked itself in 2015. Roughly 400GB of internal company emails and documents were dumped online, including invoices for zero-day purchases that traced back to Netragard. The emails were later vacuumed up by WikiLeaks, which tossed them into a searchable archive.

Tools Hackers Use for Corporate Espionage
Subtitles
  • Off
  • en

The leak revealed that Hacking Team had been peddling spyware to authoritarian regimes with abysmal human rights records. For example, the firm had sold surveillance software to a Sudanese intelligence service notorious for the systematic torture of students, journalists, and activists. Human rights workers tied the armed forces that relied on that intelligence to incidents of mass sexual violence.

Advertisement

That revelation heralded the end of Netragard’s exploit-brokering business—and several shadowy “programs” in the US along with it.

“I was disgusted,” Desautels says. “If we can’t trust the people who we’re selling to to keep these things within the right hands, we just don’t want to be apart of this.” Still, Desautels remains unapologetic about dealing in zero days “because a lot of good can come from this.” The wares brokered by Netragard had been deployed, he says, in operations tracking human traffickers, child pornography, and in various homeland security scenarios.

Advertisement

“The public doesn’t understand that,” Desautels says. “It’s not like we can jump out and explain, ‘Hey everybody, we’re doing this because it protects you, it saves lives, and it’s necessary.’ It’s something somebody is going to do at some level, somewhere. And for the other countries that aren’t our friends, we need to have the same level of capability.”

Hide Your Flaws

As if the defense of covert sale of zero days to government agencies wasn’t controversial enough, Desautels also operates under the idea that publicly disclosing vulnerabilities—which is generally seen as a security industry best practice—actually imperils customers. His reasons are difficult to dispute.

Advertisement

When a company discovers a vulnerability, it’s under increasing pressure to alert everyone it can—or else face accusations of putting people at risk in the interest of shareholders. Concealing a vulnerability is often painted as covering one’s own ass. The problem is that once bugs are disclosed, malicious hackers also become aware of their existence. Consumers and businesses, meanwhile, often fail to deploy a fix.

Statistically, less than 1 percent of hacks involve the use of zero-day exploits. In fact, the exploits affecting most companies today date back several years.

Advertisement

“I feel like if you find a bug, if you find out that something is dangerous, you find a flaw in the system, there should be a way to get that resolved without putting the world at risk,” Desautels says. “This is a serious failure on the part of the security industry.”

Advertisement

Desautels believes “egos that exist in the industry” are to blame for the idea that disclosing vulnerabilities is in the public good. “You have researchers who want their name out there. They want to be talked about. It builds up your credibility, it builds up your portfolio,” he says. “But the bad almost always outweighs the good.” A number of major ransomware attacks in several countries last year, for example, took advantage of vulnerabilities that had long been public. For months and even years, companies had simply ignored the need to patch. And it’s consumers who often suffer.

“I wish there was a way to guarantee that people would fix vulnerabilities when they are published,” he says. “But the reality is, there isn’t. It never happens that way and probably it never will.”

Advertisement

Ultimately, the number one problem with security is that often there is no security. “The majority of people responsible for protecting networks know fuck-all about protecting networks,” Desautels says. “And even those who do aren’t always allowed to.” In his experience, the vast majority of enterprise customers just want a check in a box that says they’ve been tested, that claims they’re secure. They want a gold star to slap on their website. Even among the 20 percent who do care—those who actually want real fixes and real protection—most aren’t given the resources needed to actually get the job done. “The board or the executive management doesn’t fund it, doesn’t understand it, doesn’t see the value, won’t support it,” he says. And when they do, getting a vulnerability patched is “this huge bureaucratic process that sometimes never happens.”

What inevitably follows is a data breach, or some other kind of huge data loss event, or a ransomware attack that cripples a hospital, or an airport, or a shipping company. Millions of people exposed, and what do they get? A subscription to a credit monitoring service; an identity theft insurance card.

Advertisement

“People don’t patch or they can’t patch, the security technology that’s suppose to protect them does not work as nearly as well as advertised, and boom, you have critical failures left and right,” Desautels says, pausing for a moment. “You know what doesn’t help either...,” he adds with a sigh, “most people in the security industry couldn’t hack their way out of a wet paper bag with a lightsaber.”


This story was produced with support from the Mozilla Foundation as part of its mission to educate individuals about their security and privacy on the internet.