Garve Hays, from NetIQ, talks about hacking regulation and legislation Credit: Thinkstock Garve Hays, from NetIQ, talks about hacking regulation and legislation with CSO in a series of topical discussions with industry leaders and experts. Hacked Opinions is an ongoing series of Q&As with industry leaders and experts on a number of topics that impact the security community. The first set of discussions focused on disclosure and how pending regulation could impact it. Now, this second set of discussions will examine security research, security legislation, and the difficult decision of taking researchers to court. CSO encourages everyone to take part in the Hacked Opinions series. If you would like to participate, email Steve Ragan with your answers to the questions presented in this Q&A. The deadline is October 31, 2015. In addition, feel free to suggest topics for future consideration. What do you think is the biggest misconception lawmakers have when it comes to cyber-security? Garve Hays, Solutions Architect at NetIQ (GH): To put it simply, I don’t think lawmakers can distinguish the good guys from the bad guys. The act of studying a software vulnerability does not signal the intent to commit a crime. In fact, there are many security researchers who make a living out of doing just that. Furthermore, the process usually does not happen in a vacuum and sometimes requires collaboration and at times across borders. This isn’t “Fight Club.” This is a topic we should all be talking about. What advice would you give to lawmakers considering legislation that would impact security research or development? GH: There is a quote I really like: “No problem should ever have to be solved twice.” How this pertains to security research is that we don’t want to start from scratch. Every time. In one sense, that means using tools that have already been developed, such as Nmap, OllyDbg, Kali Linux, and IDA Professional, to name a few. However, it is these very tools that are being mistakenly classified as “munitions” in upcoming legislation. Sure, they can be used to assist in the creation of “0-days” or “cyber-weapons” but the focus should be on the obviously nefarious end product and clear evidence it is being employed for wrong-doing. So my guidance would be — don’t outlaw the university or mathematics! Instead, write the law to make it illegal to employ a 0-day to trespass, or steal, or do harm to another person. It’s the tank, not the steel factory. If you could add one line to existing or pending legislation, with a focus on research, hacking, or other related security topic, what would it be? GH: Don’t make it impossible for the good gals and guys to do their jobs. Make provisions if not exclusions for research. Now, given what you’ve said, why is this one line so important to you? GH: It’s the whole standing on the shoulders of giants idea. Understanding how things work can lead to making them better. I view this as basic science (remember STEM?). Driving it underground does not improve the state-of-the-art. Making it a crime to perform legitimate security research is a guarantee that only criminals will be doing it; and guess what, it won’t be to anyone’s benefit. Do you think a company should resort to legal threats or intimidation to prevent a researcher from giving a talk or publishing their work? Why, or why not? GH: Should a company resort to legal threats or intimidation as an opening move? Definitely not. The courts have been abused in the past to suppress legitimate research that supports the public interest. Because an issue is uncomfortable, inconvenient, or awkward doesn’t predicate burying it. However, the flip side is “responsible disclosure.” Trust, but verify. If a best effort has been made to report an exploit, and it does not violate trademarks or copyright, then give the vendor a chance as most vendors appreciate the “heads-up.” However, If they do nothing or attempt to stonewall, then publishing one’s work is how the rest of the world becomes aware of it. Most importantly, the vendors are then armed with knowledge and can attempt to address the exploit themselves, or as paying customers, insist the vendor do something about it. What types of data (attack data, threat intelligence, etc.) should organizations be sharing with the government? What should the government be sharing with the rest of us? GH: I’d like to believe that social pressure and the will of the people will lead to the government behaving responsibly. Furthermore, consider disclosure as something that encompasses a government acting in the best interests of its citizens. In that case, there should be a reciprocal sharing of the five “W’s:” who, what, where, when, and why. And a bonus for how! Who: are the targets or victims? Who are the perpetrators? What: is happening? Where: is the attack taking place? From where is it originating? When: is it taking place? Just now? Or has it been ongoing for months or years? Why: did this happen? For criminal gain? A competitor conducting corporate sabotage? Is a nation-state attempting to gain an advantage? Is it politically motivated? Activism? How: are they doing that?! Related content news UK government plans 2,500 new tech recruits by 2025 with focus on cybersecurity New apprenticeships and talent programmes will support recruitment for in-demand roles such as cybersecurity technologists and software developers By Michael Hill Sep 29, 2023 4 mins Education Industry Education Industry Education Industry news UK data regulator orders end to spreadsheet FOI requests after serious data breaches The Information Commissioner’s Office says alternative approaches should be used to publish freedom of information data to mitigate risks to personal information By Michael Hill Sep 29, 2023 3 mins Government Cybercrime Data and Information Security feature Cybersecurity startups to watch for in 2023 These startups are jumping in where most established security vendors have yet to go. By CSO Staff Sep 29, 2023 19 mins CSO and CISO Security news analysis Companies are already feeling the pressure from upcoming US SEC cyber rules New Securities and Exchange Commission cyber incident reporting rules don't kick in until December, but experts say they highlight the need for greater collaboration between CISOs and the C-suite By Cynthia Brumfield Sep 28, 2023 6 mins Regulation Data Breach Financial Services Industry Podcasts Videos Resources Events SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe