The idea of bad actors stealing valuable assets brings to mind a picture of masked men breaking into a bank vault or museum and making a getaway with their illicit stash. But what if the enemy is one of us — someone who knows exactly where we keep our most valuable items, how we safeguard them and even the alarm code to disable the entire security system?

Distinguishing Malicious Insiders From Legitimate Users

Organizations hold patents, intellectual property, client data and other valuable information, and thousands of employees need access to those assets for legitimate reasons. With so much at stake, it is critical for security teams to be able to identify rogue staffers and determine whether their access credentials have been compromised by an external actor to get on the inside.

But how can security teams distinguish malicious insiders from legitimate users when suspicious activity closely resembles typical behavior? They must model the user’s normal behavior and measure this against subtle characteristic changes and anomalous activity using user behavior analytics (UBA).

Anomalous activity can include a user logging in from a different geographic location, logging in via a virtual private network (VPN) at odd hours, or transferring high volumes of data from the network to an external site or cloud storage account. Any one of these activities by itself does not necessarily indicate malicious intent, but the combination of several suspicious behaviors warrants investigation by a security operations center (SOC) analyst to determine whether the user has gone rogue or had credentials stolen. Each anomalous activity increases the user’s risk score. When it crosses a certain threshold, the user needs to be investigated or closely monitored.

Unlocking the Power of Machine Learning

Rules-based anomaly detection is a great way to identify illicit behaviors, but what if the clues are much more subtle? That’s where machine learning can help.

Let’s take a look at the activities of an employee in the marketing department, for example:

If this employee plans to quit his or her job and is looking to take proprietary data to a rival firm, he or she might exhibit the following behavior:

You’ll notice that the user does not change his or her routine drastically but exhibits certain subtle activity changes that indicate malicious intent.

A UBA solution powered by machine learning uses unsupervised learning to help model a user’s behavior in various categories, such as authentication, network access, firewall activity, application activity, port or network scans, denial-of-service events, malware or other malicious software activity. The user’s risk score is increased based on deviation from the baseline established by the model. The model also identifies deviation from normal activity versus frequency to give you a picture of the user’s risk posture.

Peer group analytics offer yet another lens into a user’s activities to help identify when a user deviates from the typical behavior of employees with similar roles and responsibilities.

Learn More

Learn more about QRadar User Behavior Analytics and try the free QRadar UBA app from the IBM Security App Exchange. You can also watch this video to learn how you can combine QRadar UBA and QRadar Advisor with Watson to investigate suspicious behavior.

If you are attending Think 2018 in Las Vegas, check out the Security and Resiliency Campus and attend these sessions on user behavior analytics:

Watch now! View the Think 2018 Security & Resiliency Sessions on-demand

More from Artificial Intelligence

AI and cloud vulnerabilities aren’t the only threats facing CISOs today

6 min read - With cloud infrastructure and, more recently, artificial intelligence (AI) systems becoming prime targets for attackers, security leaders are laser-focused on defending these high-profile areas. They’re right to do so, too, as cyber criminals turn to new and emerging technologies to launch and scale ever more sophisticated attacks.However, this heightened attention to emerging threats makes it easy to overlook traditional attack vectors, such as human-driven social engineering and vulnerabilities in physical security.As adversaries exploit an ever-wider range of potential entry points…

Are successful deepfake scams more common than we realize?

4 min read - Many times a day worldwide, a boss asks one of their team members to perform a task during a video call. But is the person assigning tasks actually who they say they are? Or is it a deepfake? Instead of blindly following orders, employees must now ask themselves if they are becoming a victims of fraud.Earlier this year, a finance worker found themselves talking on a video meeting with someone who looked and sounded just like their CFO. After the…

How to calculate your AI-powered cybersecurity’s ROI

4 min read - Imagine this scenario: A sophisticated, malicious phishing campaign targets a large financial institution. The attackers use emails generated by artificial intelligence (AI) that closely mimic the company's internal communications. The emails contain malicious links designed to steal employee credentials, which the attackers could use to gain access to company assets and data for unknown purposes.The organization's AI-powered cybersecurity solution, which continuously monitors network traffic and user behavior, detects several anomalies associated with the attack, blocks access to the suspicious domains…

Topic updates

Get email updates and stay ahead of the latest threats to the security landscape, thought leadership and research.
Subscribe today