January 18, 2024 By Sue Poremba 3 min read

The hottest technology right now is AI — more specifically, generative AI. The trend is so popular that every conference and webinar speaker feels obligated to mention some form of AI, no matter their field.

The innovations and risks that AI offers are both exciting and frightening. However, the heavy focus on this technology overshadows an important component of artificial intelligence: machine learning (ML).

For a quick overview, ML is a subset of AI  based on patterns, predictions and optimization. Cybersecurity tools rely on ML to use predictions and patterns to find anomalies and sniff out potential threats. Instead of a human spending hours reading logs, ML can do the same tasks in seconds.

Like AI, ML has been around for a long time. The reason why we’re talking about AI so much now is because generative AI is a game changer in the way we communicate with technology. But ML is also changing, and we’ll see it used in new ways in 2024.

How we use machine learning

Machine learning is all about data. ML algorithms rely on historical data to detect patterns, from software codes to customer shopping behaviors. Social media networks rely on ML algorithms to keep relevant information at the top of your feed. Self-driving cars use ML algorithms to navigate city streets and traffic laws. In cybersecurity, ML is used in areas like behavioral analytics and sending alerts for unusual usage, task automation and providing more efficient real-time threat-hunting intelligence.

There are three common types of ML currently used. Supervised learning trains ML to perform a specific task based on the presented data. Unsupervised learning relies on relationships across the data. Reinforcement learning is most similar to human learning, where the ML model learns to solve problems through trial-and-error formats.

New trends in machine learning

As AI continues to advance, so does ML, and one of the most anticipated improvements to ML in 2024 will be no-code machine learning. No-code ML relies heavily on behavioral data and plain English to get results. Instead of complicated coding language, analysts will be able to ask a question or create a command to get a report. One of the biggest benefits of no-code ML is that it allows companies of all sizes to implement ML and AI into their networks without the need to hire data analysts and engineers. The downfall is that this type of ML technology is limited and won’t be able to do deep-dive predictive analysis.

Unsupervised and reinforcement ML are both expected to expand in the coming year, in part because of no-code ML.

As ML evolves, we will likely see growth in other technologies like augmented reality and quantum computing. “Machine learning models can generate 3D objects for apps and other uses in augmented reality,” Luís Fernando Torres wrote. In addition, ML will play a role in improved facial recognition technology and interactions with generative AI.

Machine learning and security — the good and bad

As mentioned earlier, ML benefits your overall cybersecurity program by automating what were once cumbersome manual tasks. It can find threats that are otherwise missed and can cut down on false positives.

But as with any technology, there are security risks involved. Threat actors use ML and AI to launch attacks by poisoning or misleading the data to trick the system into providing false reports. Threat actors use this to bypass security systems and hijack the network.

AI and its role in security is what everyone wants to talk about today, but the ways that AI can improve your company’s security systems depend on machine learning. The time has come to get back to the basics and recognize how ML fits into your security system and how to best train ML so that your AI is even more effective.

More from Artificial Intelligence

NIST’s role in the global tech race against AI

4 min read - Last year, the United States Secretary of Commerce announced that the National Institute of Standards and Technology (NIST) has been put in charge of launching a new public working group on artificial intelligence (AI) that will build on the success of the NIST AI Risk Management Framework to address this rapidly advancing technology.However, recent budget cuts at NIST, along with a lack of strategy implementation, have called into question the agency’s ability to lead this critical effort. Ultimately, the success…

Researchers develop malicious AI ‘worm’ targeting generative AI systems

2 min read - Researchers have created a new, never-seen-before kind of malware they call the "Morris II" worm, which uses popular AI services to spread itself, infect new systems and steal data. The name references the original Morris computer worm that wreaked havoc on the internet in 1988.The worm demonstrates the potential dangers of AI security threats and creates a new urgency around securing AI models.New worm utilizes adversarial self-replicating promptThe researchers from Cornell Tech, the Israel Institute of Technology and Intuit, used what’s…

What should an AI ethics governance framework look like?

4 min read - While the race to achieve generative AI intensifies, the ethical debate surrounding the technology also continues to heat up. And the stakes keep getting higher.As per Gartner, “Organizations are responsible for ensuring that AI projects they develop, deploy or use do not have negative ethical consequences.” Meanwhile, 79% of executives say AI ethics is important to their enterprise-wide AI approach, but less than 25% have operationalized ethics governance principles.AI is also high on the list of United States government concerns.…

Topic updates

Get email updates and stay ahead of the latest threats to the security landscape, thought leadership and research.
Subscribe today