5

How Mastercard is using AI to address cyber risk

 2 years ago
source link: https://venturebeat.com/2021/07/12/how-mastercard-is-using-ai-to-address-cyber-risk/
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.
How Mastercard is using AI to address cyber risk

Transform 2021

Live now: Data, Analytics, & Intelligent Automation Summit, presented by Accenture.

July 12-16

Watch Now

Join executive leaders at the Data, Analytics, & Intelligent Automation Summit, presented by Accenture. Watch now!


As with just about every industry, AI has increasingly infiltrated the financial sector — from visual AI tools that monitor customers and workers to automating the Paycheck Protection Program (PPP) application process.

Talking at VentureBeat’s Transform 2021 event today, Johan Gerber, executive VP for security and cyber innovation at Mastercard, discussed how Mastercard is using AI to better understand and adapt to cyber risk, while keeping people’s data safe.

Welcome to Transform 2021

Lego blocks

On the one hand, consumers have never had it so easy — making payments is as frictionless as it has ever been. Ride-hail passengers can exit their cab without wasting precious minutes finalizing the transaction with the driver, while home-workers can configure their printer to automatically reorder ink when it runs empty. But behind the scenes things aren’t quite so simple. “As easy as it is for the consumer, the complexity lies in the background — we have seen the evolution of this hyper connected world in the backend just explode,” Gerber said.

Even the largest companies don’t build everything in their technology and data stacks from scratch, with countless components from different parties coming together to create the slick experiences that customers have come to expect. It’s also partly why big companies will often acquire smaller startups, as Mastercard did a few months back when it agreed to buy digital identity verification upstart Ekata for $850 million.

However, connecting all these “Lego blocks,” as Gerber calls them, is where the complexity comes in — not just from a technological standpoint (i.e. making it work), but from a data privacy perspective too.

“We’ve seen innovation happening faster than ever before, but it happens not because every company is innovating from A all the way through Z, but [because] we’ve got these third parties in the middle that are creating these wonderful experiences,” Gerber said. “Now, once I put all of this together, how do I manage security, how do I manage cyber risk, when I’ve got a hundred or thousand different third-parties connected to create that one experience for the consumer?”

In cybersecurity, there is an obvious temptation to “isolate things” to minimize the impact from cyberattacks or data leaks, but for products to work, the “Lego blocks” need to be connected. Moreover, companies need to share intelligence internally and within their industry, so that if a cyber attack is happening all their collective systems around the world are put on alert.

“Systemic risk” is what we’re talking about here, something that major financial institutions comprised of myriad Lego blocks need to address, all the while considering compliance and data privacy issues. This is particularly pertinent for global businesses that have a plethora of regional data privacy regulations to contend with, including country-specific laws around data residency.

From Mastercard’s perspective, it leans on a philosophy it calls connected intelligence, or collaborative AI, which is about connecting the dots between systems by “sharing intelligence or outcomes, and not the underlying data,” Gerber noted.

“So by not sharing the underlying data but sharing confidence levels and outcomes, I can maintain your privacy — I don’t have to say ‘this is you’ or ‘this is your card,’ I can just say ‘this person passed the first test and passed it really well,'” he said. “So the collaborative AI is basically how AI systems can share outcomes as variables, so the output of the model becomes the input variable to another model.”

Platform approach

So how does Mastercard achieve all this, so that the data is safeguarded while the systems can still derive insights from the data itself? According to Gerber, the company takes a platform approach — at the bottom end is where the raw data is ingested, upon which the company uses all manner of technologies such as Hadoop and similar tools capable of processing multiple sources of data in real time. From this raw data, Mastercard creates what it refers to as “intelligence blocks,” which are variables derived from the underlying data.

“By the time you get to the derived variable, we’ve applied a layer of compliance checking, data governance checking, [and] made sure that our models are not biased,” Gerber said. “We’ve basically done all the regulatory data scrubbing to ensure that we don’t abuse anything that goes in.”

This is the data that Mastercard can now freely use to build its AI models and products, leading to the top-end customer access layer through which third-parties such as retail stores or card issuers can query a transaction in real time through Mastercard’s API.

Above: Mastercard: Platform approach to data security and privacy

Through all of this, Mastercard doesn’t share any data with banks or retailers, but it can still greenlight a transaction on an individual level. And all this data in aggregate form can also give Mastercard valuable insights into possible attacks; for example, an unexpected spike in transactions coming from a particular retailer might indicate that something untoward is happening. Criminals have been known to procure a bunch of stolen card numbers and then try to imitate retail stores by running transactions against the cards.

Mastercard’s AI can also start imposing certain restrictions — for example, limiting specific types of card at specific retail stores to small-value purchases of less than $50 — or otherwise block any kind of transaction that it considers questionable.

So it’s clear that there is quite a lot of automation at play here — and there really needs to be, given that it would be impossible for humans alone to analyze millions of transactions in real time. The ultimate goal is to help companies improve their security and combat fraud, while ensuring that legitimate customers and retailers are affected as little as possible, as well as adhering to strict data governance rules and regulations.

VentureBeat

VentureBeat's mission is to be a digital town square for technical decision-makers to gain knowledge about transformative technology and transact.

Our site delivers essential information on data technologies and strategies to guide you as you lead your organizations. We invite you to become a member of our community, to access:

  • up-to-date information on the subjects of interest to you
  • our newsletters
  • gated thought-leader content and discounted access to our prized events, such as Transform 2021: Learn More
  • networking features, and more
Become a member

Future cloud storage demands more than NAND flash can deliver

William Van WinkleMay 25, 2021 04:30 AM
Overworked office worker, bureaucracy, archives.
Image Credit: Getty

Transform 2021

Live now: Data, Analytics, & Intelligent Automation Summit, presented by Accenture.

July 12-16

Watch Now

This article is part of the Technology Insight series, made possible with funding from Intel.

As global data volumes continue to rise, the resulting increases in storage traffic create critical infrastructure bottlenecks. Mass-scale block storage providers — organizations from big enterprise data centers to cloud service providers (CSPs) and content delivery networks (CDNs) — must find solutions for handling this data inundation.

Conventional approaches to the problem use NAND solid state drives (SSDs) to buffer data and help keep network pipelines just within their bandwidth capacities. With modern changes in network and PCI Express bandwidth, though, NAND is failing to keep pace. That means applications and services struggle to meet end-user expectations and organizational ROI objectives.

What’s needed is a new approach and technology that does not rely on overprovisioning — a popular but expensive way to increase storage performance and endurance in modern “disaggregated” environments.

Here’s a brief look at traditional approaches, and what enterprises and providers must do to position for tomorrow’s cloud storage demands.

Key Points:

  • As data volumes continue to explode, datacenters must increase bandwidth across conduits such as Ethernet fabrics and PCI Express (PCIe) channels. SSD buffers cannot cope with this increased storage load that impairs network performance.
  • Due to higher performance and endurance characteristics, Intel Optane SSDs can offer much greater efficiency and value in this buffering role than conventional NAND SSD approaches.
  • Cloud service providers, content delivery networks, and enterprises handling mass-scale block storage stand to benefit most from Optane-based buffering.

Datacenter bandwidth is squeezing storage performance

Until a few years ago, it wasn’t a problem if cloud providers placed storage next to compute in their servers. CPU and memory performance were plenty fast, and 1 GbE and 10 GbE network links sufficient for the modest amounts of data flowing into systems. This data could be written and read by NAND SSDs quickly enough to keep up with workload demands without bottlenecking PCI Express conduits.

Today, NAND SSDs have grown incrementally faster. But these improvements pale alongside the doubling of per-lane bandwidth from PCI Express 3.0 to 4.0. An x16 connection now boasts a unidirectional bandwidth of 32 GB/s. Concurrently, datacenter networking pipelines have broadened into 25 GbE, 100 GbE, 200 GbE, and even 400 GbE (although this lofty speed remains rare).

These bandwidth advances are sorely needed as data volumes continue to swell. According to Statista, the annual size of real-time data globally didn’t reach 1 zettabyte (ZB) until 2014; between 2021 and 2022, it will grow by 5ZB. By 2025, total volume in the datasphere will surpass 50ZB. This ballooning will be reflected across most major datacenters, as providers seek to deliver ever more real-time analysis, transaction processing, and other top-performance I/O services.

In short, CSPs and CDNs have too much real-time data to keep it all next to CPUs, even though that would provide the best I/O performance. The data must be spread across multiple systems. This reality popularized the idea of disaggregating storage from compute, effectively creating large “data lakes.”

The approach also lets IT in enterprises and service providers scale storage without increasing compute and memory, enabling more cost-effective capacity expansion. The faster the networking pipes, the more feasible high-performance disaggregation becomes. Otherwise, I/O demand caused by higher data volumes and bigger real-time workloads will create a bottleneck in the network fabric.

“With the [PCIe] Gen 4 interface and these faster networks, the amount of data that can go to your storage is so big that you need dozens of SSDs to absorb the data coming from the pipe,” explains Intel senior manager of product planning Jacek Wysoczynski. “Within that, you want top-performance SSDs to serve as a buffer and de-stage to the data lake. Say each storage box has 24 slots. If you only need two of those to be buffer drives, that’s one thing, but when you need 12 of them to buffer, that’s something different. Now you’re risking overflowing the storage box all the time. If that happens, data can’t be written to storage, which will pause the networking traffic, which temporarily stops the datacenter. That’s a ‘sky-is-falling’ moment.”

The situation Wysoczynski describes involves SSD overprovisioning, typically done to improve storage performance and/or endurance. Imagine having an 800GB SSD, but only making 400GB visible to the host. The invisible space can be allocated to activities such as additional garbage collection, which will help improve write performance. It can also help keep usage below the 50% capacity threshold, above which drive speeds can start to decline. An Intel white paper details how SSD overprovisioning can also significantly improve drive endurance. The downside, of course, is the cost of potentially massive amounts of unused capacity. Without a better alternative, overprovisioning was the highest-performance (if costly) option for storage buffering. Fortunately, that’s now changing.

The Optane alternative

Since their arrival in 2017, Intel Optane SSDs have provided a higher-performance alternative to even enterprise-class NAND SSDs on both write (especially random workloads) and endurance metrics. In write-intensive, real-time storage application settings such as CSPs or any sizable datacenter implementing elastic block storage at scale, Optane SSDs excel in buffering roles. However, the growing bandwidth in datacenter networks, now coupled with rising PCIe bandwidths, have changed the dynamics of how storage should be deployed.

Consider the following figures from Intel’s white paper “Distributed Storage Trends and Implications for Cloud Storage Planners.” Note the emphasis on achieving 90% network bandwidth, which is what datacenter admins often consider the “sweet spot” for maximizing bandwidth value.

NANP-1-icture1.jpg?w=424&resize=424%2C366&strip=all

Above: Then: With components and connectivity from circa 2018, Optane and NAND SSDs were roughly similar in their ability to fill a network pipe.

Image Credit: Intel

Given the prevalent technologies of the era, 90% saturation could be achieved on a 25 GbE connection by only two Optane P4800X drives on PCIe Gen 3. A then-high-performance SSD like the P4610 couldn’t supply as much I/O as the P4800X, but the two weren’t miles apart.

NAND-2-Picture1.jpg?w=420&resize=420%2C362&strip=all
NAND-3-Picture1.jpg?w=422&resize=422%2C366&strip=all

Above: Now: Updated to 2021 technologies, it becomes clear how difficult it is for NAND storage to make effective utilization of networking bandwidth.

With 100 GbE and PCIe Gen 4, the situation changes significantly. Keep in mind that the new 400 GB P5800X offers impressive leaps in performance over the 3.75GB P4800X across several key metrics, including 100% sequential write bandwidth (6200 vs. 2200 MB/s, respectively), random write IOPS (1,500,000 vs. 550,000), and latency (5 vs. 10 µs), all of which contribute to much more storage traffic. Thus, despite the quadrupling of network bandwidth, it only takes three second-gen P5800X Optane SSDs on a PCIe Gen 4 bus to nearly fill that Ethernet link. In contrast, up to 13 current-gen NAND SSDs are needed to supply the same network I/O, depending on the workload. Not surprisingly, the numbers roughly double when stepping up to 200 GbE.

The big point: Optane can use low-capacity drives and still achieve huge performance and endurance leads over drives four or eight times bigger – 1.6TB/3.2TB NAND SSDs.

Another factor to consider is performance consistency. High-volume mixed read/write loads can be particularly grueling for SSDs. Numerous Intel studies have examined how Optane media retains consistent, very low I/O latency over time when stressed under complex, heavy workloads. In contrast, NAND SSD responsiveness tends to deteriorate over time with similar conditions, making application quality of service difficult to maintain.

Real-time implications

Intel’s “Distributed Storage Trends” paper discusses a common datacenter scenario with dense storage racks, 100 Gb/s Ethernet, and 90% I/O saturation. The bottom line is that three Optane P5800X SSDs can do the buffering work of 13 TLC NAND SSDs, leaving room for many more bulk storage drives per enclosure. Intel claims this leads to a “12.6% improvement in cost per GB of raw storage,” including both Capex and Opex savings over three years of power use.

This strategy of using Optane SSDs to provide sufficient buffering performance for current and coming data volumes flowing over expanded I/O conduits will interest CSPs, CDNs, and infrastructure-as-a-service (IaaS) providers offering storage. That said, the performance and cost advantages of Optane SSDs in this scenario could also apply to compute server clusters with considerable local-attached storage, provided the cluster was handling multiple large data sources at once for real-time processing. The Optane SSD can offer greater cost-efficiency, higher total performance, and fewer sources of admin frustration.

“It’s very common for people to try to optimize workloads and be kind to their SSDs,” says Andrew Ruffin, Intel strategy and business development manager. “You can try to only stream sequentially, do the deep queues, and so on. But when you have multiple nodes hitting the same data lake, it just becomes random traffic with zero sequentiality. No matter how much you over-provision or whatever, when you have those multi-tenant environments, it will be hard on the storage devices. This is why it’s essential to understand the need to optimize your device for the traffic.”


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK