UPDATED 12:00 EDT / APRIL 20 2018

BIG DATA

Stream processing turning the database inside out

When stream processors began, they were considered by the computing industry as a mere addition to the batch data stack or other stacks being utilized by information technology administrators. But stream processing has come a long way, and it is beginning to shift the way people collect data and upgrade applications. This evolution is shifting the principles applied to stream processing as the technology’s limitations are lifted.

“When you start to take this streaming style approach to things … it’s been called turning the database inside out, unbundling the database,” said Stephan Ewen, chief technology officer of data Artisans GmbH. “Your input sequence of event is arguably the ground truth, and what the stream processor computes is as a view of the state of the world.”

Ewen spoke with George Gilbert (@ggilbert41), co-host of theCUBE, SiliconANGLE Media’s mobile livestreaming studio, at the Flink Forward event in San Francisco. They discussed how stream processing has evolved and what Flink brings to the table through stream processing. (* Disclosure below.)

Stream processing for real-time applications

When stream processors thought of state with a different consistency model, as Ewen explained, the way they thought of time was actually different than batch processors of the databases. So early stream processors couldn’t complement the stack with that interpretation; a batch job could not interpret a streaming job.

“So once the stream processors adopted a stronger consistency model, a time model that is more compatible with reprocessing … all of these things all of a sudden fit together much better,” Ewen said.

As stream processing evolved, platforms like Apache Flink were able to take a concept of distributed, consistent snapshots and create one big local snapshot. When consistent snapshots are in place, then upgrading applications — which usually involves running two version of the same application — becomes much easier to switch between one application to the other as if nothing was shared.

“Now that is what is … kind of the core idea and what helps Flink generalize from batch processing to … real-time stream processing to event-driven applications,” Ewen concluded.

Watch the complete video interview below, and be sure to check out more of SiliconANGLE’s and theCUBE’s coverage of the Flink Forward 2018 event(* Disclosure: TheCUBE is a paid media partner for Flink Forward 2018. Neither data Artisans GmbH, the event sponsor, nor other sponsors have editorial control over content on theCUBE or SiliconANGLE.)

Photo: SiliconANGLE

Since you’re here …

… We’d like to tell you about our mission and how you can help us fulfill it. SiliconANGLE Media Inc.’s business model is based on the intrinsic value of the content, not advertising. Unlike many online publications, we don’t have a paywall or run banner advertising, because we want to keep our journalism open, without influence or the need to chase traffic.The journalism, reporting and commentary on SiliconANGLE — along with live, unscripted video from our Silicon Valley studio and globe-trotting video teams at theCUBE — take a lot of hard work, time and money. Keeping the quality high requires the support of sponsors who are aligned with our vision of ad-free journalism content.

If you like the reporting, video interviews and other ad-free content here, please take a moment to check out a sample of the video content supported by our sponsors, tweet your support, and keep coming back to SiliconANGLE.