Développez l'apprentissage automatique prédictif avec Flink | Atelier du 18 déc. | S'inscrire
Change data capture is a popular method to connect database tables to data streams, but it comes with drawbacks. The next evolution of the CDC pattern, first-class data products, provide resilient pipelines that support both real-time and batch processing while isolating upstream systems...
Learn how the latest innovations in Kora enable us to introduce new Confluent Cloud Freight clusters, which can save you up to 90% at GBps+ scale. Confluent Cloud Freight clusters are now available in Early Access.
Learn how to contribute to open source Apache Kafka by writing Kafka Improvement Proposals (KIPs) that solve problems and add features! Read on for real examples.
For many, microservices are built on a protocol of requests and responses. REST etc. This approach is very natural. It is after all the way we write programs: we make […]
We are very excited to share a wealth of streaming news from the past month! If you are looking for an ideal streaming data service that delivers the resilient, scalable […]
Introduction What’s great about the Kafka Streams API is not just how fast your application can process data with it, but also how fast you can get up and running […]
Kafka Summit New York City was the largest gathering of the Apache Kafka community outside Silicon Valley in history. Over 500 “Kafkateers” from 400+ companies and 20 countries converged in […]
Apache Kafka® is the best enterprise streaming platform that runs straight off the shelf. Just point your client applications at your Kafka cluster and Kafka takes care of the rest: […]
Today, I’m really excited to announce Confluent CloudTM, Apache Kafka® as a Service: the simplest, fastest, most robust and cost effective way to run Apache Kafka in the public cloud. […]
The team at Confluent, along with the Apache KafkaTM community, are excited the day is finally here – it’s time for Kafka Summit NYC! We’ll be at the Midtown Hilton […]
Ever wondered what it’s like to run Kafka in production? What about building and deploying microservices that process streams of data in real-time, at large scale? Or, maybe just the […]
In Q1, Confluent conducted a survey of the Apache Kafka® community and those using streaming platforms to learn about their application of streaming data. This is our second execution of […]
The Google Dataflow team has done a fantastic job in evangelizing their model of handling time for stream processing. Their key observation is that in most cases you can’t globally […]
In our previous post on the Streaming Pipelines track, we highlighted some of the sessions not to be missed at Kafka Summit NYC. As a follow on to that, let’s […]
Here at Confluent, our goal is to ensure every company is successful with their streaming platform deployments. Oftentimes, we’re asked to come in and provide guidance and tips as developers […]
Pandora began adoption of Apache Kafka® in 2016 to orient its infrastructure around real-time stream processing analytics. As a data-driven company, we have a several thousand node Hadoop clusters with hundreds of Hive tables critical to Pandora’s operational and reporting success...
It is just a few weeks out until Kafka Summit NYC! Since we’re on the program committee for this event and are also the track leads for the Streaming Pipelines track, […]