Développez l'apprentissage automatique prédictif avec Flink | Atelier du 18 déc. | S'inscrire
We are pleased to announce the release of KSQL v0.5, aka the February 2018 release of KSQL. This release is focused on bug fixes and as well as performance and stability improvements.
Our focus for KSQL has been on improving its operational experience and on addressing issues and bugs reported by the KSQL user community or surfaced through KSQL’s rigorous testing strategies. With the general availability announcement today, we’re excited to see KSQL in production soon. Our efforts have centered around soak testing with eye towards supporting continuous workloads and operation, bounds testing, and performance testing, much of which runs on a daily basis through automated test suites. For example, our soak testing cluster has racked up over 1,000 hours and runs KSQL workloads 24×7. The performance tests we conduct allow us to understand performance characteristics of stateless and stateful KSQL queries. We currently run over 42 different tests that collect more than 700 metrics.
As they say, April showers bring May flowers. In this case, a relentless February focus on quality and stability will bring a trustworthy, production-ready product when KSQL sees its GA release next month—a future poised for a regular cadence of new features that help you take your KSQL-based systems in the direction you want to go.
If you have enjoyed this article, you might want to continue with the following resources to learn more about KSQL:
If you are interested in contributing to KSQL, we encourage you to get involved by sharing your feedback via the KSQL issue tracker, voting on existing issues by giving your +1 or opening pull requests. Use the #ksql channel in our public Confluent Slack community to ask questions, discuss use cases or help fellow KSQL users.
In this final part of the blog series, we bring it all together by exploring data streaming platforms (DSPs), event-driven architecture (EDA), and real-time data processing to scale AI-powered solutions across your organization.
In Part 2 of the series, we take things a step further by enhancing GenAI with the tools it needs to deliver smarter, more relevant responses. We introduce retrieval-augmented generation (RAG) and vector databases (VectorDBs), key technologies that provide LLMs with the context they need.