On this page
Sustainability is an essential topic in the energy sector. Eneco’s mission is “Sustainable energy for everyone” and wants to help its customers take the next step in their energy transition. This requires knowledge in two areas, specifically: knowledge of the market and of the customer. This knowledge is obtained from data. More data arise every year, and this data must be available and processed as fast as possible. Because “real-time data” plays an increasingly important role in energy transformation, Eneco decided in 2019 to set up an Event Streaming Platform. This project was awarded to the partnership of Conclusion Mission Critical and Axual.
Read the full use case about Eneco's Event Streaming Platform for energy transition
Answers to your questions about Axual’s All-in-one Kafka Platform
Are you curious about our All-in-one Kafka platform? Dive into our FAQs
for all the details you need, and find the answers to your burning questions.
An event streaming platform is a system that allows for the real-time processing and management of event data as it flows through a network. It enables organizations to capture, store, process, and analyze streams of events or data changes, facilitating the creation of data-driven applications and enabling real-time analytics.
Eneco is a Dutch energy company focused on the supply and production of sustainable energy. Founded in 1995 and headquartered in Rotterdam, it operates primarily in the Netherlands and Belgium and has a significant commitment to renewable energy sources. Eneco provides energy to residential and business customers, offering electricity, natural gas, and heat services. The company has focused on sustainable energy solutions, aiming to help customers transition to renewable energy.
The Energy Transition refers to the global shift from fossil fuel-based energy systems to more sustainable, low-carbon energy sources, particularly renewable energy. This transition is driven by the need to address climate change, reduce greenhouse gas emissions, and create a more resilient and sustainable energy infrastructure.
Related blogs

Consumer group offsets are essential components in Apache Kafka, a leading platform for handling real-time event streaming. By allowing organizations to scale efficiently, manage data consumption, and track progress in data processing, Kafka’s consumer groups and offsets ensure reliability and performance. In this blog post, we'll dive deep into these concepts, explain how consumer groups and offsets work, and answer key questions about their functionality. We'll also explore several practical use cases that show how Kafka’s consumer groups and offsets drive real business value, from real-time analytics to machine learning pipelines.

Apache Kafka is a powerful event-streaming platform, but does your enterprise need to go all in from day one? In this blog, we explore why starting small with Kafka is the best strategy. Learn how an incremental approach can help you reduce complexity, and scale efficiently as your needs grow. Whether you're new to Kafka or looking for a practical implementation strategy, this guide will set you on the right path.

Kafka Consumer Configuration is at the heart of building efficient, scalable, and reliable data streaming applications. Whether you’re working with event-driven architectures, batch data ingestion, or real-time stream processing, the right configurations can make all the difference. In this guide, we’ll explore the most important Kafka consumer settings, break down their impact, and showcase practical use cases to help you optimize performance. By the end, you’ll have a clear roadmap to fine-tune your Kafka consumers for maximum efficiency.