Waterstream.io | Simplify MQTT Data Integration https://waterstream.io/ High performance MQTT Broker Fri, 20 Feb 2026 15:52:41 +0000 en-US hourly 1 https://wordpress.org/?v=6.9.1 https://waterstream.io/wp-content/uploads/2022/06/waterstreamfavicon-01.png Waterstream.io | Simplify MQTT Data Integration https://waterstream.io/ 32 32 MQTT Beyond IoT: Where Protocol Efficiency Becomes a Competitive Advantage https://waterstream.io/2026/02/20/mqtt-beyond-iot/ Fri, 20 Feb 2026 15:52:39 +0000 https://waterstream.io/?p=6933 When we talk about MQTT, we immediately think of the sensors, smart cities, and connected devices that define the IoT landscape. This is only natural: MQTT was designed for these environments, where protocol efficiency, low bandwidth consumption, and resilience against intermittent connections are critical. However, stopping there means missing out on a significant opportunity. MQTT […]

The post MQTT Beyond IoT: Where Protocol Efficiency Becomes a Competitive Advantage appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

When we talk about MQTT, we immediately think of the sensors, smart cities, and connected devices that define the IoT landscape. This is only natural: MQTT was designed for these environments, where protocol efficiency, low bandwidth consumption, and resilience against intermittent connections are critical. However, stopping there means missing out on a significant opportunity.

MQTT represents a valuable resource whenever a software architecture needs to manage efficient, scalable, and reliable real-time communications, regardless of the application domain. Its publish-subscribe model, combined with sophisticated Quality of Service management, makes it ideal for scenarios well beyond traditional IoT.

The Hidden Power of Protocol Efficiency

In modern enterprise environments, where distributed applications require real-time communication across complex networks, protocol overhead quickly becomes a bottleneck. MQTT addresses this problem fundamentally: its minimal header and efficient handling of persistent connections make it ideal for high-frequency messaging, where every millisecond and every byte matters.

Consider financial applications managing real-time quote streams, where thousands of updates per second must reach hundreds of clients simultaneously. Or enterprise notification systems, where the ability to maintain open connections with thousands of users without exhausting server resources is critical.  In these scenarios, the efficiency of the MQTT protocol translates directly into reduced infrastructure costs and superior performance.

Waterstream amplifies these benefits by bridging MQTT directly into modern Kafka-based architectures. This means you can leverage protocol efficiency while maintaining the power of a distributed streaming system, complete with its guarantees of persistence, horizontal scalability, and seamless integration with the enterprise data ecosystem.

When Resilience Makes all the Difference

Another often overlooked aspect of MQTT is its handling of unstable connections. In a perfect world, every application would operate on predictable, always-available networks. The reality is far different: mobile applications handing over between Wi-Fi and cellular data, clients operating under volatile network conditions, and scenarios where intermittent connectivity is the norm rather than the exception.

With its built-in QoS (Quality of Service) mechanisms and native reconnection logic, MQTT offers resilience by design. A message published with QoS 1 or 2 is guaranteed to arrive even despite temporary disconnections, abstracting away the need for complex retry mechanisms at the application layer. This translates into simpler code, fewer bugs stemming from network error handling, and a smoother user experience.

Whether for real-time collaboration apps, push notification systems, or multiplayer gaming platforms, this functionality is a key differentiator. Waterstream scales this resilience into an enterprise-grade context, where message loss can lead to a quantifiable impact on both revenue and customer satisfaction.

The Integration that completes the Picture

Historically, the real limitation of MQTT in non-IoT contexts has been its isolation from the broader enterprise data ecosystem. While the protocol is highly efficient, it has largely existed in a silo, separated from the analytics pipelines, stream processing engines, and business intelligence tools that drive modern operations.

Waterstream eliminates this dichotomy by acting as a native bridge between MQTT and Kafka. Every message published to an MQTT topic is immediately ingested as an event into a Kafka topic ready to be processed, aggregated, analyzed, or routed to other systems. This removes the need for custom adapters, complex ETL pipelines, or fragile workarounds: the integration is native, seamless, and bidirectional.

This unlocks scenarios that were previously impractical. A mobile application can publish events via MQTT, benefiting from its extreme efficiency, while downstream back-end services consume those events from Kafka to feed real-time dashboards, trigger automatic workflows, or train machine learning models.

Scaling Without a Complete Redesign

Scalability is often the elephant in the room when designing real-time systems. An application that starts with a few hundred users can quickly scale to tens of thousands, with sudden spikes driven by product success or specific events. At this stage, re-architecting the messaging infrastructure becomes expensive, risky, and prohibitively complex.

Waterstreamscales linearly alongside Kafka, providing a natural path for growth. Increasing capacity merely requires adding brokers to the Kafka cluster, eliminating the need to rethink application logic or migrate MQTT clients. Growth evolves into a manageable operational task rather than an architectural bottleneck.

For companies developing digital products with high growth potential, this architecture eliminates a significant source of technical debt. The messaging infrastructure scales incrementally, fueling business growth instead of stalling it with costly migrations.

Conclusion

When an architecture must manage real-time communications with stringent efficiency requirements or when resilience over unstable networks and seamless integration with modern data pipelines are paramount, MQTT transcends its role as an IoT protocol and becomes a highly pragmatic technological choice.

Waterstream makes this choice even more compelling by eliminating the friction that has historically limited MQTT adoption in enterprise contexts: data silos, the complexity of managing scalable infrastructures, and the need for custom integrations to bridge disparate systems.

The result is a leaner, more efficient architecture that is perfectly equipped to evolve alongside business needs.

The post MQTT Beyond IoT: Where Protocol Efficiency Becomes a Competitive Advantage appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
From Sensor to Artificial Intelligence: Simplifying Large-Scale IoT Integration https://waterstream.io/2025/11/10/from-sensor-to-artificial-intelligence-simplifying-large-scale-iot-integration/ Mon, 10 Nov 2025 09:47:38 +0000 https://waterstream.io/?p=6069 The evolution of digital ecosystems makes data management a central strategic challenge. For companies aiming to fully leverage the Internet of Things (IoT) and Artificial Intelligence (AI), the ability to integrate, process, and analyze real-time data streams is not just an advantage, but an operational imperative. Organizations face growing volumes of information from heterogeneous sources, […]

The post From Sensor to Artificial Intelligence: Simplifying Large-Scale IoT Integration appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

The evolution of digital ecosystems makes data management a central strategic challenge. For companies aiming to fully leverage the Internet of Things (IoT) and Artificial Intelligence (AI), the ability to integrate, process, and analyze real-time data streams is not just an advantage, but an operational imperative.

Organizations face growing volumes of information from heterogeneous sources, which must be conveyed to the decision-making center with guaranteed speed, quality, and security. Waterstream enters this context as an innovative solution, designed to eliminate complexity in the integration of MQTT data at large scale.

Addressing the Complexity of Big Data in Motion

Connecting data generated by sensors to Artificial Intelligence is often complicated by architectural obstacles. Enterprise platforms daily face data fragmentation due to information silos. Specifically, IoT data on separate brokers from enterprise data on legacy databases prevent a unified, real-time vision.

This disconnection is particularly critical when dealing with big data in motion, which is the management of massive volumes of constantly flowing data that require ultra-low latency processing. The complexity is aggravated by the need to:

  • Guarantee Integrity and Security: Every message, even in a continuous flow and in scenarios of intermittent connectivity, must be authenticated, encrypted, and not lost.
  • Manage Scalability and Resource Efficiency: Traditional integration pipelines (the so-called MQTT-Kafka bridges) are expensive to write and maintain, creating a bottleneck that limits expansion and requires double data persistence (first on the MQTT broker and then on Kafka).

Addressing these challenges requires a sophisticated approach to data integration, one that does not just “connect” systems, but unifies them at the foundational level, ensuring the agility necessary to adapt quickly to changing business needs.

Eliminating the Architectural Paradox

The MQTT protocol is fundamental for the IoT, being lightweight and designed for unstable connectivity and resource-constrained devices. Apache Kafka is the chosen architecture for reliability, durability, and the stream processing ecosystem. 

Traditionally, their integration has required the introduction of complex architectures including: an external MQTT broker, dedicated integration pipelines (often based on Kafka Connect), and duplication of storage and replica.

Waterstream intervenes to eliminate this “integration paradox” by transforming the Kafka platform into a genuine native MQTT broker. Waterstream is not a simple connector, but a stateless implementation that acts as a bidirectional layer, encapsulating MQTT messaging directly into Kafka, using it as the sole storage and distribution engine.

This strategic architectural choice offers a paradigm shift:

  • Architecture Simplification: The need to maintain and operate a separate MQTT broker and its integration pipelines is completely eliminated.
  • Robustness: Every IoT data immediately benefits from persistence, high availability, and native fault tolerance of Kafka. The management of unstable networks, typical of edge or mobile implementations, is absorbed by the reliability of the Kafka log.
  • Centralization: By leveraging native integration with tools like the Schema Registry, Waterstream ensures data validation and consistency from the source through the entire pipeline, a fundamental prerequisite for analytics and AI.

The impact of this architectural choice directly translates into a reduction in operational costs, allowing IT teams to re-direct resources from infrastructure management to data innovation.

Scalability and Resilience: The Waterstream Architecture

Waterstream’s architecture is based on a stateless design and allows MQTT messages to inherit Kafka’s robustness.

The system is designed to offer linear scalability, supporting millions of simultaneous connections, and dynamically adapting to continuously growing workloads. This flexibility is essential, especially in a context of accelerated adoption of IoT devices across every sector. Since Waterstream functions as an application that natively interacts with Kafka, it can be run wherever the Kafka cluster is present, ensuring the necessary deployment flexibility for edge, on-premises, or multi-cloud and hybrid architectures.

Furthermore, Waterstream is optimized for the most difficult operational conditions. Leveraging the inherent lightness of MQTT, it ensures that data reaches its destination even in scenarios of high latency and intermittent connectivity. This translates into greater reliability and resilience of the overall system, allowing uninterrupted data streams to be transformed into actionable information.

The integration is bidirectional: not only does data flow from devices to Kafka, but Waterstream also allows commands and notifications to be sent from Kafka to the MQTT clients, closing the loop of real-time control and monitoring.

Application Cases

The adoption of a unified data architecture is not limited to optimizing the infrastructure; it enables a wide range of application cases that transform data potential into tangible business value.

Waterstream not only solves an integration problem but unlocks the capability to extend MQTT data with all the advanced features of the Kafka ecosystem, such as stream processing, real-time analytics, and connectors to databases and downstream systems.

Consider, for instance, industrial monitoring (industrial IoT): in complex production environments, Waterstream manages thousands of connected sensors and machinery in real time. Direct integration with Kafka immediately feeds predictive maintenance models capable of identifying anomalies and potential failures hours or days in advance. This not only optimizes production quality but also reduces costly unplanned downtime.

Another critical sector is smart energy management. In the scenario of Smart Grid and Smart Metering, Waterstream enables dynamic monitoring and optimization of consumption. Real-time analysis of data from millions of smart meters allows energy companies not only to balance the grid and prevent overloads but also to provide personalized optimization strategies directly to the final consumer.

The gaming and IoT applications sector represents a further field of application. In gaming, the solution supports real-time communications among millions of players, ensuring the low latency and high reliability necessary for the user experience. At the same time, for IoT applications, it offers a robust platform for managing networks of distributed sensors and edge devices, ensuring that crucial data is collected and processed efficiently, feeding large-scale diagnostics and performance analysis systems.

In urban infrastructures (smart cities), Waterstream allows the real-time flow of data from environmental, traffic, or security sensors, supporting more efficient and reactive city services. 

Last but not least, in the healthcare sector, Waterstream supports remote patient monitoring and the management of networks of distributed medical devices, where reliability and low latency are non-negotiable requirements for immediate intervention and the provision of personalized services.

The Future of Integration is Unified and Stateless

The investment in Waterstream represents a decisive step toward creating a high-performance data infrastructure. It is not simply about replacing a component, but about adopting an architectural vision that maximizes efficiency and the ability to scale, reducing friction between the IoT layer and the enterprise analytics layer.

Waterstream offers the best of both worlds: the lightness and efficiency of MQTT for edge data collection, combined with the robustness, persistence, and mature ecosystem of Apache Kafka for large-scale processing. Adopting Waterstream means eliminating data silos at the root, ensuring that every single byte from sensors is a high-quality asset, immediately available for Machine Learning models and real-time decisions. In an era defined by the speed of data, Waterstream’s unified and stateless architecture is the key to maintaining competitiveness and driving the Artificial Intelligence based transformation.

Discover how Waterstream can revolutionize your data streaming architecture, facilitating the fluid and scalable integration of IoT data with Kafka. Contact us for more information.

The post From Sensor to Artificial Intelligence: Simplifying Large-Scale IoT Integration appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
Kafka vs. IoT: Why It Doesn’t Work and the new Solution https://waterstream.io/2025/10/06/kafka-vs-iot-why-it-doesnt-work-and-the-new-solution/ Mon, 06 Oct 2025 14:53:24 +0000 https://waterstream.io/?p=6019 We often observe companies using Kafka for IoT just because it’s a ‘standard.’ But what happens when this ‘standard’ only adds unnecessary complexity, high costs, and stringent constraints to your projects? Discover how to simplify your IoT architecture and escape cloud lock-in.” The Realization It was one of those days every software architect knows well. […]

The post Kafka vs. IoT: Why It Doesn’t Work and the new Solution appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

We often observe companies using Kafka for IoT just because it’s a ‘standard.’ But what happens when this ‘standard’ only adds unnecessary complexity, high costs, and stringent constraints to your projects? Discover how to simplify your IoT architecture and escape cloud lock-in.”

The Realization

It was one of those days every software architect knows well. The client sitting across from me had just finished describing their IoT infrastructure: 50,000 industrial sensors, millions of messages per second, and a latency requirement of under 100ms. “Obviously, we’ll use Kafka,” he concluded, as if it was the only option. I nodded, but something didn’t feel right. Not for the first time, I was facing what I call the “Kafka Default Syndrome” – the automatic assumption that Kafka is the solution for any streaming problem.

That evening, heading home, I did the math. To manage those 50,000 devices with Kafka, we would have needed:

  • A Kafka cluster with at least 5 brokers for redundancy
  • Zookeeper or KRaft (which adds its own complexity)
  • A custom MQTT-to-Kafka translation layer
  • A dedicated team just to keep it all running
  • A cloud budget that would make the CFO cry

But the real problem wasn’t the cost. It was the pointless complexity we were about to sell to the client.

The Problem Everyone Knows But No One Admits

Let me say this clearly: Kafka was not designed for IoT. It was designed for data pipelines between enterprise systems. It’s like using a Formula 1 car to go grocery shopping – technically possible, but practically absurd.

IoT devices speak MQTT; it’s their native language.

MQTT is lightweight and efficient, built for unstable connections and resource-constrained devices.

Kafka speaks… Kafka. It’s powerful but heavy, designed for servers with GBs of RAM, not for sensors with KBs of memory.

At Bitrock, we’ve implemented dozens of these bridges over the years, so we know exactly what it entails. The process looks like this:

  • Devices send MQTT messages to an MQTT broker.
  • A custom component reads from MQTT and writes to Kafka.
  • Applications consume data from Kafka.
  • Another component reads from Kafka and responds via MQTT.

Every step adds latency, every component adds a point of failure, and every translation loses something in the process. And we kept selling this complexity as a “best practice”.

The Cloud Lock-in Trap (August 2025 Edition)

In 2025, every cloud provider has its own “solution”.

  • AWS IoT Core + MSK: They promise seamless integration, but the reality is you pay for GBs of data ingested into IoT Core, then for streaming to MSK, and then for processing. One client saw their bill jump from $5K to $45K per month just by tripling their sensors. And trying to migrate? Good luck with their proprietary APIs.
  • Azure IoT Hub + Event Hubs: Microsoft sells “synergy” with the rest of their Azure stack. However, their proprietary SDKs, custom message formats, and the inability to cleanly export data hold you hostage. One of our clients took eight months to migrate away.
  • Google Cloud IoT Core + Pub/Sub: Oh, wait, they discontinued IoT Core in 2023. If you built on that, congratulations. Now they tell you to use Pub/Sub directly, but guess what? It doesn’t speak MQTT natively; you need a bridge. We’re back to square one.
  • Oracle Cloud IoT + Streaming: I won’t even comment. If you end up there, you have bigger problems than MQTT vs. Kafka.

The pattern is always the same: they lure you in with entry-level pricing, and when you have 100K devices in production, costs explode and migration becomes impossible. It’s lock-in by design.

The Discovery That Changed Everything

It was during a due diligence for the acquisition of Waterstream by Fortitude Group that I truly understood what they had built. It wasn’t just “another MQTT broker.” It was the solution to the problem everyone pretended not to see.

Waterstream is a broker that speaks both MQTT and Kafka natively. It doesn’t translate or use a bridge. It speaks both protocols as a native language. Devices connect via MQTT, applications consume via the Kafka API, and in between… there’s nothing.

Literally nothing to manage, debug, or break at 3 a.m.

From Users to Maintainers: The Turning Point

When Waterstream joined the Fortitude Group, we didn’t just acquire a product. The original team passed the torch entirely to us at Bitrock. Today, we maintain, evolve, and support Waterstream directly.

This means that when a client has a problem at 2 a.m., they don’t have to open a ticket and pray. They call us. When a specific use case requires a new feature, they don’t have to convince a product manager in Silicon Valley. We discuss it over coffee and implement it.

In recent months, we have:

  • Released full support for MQTT 5.0
  • Optimized performance for edge deployments (sub-millisecond latency)
  • Added native integration with OpenTelemetry
  • Implemented enterprise-grade multi-tenancy

But above all, we’ve kept the original promise: simplicity. Every feature added must simplify, not complicate.

Concrete Results in 2025

Logistics Client (migrated from AWS IoT Core):

  • Before, they paid $38K/month to AWS and had total vendor lock-in.
  • After, their on-prem infrastructure and Waterstream cost $8K/month.
  • The migration took two weeks (10 of which were spent convincing management it was really that simple).

Energy Client (migrated from Azure IoT Hub)

  • Before, they had six Azure components and proprietary SDKs everywhere.
  • After, they had one Waterstream deployment on Kubernetes, with their choice of cloud provider.
  • The freedom to move workloads went from zero to total.

Manufacturing Client (migrated from a custom architecture)

  • Before, they had 12 microservices to manage MQTT-to-Kafka.
  • After, they had a single Waterstream deployment.
  • Their DevOps team went from four people to 0.5 FTE.

Why Not the Alternatives (The Brutally Honest Version)

  • HiveMQ: It costs as much as a luxury German car. An enterprise license for 50K devices? Get ready to cry. And you still need a separate Kafka instance.
  • EMQ X: The open-source version is okay, but then you go into production, need support, and find it costs as much as HiveMQ. Oh, and the Kafka integration? It’s a plugin no one wants to touch.
  • Confluent MQTT Proxy: It’s. A. Proxy. It adds latency, it’s another component to manage, and it costs like the Confluent Platform. No thanks.
  • Eclipse Mosquitto + Custom Bridge: We’ve all done it. It works for 1,000 devices. At 10K, it starts to creak. At 50K, it’s a maintenance nightmare.
  • RabbitMQ with MQTT Plugin: Cute for hobby projects. In production with millions of messages? Good luck.

The Future We’re (Literally) Building

Now that Waterstream is in our hands at Bitrock, the 2025-2026 roadmap is clear:

  • Q3 2025: WebAssembly plugins for custom transformations without forking
  • Q4 2025: Edge computing mode for Raspberry Pi deployments
  • Q1 2026: Native AI-powered anomaly detection
  • Q2 2026: Multi-region federation with automatic consensus

But the philosophy remains: every feature must eliminate complexity, not add it.

The Lesson for the Industry

After years of “best practices” that were just accepted workarounds, here’s what I’ve learned:

  • Lock-in isn’t inevitable: You can have cloud convenience without vendor prison.
  • Complexity isn’t sophistication: It’s technical debt in disguise.
  • Bridges are admissions of failure: If you need a bridge, the architecture is wrong.
  • True TCO includes freedom: How much does it cost to not be able to change?

Today, when a client tells me, “We’ll use Kafka for IoT,” my answer is simple: “Perfect. Let me show you how to do it without going crazy”. And then we deploy Waterstream. One container. Devices on MQTT. Apps on Kafka. Zero drama.

The revolution isn’t always noisy. Sometimes, it’s just a system that works the way it should.

Main Author: Franco Geraci, Head of Engineering @ Bitrock

Franco Geraci is the Head of Engineering at Bitrock, where, together with his Team, he maintains and evolves Waterstream – while also helping companies escape cloud lock-in. When he’s not freeing Clients from hostage architectures, he’s probably explaining why, no, blockchain won’t solve this problem either.

The post Kafka vs. IoT: Why It Doesn’t Work and the new Solution appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
An IoT Architecture with Waterstream as a Gateway: Flexible, Scalable, and AI-Ready https://waterstream.io/2025/09/22/an-iot-architecture-with-waterstream-as-a-gateway-flexible-scalable-and-ai-ready/ https://waterstream.io/2025/09/22/an-iot-architecture-with-waterstream-as-a-gateway-flexible-scalable-and-ai-ready/#comments Mon, 22 Sep 2025 09:11:51 +0000 https://waterstream.io/?p=5916 The ability to orchestrate complex information flows from the edge to the cloud and extract predictive insights has become a strategic imperative for many companies.  In this context, the synergistic integration of Waterstream (a Kafka-native MQTT broker), Apache Kafka (an event streaming platform), Apache Flink (a stream processing engine), and Artificial Intelligence platforms like Radicalbit represents a true enabler of business transformation. This […]

The post An IoT Architecture with Waterstream as a Gateway: Flexible, Scalable, and AI-Ready appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

The ability to orchestrate complex information flows from the edge to the cloud and extract predictive insights has become a strategic imperative for many companies. 

In this context, the synergistic integration of Waterstream (a Kafka-native MQTT broker), Apache Kafka (an event streaming platform), Apache Flink (a stream processing engine), and Artificial Intelligence platforms like Radicalbit represents a true enabler of business transformation. This robust and scalable architecture allows for the transformation of raw IoT data into actionable intelligence, supporting strategic decisions and optimizing critical processes. 

This article will examine several business cases that exemplify the potential of such synergy.

Predictive Maintenance for Industrial Assets 

Industrial maintenance is a crucial area where IoT and AI can generate a significant impact. One of the main challenges in this sector is minimizing unplanned downtime and maintenance costs while maximizing the uptime of critical assets. 

To address this challenge, advanced sensors can be installed directly on industrial machinery to collect high-frequency data such as vibrations, temperature, and pressure. This data is efficiently transmitted via MQTT to Waterstream, which acts as a Kafka-native broker, channeling the information flows directly into Apache Kafka, ensuring intrinsic scalability and durability. Subsequently, Apache Flink processes these data streams in real-time, promptly identifying anomalous patterns or significant deviations from standard operational profiles. Machine Learning algorithms, hosted on dedicated AI platforms like Radicalbit, then analyze historical and real-time data to build predictive failure models capable of accurately forecasting the remaining useful life (RUL) of components.

The added value of this solution is a strategic transition from reactive or preventive maintenance to predictive maintenance, leading to a drastic reduction in unplanned machine downtime, optimized maintenance scheduling, extended asset life, and a significant reduction in operational costs and safety risks.

Smart Energy Management for Smart Grids and Industrial Consumption

The agricultural sector can also benefit enormously from the application of these technologies, with the goal of optimizing resource use, predicting harvests, and improving crop health while reducing environmental impact. IoT sensors strategically distributed in fields (to detect soil moisture, pH, nutrients, temperature, etc.), monitoring drones, and cameras send georeferenced data via MQTT to Waterstream and then to Kafka.

Flink is fundamental in this phase, streaming, correlating, and aggregating this information to build a real-time view of agronomic conditions. Advanced Computer Vision and Machine Learning algorithms, running on the Radicalbit Platform, then analyze images captured by drones to assess plant health and predict yield. In this context, AI can suggest highly targeted irrigation and fertilization plans, optimizing the use of water and nutrient resources. 

All of this translates into an overall increase in crop yield, a significant reduction in operating costs (thanks to less waste of water and fertilizers), an improvement in the quality of agricultural products, and the adoption of more sustainable agricultural practices.

Smart City and Intelligent Traffic Management 

In the context of smart cities, technology offers advanced solutions to address complex urban challenges like traffic congestion, pollution, and the need to optimize public services. Traffic sensors (inductive, cameras, etc.), data from connected vehicles, environmental sensors, and IoT devices for parking monitoring transmit real-time data streams via MQTT to Waterstream and then to Kafka. Flink analyzes these streams to detect incidents and violations in real-time or to identify movement patterns. AI algorithms, supported by Radicalbit, can predict transport demand, suggest alternative routes in real-time, and support the efficient management of emergency fleets. 

Tangible benefits for cities include improved urban mobility, reduced travel times, greater safety for citizens, and optimized resources for municipal administrations.

Health and Wellness Monitoring 

In the healthcare sector, the application of these technologies is enabling a new frontier of telemedicine and health. The challenge is to provide continuous and personalized health monitoring, enable early diagnoses and timely interventions, improve patients’ quality of life, and reduce overall healthcare costs. 

Wearable devices, medical sensors for detecting vital parameters (such as ECG, blood pressure, oxygen saturation, glucose), and smart home devices collect crucial data on health and the surrounding environment. This data is transmitted in real-time via MQTT to Waterstream and then to Kafka, with strict security and privacy guarantees in line with current regulations. Flink processes this streaming data, normalizing it, aggregating it, and detecting anomalies compared to personalized baselines for each individual. AI and Machine Learning models, operating on Radicalbit, then analyze patterns to identify potentially dangerous medical conditions early, generate automatic alarms for caregivers or medical staff, and suggest personalized care pathways or preventive interventions. The result is a significant improvement in the prevention and management of chronic diseases, robust support for telemedicine, greater autonomy for patients, and overall optimization of healthcare resources.

Intelligent Inventory and Supply Chain Management 

Inventory and entire supply chain management can also be radically transformed through a strategic application of IoT and AI. The primary objective here is not only to optimize stock levels and reduce waste but also to drastically improve product availability and, above all, increase the overall resilience of the supply chain in an increasingly volatile market context. To achieve this, RFID sensors, beacons, smart cameras, and a variety of other IoT devices are employed in warehouses, distribution centers, and retail outlets, monitoring inventory levels, product movements, environmental conditions, and even emerging customer demand in real-time.

All this data, which is crucial for decision-making, is sent in real-time via MQTT to Waterstream, which acts as an efficient and scalable gateway for the Kafka ecosystem, and from there directly to Kafka. Flink comes into play at this stage, streaming and processing these complex data flows to provide a granular and real-time view of inventory, from raw materials to finished products, tracing every single movement and immediately identifying any discrepancies. To complete the picture, advanced Artificial Intelligence algorithms, often operating on platforms like Radicalbit, analyze historical and real-time data to more accurately predict future demand, optimize reorder points and stock levels, suggest optimal allocation strategies, and, crucially, proactively identify and mitigate potential disruptions or bottlenecks within the supply chain. 

This integrated approach translates into tangible benefits, such as a significant reduction in storage costs and obsolescence, a substantial improvement in product availability (minimizing “out of stock” situations), greater responsiveness of the entire supply chain to demand fluctuations, and, ultimately, an increase in customer satisfaction.

Conclusions

At Bitrock, our commitment is focused on providing technologies and architectures that are not just simple tools, but true catalysts for value. Through the integration of Waterstream, Apache Kafka, Apache Flink, and Artificial Intelligence platforms like Radicalbit, we aim to equip companies with the necessary tools to navigate the complexity of IoT data, transforming it into strategic insights and proactive decisions. 

These business cases clearly demonstrate how a robust and intelligent data infrastructure is no longer an option but a fundamental requirement for growth and competitiveness in today’s industrial and technological landscape. Our professionals can accompany your company on this transformation journey, unlocking the full potential of your data: contact us for a free consultation.

The post An IoT Architecture with Waterstream as a Gateway: Flexible, Scalable, and AI-Ready appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
https://waterstream.io/2025/09/22/an-iot-architecture-with-waterstream-as-a-gateway-flexible-scalable-and-ai-ready/feed/ 87
Waterstream: Simplify your MQTT Data Integration at Scale https://waterstream.io/2025/09/15/waterstream-simplify-your-mqtt-data-integration-at-scale/ https://waterstream.io/2025/09/15/waterstream-simplify-your-mqtt-data-integration-at-scale/#comments Mon, 15 Sep 2025 07:37:48 +0000 https://waterstream.io/?p=5888 Waterstream is a cutting-edge solution for managing data in real-time. Waterstream is an MQTT broker that utilizes Apache Kafka as its storage and distribution engine, merging the top communication protocol in the IoT industry with broadly adopted streaming APIs. This incorporation enhances the usability of Waterstream and enables it to manage the connection of numerous clients effortlessly, expanding proportionally and providing […]

The post Waterstream: Simplify your MQTT Data Integration at Scale appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

Waterstream is a cutting-edge solution for managing data in real-time. Waterstream is an MQTT broker that utilizes Apache Kafka as its storage and distribution engine, merging the top communication protocol in the IoT industry with broadly adopted streaming APIs. This incorporation enhances the usability of Waterstream and enables it to manage the connection of numerous clients effortlessly, expanding proportionally and providing exceptional adaptability for various real-time data requirements.

In this article, we will explore in detail what Waterstream is, the benefits it offers, and how its use goes far beyond just the Internet of Things (IoT) sector.

Waterstream and Apache Kafka: a successful match 

Waterstream was created to efficiently handle and distribute real-time data by utilizing the features of Apache Kafka, providing high availability, high throughput, and low latency. In a practical sense, Waterstream functions as a stateless proxy within Kubernetes setups, translating MQTT topics into Kafka. This implies that all data produced by a sensor or device can be sent straight to Kafka, enabling applications to readily make use of this data without any additional steps.

One key feature of Waterstream is its ability to convert protocols bidirectionally, writing MQTT data to Kafka and extracting it back to MQTT for seamless interaction between applications and devices. Moreover, this design eliminates many of the typical challenges in handling an MQTT server, data replication, and specialized integration pipelines.

Application Areas and Benefits of Waterstream

Waterstream yields major advantages across different fields. In the IoT industry, it makes it easier to handle countless connected devices that send live data, seamlessly incorporating data into intricate systems. Let’s explore some of the key advantages provided by this technology: 

  • Scalability and High Availability: Waterstream architecture can handle millions of clients and scale according to the company’s needs, with minimal infrastructure adjustments necessary. The core of this solution is Apache Kafka, which already provides robust features for high availability and fault tolerance to ensure continuous access to secure data.
  • Integration simplicity and decreased complexity: Waterstream eliminates the need for a separate MQTT server to integrate IoT devices with Apache Kafka. The platform simplifies the company’s IT architecture by reducing data duplication and the challenges of creating integration pipelines.
  • Bidirectional Interaction Between Devices and Applications: Waterstream’s capability to convert data back and forth between Kafka and MQTT enables new ways for devices and applications to interact, facilitating ongoing data syncing. This is especially valuable in situations where immediate data analysis and utilization are crucial.
  • Flexibility of the Architecture and Simplified Management: Waterstream is able to work with any platform that Kafka supports, offering adaptability to different business setups and requirements with simplified management. Furthermore, the option to utilize WebSocket or MQTT instead of HTTP for communication enhances the versatility and simplicity of data access.

These advantages make Waterstream the ideal solution for companies wishing to manage high volumes of data from distributed devices in various sectors, not limited to the IoT domain.

Waterstream  beyond IoT

Although Waterstream was created with the aim of making the management of IoT devices easier and more efficient, it can be used for a variety of other purposes as well. Waterstream provides a customizable and adaptable option for businesses across all sectors needing effective handling and transfer of live data. There are more applications than sensor devices alone.

The technology of Waterstream can be used in various situations that require real-time data management, like messaging and gaming, thanks to its infrastructure and features. Kafka’s characteristic quick communication ability further supports this, guaranteeing speedy response times.

Conclusions

In a competitive environment where data and its real-time control are crucial, Waterstream is a valuable tool for companies looking to enhance operational efficiency and streamline device and application integration. Its flexibility makes it a great option for any company looking to maximize the benefits of streaming technologies while ensuring smooth information flow without added complications.

Main Author: Franco Geraci, Head of Engineering @Bitrock

The post Waterstream: Simplify your MQTT Data Integration at Scale appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
https://waterstream.io/2025/09/15/waterstream-simplify-your-mqtt-data-integration-at-scale/feed/ 71
Waterstream Joins the Connect with Confluent Partner Program https://waterstream.io/2023/07/18/waterstream-joins-the-connect-with-confluent-partner-program/ Tue, 18 Jul 2023 20:53:59 +0000 https://waterstream.io/?p=5481 The easiest way to directly integrate data streaming with the Internet of Things (IoT) leveraging the standard MQTT protocol  [Milan] – July 18, 2023 – Waterstream today announced it has joined the Connect with Confluent partner program. This new program helps organizations accelerate the development of real-time applications through a native integration with Confluent Cloud. […]

The post Waterstream Joins the Connect with Confluent Partner Program appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

The easiest way to directly integrate data streaming with the Internet of Things (IoT) leveraging the standard MQTT protocol 

[Milan] – July 18, 2023 – Waterstream today announced it has joined the Connect with Confluent partner program. This new program helps organizations accelerate the development of real-time applications through a native integration with Confluent Cloud. Organizations now have the best experience for working with data streams within MQTT clients, paving a faster path to powering next generation customer experiences and business operations with real-time data. 

“Joining the Connect with Confluent partner program is an exciting milestone for Waterstream. This partnership allows us to directly integrate Waterstream Cloud with Confluent Cloud, opening up endless possibilities for our customers.
By leveraging Confluent Cloud’s powerful data streaming platform, we provide a cost  efficient solution to connect millions of MQTT clients to the Apache Kafka® ecosystem with just a few clicks. Waterstream brings exciting new features like replaying MQTT messages on the client or data validation with Confluent’s fully managed Schema Registry.
We are thrilled to be a part of the initial launch of this program and we are confident this partnership will deliver great value to customers.”
– Paolo Castagna, Strategic and Business Advisor at Waterstream

Connect with Confluent gives organizations direct access to Confluent Cloud, the cloud-native and complete data streaming platform that processes more than an exabyte of data per yearIt’s now easier than ever for organizations to stream data from anywhere with MQTT clients like cars, smart bands, scooters or even browsers with a fully managed Kafka service that spans hybrid, multi-cloud, and on-premises environments. Moreover, the program supercharges partners’ go-to-market efforts with access to Confluent engineering, sales, and marketing resources. This ensures customer success at every stage from onboarding through technical support. 

Data streaming is now a critical business requirement as companies shift toward a digital-first approach to everything,” said Paul Mac Farland, vice president, partner and innovation ecosystem, Confluent. “However, many companies don’t have the resources needed to successfully bring a complete set of data streaming capabilities to their applications and end users. Connect with Confluent solves this problem, helping Confluent technology partners accelerate their customers’ data-driven ambitions so they can win in the modern digital era.” 

Get started with Waterstream.

The easiest way to get started with using Waterstream is with Waterstream Cloud, with Waterstream Cloud you can create an MQTT Broker connected with Confluent and Kafka data streams with just a few clicks. If you prefer a self-managed approach with greater control and flexibility, Waterstream also provides this option. To get started with the self-managed solution, follow this link.

About Waterstream

Waterstream is a leading technology company specializing in data streaming solutions. Waterstream innovative MQTT Broker expands the Kafka ecosystem with IoT use cases, supporting the most popular IoT protocol with simple and efficient integration. Waterstream is available on edge, on-prem, public and private clouds, and as a fully managed solution with Waterstream Cloud.

Waterstream has formed a strategic partnership with Confluent, the creators of Apache Kafka, to leverage the power of Confluent Cloud. This integration combines the most popular streaming platform with MQTT allowing businesses to process and analyze real-time data seamlessly from any device.

Waterstream has received investment and support from CDP Venture Capital and Digital Magics. These partnerships provide Waterstream with resources and expertise to continue to enhance and evolve its platform. 

The post Waterstream Joins the Connect with Confluent Partner Program appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
The Future of Decision-Making: Collaborative Intelligence with Waterstream Cloud, Confluent Cloud, and ChatGPT https://waterstream.io/2023/07/10/the-future-of-decision-making-collaborative-intelligence-with-waterstream-cloud-confluent-cloud-and-chatgpt/ Mon, 10 Jul 2023 16:17:21 +0000 https://waterstream.io/?p=5456   This blog post is co-authored with Donnie Kerr, Principal Technology Architect at Centric Consulting   Introduction In today’s fast-paced technological landscape, organizations are continuously exploring opportunities to harness the power of real-time data streaming and artificial intelligence (AI) to drive productivity and streamline processes. This blog post presents an innovative concept called the Collaborative […]

The post The Future of Decision-Making: Collaborative Intelligence with Waterstream Cloud, Confluent Cloud, and ChatGPT appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

 

This blog post is co-authored with Donnie Kerr, Principal Technology Architect at Centric Consulting

 

Introduction

In today’s fast-paced technological landscape, organizations are continuously exploring opportunities to harness the power of real-time data streaming and artificial intelligence (AI) to drive productivity and streamline processes. This blog post presents an innovative concept called the Collaborative Intelligence Platform (CIP) by our partner Centric Consulting, which revolutionizes decision-making by seamlessly integrating AI, private enterprise data, and human guidance. Central to CIP are cutting-edge technologies such as Waterstream Cloud, Confluent Cloud, and advanced language models (LLMs) like ChatGPT.

 

From Waterstream to Waterstream Cloud

Waterstream was created as a straightforward yet sophisticated solution to bridge the gap between the Internet of Things (IoT) world, which utilizes MQTT, and the Apache Kafka® data streaming ecosystem. Despite its conceptual simplicity, Waterstream offers a multitude of configurations that users can customize. These configurations include mapping MQTT topics to Kafka topics and establishing authorization and authentication settings. Even seasoned DevOps professionals may require time to grasp and manage the details.

 

To address this, Waterstream has evolved from being solely available as a Docker image with accompanying scripts to becoming a software-as-a-service platform called Waterstream Cloud. Currently, Waterstream Cloud is accessible only by invitation (contact us if you want to be included in the waiting list), but it will soon transition to the General Availability status. Through a user-friendly interface, users can effortlessly create one or multiple Waterstream clusters in a specific region on AWS and connect them with existing Confluent or open source Kafka clusters with just a few clicks.

 

Once the connection to Kafka is set up, users can focus more on defining the business case, determining how Waterstream maps MQTT topics to Kafka topics, and establishing the appropriate authorization and authentication rules. Waterstream Cloud further enhances the user experience by integrating a straightforward UI, enabling users to send and receive messages from the Waterstream Cluster or any MQTT broker, provided that the cluster has WebSocket functionality enabled. Additionally, if the Kafka cluster is temporarily unavailable or inaccessible, users can still test Waterstream by creating a single-node sidecar Kafka cluster that allows Waterstream to operate. However, it’s important to note that the sidecar Kafka cluster is not designed for production use and cannot be accessed by external producers or consumers.

Waterstream Cloud and Confluent Cloud, a perfect match

Waterstream Cloud simplifies the creation of a Waterstream MQTT broker, and it pairs perfectly with Confluent Cloud. Confluent Cloud is a fully managed, cloud-native, and complete dataevent streaming platform offered by Confluent, the company founded by the original creators of Apache Kafka. It provides a scalable and reliable infrastructure for developing real-time data streaming applications and implementing event-driven architectures in the cloud.

 

By leveraging Confluent Cloud, users can harness the power of Kafka without the burden of managing the underlying infrastructure. To facilitate the seamless integration of Waterstream Cloud and Confluent Cloud, a dedicated configuration form is available. This form includes the necessary Kafka parameters from Confluent Cloud, making it effortless for users to connect the two platforms. Additionally, Waterstream offers the option to validate the payload of incoming MQTT messages, formatted as JSON, using the fully managed Schema Registry included in Confluent Cloud. It is worth mentioning that the integration of Waterstream with the Confluent Schema Registry is a powerful combination that brings unified validation for MQTT messages and Kafka records. This integration simplifies development because it ensures consistent validation rules, and enhances interoperability between MQTT-based devices and Kafka-based services.

Confluent Cloud options in Waterstream Cloud

Here is a short video showcasing the quick and straightforward process of starting a Waterstream Broker with Waterstream Cloud. It highlights the seamless connection to Confluent Cloud, enabling users to send MQTT messages directly from their browser, which are then written into Kafka.

The Collaborative Intelligence Platform

At this pivotal moment, generative AI technologies offer the potential for remarkable productivity gains that organizations desire. However, only 27% of companies currently use AI tools, despite the 90% higher productivity reported by AI users. To foster the success of generative AI in the workplace, it must operate in a trusted environment where users have control of their own data and unwavering confidence in the relevance of its outputs.

 

Donnie Kerr, Principal Architect at Centric Consulting, has been exploring the capabilities of this approach and introduced the paradigm of Collaborative Intelligence as a new methodology:

 

“Collaborative Intelligence leverages AI to mature a business process from little-to-no automation up to fully autonomous or somewhere in between. The end state of each process will vary depending on the goal. In some cases, a process can become fully autonomous, not requiring any human intervention. In other cases, a process may be very manual at first.”

 

The Collaborative Intelligence Platform (CIP) seamlessly incorporates cutting-edge technologies essential for the comprehensive implementation of the CI methodology. The platform architecture harmoniously integrates prominent cloud-native technologies, as depicted below:

The Collaborative Intelligence Platform, courtesy of Centric Consulting

To illustrate a practical application of CIP, let’s consider an example. Imagine an operator working in a logistics center who receives a notification through a conversational user interface powered by ChatGPT. In real-world data, the shipment system detects a delay of 30 minutes. The notification informs the operator that a shipment they are responsible for will be late. The operator is then prompted to decide whether they will handle the delayed shipment or if it should be reassigned to the next shift. As time progresses, the AI learns from patterns and identifies that parcels originating from a specific location often encounter delays, leading to automatic reassignment to the subsequent shift.

The core essence of CIP lies in the ability to seamlessly combine an AI platform with private enterprise data in real time, along with user decisions. In this scenario, Confluent Cloud plays a vital role by facilitating the merging of real-world events, user actions, and AI-generated outcomes. This integration enables a smooth and dynamic exchange of information, leading to improved collaboration and decision-making within the CIP framework letting the human control the full process.

Let’s delve into the technical workflow of the Collaborative Intelligence Platform (CIP) by examining a simple Proof of Concept (PoC) that highlights its essential components. This PoC leverages Waterstream Cloud and Confluent Cloud. Both technologies utilize the concept of topics, MQTT or Kafka.  MQTT topics and Kafka topics share the same name to facilitate seamless mapping.

The process initiates with user interactions on the UI, where they make information requests using ChatGPT prompts. To streamline testing, the UI offers some examples to get the user started.

The ChatGPT Prompt

Within the UI front end, an MQTT client communicates with Waterstream Cloud to create new chats, send the user’s prompt, and remember the conversation history. Anyone experimenting with Chat GPT APIs knows it doesn’t maintain context. Your application will need to manage that, and we think of this in terms of state.

Usually, the app will have to send the call with the initial prompt payload (plus some hidden prompt engineering) and get the response, then send the initial prompt, plus the response, plus the new prompt for ChatGPT. This means incrementally larger payloads and slower response times as you iterate. Not to mention, the history is trapped in that session or state. But, if we use Confluent Cloud and ksqlDB, for example, we can preserve both state and history.

In our case, ksqlDB is used to load the user’s chat history and the ongoing conversation flow within each chat. This also allows for opening previous chats and continuing the conversation in real time. You could add a tumbling window so that the stream presented by ksqlDB only has the most recent chats or for a specific retention window. You could also use ksqlDB to present a single materialized view of the user’s prompt template. For example, a control operator would ask different questions than a diagnostician or machine programmer. In any of these cases these same streams are persisted in Kafka for learning and training the back end models, the results of which which can then be merged back in to these contexts, and get better over time.

Waterstream Cloud enables the direct saving of new Chats into a “Chats” topic for each user and user prompts as MQTT messages into a Confluent Cloud Kafka topic named “chatMessages”, eliminating the need for intermediaries. From there, the platform leverages ksqlDB stream processing to orchestrate the flow of events.

ksqlDB and MQTT topics are a match made in heaven.  MQTT topic structure is ideal for routing flows of events from devices, UI, and the AI assistant into various ksqlDB stream pathways depending on the topic and/or value of the message. For example, the ChatGPT app supports Azure OpenAI GPT-3.5-turbo or OpenAI GPT-4.  A single prompt gets routed from the “chatMessages/{chatId}” MQTT topic to either Azure OpenAI or OpenAI Azure Function Sink connectors.  “ChatMessages_stream_aoai” and “ChatMessages_stream_oai” are the input streams that trigger the api calls. 

Similarly, to make the prompt aware of the enterprise context, chatMessages can be joined with data obtained from private data streams using ksqlDB. The MQTT topic effectively steers the prompt to be appended with the correct private data source(s). These private streams can be generated using connectors available in Confluent Cloud.  The augmented prompt topic is then forwarded to ChatGPT for processing.

 

The PoC within Confluent Cloud’s Stream Lineage

ChatGPT generates responses in the form of chunks, which are groups of one or more words. These chunks are stored in a stream called “chatgptResponseStream” as they are generated.

The chatGptResponseStream in Confluent Cloud

At this stage, Waterstream Cloud plays a role once again by transmitting the answer chunks from the Kafka topic as messages to a specific MQTT topic. The frontend client subscribes to the MQTT topic “chatgptResponseStream.” As ChatGPT generates responses, the UI dynamically updates to display the words in real time. This allows users to observe the gradual construction of the response, with new words appearing on the screen, similar to the familiar OpenAI application. It is worth mentioning that other topics are used in this PoC to save chat history and user information, but aren’t  discussed in detail here.

One of the advantages offered by Waterstream is the ability to write directly from the UI into Kafka and vice versa. ChatGPT chunks and other events can be easily communicated to the UI as notifications. Utilizing an MQTT broker instead of plain WebSockets enables the organization of data flow into multiple streams, simplifying the creation of a reactive UI. Moreover, Waterstream Cloud provides authentication and authorization capabilities, which are crucial for enterprise applications. User information can also be utilized to organize MQTT topics, leading to a well-structured data system.

 

For further details about the Collaborative Intelligence Platform (CIP), please refer to the “Collaborative Intelligence: An AI-Powered System That Puts Humans in Control” white paper available on the Centric Consulting website. Another interesting reading on these themes is a blog post titled “GPT-4 + Streaming Data = Real-Time Generative AI” by Michael Drogalis, Principal Technologist at Confluent. 

Conclusion

In conclusion, in this blog post we discussed the journey from Waterstream to Waterstream Cloud, the integration with Confluent Cloud, and the idea and implementation of a Collaborative Intelligence Platform (CIP). Together, these technologies form a unified system that combines user prompts, actions, and private enterprise data to form Collaborative Intelligence. Although CIP is still in its early stages, it holds great potential by utilizing a streaming data approach to merge real-world events, user knowledge, control, and AI assistance. This integrated approach shows promise for enhancing collaboration and decision-making processes with the human still in the center.

 

The post The Future of Decision-Making: Collaborative Intelligence with Waterstream Cloud, Confluent Cloud, and ChatGPT appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
Simplify IoT usecases at the Schiphol Airport with MQTT, Kafka and Waterstream https://waterstream.io/2022/10/03/simplify-iot-usecases-at-the-schilphol-airport-with-mqtt-kafka-and-waterstream/ Mon, 03 Oct 2022 18:00:59 +0000 https://waterstream.io/?p=5320 The mission of the IoT initiative at the Schiphol airport is to provide an infrastructure as a service that contractors (e.g., B2B companies) and the airport can leverage to share data with the Schiphol Group.The goal is to accelerate and simplify the onboarding of new contractors, ideally up to a self-service fashion and allow them […]

The post Simplify IoT usecases at the Schiphol Airport with MQTT, Kafka and Waterstream appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

The mission of the IoT initiative at the Schiphol airport is to provide an infrastructure as a service that contractors (e.g., B2B companies) and the airport can leverage to share data with the Schiphol Group.
The goal is to accelerate and simplify the onboarding of new contractors, ideally up to a self-service fashion and allow them to improve their own services with data from sensors installed at the airport.

The cornerstone of the strategy is the messaging backbone that the IoT team at Schiphol created using Kafka at its core. At one end, it collects data from sensors and assets deployed in the airport, usually leveraging LoRaWAN in combination with HTTP and MQTT. On the other side, services on Kubernetes process the ingested data to create events or messages that, in the same way, are stored in Kafka. B2B customers and contractors receive such events or notifications again via the MQTT protocol. With this knowledge, third parties can manage their on-site assets and services more independently thus enhance their offerings. Current use cases include predictive maintenance, passenger feedback, and asset tracking.

In this context, Waterstream will play a key role. Waterstream works on both ends, providing an MQTT broker that translates the data from MQTT to Kafka and vice versa without requiring an additional integration layer with Kafka. Waterstream makes the integration easy so the team can focus more on the business case instead of solving integration problems.

The benefits of using Waterstream will grow as the IoT team works to reduce or eliminate the use of HTTP in favor of the MQTT protocol to increase scalability and integration with multiple sensors.

This blog post is co-authored with Werner de Bruijn and Tristan Godfrey from the Schiphol Group

The post Simplify IoT usecases at the Schiphol Airport with MQTT, Kafka and Waterstream appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
Waterstream Offers Cloud MQTT Broker Through AWS Marketplace https://waterstream.io/2021/07/28/waterstream-offers-cloud-mqtt-broker-through-aws-marketplace/ Wed, 28 Jul 2021 08:51:14 +0000 https://waterstream.io/?p=3404 By extending its relationships with Amazon AWS cloud platform provider Waterstream is making it easier for partners and customers to build and operate high-performance IoT applications in hybrid-cloud environments. Waterstream with this parnership makes it easier to build and operate MQTT-connected applications in cloud environments. If you wish to try our AWS-based MQTT Broker for […]

The post Waterstream Offers Cloud MQTT Broker Through AWS Marketplace appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

By extending its relationships with Amazon AWS cloud platform provider Waterstream is making it easier for partners and customers to build and operate high-performance IoT applications in hybrid-cloud environments.

Waterstream with this parnership makes it easier to build and operate MQTT-connected applications in cloud environments.

If you wish to try our AWS-based MQTT Broker for free. You can find it on the AWS Marketplace or you can use our self-hosted solution that is recommended for edge computing.

The post Waterstream Offers Cloud MQTT Broker Through AWS Marketplace appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>
What if the Waterstream node goes down? https://waterstream.io/2021/04/16/what-if-the-waterstream-node-goes-down/ Fri, 16 Apr 2021 06:26:39 +0000 https://waterstream.io/?p=2994 If you’ve ever wondered how Waterstream cluster behaves if one of its members goes down – here are some tests we’ve done. The test setup consists of Kafka provided by Confluent Cloud, 5 nodes of Waterstream and a load balancer running on Google Cloud Platform, and 5 nodes of MQTT load simulator – also on GCP. The scripts we […]

The post What if the Waterstream node goes down? appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>

If you’ve ever wondered how Waterstream cluster behaves if one of its members goes down – here are some tests we’ve done.

The test setup consists of Kafka provided by Confluent Cloud, 5 nodes of Waterstream and a load balancer running on Google Cloud Platform, and 5 nodes of MQTT load simulator – also on GCP.

The scripts we used for creating topics in Confluent Cloud and for running Waterstream in Google Cloud are here: https://github.com/simplematter/waterstream-gcp-terraform. This setup has a single Kafka topic for MQTT messages with 10 partitions (in Confluent Cloud cluster capacity depends on number of partitions). For Waterstream deployment, we’ve used 5 n1-standard-1 nodes (1 CPU, 3.75 GB RAM). A separate VM hosts Prometheus and Grafana, which we’re going to use for monitoring Waterstream behavior during a simulated node failure.

The scripts we used for creating topics in Confluent Cloud and for running Waterstream in Google Cloud are here: https://github.com/simplematter/waterstream-gcp-terraform. This setup has a single Kafka topic for MQTT messages with 10 partitions (in Confluent Cloud cluster capacity depends on number of partitions). For Waterstream deployment, we’ve used 5 n1-standard-1 nodes (1 CPU, 3.75 GB RAM). A separate VM hosts Prometheus and Grafana, which we’re going to use for monitoring Waterstream behavior during a simulated node failure. MQTT load generator also has scripts for launching it on Google Cloud.

We’ve configured it to run 5 nodes on the same machine type – n1-standard-1. Each node spawns 20k clients with ramp-up time 120 seconds. Together that makes 100k clients. When ramp-up completes, each clients sends 0.8… 1.2 KB QoS 2 (exactly once) PUBLISH message every 10 seconds. Clean Session flag for all the clients is false, so that we could also test the loading of the session data upon client reconnect.

Having all this infrastructure started, we’ve waited few minutes to see all the clients connected and produce messages and expected. 

The load simulator before stopping a node - see the red line
Waterstream connected clients before stopping a node - see the red line

Then we opened a console of one of the Waterstream nodes and shut down the Docker container of the Waterstream and watch Waterstream Grafana dashboard to see the effect:

Waterstream rebalancing the load of survivor nodes - see the red line

As you can see, after a while clients started to notice that connections are broken and they need to re-connect. 1 minute 45 seconds after the start of the simulated incident all the clients have successfully connected to the survivor nodes. Looking at the connection details in the Load Simulator Waterstream dashboard we can see that there were 19.9k connections with the node that went down and that there were some unsuccessful attempts to re-connect while the load balancer hadn’t yet detected the node failure:

The load simulator shows restoring clients on the survivor Waterstream nodes - see the red line

And here is the part of the Waterstream dashboard that shows the session loading metrics. You may see that it has successfully loaded existing sessions for the clients that were re-connecting.

As the tests are complete, shut down load generator and Waterstream, and remove topics from Confluent Cloud to stop being charged.

 

 

This test demonstrates how survivor Waterstream nodes may take over the load of the failed one, keep the cluster running and client sessions available. If you want to repeat these tests yourselves you can ask evaluation license here and get support on our forum.

Enjoy your IoT!

Author: Paul Lysak

The post What if the Waterstream node goes down? appeared first on Waterstream.io | Simplify MQTT Data Integration.

]]>