Streamdal

Detect and resolve data quality incidents faster

5.0
1 review

504 followers

Data observability that drives action. Detect and resolve data quality incidents faster by viewing data flowing through your systems and acting on them in real-time with Streamdal, the open-source data observability and governance tool.
Free
Launch Team
NMI Payments
NMI Payments
Don’t Integrate Payments Until You Read This Guide
Promoted

What do you think? …

Ustin Zarubin
We're so excited to share the next evolution of Streamdal with PH! Thanks @mwseibel for Hunting us. Streamdal is an open-source data observability tool that drives action. Data engineering teams will be able to detect and resolve data incidents faster by viewing data flowing through their systems and acting on it in real-time. We built Streamdal because we've experienced a growing divide between the systems that handle data and the tooling used to detect and resolve incidents. Observing data happens downstream once data is already in data stores, while at the same time detecting and resolving data incidents typically requires you to trace the issue back upstream. We believe both observability, detection, and resolution should exist upstream. Streamdal helps teams overcome this divide by generating a real-time, dynamic visualization of the data flow: we call it “The Data Graph”. This means teams will get a dynamic bird's-eye view of their producers and consumers as they scale up and down in real-time. From there, they’ll be able to watch a live view of data flowing throughout their system. By importing and wrapping a few lines of code with the Streamdal SDK, they’ll be able to detect data incidents the moment they happen. Along with the dynamic data graph, they’ll get throughput metrics, schema inference, and the ability to zoom in and observe real-time data - it's essentially a `tail -f`, but for data and in human readable format. Today, we're excited to share our open-source product with the PH community! In this beta phase of our launch, we currently have SDK support for Golang, Python and Node.js. The UI and server components can be deployed in containers anywhere: on-prem or in cloud. Coming very soon is a host of data quality management features allowing you to proactively interact with real-time data, and prevent downstream issues! Think of it as a `firewall` for your data. We'd love to hear your thoughts, ideas, and feedback. We'll be here to answer questions in the comments.
Daniel Feles
@ustin cool product! congrats and good luck with the launch!
Eugene Yarovoii
@ustin Oh boy, your launch is a total game-changer! It's exactly what we've been hoping for. Sending you all the best as your project continues to thrive. Followed you on Twitter because we know greatness when we see it! 🌟🐦
Gadir
@mwseibel @ustin The ability to detect and resolve data quality incidents in real-time is crucial in a data-driven world. Streamdal's focus on providing visibility into the flow of data through systems and the capability to take immediate action can significantly improve data reliability and integrity Coooool!
Ustin Zarubin
@eugeneyarovoii Thank you for the kind words!!
Ustin Zarubin
@mwseibel @gipetto This is what we are going for! Upstream data validation, reliability, and integrity.
Oleg Naumenko
Congrats on the launch! Your approach to observability and data incidents resolution truely stands out. Keep up the good work!
Paul Gorval
Hey @ustin, I've just had a look at Streamdal and I must say, I'm intrigued! 😃 The divide between data handling systems and incident resolution is a real challenge, and seeing your product attempt to bridge this gap is so refreshing. I love the concept of "The Data Graph" – dynamic, real-time visualization sounds like a game changer. 🚀 Just a quick question, when will the data quality management features you mentioned be rolled out? I can't wait for it. Great job and best of luck with your beta phase! 👏
Dan Silber
@ustin @paul_gorval Hey Paul! Some of the data quality management features are already available in beta within the console (we haven't put them through their paces yet) in the Pipelines section of the UI. We've outlined more of our features that are coming soon in our roadmap! https://github.com/orgs/streamda...
André J
So I guess the differentiator here is the "realtime observability" How are the competitors in this space? Are they not realtime?
Daniel Selans
@sentry_co you got it! We actually talk about exactly this in our manifesto - real time observability is not actually real time .. it’s closer to “fairly recent”. Our approach to this though is not just “make it faster” - it’s about providing a way to expose data in a way that none of our competitors have done yet. Specifically - allowing folks to look at the actual bits that the app is reading or writing. Or less abstract - our competitors are doing everything via metrics, traces and logs - we are saying “that’s cool, but here’s another angle that may work better”.
Daniel Selans
@sentry_co forgot to link manifesto which talks exactly about this: https://streamdal.com/manifesto
Zac Harris
Congrats on the launch, team! I love everything I saw from the demo. So much so, that I'm trying to get the Copy.ai team onboard with using y'all!
Dan Silber
@zac_t_harris Love to hear it!
VALENTINE JOHN OGOSHI
This will come in handy for my data analysis friends.... I'll sure recommend
Daniel Selans
@valentine_john_ogoshi Definitely - data eng is a _perfect_ use-case for this.
Ken Savage
Congrats on your launch Y'all. How much overhead does this add to something like a NodeJS script? This is all that's needed? import { OperationType, Streamdal } from "@streamdal/node-sdk/streamdal"; export const example = async () => { const streamdal = new Streamdal({ streamdalUrl: "localhost:8082", streamdalToken: "1234", serviceName: "test-service-name", pipelineTimeout: "100", stepTimeout: "10", }); const result = await streamdal.processPipeline({ audience: { serviceName: "test-service", componentName: "kafka", operationType: OperationType.PRODUCER, operationName: "kafka-producer", }, data: new TextEncoder().encode(JSON.stringify({ key: "value" })), }); };
Jacob
Hi Ken, I wrote the node sdk! There is some minimal overhead in that example for stringifying and encoding the data that can be optimized. As far as the pipeline rules are concerned, we implemented the pipeline rules in wasm and the overhead is very low, about 1% or less. I haven't benchmarked the node sdk yet, but we have for the go sdk and you can see it here: https://github.com/streamdal/go-...
1234
Next
Last