Explore our expert-made templates & start with the right one for you.
Upsolver's data ingestion platform enables high quality, observable data ingestion from streams, files, and operational databases for the Snowflake Data Cloud. Roy Hasson, head of product at Upsolver, shows how you can now easily build dbt Core models that deploy data movement tasks on Upsolver, bringing software engineering best practices to data ingestion for the first time.
Learn how Upsolver has adapted best practices for disaster recovery to streaming data from sources such as Kafka and Kinesis so your data architecture is always resilient.
Learn how Upsolver performs joins and aggregations efficiently on massive data in motion.
Upsolver CTO and Co-Founder Yoni Eini and Head of Data Santona Tuli break down incremental models — why we need them and why they’re difficult to write.
Data ordering is a prerequisite for exactly once processing. Watch to find out how Upsolver solves this difficult problem.
Yoni and Santona explain how Upsolver has implemented effective exactly once processing by leveraging idempotency — the concept that a repeated process should always produce the same result — and at least once processing.
Upsolver CTO and Co-Founder Yoni Eini and Head of Data Santona Tuli explore the most common types of streaming data organizations encounter.
Upsolver CTO and Co-Founder Yoni Eini and Head of Data Santona Tuli dive into the definitions of realtime and how they relate to data ingestion and analytics.
Learn practical approaches you can implement today to help your company start benefiting from Data Mesh.
Join Upsolver CTO Yoni Eini to learn how we automate synchronizing pipeline tasks to guarantee strong consistency.
Join Upsolver CTO Yoni Eini to learn how to ensure events are properly ordered in Upsolver's stream processing engine.
Join Upsolver CTO Yoni Eini to learn how we automate state score management at scale.
Learn from Upsolver CTO Yoni Eini about automating orchestration of pipeline tasks as data velocity increases fro daily to minutes.
Building Self-Orchestrating Pipelines for Presto
Learn to deliver analytics-ready data from batch & streaming sources with no manual orchestration
This brief demonstration walks through the first part of building a declarative data pipeline: Ingesting data into SQLake.
ironSource managed to transform 500K events per second, using only a visual interface and SQL, saving thousands of engineering hours.
In this brief video, we demonstrate in real-time how easy it can be to build, enrich, preview, and go live with a complete streaming data pipeline
Discover how Upsolver helps you take streaming nested data containing arrays and prepare it for Amazon Athena.
Watch this video to see how quickly you can build a data pipeline and begin analyzing data from your data lake
See how you can combine data lake economics with database simplicity and speed, and simplify your cloud analytics initiatives
We walk you through the architecture and detailed steps to ingest data from Salesforce to an Upsolver-enabled Amazon S3 bucket.
Take this quick tour of Upsolver's capabilities and see how you can combine data lake economics with database simplicity and speed.
Upsolver can load data into many databases including Snowflake. This video guide shows you how to create a Snowflake data output.
Many users are looking for ways to reduce their Splunk cost. This video provides an example of how to index less data into Splunk.
Learn how to set up & operate a data lake using Upsolver, query it from Athena, & visualize anomalies with Looker - with no code.
Explore our expert-made templates & start with the right one for you.