I have one very practical question, having hit frustrations pushing data from our data warehouse into Salesforce (I see you have a Salesforce adapter).
When you push data into Salesforce, and Salesforce (API) returns error messages indicating some row in the target can't be modified because it is busy, is your tool able to detect which rows couldn't get loaded and retry those rows N times (hopefully with some exponential backoff delay or perhaps a fixed delay between retries), and then if those retries fail, report back up permanent failures of which rows/IDs didn't make it to Salesforce? (I haven't run across a tool that does this but this was my pain point.)
Cool! I'm working on something on the other side of the same pipeline. Real time data ingestion and identity resolution, delivering into a data lake.
Have you considered adding source connectors for S3 based data lakes? For example Parquet files or Delta Lake? Maybe via AWS Athena to make it similar to the Red Shift connector?
this looks awesome! Cant wait to try!
Will this stay open source or will it end up as another limited open core data product?
[flagged]
[flagged]
I understand Reverse ETL as a concept, but why does it require different software to solve? Aren't most ETL/ELT tools designed to move data from any source to any other source? We've got a pretty vanilla requirement of pulling data from CRM to data warehouse for reporting and then pushing rollups back to CRM. And we use the same tool for both flows.