This a great feature. We started using it right away when it was available as a private preview for production jobs - that’s how great it is. I only have two core complaints: 1. It’s not possible to setup unlimited retries of child tasks - so it’s not good for launching multiple autoloader tasks that are going to run continuously. 2. As you are stepping through a previous or currently running workflow, you cannot go back to the parent job easily. It trips me up everytime.
@AdvancingAnalytics3 ай бұрын
Super useful feedback! I'm getting some feedback over to the product team, I'll bring these points up in that session - Simon
@rbharath893 ай бұрын
Interesting…how well does it handle failure and restarts after failure…say the job fails in the second of the inputs…or the second of the process inside the input how does it handle it? Does repair run start a new run from the beginning ?? Does it fail immediately if if the concurrent job fails ?
@GraemeCashАй бұрын
It's a really good feature and brings the prospect of removing tools like ADF. However, I tried running an ingest notebook using autoloader where I pass in a dictionary of values via a task value reference. When I tried doing it for a large number of tables I hit the 48 KiB limit. So, I will have to revert to using a threading function or work out how I can chunk up the data being passed to the child notebook and have another loop.
@StartDataLate18 күн бұрын
by putting subflows in one workflow, that means sharing cluster between workflows is possible? i mean the normal cluster not the serverless
@colinmanko70022 ай бұрын
Thanks for the video! I’m stuck in something I can’t quite figure out. Maybe you’d be willing to help! How do you run sequential backfills in Databricks? The concept is, you have a scd type 2 / cumulative table. And you need to backfill it to the same state that it currently is with some adjustment using the dependencies’ delta table versioning. So with airflow and without delta tables, you’d have a data lake table that has a daily dump as a date partition within the same table. So say, using airflow and this snapshot style table, you would simply compare the last two snap shots and when you set it in airflow, you say “depends_on_past”. In this way, you would go back and on a daily basis, for each day, do your compare. What I can’t figure out is an elegant way to do this in Databricks with a delta table. In particular because the delta table does not have a daily dump partition (I guess I could add it but trying to save space!) The closest thing I can think, which seems awkward, is to do this for each loop and have a metadata task like you have, but to set the version dates I want to run off in a list. So if you imagine a scd type 2 table I’m trying to backfill, you 1. set a date for starting sequential backfill, 2. get the previous update/insert timestamps through the scd type 2 table version history, 3. concaténate it with say a daily date for the dependency for previous versions of that delta table that are prior to the new table’s creation date. Hopefully this makes sense, but then you can play the backfill over a deterministic set of dates, which would give you the same state back for the backfill. Can you think of a more elegant way to do this with delta tables? It seems very complicated 😅
@pic1013 ай бұрын
Horray, Databricks Workflows does ForEach! [Queue side-eye from ADF.]
@SheilaRuthFaller3 ай бұрын
Great Content. I hope you can have another vid showcasing the use of shared cluster. 😊
@brads20413 ай бұрын
Interesting. We used ADF in our project for orchestration after finding out that dlt does not handle dynamic ingest processing. Will have a look at this. We are using a dedicated interactive cluster.
@montekverma17273 ай бұрын
Hello there. Did you try nesting a DLT Pipeline in a DBX For Each loop? I have been trying to do it but am unable to pass value from the loop to DLT Pipeline's parameters.
@jeanheichman41133 ай бұрын
Are scripts available in a repo? Using ForEach in ADF streamlines things brilliantly and can't wait to use it Databricks Workflows