Join us at FabCon Atlanta from March 16 - 20, 2026, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.
Register now!Calling all Data Engineers! Fabric Data Engineer (Exam DP-700) live sessions are back! Starting October 16th. Sign up.
Hi all, my Fabric Dataflow gen2 failed after 4 hrs, and I am getting this error message. Is there some time limit, and I'm not sure. Please help with this part.
Source- Databricks
Destination- Fabric Lakehouse
Medium- Dataflow Gen2 (No transformation, only full load)
1 Table - 120M+ records, 160+ columns, Size~ 18GB
Fabric SKU- F128
Thanks!
Solved! Go to Solution.
Hi @avisri,
Thanks for sharing the update. The error you are seeing "Lakehouse036: conflicting metadata change" is happening because multiple Dataflow Gen2 jobs are writing to the same Lakehouse table in parallel. When this happens, each dataflow tries to commit changes to the table metadata, which leads to conflicts, and some of them fail.
To resolve your issue try avoiding parallel writes to the same table. Instead of 5 Dataflows writing into the same final table at the same time, run them sequentially. This will ensure there are no metadata conflicts when committing changes.
Or you can use staging tables. Point each yearly Dataflow to its own staging table in the Lakehouse (e.g., Table_2019, Table_2020, etc.). After that, you can either use a notebook or Pipeline activity to merge these staging tables into your final table, or create a lakehouse shortcut or union query on top of them.
If you prefer keeping the Dataflows, orchestrate them using a Fabric Data Pipeline to run one after another (or write to staging first, then merge).
This approach removes the metadata conflict and will make your loads more stable.
Best Regards,
Hammad.
Hi @avisri,
Thanks for reaching out to the Microsoft fabric community forum and yes, Fabric Dataflow Gen2 currently has a affective execution time limit of 4 hours per run, even with higher SKUs like F128. Since your job ran for exactly 4:00:43 and then failed, itโs very likely that it hit this hard timeout limit.
Given your scenario (18GB, 120M+ rows, 160+ columns, full load), even without transformations, this volume can push Dataflow Gen2 close to or beyond the timeout window, especially when moving from external sources like Databricks.
If I misunderstand your needs or you still have problems on it, please feel free to let us know.
Best Regards,
Hammad.
Sp I wanted to you whats the best use case to transfer this amount of data through Dataflow?
But I am also seeing on the internet that its timeout is 8hrs. @v-mdharahman Can you please help on this?
Thanks
Hi @avisri,
Yes right the execution time limit is up to 8 hours, depending on factors like workload, SKU, and internal resource management. However, in real-world usage, especially in production environments (including with high SKUs like F64 or F128), many users have observed a hard stop at exactly 4 hours like in your case.
Also the best use case to transfer this amount of data is by spliting the load. Likje you can break the source table into smaller partitions ( like by date range, primary key buckets, or any natural partitioning column). Then create multiple parameterized dataflows or loops using pipeline to load each partition individually and avoid hitting the execution time limit.
You can also use Data Pipelines with Copy Activity. Since you're only copying data without transformations, using Copy Activity in Fabric Data Pipelines from Databricks to Lakehouse can be more efficient and does not have execution time restriction. It will also handles larger datasets more gracefully.
If I misunderstand your needs or you still have problems on it, please feel free to let us know.
Best Regards,
Hammad.
In addition to this, I have divided the data according to the year, and I am using 5 parallel Dataflows now and targeting them to one Final table in Lakehouse. While trying this, some of my Dataflows failed, and I am getting a Dataflow Gen2 error.
semantic_table: There was a problem refreshing the dataflow: 'Couldn't refresh the entity because of an issue with the mashup document MashupException.Error: Error in comitting version., Underlying error: conflicting metadata change Details: Reason = DataSource.Error;ErrorCode = Lakehouse036;Message = conflicting metadata change;Message.Format = conflicting metadata change;Microsoft.Data.Mashup.Error.Context = User GatewayObjectId: e58a7629-e2e0-412b-8688-6f0ebc4f1e50'. Error code: 104100. (Request ID: 3cbf7621-0a97-481e-be25-97ad4ab717f6).
Can you please let me know about this part? And what will be the solution for this?
Thanks!
Hi @avisri,
Thanks for sharing the update. The error you are seeing "Lakehouse036: conflicting metadata change" is happening because multiple Dataflow Gen2 jobs are writing to the same Lakehouse table in parallel. When this happens, each dataflow tries to commit changes to the table metadata, which leads to conflicts, and some of them fail.
To resolve your issue try avoiding parallel writes to the same table. Instead of 5 Dataflows writing into the same final table at the same time, run them sequentially. This will ensure there are no metadata conflicts when committing changes.
Or you can use staging tables. Point each yearly Dataflow to its own staging table in the Lakehouse (e.g., Table_2019, Table_2020, etc.). After that, you can either use a notebook or Pipeline activity to merge these staging tables into your final table, or create a lakehouse shortcut or union query on top of them.
If you prefer keeping the Dataflows, orchestrate them using a Fabric Data Pipeline to run one after another (or write to staging first, then merge).
This approach removes the metadata conflict and will make your loads more stable.
Best Regards,
Hammad.
Hi @avisri,
As we havenโt heard back from you, so just following up to our previous message. I'd like to confirm if you've successfully resolved this issue or if you need further help.
If yes, you are welcome to share your workaround so that other users can benefit as well. And if you're still looking for guidance, feel free to give us an update, weโre here for you.
Best Regards,
Hammad.