Skip to main content
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Calling all Data Engineers! Fabric Data Engineer (Exam DP-700) live sessions are back! Starting October 16th. Sign up.

Reply
avisri
Helper I
Helper I

Dataflow Gen2 error in Fabric

avisri_0-1754566190695.png

avisri_0-1754566646793.png

Hi all, my Fabric Dataflow gen2 failed after 4 hrs, and I am getting this error message. Is there some time limit, and I'm not sure. Please help with this part.

Source- Databricks

Destination- Fabric Lakehouse

Medium- Dataflow Gen2 (No transformation, only full load)

1 Table - 120M+ records, 160+ columns, Size~ 18GB

Fabric SKU- F128

 

Thanks!

 

 

1 ACCEPTED SOLUTION

Hi @avisri,

Thanks for sharing the update. The error you are seeing "Lakehouse036: conflicting metadata change" is happening because multiple Dataflow Gen2 jobs are writing to the same Lakehouse table in parallel. When this happens, each dataflow tries to commit changes to the table metadata, which leads to conflicts, and some of them fail.

To resolve your issue try avoiding parallel writes to the same table. Instead of 5 Dataflows writing into the same final table at the same time, run them sequentially. This will ensure there are no metadata conflicts when committing changes.

Or you can use staging tables. Point each yearly Dataflow to its own staging table in the Lakehouse (e.g., Table_2019, Table_2020, etc.). After that, you can either use a notebook or Pipeline activity to merge these staging tables into your final table, or create a lakehouse shortcut or union query on top of them.

If you prefer keeping the Dataflows, orchestrate them using a Fabric Data Pipeline to run one after another (or write to staging first, then merge).

This approach removes the metadata conflict and will make your loads more stable.

 

Best Regards,

Hammad.

View solution in original post

8 REPLIES 8
v-mdharahman
Community Support
Community Support

Hi @avisri,

Thanks for reaching out to the Microsoft fabric community forum and yes, Fabric Dataflow Gen2 currently has a affective execution time limit of 4 hours per run, even with higher SKUs like F128. Since your job ran for exactly 4:00:43 and then failed, itโ€™s very likely that it hit this hard timeout limit.

Given your scenario (18GB, 120M+ rows, 160+ columns, full load), even without transformations, this volume can push Dataflow Gen2 close to or beyond the timeout window, especially when moving from external sources like Databricks.

 

If I misunderstand your needs or you still have problems on it, please feel free to let us know.  

Best Regards,
Hammad.

avisri
Helper I
Helper I

Sp I wanted to you whats the best use case to transfer this amount of data through Dataflow? 

avisri
Helper I
Helper I

But I am also seeing on the internet that its timeout is 8hrs. @v-mdharahman Can you please help on this?

Thanks

 

 

 

 

 

 

Hi @avisri,

Yes right the execution time limit is up to 8 hours, depending on factors like workload, SKU, and internal resource management. However, in real-world usage, especially in production environments (including with high SKUs like F64 or F128), many users have observed a hard stop at exactly 4 hours like in your case.

Also the best use case to transfer this amount of data is by spliting the load. Likje you can break the source table into smaller partitions ( like by date range, primary key buckets, or any natural partitioning column). Then create multiple parameterized dataflows or loops using pipeline to load each partition individually and avoid hitting the execution time limit.

You can also use Data Pipelines with Copy Activity. Since you're only copying data without transformations, using Copy Activity in Fabric Data Pipelines from Databricks to Lakehouse can be more efficient and does not have execution time restriction. It will also handles larger datasets more gracefully.

 

If I misunderstand your needs or you still have problems on it, please feel free to let us know.  

Best Regards,
Hammad.

 

avisri
Helper I
Helper I

In addition to this, I have divided the data according to the year, and I am using 5 parallel Dataflows now and targeting them to one Final table in Lakehouse. While trying this, some of my Dataflows failed, and I am getting a Dataflow Gen2 error.


semantic_table: There was a problem refreshing the dataflow: 'Couldn't refresh the entity because of an issue with the mashup document MashupException.Error: Error in comitting version., Underlying error: conflicting metadata change Details: Reason = DataSource.Error;ErrorCode = Lakehouse036;Message = conflicting metadata change;Message.Format = conflicting metadata change;Microsoft.Data.Mashup.Error.Context = User GatewayObjectId: e58a7629-e2e0-412b-8688-6f0ebc4f1e50'. Error code: 104100. (Request ID: 3cbf7621-0a97-481e-be25-97ad4ab717f6).

 

Can you please let me know about this part? And what will be the solution for this?
Thanks!

Hi @avisri,

Thanks for sharing the update. The error you are seeing "Lakehouse036: conflicting metadata change" is happening because multiple Dataflow Gen2 jobs are writing to the same Lakehouse table in parallel. When this happens, each dataflow tries to commit changes to the table metadata, which leads to conflicts, and some of them fail.

To resolve your issue try avoiding parallel writes to the same table. Instead of 5 Dataflows writing into the same final table at the same time, run them sequentially. This will ensure there are no metadata conflicts when committing changes.

Or you can use staging tables. Point each yearly Dataflow to its own staging table in the Lakehouse (e.g., Table_2019, Table_2020, etc.). After that, you can either use a notebook or Pipeline activity to merge these staging tables into your final table, or create a lakehouse shortcut or union query on top of them.

If you prefer keeping the Dataflows, orchestrate them using a Fabric Data Pipeline to run one after another (or write to staging first, then merge).

This approach removes the metadata conflict and will make your loads more stable.

 

Best Regards,

Hammad.

Hi @avisri,

As we havenโ€™t heard back from you, so just following up to our previous message. I'd like to confirm if you've successfully resolved this issue or if you need further help.

If yes, you are welcome to share your workaround so that other users can benefit as well.  And if you're still looking for guidance, feel free to give us an update, weโ€™re here for you.

 

Best Regards,

Hammad.

Hi @v-mdharahman, Thanks for helping. The issue is resolved

Helpful resources

Announcements
November Fabric Update Carousel

Fabric Monthly Update - November 2025

Check out the November 2025 Fabric update to learn about new features.

Fabric Data Days Carousel

Fabric Data Days

Advance your Data & AI career with 50 days of live learning, contests, hands-on challenges, study groups & certifications and more!

FabCon Atlanta 2026 carousel

FabCon Atlanta 2026

Join us at FabCon Atlanta, March 16-20, for the ultimate Fabric, Power BI, AI and SQL community-led event. Save $200 with code FABCOMM.

Users online (27)