web
You’re offline. This is a read only version of the page.
close
Skip to main content

Notifications

Announcements

No record found.

Community site session details

Community site session details

Session Id :
Finance | Project Operations, Human Resources, ...
Suggested Answer

PowerBI refreshers are failing since the sourcefile in datalake is getting updated the sametime

(0) ShareShare
ReportReport
Posted on by 37

Hi Everyone,

To build Power BI reports for F&O, we are exporting the F&O data to data lake and scheduled Power BI refreshes to pull the data from Data lake to Power BI. The challenge we are seeing is, refreshes are failing sometimes because the files in the datalake are getting refreshed at the same time.

I want to understand how other projects are handling this. We see there is no way to control when the data should get refreshed to the datalake. If we can say datalake refresh to happen at 11pm, we will be able to schedule the Power BI refresh at 3am so that we don't get into this problem.

Appreciate your thoughts.

Regards,
Kumar

I have the same question (0)
  • Suggested answer
    huijij Profile Picture
    19,811 on at

    Hi Kumar,

    It is recommended that you post your question to the Power BI forum for more professional help.

    https://community.powerbi.com/.

    You can troubleshoot by referring to the following link, Checking refresh status and history in a workspace.

    learn.microsoft.com/.../refresh-data

    learn.microsoft.com/.../refresh-troubleshooting-refresh-scenarios

  • Suggested answer
    Shawnsauve Profile Picture
    1,093 on at

    There are a few possible solutions to your challenge of failed refreshes due to simultaneous updates to the data lake.

    1. Implement a staggered refresh schedule: Instead of scheduling both the data lake and Power BI refreshes at the same time, you could stagger them to occur at different times. For example, you could schedule the data lake refresh to happen at 11pm and the Power BI refresh to happen at 3am, as you suggested. This way, there will be a time gap between the two refreshes, which should reduce the chances of any conflicts.

    2. Use version control: Another approach is to use version control for your data in the data lake. This involves creating a new version of the data each time it is updated, rather than overwriting the existing data. This way, you can ensure that the Power BI reports are always using a consistent version of the data, regardless of when the refreshes happen. You can use tools like Azure Data Factory to implement version control for your data in the data lake.

    3. Implement a locking mechanism: A locking mechanism can be used to prevent simultaneous updates to the data lake. This involves locking the file or folder that is being updated during the refresh, which prevents other processes from accessing it until the refresh is complete. This can be done using tools like Azure Data Lake Storage or Azure Blob Storage.

    4. Optimize your refreshes: It's also important to optimize your refreshes to minimize the amount of time they take. This can involve using incremental refreshes, which only refresh the data that has changed since the last refresh, rather than refreshing the entire dataset. You can also use compression and partitioning techniques to reduce the amount of data that needs to be refreshed.

    Overall, it's important to have a robust data governance strategy in place to ensure the reliability and consistency of your data lake and Power BI reports. This may involve a combination of the above solutions, as well as regular monitoring and maintenance of your data infrastructure

  • bbb0777 Profile Picture
    158 on at
    @SantoshK
     
    I think same issue here, with Dynamics F&O.  Did you find any solution(s)?  

    ------------------

    Our situation: 
    • F&O Setup with Export to (Synapse) Datalake.  i.e. it's constantly updating the files in the Datalake
    • Synapse Serverless DB, with views that hit those CSV files in the datalake
     
    I think the issue is the updating files as well, because
    • Last at night updates: 100% go through fine
    • Evening updates: 90% go through
    • Middle of day: 50% go through - i.e. when those CSV files are getting the most updates
    • Every failure is the same error, but tends to reference a different high-activity CSV file
     

    Most (I think) relevant bits of the error message.  The first part is always consistent, the second part tends to reference a different CSV each time. 
    'Unexpected end-of-input within record... 'Tables/General/Miscellaneous/PMIPBillingScheduleTable/PMIPBILLINGSCHEDULETABLE_00001.csv'.

Under review

Thank you for your reply! To ensure a great experience for everyone, your content is awaiting approval by our Community Managers. Please check back later.

Helpful resources

Quick Links

Responsible AI policies

As AI tools become more common, we’re introducing a Responsible AI Use…

Neeraj Kumar – Community Spotlight

We are honored to recognize Neeraj Kumar as our Community Spotlight honoree for…

Leaderboard > Finance | Project Operations, Human Resources, AX, GP, SL

#1
Martin Dráb Profile Picture

Martin Dráb 664 Most Valuable Professional

#2
André Arnaud de Calavon Profile Picture

André Arnaud de Cal... 522 Super User 2025 Season 2

#3
Sohaib Cheema Profile Picture

Sohaib Cheema 303 User Group Leader

Last 30 days Overall leaderboard

Product updates

Dynamics 365 release plans