How to troubleshooting Bring Your Own Storage (BYOS) issues on App Service Linux

9 minute read | By Anthony Salemo

This post will cover troubleshooting issues that may arise when using Bring Your Own Storage with App Service Linux for both “Blessed” images and Web Apps for Containers. This is for scenarios that render the application and Kudu container unavailable.


You can mount volumes to your container on App Service Linux or Web Apps for Containers, which map to a specified Azure Storage Account - and either use Azure Blob (read-only) or Azure File Shares (read/write) - which is named “Bring Your Own Storage”.

We’ll go over some of the more common scenarios that may arise when an issue occurs.

What kind of issues will I see?

If a volume from BYOS is mounted to your container and fails for any of the reasons below - this may cause your application to not show any logging - this will also show the typical “Application Error :(“ page. This wil show both for the application container and the Kudu container.

The reason for the lack of application specific logging is that the volume is created and mounted very early on in the container lifecycle, prior to any point of where application logic is able to be invoked. Since it fails before any of this, which normally is what would generate stdout/stderr - there will likely be no logging.

The container will be unable to start until either the problem with BYOS is corrected, or, you remove the BYOS mount. You can change or remove a mount through the Azure Portal under Configuration -> Path Mappings or using something such as the Azure CLI.


Although no application logs are generated - you can still use the Diagnose and Solve Problems blade to make use of the Container Issues detector to determine if there is an issue. If so, you will see some data like the below:

BYOS failures

In a scenario where the volume is able to be mounted, you can review this from the application or Kudu container by running df -h. This will show the Storage Account and File Share or Blob (mounted with blobfuse) and then which path this is mounted to under the “Mounted on” column of the output.

BYOS output


Incorrect or rotated Access Keys

Incorrect keys:

When adding a Storage Mount, if using the “Basic” option, this will automatically handle adding in the Primary Access Key for you.

If you’re using the “Advanced” option, you will need to manually copy and paste this key in.

If the key in incorrectly copied or mistyped - this will fail authentication for adding the mount. Thus, causing the container to crash.

Keys that are rotated/refreshed:

If you rotate an Access Key on the Storage Account side - the application will only pick this change up where the container is needed to be restarted, such as:

  • Restarting this specific application manually
  • Other restart reasons such as deployment or site configuration updates
  • Instance movement, such as scaling or typical PaaS instance movements

By default, the client (the App Service) has no self-awareness when a key is rotated on the Storage side other than after a restart.

Therefor when the volume is attempted to be remounted (after the restart), this issue can occur since the key used on the Storage side will now differ from the key being used on the application side.

In these scenarios, double check that the “Advanced” configuration blade has the correct Access Key and update it as needed.

Deleted Storage Account or Blob Container/Azure File Share

In the same vein as the Keys that are rotated/refreshed section, if either the Storage Account, blob, or file share within this is deleted - the container will fail to start after a restart - since it will now try to mount a non-existent file share.

However, read/write operations to the blob or share may fail even if the application isn’t restarted - but while the share or blob is deleted still. This can be tested by going to an SSH session on the application or Kudu container and running df -h after deleting the blob or share (but not restarting yet), you will notice this will disappear from available volumes shortly after.

If the Blob or File Share is deleted - you can try to recreate these with the same name. A recreated Storage Account will need to use a different Access Key than the original.

Networking and DNS

If the Storage Account cannot be accessed, then mounting the volume will fail.

Consider the following if your application is in a VNET and this is a possibility in case the application is down:

  • Was there any networking changes on the Storage Account Side?
    • Networking changes, such as changing traffic flow or any reconfiguration that may affect DNS resolution may not show up as a change on the application side.
  • Storage firewall is supported only through service endpoints and private endpoints (when VNET integration is used).
    • If the app and Azure Storage account are in same Azure region, and if you grant access from App Service IP addresses in the Azure Storage firewall configuration, then these IP restrictions are not honored.
  • Is custom DNS being used? Unresolvable DNS for the Storage Account will result in failure to mount the volume
  • When VNET integration is used, ensure the following ports are open:
    • Azure Files: 80 and 445.
    • Azure Blobs: 80 and 443.

If applicable, a relatively easy test for this is to disconnect the application from the VNET and open the Storage Account to public traffic. If this succeeds, then this can be a point of focus.

You can check connectivity and resolution with commands like the following - the below is using a Storage Account with an Azure File Share as an example:

  • nslookup
  • dig
  • tcpping

NOTE: If you enable a Storage Firewall after succesfully mounting storage - and try to tcpping the Storage Account, you may be unable to reach it.

If you try to list the contents while still in the mapped directory, you may recieve the following message:

root@8576066a2932:/home/site/files# ls -ltra
ls: cannot open directory '.': Host is down

Mount Paths

/home or / cannot be mounted to (as well as from the az cli).

Mounting to /home or /, if it was possible, would likely cause the container to not start, as this directory and all subdirectories is already mounted via App Service built-in storage.

Mounting to /tmp or directories under /tmp is not recommended as this has the potential to cause the container to not start.

File Locking

Depending on what the application is doing, file locking for certain services running over a volume may cause the application to crash (or not start) - depending on the application logic.

Below are some examples:


Trying to run MongoDB over mounted storage:

  • Example error: connection: /data/db/: handle-open: open: Operation not permitted
    • MongoDB has limitations surrounding trying to mount a volume that is mapped to a Windows File Share (which in this case, is what Azure Files is)
    • There may be other limitations with Windows File Shares, MongoDB, and non-root users.
    • There may be lower level API’s that are not supported with MongoDB when mapped to a Windows File Share.


  • Use a managed MongoDB instance like Mongo Atlas or the MongoDB API For CosmosDB.


Trying to run PostgreSQL over mounted storage:

  • Example error: data directory "/var/lib/postgresql/data" has wrong ownership:
    • PostgreSQL may have limitations with being used over a Windows File Share mounted to the container due to the user being different in regards to the volume (and data ownership) - reference article - pgdata has wrong ownership - forums.docker


  • Use a managed PostgreSQL instance like Azure Database for PostgreSQL


  • Trying to use SQLite over mounted storage:
    • As called out in Best Practices - It’s not recommended to use storage mounts for local databases (such as SQLite) or for any other applications and components that rely on file handles and locks.
    • SQLite is not recommended for production scenarios. Using a managed database offering should be used instead.
    • Since SQLite is a file-based database implementation, there are various reports throughout the years of SQLite file locking issues, including over volume mounts.

Best practices and limitations

For general best practices and limiaations not within the scope of this post, review the below: