We are in the process of upgrading our ARM Templates to Bicep, one of which is for storage. In the initially ported Bicep file, everything worked fine and then
im trying to add new AD Object ID in ACL using ace block for an already existing subfolders inside the azure datalake container using terraform code and im gett
I have an ETL pipeline that gets triggers when file is uploaded into Datalake location "Processing" folder, Once data is processed I want to move the file to d
I have a copy pipeline set up that connects to an SFTP server (on Azure Synapse). I have used it to copy csv files and this works fine, but now I have the compl
I have a databricks notebook that is writing a dataframe to a file in ADLS Gen2 storage. It creates a temp folder, outputs the file and then copies that file to
I'm using Power BI Dataflows to access spreadsheets I have in blob storage. I have configured IAM permissions on the storage account for myself and the Power BI
Currently, I am generating SAS token from Azure file share and trying to access the file. But when I hit that url on browser, file automatically getting downloa
We have an Azure Data Lake Gen 2 which contains 100's of thousands of JSON messages that come in on a continuous basis. These files are stored in a folder struc
I have created an external table in Azure Synapse from a parquet file stored in an ADLS Gen2 container. I have used the following three queries to create the da
I am trying to find some integration to use iceberg table format on adls /azure data lake to perform crud operations. Is it possible to not use any other comput