After executing my mapping dataflow I would like to run a clean-up script using the Post SQL Script option of the Sink activity. However I'm not having much su
I am getting below error while running my dataflow. This dataflow was running fine till yesterday. From today onwards I am getting error like this Operation on
I am hitting a query through lookup activity on DB2 Database using ADF. Query is: But when I execute the query the '\' count gets double. The output is provide
I get following result from my script task, How should I build expression to check if the result contains 'exceed' { "resultSetCount": 1, "recordsAffected": 0,
I have to pass multiple records from a SQL source to the body of a web call in Data Factory in JSON format. eg. {"EmployeeCode":"1234",""FirstName":"Joe","LastN
My pipeline receives the Path and the name of a JSON file. This is part of the flow that I created: The lookup step is used to read the JSON File and later I n
Have had a look around but can't see any concrete information. Essentially, if anyone could help it would be great. We are building reporting in the cloud and l
I have a copy activity in Datafactory that dynamically maps the columns between files in tables A and B. Both tables, A and B are .parquet. Table A has 8 column
I currently have a Data Pipeline within ADF that pulls data from multiple REST APIs, transforms the data and stores it in an Azure SQL Database and from there i
I have a question, hopefully someone in the forum could give some help here. I am able to pull data from Soap API call to SQL Server table (xml data type field
I have set a storage event trigger for the pipelines in the azure data factory Dev environment. the blob path begins with is different in the dev environment an
I have a an api that contains some data and another api url named "nextapi" I want to loop through each api under api and store the data of each api page to azu
I am deploying ADF linked templates from ADLS through GitHub Actions and pipeline fails withy generic Python error: ERROR: 'str' object has no attribute 'get' I
I'm having issues using parameter within a storage event trigger for the typeProperty "blobPathBeginsWith". Per default when using a storage event trigger, the
I am new to data flows in adf. I have a set of json files in a folder which i would like to parse and flatten two arrays into a csv. The json structure is as fo
I have a requirement to POST json data (stored as json files in ADLS) to REST API endpoint. I have to use Azure Data factory for this. I have gone through Micr
I have a strange source CSV file where it contains a trailing column delimiter at the end of each record just before the carriage return/new line. When ADF is
I am trying to import Google Analytics data into Azure Blob or Data Lake storage for analysis or reporting. But I don't see a Google Analytics connector in Azur
I am trying to do a lookup from a sql server table. I am able to connect to the sql server instance via MSSQL management studio, but in ADF I get the below erro
I have a pipeline that copies data from source (dynamics) to SQL server datawarehouse. There is a ForEach activity which iterates over the list of all the table