TrustRadius: an HG Insights company

Azure Data Factory Reviews and Ratings

Rating: 8.2 out of 10
Score
8.2 out of 10

Community insights

TrustRadius Insights for Azure Data Factory are summaries of user sentiment data from TrustRadius reviews and, when necessary, third party data sources.

Pros

Connector Variety: Users have expressed satisfaction with the extensive range of connectors supported by Azure Data Factory, facilitating smooth data integration across various sources within and beyond the Microsoft ecosystem.

Visual Transformation Design: The feature enabling visual data transformation design through Mapping Data Flows in Azure Data Factory is highly prized by users for simplifying intricate ETL and ELT processes without requiring coding.

Monitoring Dashboard: The unified monitoring dashboard in Azure Data Factory is valued by users for offering a holistic view of pipeline activities, streamlining job status tracking, failure identification, and bottleneck location. This comprehensive oversight enhances efficiency in managing data workflows.

Reviews

10 Reviews

Azure Data Factory an Universal pipe

Rating: 6 out of 10
Incentivized

Use Cases and Deployment Scope

We live in a world where half of the data for analytics come from SAP and half from non SAP sources. We use Azure Data Factory to load non SAP data from different source systems into Azure lake house. The project follows medallion architecture where Azure Data Factory takes data from multiple sources and stores them in the bronze layer of the medallion architecture. Since our SAP datasphere has limitations connecting to non SAP sources as good and native like Azure Data Factory, we use Azure Data Factory for these scenarios. Further modelling of data in the next layers (silver layer and gold layer) is done using Azure Data Bricks, where the final data product is created. The Azure Data Factory also helps in applying transformations which loading the data from different source systems. Datasphere often relies on ODBC/JDBC/OData connectivity, whereas Azure Data Factory provides maintenance-free connectors for our web applications, like partner portal, cloud applications like one crm, on-prem Oracle systems, and also to NoSQL dbs like MongoDB. To summarize Azure Data Factory is used in our organisation to ingest non SAP data from different sources into our Bronze layer for the Databricks to further clean and curate the data for data product creation. Without Azure data factory connecting the data from different source wouldnt been possible because SAP Datasphere has limitations when it comes to connection with non SAP source systems

Pros

  • Connectivity with other cloud environment like Salesforce
  • Connectivity with non structured data and big data systems
  • Reduces data islands
  • Azure Data Factory handles perfectly the huge volume of data in JSON format from our global apps and services

Cons

  • The error details where there is an error while processing the files is not clear
  • Connectivity with s4 system is not so good compared to Datasphere
  • Since Azure Data Factory just transfers data it lacks the capacity to identify the wrongness in the data. It is just a dumb data transfer tool from point A to B

Likelihood to Recommend

Best scenario is for ETL process. The flexibility and connectivity is outstanding. For our environment, SAP data connectivity with Azure Data Factory offers very limited features compared to SAP Data Sphere. Due to the limited modelling capacity of the tool, we use Databricks for data modelling and cleaning. Usage of multiple tools could have been avoided if adf has modelling capabilities.

Azure Data Factory - data integration tool to build your Cloud Data Platform.

Rating: 9 out of 10
Incentivized

Use Cases and Deployment Scope

Azure Data Factory is a data integration technology used to integrate data to the cloud, especially. Azure cloud. Today, any organization wants to integrate its data into a cloud data platform. This cloud platform serves data for various purposes, such as sharing it with downstream applications, developing analytics, and even building AI applications on it. Azure Data Factory integrates data from various sources, such as on-prem applications, databases, and files, into this data platform.

Pros

  • Data Ingestion - it works very well with numerous data sources.
  • Data pipeline orchestration: It is a generic, popular tool for orchestrating data pipelines.
  • Works well in Azure ecosystem, Azure services and data platforms like Databricks.
  • It is a serverless and scalable solution for cloud data integration.

Cons

  • Data transformation has provided a data flow. But it is not ideal for complex data transformation.
  • Cost of Data Factory depends on number of pipelines and transformations used.
  • Azure Data Factory is efficient and good for parallel data pipeline runs. But not ideal for a large volume of data.

Likelihood to Recommend

Azure Data Factory is a great data integration tool for developing a cloud data platform, especially within the Azure ecosystem. Azure Data Factory is very good for the Data Ingestion part. It can work for simple data transformation with its Data Flow, but it will also need cluster configuration, and there is some cost. Also, it is an excellent tool for orchestrating data pipelines. But for complex data transformations, you may need to use technologies like Databricks and PySpark.

Overall helpful product that works as advertised.

Rating: 7 out of 10
Incentivized

Use Cases and Deployment Scope

Using SHIR to pull records from on-premise databases and storing in ADLS storage. From ADLS storage, bringing data into databricks for analytics use. Roughly 50 different pipelines in each environment, with 3 separate environments. Code is stored and deployed from Azure Dev ops. Alerting is handled via LogicMonitor and Azure Functions.

Pros

  • Step by step processes.
  • Storing infrastructure as code.
  • Alerting on job failures.
  • SHIR

Cons

  • Learning curve for pipeline creation interface.
  • Alerting isn't necessarily built in. Had to work around this to meet team needs.
  • With GIT enabled, some features can only be done via git, while some need to be done via the portal.

Likelihood to Recommend

Pulling data from Databases in a scheduled and controlled pattern. We heavily use the self hosted SHIR option to pull from existing sources. In addition, we have several API's directly from Azure Data Factory using keyvault for passphrase storage. Utilizing GIT in Devops helps promote changes within our multiple environments in a timely matter.

Vetted Review
Azure Data Factory
3 years of experience

The go-to ETL tool for most situations

Rating: 9 out of 10
Incentivized

Use Cases and Deployment Scope

Data Integration: We harness Azure Data Factory's capabilities to move data from various sources – both on-premises databases and cloud storage – into our Azure data storage solutions like Azure SQL Database, Azure Blob Storage, and Azure Data Lake Store. This ensures all our data, regardless of its origin, is consolidated in one place.

Transformations: Azure Data Factory's data flow transformations help us clean, transform, and enrich our data before loading it to the destination. This is crucial for maintaining data quality, especially when dealing with diverse datasets.

Pros

  • Azure Data Factory supports a vast array of source and destination connectors, both from within the Microsoft ecosystem (like Azure Blob Storage, Azure SQL Database, Azure Cosmos DB) and external platforms (like Amazon S3, Google Cloud Storage, SAP, Salesforce, and many more).
  • Azure Data Factory's Mapping Data Flows provides a code-free environment to design data transformations visually. Users can drag and drop elements to create complex ETL (Extract, Transform, Load) and ELT (Extract, Load, Transform) processes without needing to write any code.
  • Azure Data Factory provides a unified monitoring dashboard that offers a holistic view of all pipeline activities. I think this makes it easier for users to track the status of various jobs, identify failures, and pinpoint bottlenecks.

Cons

  • Granularity of Errors: Sometimes, Azure Data Factory provides error messages that are too generic or vague for us, making it challenging to pinpoint the exact cause of a pipeline failure. Enhanced error messages with more actionable details would greatly assist us as users in debugging their pipelines.
  • Pipeline Design UI: In my experience, the visual interface for designing pipelines, especially when dealing with complex workflows or numerous activities, can become cluttered. I think a more intuitive and scalable design interface would improve usability. In my opinion, features like zoom, better alignment tools, or grouping capabilities could make managing intricate designs more manageable.
  • Native Support: While Azure Data Factory does support incremental data loads, in my experience, the setup can be somewhat manual and complex. I think native and more straightforward support for Change Data Capture, especially from popular databases, would simplify the process of capturing and processing only the changed data, making regular data updates more efficient

Likelihood to Recommend

Well-suited Scenarios for Azure Data Factory (ADF):

When an organization has data sources spread across on-premises databases and cloud storage solutions, I think Azure Data Factory is excellent for integrating these sources.

Azure Data Factory's integration with Azure Databricks allows it to handle large-scale data transformations effectively, leveraging the power of distributed processing.

For regular ETL or ELT processes that need to run at specific intervals (daily, weekly, etc.), I think Azure Data Factory's scheduling capabilities are very handy.

Less Appropriate Scenarios for Azure Data Factory:

Real-time Data Streaming - Azure Data Factory is primarily batch-oriented.

Simple Data Copy Tasks - For straightforward data copy tasks without the need for transformation or complex workflows, in my opinion, using Azure Data Factory might be overkill; simpler tools or scripts could suffice.

Advanced Data Science Workflows: While Azure Data Factory can handle data prep and transformation, in my experience, it's not designed for in-depth data science tasks. I think for advanced analytics, machine learning, or statistical modeling, integration with specialized tools would be necessary.

Vetted Review
Azure Data Factory
3 years of experience

One of the best and reliable ETL & ELT platforms for pulling data from multiple sources

Rating: 10 out of 10
Incentivized

Use Cases and Deployment Scope

One of the best Data Integration tools for both ETL and ELT. I have been using ADF for the last 6+ years and it helped me in extracting several data feeds within our organization that meets our specific business needs. The tool provides many features such as Move and Transform, Data explorer, Azure Functions, Data bricks, Data Lake Analytics, Blob Storage, Linked services, Machine Learning, and Power Query.

Pros

  • It allows copying data from various types of data sources like on-premise files, Azure Database, Excel, JSON, Azure Synapse, API, etc. to the desired destination.
  • We can use linked service in multiple pipeline/data load.
  • It also allows the running of SSIS & SSMS packages which makes it an easy-to-use ETL & ELT tool.

Cons

  • For complex JSON when it comes to mapping nested attribute it's not easy to flatten out
  • Data Factory V1 does not have a good implementation experience as compared to V2
  • Work with on premise solutions sometimes is not too friendly because you will need to set a VPN

Likelihood to Recommend

In a data pipeline, you will be able to add different kinds of activities for example connect from your on-premise SFTP and move CSV files to storage accounts. As well data factory has its own data flow if you are an ETL developer who experimented with maybe you have worked with SSIS, thus, you will start quickly with this new feature of the data factory.

Azure Databricks

Rating: 9 out of 10
Incentivized

Use Cases and Deployment Scope

Orchestration platform for the Databricks notebooks. Have used an ETL for loading csv files into SQL server based database.

Pros

  • Orchestration engine
  • Low code Data pipeline
  • Logic apps integration

Cons

  • Error Flagging, Details of the error code is not specific especially faced this during Azure Table load
  • Missing feature of Data exploration functionality similar to Synapse Data explorer
  • missing access to orchestrate/create stream analytics job

Likelihood to Recommend

Well Suited:

<ul><li>Offers low code/no code features executes against spark pool.</li><li>Batch processing features, Tight coupling with Databricks &amp; ETL jobs. </li><li>Offers Logic apps &amp; Azure functions invoking API. </li></ul>Less Appropriate:

<ul><li>Not much inherent features of Stream analytics (Liasing Azure Stream analytics to DF might be good option).</li><li>Advanced load options viz . Upsert type operations missing. </li></ul>

Simple Solution to Data Migration

Rating: 8 out of 10
Incentivized

Use Cases and Deployment Scope

We use Azure Data Factory to orchestrate ETL and ELT pipelines in our projects and it has a wide range of data connectors right from classic(FTP) and modern data lake storage like ADLS and AWS S3. Using the Dataflow and SSIS integration runtime server, it can perform complex transformations without the need for another tool. Executing and Monitoring ETL loads the data factory is very simple and user-friendly.

Pros

  • Orchestration
  • On premises support
  • Support to vast no of data connectors
  • Cloud migration

Cons

  • Native transform functions missing.
  • Pricing
  • Limited trigger functions.

Likelihood to Recommend

Azure Data Factory can perform better with Azure services and can easily do cloud migrations from on-premises services like SQL Server. It has a limited set of functionalities to transform data using SSIS integration services and data flow.

Vetted Review
Azure Data Factory
5 years of experience

Database management and ETL tool for big data that is smart and reliable

Rating: 9 out of 10
Incentivized

Use Cases and Deployment Scope

Get full data integration at scale with Microsoft Azure Data Factory's management service for serverless data integration,‌ hybrid data integration is easily and agilely possible through this software. We are creating ETL and ELT workflows as well as orchestrating and monitoring pipelines without writing any code. Full management and serverless integration with default features installed on the system and various connectors reduce costs. It is used in all company departments and project management units of our customers. Since ADF adopts an intelligent intent-driven mapping methodology, it enables copy activities to be automated.

Pros

  • Creating ETL and ELT workflows as well as orchestrating and monitoring pipelines without writing any code.
  • Hybrid data integration is easily and agilely possible through this software.
  • It has lot of various useful components

Cons

  • It should integrate more ETL and audit functionality.
  • Pipelines lack flexibility because moving Data Factory pipelines between different environments, such as for development or testing, require increased security and flexibility.
  • The number of pre-defined templates is small and they should have more variety.

Likelihood to Recommend

If you know a bit about database management everything is pretty easy, based on my personal experience. You can build a lot of things entirely in design, even if you do not know all the syntax, [since] having ready-made templates simplifies everything to get started and build the first pipeline.

ADF is awesome!

Rating: 10 out of 10
Incentivized

Use Cases and Deployment Scope

I've used it to perform PoC's and work with data transformation processes that interact with other applications or tools.

Pros

  • Cloud-based
  • Fast
  • Reliable

Cons

  • Some features exist on the UI but are not implemented
  • Its always changing

Likelihood to Recommend

It works better than other tools from the same range, it has a beautiful UI and it makes work easy. Its also very easy to integrate with other tools, tools, apps and ecosystems.

Azure Data Factory - Don't Abandon SSIS Just Yet

Rating: 7 out of 10
Incentivized

Use Cases and Deployment Scope

Azure Data Factory, particularly V2, offers a good option as a cloud-based ETL tool if you are leveraging the Azure cloud. We are using it as we begin to hybridize our on-prem data warehouse and applications with Azure. Up until now, we have leveraged SSIS for these purposes, but are beginning to migrate ETL and other data movement functions to the cloud, with Azure Data Factory as the primary utility.

Pros

  • Easy to set up and get started.
  • Runtimes make integration with on-prem data simple and also allow for support of existing investments in SSIS.

Cons

  • Limited source/sink (target) connectors depending on which area of Azure Data Factory you are using.
  • Does not yet have parity with SSIS as far as the transforms available.

Likelihood to Recommend

If you are just getting started and all your data is resident in the Azure cloud, then Azure Data Factory is likely to work fine without having to jump through too many hoops. However, in a hybrid environment (which is most of them these days), ADF will likely need a leg up. It works well for scheduling and basic scheduling/orchestration tasks, but the feature set is not at a level with SSIS (which has been around for 15 years so...). As ADF now supports deploying SSIS, it is also a good candidate if large amounts of your data are resident in the Azure cloud and you have an existing SSIS investment in code and licensing. We are using it in a hybrid fashion for the data warehouse and will slowly transition over to ADF as the feature set improves. We are also using it for cloud-native applications that only require supplemental data from on-prem resources.

Vetted Review
Azure Data Factory
1 year of experience