data replication Articles / Blogs / Perficient https://blogs.perficient.com/tag/data-replication/ Expert Digital Insights Tue, 25 Feb 2025 15:38:08 +0000 en-US hourly 1 https://blogs.perficient.com/files/favicon-194x194-1-150x150.png data replication Articles / Blogs / Perficient https://blogs.perficient.com/tag/data-replication/ 32 32 30508587 Accelerate the Replication of Oracle Fusion Cloud Apps Data into Databricks https://blogs.perficient.com/2025/02/25/databricks-accelerator-for-oracle-fusion-applications/ https://blogs.perficient.com/2025/02/25/databricks-accelerator-for-oracle-fusion-applications/#respond Tue, 25 Feb 2025 15:34:08 +0000 https://blogs.perficient.com/?p=377771

Following up on my previous post which highlights different approaches of accessing Oracle Fusion Cloud Apps Data from Databricks, I present in this post details of Approach D, which leverages the Perficient accelerator solution. And this accelerator applies to all Oracle Fusion Cloud applications: ERP, SCM, HCM and CX.

As demonstrated in the previous post, the Perficient accelerator differs from the other approaches in that it has minimal requirements for additional cloud platform services. The other approaches of extracting data efficiently and in a scalable manner require the deployment of additional cloud services such as data integration/replication services and an intermediary data warehouse. With the Perficient accelerator, however, replication is driven by techniques that are solely reliant on native Oracle Fusion and Databricks. The accelerator consists of a Databricks workflow with configurable tasks to handle the end-to-end process of managing data replication from Oracle Fusion into the silver layer of Databricks tables. When deploying the solution, you get access to all underlying python/SQL notebooks that can be further customized based on your needs.

Why consider deploying the Perficient Accelerator?

There are several benefits to deploying this accelerator as opposed to building data replications from Oracle Fusion from the ground up. Built with automation, the solution is future-proof and enables scalability to accommodate evolving data requirements with ease. The diagram below highlights key considerations.

Databricks Accelerator Solution Benefits

A Closer Look at How Its Done

In the Oracle Cloud: The Perficient solution leverages Oracle BI Cloud Connector (BICC) which is the preferred method of extracting data in bulk from Oracle Fusion while minimizing the impact to the Fusion application itself. Extracted data and metadata is temporarily made available in the OCI Object Storage buckets for downstream processing. Archival of exported data on the OCI (Oracle Cloud Infrastructure) side is also automatically handled, if required, with purging rules.

Oracle Fusion To Databricks Data Replication Architecture

In the Databricks hosting cloud:

  • Hosted in one of: AWS, Azure or GCP, the accelerator’s workflow job and notebooks are deployed in the Databricks workspace. The Databricks delta tables schema, configuration and log files are all hosted within the Databricks Unity Catalog.
  • Notebooks leverage parametrized code to programmatically determine which Fusion view objects get replicated through the silver tables.
  • The Databricks workflow triggers the data extraction from Oracle Fusion BICC based on a predefined Fusion BICC job. The BICC job determines which objects get extracted.
  • Files are then transferred over from OCI to a landing zone object store in the cloud that hosts Databricks.
  • Databricks AutoLoader handles the ingestion of data into bronze Live Tables which store historical insert, update and delete operations relevant to the extracted objects.
  • Databricks silver Live Tables are then loaded from bronze via a Databricks managed DLT Pipeline. The silver tables are de-duped and represent the same level of data granularity for each Fusion view object as it exists in Fusion.
  • Incremental table refreshes are set up automatically leveraging Oracle Fusion object metadata that enables incremental data merges within Databricks. This includes inferring any data deletion from Oracle Fusion and processing deletions through to the silver tables.

Whether starting small with a few tables or looking to easily scale to hundreds and thousands of tables, the Perficient Databricks accelerator for Oracle Fusion data handles the end-to-end workflow orchestration. As a result, you end up spending less time with data integration and focus efforts on business facing analytical data models.

For assistance with enabling data integration between Oracle Fusion Applications and Databricks, reach out to mazen.manasseh@perficient.com.

]]>
https://blogs.perficient.com/2025/02/25/databricks-accelerator-for-oracle-fusion-applications/feed/ 0 377771
How to Access Oracle Fusion Cloud Apps Data from Databricks https://blogs.perficient.com/2025/02/24/databricks-for-oracle-fusion-cloud-applications/ https://blogs.perficient.com/2025/02/24/databricks-for-oracle-fusion-cloud-applications/#respond Mon, 24 Feb 2025 15:00:09 +0000 https://blogs.perficient.com/?p=377620

Connecting to Oracle Fusion Cloud Applications data from external non-Oracle systems, like Databricks, is not feasible for bulk data operations via a direct connection. However, there are several approaches to making Oracle apps data available for consumption from Databricks. What makes this task less straightforward is the fact that Oracle Fusion Cloud Applications and Databricks exist in separate clouds. Oracle Fusion apps (ERP, SCM, HCM, CX) are hosted on Oracle Cloud while Databricks leverages one of AWS, Azure or Google Cloud. Nevertheless, there are several approaches that I will present in this blog on how to access Oracle Apps data from Databricks.

While there are other means of performing this integration than what I present in this post, I will be focusing on:

  1. Methods that don’t require 3rd party tools: The focus here is on Oracle and Databricks technologies or Cloud services.
  2. Methods that scale to large number of objects and high data volumes: While there are additional means of Fusion data extraction such as using REST APIs, OTBI, or BI Publisher, these are not recommended methods for handling large bulk data extracts from Oracle Fusion and are therefore not part of this analysis. One or more of these techniques may still be applied though, when necessary, and may co-exist with the approaches discussed in this blog.

The following diagrams summarize four different approaches on how to replicate Oracle Fusion Apps data in Databricks. Each diagram highlights the data flow, and the technologies applied.

  • Approach A: Leverages Oracle Autonomous Data Warehouse and an Oracle GoldenGate Replication Deployment
  • Approach B: Leverages Oracle Autonomous Data Warehouse and the standard Delta Sharing protocol
  • Approach C: Leverages Oracle Autonomous Data Warehouse and a direct JDBC connection from Databricks.
  • Approach D: Leverages a Perficient accelerator solution using Databricks AutoLoader and DLT Pipelines. More information is available on this approach here.

Oracle Fusion data flow to Databricks with Oracle Autonomous DW and GoldGateOracle Fusion data flow to Databricks with Oracle Autonomous DW and Delta Sharing Oracle Fusion data flow to Databricks with Oracle Autonomous DW and JDBCOracle Fusion data flow to Databricks with Perficient Accelerator Solution

Choosing the right approach for your use case is dependent on the objective of performing this integration and the ecosystem of cloud platforms that are applicable to your organization. For guidance on this, you may reach Perficient by leaving a comment in the form below. Our Oracle and Databricks specialists will connect with you and provide recommendations.

]]>
https://blogs.perficient.com/2025/02/24/databricks-for-oracle-fusion-cloud-applications/feed/ 0 377620