> Blog >
Oracle to Postgres Migration: Complete Guide with Tools, Steps & Zero-Downtime Strategy
Plan your Oracle to Postgres migration with confidence. Explore tools like Ora2Pg and AWS DMS, CDC-based zero-downtime strategy, and steps to avoid risks.

Oracle to Postgres Migration: Complete Guide with Tools, Steps & Zero-Downtime Strategy

4 mins
April 17, 2026
Author
Aditya Santhanam
TL;DR
  • Moving from Oracle to Postgres is rarely about just copying data over. The real work sits in refactoring PL/SQL to PL/pgSQL, remapping data types like NUMBER and VARCHAR2, and redesigning schemas to take advantage of Postgres-native features like JSONB.
  • Most enterprise Oracle to Postgres migrations take anywhere from 3 to 9 months, depending on the volume of stored procedures, triggers, and application-level dependencies. Heavily customized systems almost always land at the upper end.
  • Zero-downtime migration is achievable using Change Data Capture (CDC) tools like Debezium, AWS DMS, and BryteFlow. These stream live changes from Oracle redo logs into Postgres so the final cutover takes only seconds of disruption.
  • The biggest drivers pushing teams off Oracle are licensing costs, vendor lock-in, and weak cloud-native flexibility. But if your system leans hard on Oracle RAC, deep PL/SQL packages, or vendor-certified ERPs, staying on Oracle may still cost less than refactoring everything.
  • Considering a move towards cloud-based solutions? Your database should also be shifted to the cloud. Oracle is a trusted database that has been used for years. But with its high yearly licensing costs, rising maintenance costs, and decreased flexibility, people are considering moving to databases that can save them money. One such database is PostgreSQL (often referred to as Postgres), an open, scalable database that offers cloud-native architecture, flexibility, and cost savings without compromising quality. It is mostly a developer database with JSONB performance that rivals NoSQL and a vast ecosystem of extensions.

    In this article, we will explore the Oracle to Postgres migration steps and how to achieve zero downtime.

    Table of Contents

      When to Migrate from Oracle to PostgreSQL (and When Not To)

      Moving from an Oracle to Postgres is one of the most common digital transformation trends today. But deciding to move is based on technical fit and business needs. 

      Signs You've Outgrown Oracle

      • Licensing costs and maintenance costs: Oracle’s licensing costs have increased, and have increased financial anxiety. It becomes an obstacle to innovations and growth.
      • Vendor lock-in, Flexibility, and Customization: Oracle heavily relies on specific features, and this creates a vendor lock-in and creates more control over the database. This can restrict flexibility and cloud portability. Organizations that require hybrid setups often prefer open-source databases such as PostgreSQL, which offers extensive extensions and customization options.
      • Cloud-native: Nowadays, new trends are moving towards Docker, Kubernetes, and agile DevOps to reduce dependency on proprietary systems. Oracle’s footprint may feel like a bottleneck, whereas Postgres integrates well within modern DevOps and containerized environments. 
      • Over-Provisioned Enterprise features: If your business needs modern applications that require specialized data types such as vector searches for AI, advanced JSON handling, or Geospatial data, Postgres offers a more flexible, plugin-friendly architecture without any expensive add-on packs. Moreover, Oracle provides many advanced features that may not be needed. But for that, the organizations need to pay the costs in that case. Postgres serves as a better alternative.

      When Oracle Is Still the Right Choice

      But Oracle may also be a dominant force for a reason. In that case, migrating from Oracle to Postgres might become a risk.

      • Dependency on Oracle features: If the database contains thousands of complex business logic, packages, and triggers, the cost may exceed the savings. Applications heavily using PL/SQL, RAC, or proprietary tools may be costly to refactor.
      • Strong Vendor Support Requirements: Organizations that rely on dedicated enterprise support may prefer Oracle’s service model. Postgres support is available, but often through third-party vendors.
      • Compliant needs: Some organizations may need Oracle’s certification and long-standing compliance track record. Switching to Postgres may require validation again and auditing.
      • No cost pressure: If one has a stable setup and costs are not a concern, an Oracle to Postgres migration might not be necessaryVendor-Locked Third-Party Software: Many enterprise ERP, CRM, and financial systems are certified only for Oracle. So moving from Oracle to Postgres will create a significant compliance risk.
      • Vendor-Locked Third-Party Software: Many enterprise ERP, CRM, and financial systems are certified only for Oracle. So moving from Oracle to Postgres will create a significant compliance risk.
      S.No When to be with Oracle When to Migrate to PostgreSQL
      1 In the case of mission-critical, high-workload applications Modern applications or cost-sensitive workloads
      2 Expensive licensing costs Open-source
      3 Dependent on Oracle-specific features Give high preference to extensibility and open standards
      4 When one needs Oracle or a tightly integrated stack When one needs a multi-cloud or cloud-native architecture

      Oracle vs PostgreSQL - Key Differences

      Choosing between Oracle and Postgres is not between “enterprise vs. open-source.” It depends on the features, cost, and architectural philosophy.

      S.No Feature Oracle PostgreSQL
      1 Licensing Expensive Open-source
      2 Architecture Tightly integrated system Open standards, highly extensible
      3 SQL standards Supports SQL with Oracle-specific extensions Strong adherence to SQL standards with custom extensions
      4 Procedural language PL/SQL PL/pgSQL
      5 Security Transparent data encryption Row-level security
      6 Extensions Limited (Oracle-only) Vast ecosystem (PostGIS, Timescale DB)
      7 Community support Vendor-backed enterprise support Large open-source community and also third-party vendors
      8 Cost of Ownership High cost Lower total cost

      Pre-Migration Assessment - What to Audit Before You Start

      Pre-migration assessment and planning are critical steps for a successful Oracle to Postgres migration. The critical areas to be concentrated on while doing an Oracle to Postgres migration are

      Schema Analysis & Document-to-Table Mapping

      • Schema and Data types conversion: Mapping database schemas, including tables, views, indexes, and constraints, is important to prevent data loss. Both Oracle and PostgreSQL have differences in their data type syntax. 
        For example, Oracle’s NUMBER datatype may get mapped to either NUMERIC, INTEGER, or BIGINT in Postgres, and Postgres separates DATE and TIMESTAMP; those need to be corrected manually. Create a mapping list to track how each data type in Oracle should be mapped in Postgres. Converting Oracle-specific features, like BLOB, CLOB, and LONG datatypes, requires careful planning. 
      • Stored procedures and functions: Oracle uses PL/SQL, and it is not compatible with PostgreSQL‘s PL/pgSQL; its syntax and structures of stored procedures differ in PostgreSQL. We need to rewrite these parts manually according to Postgres.
      • Document-to-Table Mapping: The Oracle system stores semi-structured or document-like data. If the document data has a constituent structure, you need high-speed joins and map the document attributes to standard Postgres columns. This is called the Flattening approach. Postgres is famous for its JSONB support. Then map the Oracle’s JSON or XML blobs directly into JSONB columns.

      Risk Assessment and Mitigation: 

      • Selecting the migration tool: Several migration tools are available, such as Ora2pg, SQLines, pgLoader, and AWS Database Migration Service (DMS). Ora2Pg is a free tool for migrating data from Oracle to PostgreSQL. It connects to Oracle DB, extracts the data, and generates SQL scripts that can be imported into Postgres DB. SQLines translates SQL code across the databases, and pgLoader handles the data migrations. Choose the migration tool based on database size, downtime tolerance, and real-time needs accordingly. 
      • Performance tuning and query optimization: After the Oracle to Postgres migration, there will be some performance degradation in some of the queries due to the changes and incompatibility in the indexes. Carefully look into the SQL queries and tune them accordingly so that their performance is not affected.
      • Mapping access and its permissions: Analyze the existing users, roles, and their privileges in the Oracle database. Using Postgres Role-based Access Control (RBAC), Row-level Security (RLS), and other authentication features, make sure to replicate the users’ access and align it with Oracle’s security policies.
      • Minimizing the downtime: As Oracle can handle large volumes of data, migrating it requires a significant amount of downtime and might affect the business. Ensure that the business operations are monitored 24/7. Utilizing tools like AWS Database Migration Service (DMS) for replication can minimize downtime.

      Oracle to Postgres Migration Tools - ETL & CDC Options

      Choosing the right migration tool depends on database size, the complexity of the logic, and downtime in the migration process. The tools and methods for moving from Oracle to PostgreSQL are

      Built-In Utilities (Manual / DIY)

      This traditional method is useful if the database size is smaller and has a highly skilled DBA team. These are basically open-source or native tools that give full control of the Oracle to Postgres migration.

      • Ora2Pg: It extracts schema, converts Oracle objects, and performs bulk data migration. 
      • PgLoader: High-performance data loader with parallel processing.
      • CSV Export/Import (SQL*Plus +COPY): In this method, it exports data from Oracle to flat files and uses the Postgres COPY command for high-speed ingestion.

      When to use

      • Small or medium datasets with flexible maintenance windows.
      • Teams with strong DBA expertise.

      No-Code ETL Platforms 

      No-Code or low-code ETL tools simplify the data migration process without complex scripts. They give faster results with less manual effort.

      • Fivetran: Cloud-based connectors for continuous data sync. It requires volume-based pricing.
      • Airbyte: It is an open-source alternative that offers over 600 connectors and allows you to run your migration both on-premises and in the cloud. It uses usage-based pricing.
      • AWS Database Migration Services (DMS): Managed service with schema conversion and continuous replication.

      When to use

      • One needs a simple interface and faster execution.

      CDC (Change Data Capture) Tools for Zero-Downtime

      CDC tools are essential when one needs a zero-downtime migration. These tools read Oracle’s Redo Logs or Archive Logs and “replay” those transactions on the Postgres target in real-time.

      • Debezium: It is an open-source and highly flexible CDC tool. It streams real-time changes via log-based CDC and is ideal for custom, event-driven architectures.
      • AWS DMS: It is a highly popular cloud-native tool. It handles both the initial “Full Load” and the ongoing CDC phase. It has built-in monitoring and validation with minimal downtime.
      • BryteFlow: It is a no-code CDC tool with automated reconciliation. It can handle both initial load and incremental sync.

      When to use

      • Choose CDC when downtime is not to be considered.

      How to choose the Migration Tool

      Choosing a migration tool is an important decision. It can be based on 

      • Data volume: For small datasets, choose Built-in utilities, and for medium to large datasets, choose CDC tools or an ETL platform.
      • Downtime tolerance: If zero downtime is preferred, choose the CDC approach; if downtime can be tolerated, the ETL (Extract, Transform, and Load) approach can be chosen.
      • Complexity: For a simple schema, choose DIY or no-code ETL, and for a complex transformation, choose advanced ETL or CDC pipelines.

      Best Practices for Selecting a Tool

      Before sticking with a tool, it is better to follow the Oracle to Postgres best practices.

      • Test tools with a sample project with a minimal dataset before full migration.
      • Ensure that the tool is compatible with both Oracle and Postgres.
      • Evaluate scalability, monitoring, and support features.
      • Consider all the costs associated with the tool, including setup, maintenance, and licensing.

      Step-by-Step Oracle to Postgres Migration Process

      Migrating from Oracle to Postgres needs a structured approach to ensure data accuracy, minimal downtime, and optimal performance. The Oracle to Postgres migration step-by-step process is

      Step 1: Schema Design & Data Mapping

      Before writing any code, do an audit assessment and identify risks and constraints. Then start designing your target schema. Perform an audit on the Oracle database and collect the details about schemas, objects, tables, data types, Oracle-specific features, stored procedures, functions, and application dependencies. Identify nested arrays and objects in your documents that should become separate, linked tables in PostgreSQL. 

      Be clear in the migration objectives on what needs to be migrated and whether it is required in the Postgres database or not. This step will help us identify potential risks and complexities involved in the Oracle to Postgres migration. Use tools such as Ora2Pg to convert Oracle-specific objects into Postgres. Map Oracle-specific types such as Number to Numeric, VARCHAR2 to VARCHAR, and CLOB to TEXT.

      Step 2: Set Up PostgreSQL Environment

      Install Postgres on-premises, cloud, or managed like AWS RDS, Azure, or GCP. Adjust configurations such as max_connections, shared buffers, and work_mem based on Oracle AWR reports. Create users, roles, and configure them.

      Step 3: Choose Your Migration Method

      Select the migration strategy that aligns with your uptime requirements. For smaller and non-critical DBs, choose Offline Migration. For mission-critical systems, choose CDC migrations to sync changes in real-time until the final cutover.

      Step 4: Export & Transform Data

      Start extracting data from the source by ensuring the Postgres environment is ready. Export data using tools such as Ora2Pg or native utilities. Transform data types and formats. This will reduce the migration errors and ensure compatibility. 

      Step 5: Load Data in Postgres

      Once the data (objects, tables) is exported, load it into PostgreSQL using the COPY command, even if you are using Ora2Pg.

      Step 6: Validate data

      Validate row counts, checksums, and constraints. Validate Oracle and Postgres to ensure no data was lost or corrupted during the migration process. Compare sample queries between the Oracle and Postgres databases. If any mismatches occur, resolve them immediately. Run validation scripts to compare the data integrity between the two databases.

      Step 7: Application code Updates

      Checking application complexity is one of the most time-consuming steps. Modify application logic to work with Postgres by replacing Oracle-specific SQL syntax. Update database drivers and connection strings. Refactor PL/SQL logic into PL/pgDQL or application code.

      Step 8: Test Application and Cutover Planning

      Run parallel testing on both Oracle and Postgres environments. Compare the results of queries from both databases. Testing is essential to identify any migration-related issues before going live. Set up a separate test environment and test the data migration process. We should check the consistency of the transferred data and verify that the migrated data behaves in the same way as it did in the Oracle database. Do functional testing, performance testing, and regression testing to ensure that there are no data losses.

      Ensure all services and APIs work seamlessly with PostgreSQL. Schedule a downtime period for the final cutover of the Oracle. Stop writing to Oracle and decommission the old database only after a sufficient observation period. Closely monitor the Postgres and application after the cutover to identify and address any issues.

      Step 9: Monitor & Optimize Post-Migration

      Redirect application traffic from Oracle to Postgres. Continuously monitor application and database performance. Track query performance, index utilization, and optimize the workload. For the slow query log, tune them using Postgres-specific indexing such as GIN or BRIN.

      Common Data Mapping Challenges and How to Solve Them

      Data mapping is one of the time-consuming parts of the Oracle to Postgres migration challenges. Some of the challenges faced during the Oracle to Postgres migration are

      • Integer confusion: Oracle’s NUMBER type represents integers, decimals, and large values. But PostgreSQL separates them into different types. To solve this, use INTEGER or NUMERIC. For whole numbers, use INTEGER or BIGINT, and for precision-based values, use NUMERIC.
      • VARCHAR2 vs TEXT Behaviour: Oracle’s VARCHAR has size limits and behaves differently compared to Postgres. To solve this, use TEXT in Postgres for flexibility and VARCHAR only when the length is constrained.
      • DATE and TIMESTAMP differences: Oracle uses both date and time in its DATE type, while Postgres displays only date in the DATE datatype, and TIMESTAMP handles the time. To solve this, always map Oracle’s DATE columns to Postgres TIMESTAMP.
      • EmptyString vs. NULL Debate: In Oracle, an empty string is treated as NULL. In PostgreSQL, an empty string is a valid, non-null value. To solve this, during the ETL process, decide whether to convert all empty strings to NULL or handle empty strings in your application code. Add CHECK constraints to ensure consistency if one wants to forbid empty strings.
      • Boolean Data Type Differences: Oracle does not have a native BOOLEAN type in tables, whereas Postgres does. To solve this, map Oracle flags NUMBER, CHAR to BOOLEAN. Update queries and application logic accordingly.

      Zero-Downtime Oracle to PostgreSQL Migration with CDC

      Perfect planning in Oracle to Postgres migration ensures zero downtime or near-zero downtime cutover to minimize business disruption. The migration begins with a staging environment to validate data mappings, relational schemas, and application query behavior. Business critical applications often require continuous availability, which is where Change Data Capture (CDC) enables a zero-downtime transition from Oracle to Postgres.

      How CDC Works for Oracle Migrations

      Change Data Capture (CDC) is a software pattern that identifies and tracks changes in a source database so that action can be taken using change data. It captures the real-time changes for inserts, updates, and deletes in real time. These changes are streamed into Postgres, which ensures both databases stay in sync. Most commonly used tools are Debezium and Kafka.

      CDC works in the following flow

      • Do a full export of existing data from Oracle and load it into PostgreSQL using bulk migration tools.
      • CDC tools read redo/transaction logs, capture their ongoing changes (INSERT, UPDATE, DELETE), and convert the stream into Postgres.
      • Now apply the captured changes to Postgres in near real time, monitor for lag and consistency.

      Benefits of CDC for Migration

      • Minimal zero downtime
      • Reduced Business risk
      • Parallel Testing
      • Scalable and Flexible

      Dual-Write Strategy

      The dual-write strategy involves writing data on both Oracle and Postgres simultaneously during migration. Both systems stay in sync at the application level and are used alongside or instead of CDC in some cases. Choose it when CDC is not feasible.

      Migration Window vs. Live Replication

      Choosing between a migration window and live replication depends on business needs. When the cutover is planned, we must decide between traditional migration and live replication.

      Feature Migration Window Live Replication
      Downtime High downtime. The app must be offline during the entire transfer. Less downtime. Only a few seconds or minimal disruption
      Data Volume Suited for small datasets(<10GB) Essential for large datasets (>100 GB)
      Complexity Simple script-based export/import - Low Requires CDC infrastructure (Kafka, Debezium)
      Use case Internal tools are used for non-critical legacy systems Use mostly for critical systems, 24/7 user-facing applications
      Risk High. Again, one must start migration from the start Low. Data validation can be done in parallel for weeks.
      Which one to choose For smaller systems with flexible downtime, a migration window may still be sufficient When downtime is not important

      Best Practices for ensuring Zero-downtime migration

      • Start with a pilot project with a small dataset to validate the pipeline.
      • Monitor replication lag and system performance.
      • Ensure to follow the data validation mechanism.
      • Use feature flags or routing layers for controlled cutover.
      • Always have a backup plan to rollover.
      • Implement a rollback strategy even with CDC in place.

      Testing, Validation, and Post-Migration Performance Tuning

      The success of the Oracle to Postgres migration depends on how thoroughly you test, validate, and optimize the Postgres environment. Without rigorous validation and strategic tuning, it leads to decreased performance. 

      Continuous monitoring of the new PostgreSQL environment is needed. Check for error logs and tune the Postgres according to workload requirements. Update the applications and systems to connect to the newly configured PostgreSQL database by modifying the connection strings, updating configuration files, or rewriting SQL queries. Differences in database behaviour, query execution, and indexing can introduce subtle issues that only appear under real workloads. 

      Confirm data migrated is accurate and complete. Use checksums or hashes for data verification. This step ensures no data loss or corruption.

      Post-migration validation checklist:

      • The priority is analyzing query performance, reviewing executing plans, and tuning indexes based on real usage patterns. Continuously monitor applications for bottlenecks and anomalies. Utilize the cloud-native features for better scalability. Ensure row counts match between Oracle and Postgres.
      • Critical applications workflows are tested with proper security roles and permissions.
      • Postgres settings are generic. Database administrators should review foreign key constraints, update statistics, and enable routine vacuuming. Do a checksum on critical tables. Connection pooling configured. Indexes created for high-traffic query patterns.
      • Regular backup and snapshot policies must be updated to reflect the new environment, along with automated alerting for query latency. Continuously monitoring ensures long-term performance stability and data reliability.

      Proper testing and validation delivers its promise of scalability, reliability, and cost efficiency.

      Common Oracle to PostgreSQL Migration Mistakes to Avoid

      The most critical mistakes to avoid when doing an Oracle to Postgres migration are

      • PL/SQL to PL/pgSQL: Oracle’s PL/SQL is incredibly deep with features such as autonomous transactions and structures that don’t exist natively in Postgres. To overcome this, don’t use automated tools, manually rewrite complex logic, and convert PL/SQL to PL/pgSQL carefully. Refactor complex logic into the application layer.
      • No “Lift and Shift”: Copying the schema and data directly without redesigning for Postgres. To solve this, redesign the schema wherever needed, optimize accordingly to Postgres features.
      • Ignoring Case Sensitive and Quoting: Oracle does not identify the identifiers whereas Postgres identifiers are folded to lowercase by default. To solve this problem, convert all tables and columns to lowercase during the Oracle to Postgres migration. This makes SQL remain clean and standard-compliant.
      • Data Type Mapping Decisions: Using incorrect data type mappings impacts performance and storage. It causes data precision issues. To solve this problem, analyze actual data usage before mapping. Use appropriate Postgres type for INTEGER, BIGINT, NUMERIC, TEXT, etc., for Oracle’s NUMERIC data type.
      • Application changes: One should also check the application-level impact, not only considering databases. To mitigate this, audit the environment beforehand before starting the migration and test for compatibility thoroughly.
      • Validating data: Not doing the data verification once the migration has happened. To avoid this, validate row counts, checksums, and key datasets and reconciliation.
      • Downtime: Downtime should be considered when planning for migration. To avoid this use, CDC for near-zero downtime and do a test cutover scenario in advance.

      Why Choose Entrans for Your Oracle to Postgres Migration

      Though Oracle is solid, the future now belongs to Postgres. Migrating is not just copying data to another database; it is about performance tuning and application-level changes. Oracle to Postgres migration is a complex process that demands substantial investment in skilled personnel and dedicated development time, and Entrans provides all of that. 

      By meticulously planning, perfect execution, and optimizing each phase, Entrans unlocks the full potential of the Postgres database, thereby giving a strong foundation for future growth. Hiring a team of migration experts like the team from Entrans takes away the risks that come along with it; we optimize your data integration seamlessly.

      Planning to migrate from Oracle to Postgres, Entrans is here to help make the transition smoother and more predictable. Book a 20-minute Consultation today.

      Share :
      Link copied to clipboard !!
      Ready to Cut Oracle Licensing Costs Without Breaking Production?
      Entrans delivers end-to-end Oracle to Postgres migrations with near-zero downtime, full schema redesign, and hands-on PL/SQL to PL/pgSQL refactoring.
      20+ Years of Industry Experience
      500+ Successful Projects
      50+ Global Clients including Fortune 500s
      100% On-Time Delivery
      Thank you! Your submission has been received!
      Oops! Something went wrong while submitting the form.

      Frequently Asked Questions

      1. How long does it take to migrate from Oracle to Postgres?

      The time taken to migrate from Oracle to Postgres depends on the volume of PL/SQL code, the number of triggers, and the dataset size. Typically, it may take 3 to 9 months, depending on complexity. Large enterprise systems with heavy customization and downtime constraints take longer due to testing and validation.

      2. How to ensure zero downtime during Oracle to Postgres migration?

      Zero downtime is achieved by using Change Data Capture (CDC) and real-time replication and parallel run environments. Cutover is handled via phased switching or blue-green deployment to avoid service interruption.

      3. Are Oracle to Postgres migration tools available, or should we go with a manual script?

      Tools such as AWS DM and Ora2pg are available for automating schema and data migration. Manual scripting is only recommended for highly customized logic or handling incompatible data types and tuning specific performance constraints.

      4. How does Oracle compare to Postgres?

      Oracle is a proprietary RDBMS known for advanced features such as Real Application Clusters (RAC). Postgres is a highly extensible open-source ORDBMS that offers comparable performance for most applications for enterprise workloads.

      5. When to use Oracle vs. Postgres and how to decide?

      Choose Oracle when one needs niche enterprise features such as RAC and has a large existing investment in Oracle licenses. So if one wants to reduce or eliminate licensing costs, cloud-native flexibility, choose Postgres.

      6. What is Oracle to PostgreSQL ETL?

      Oracle to PostgreSQL refers to extracting data from Oracle, transforming it to fit Postgres schemas, and loading it into the target database. It often involves modernizing legacy data structures and cleaning records to ensure integrity in the new environment.

      7. Is migrating from Oracle to PostgreSQL hard?

      Sometimes migrating from Oracle to Postgres might be complex due to differences in SQL syntax, data types, and procedural logic (PL/SQL). Choosing the right tool and doing effective planning and testing, the process becomes manageable.

      8. Will the Oracle to Postgres migration affect my application’s performance?

      Yes. Oracle to Postgres migration may fluctuate performance because Postgres uses a Query planner and execution strategy than Oracle. With proper optimization, Postgres can match or even outperform Oracle.

      Hire Senior Postgres and Migration Engineers on Demand
      Tap into Entrans' bench of pre-vetted database specialists with production experience in Ora2Pg, AWS DMS, Debezium, and large-scale CDC pipelines.
      Free project consultation + 100 Dev Hours
      Trusted by Enterprises & Startups
      Top 1% Industry Experts
      Flexible Contracts & Transparent Pricing
      50+ Successful Enterprise Deployments
      Aditya Santhanam
      Author
      Aditya Santhanam is the Co-founder and CTO of Entrans, leveraging over 13 years of experience in the technology sector. With a deep passion for AI, Data Engineering, Blockchain, and IT Services, he has been instrumental in spearheading innovative digital solutions for the evolving landscape at Entrans. Currently, his focus is on Thunai, an advanced AI agent designed to transform how businesses utilize their data across critical functions such as sales, client onboarding, and customer support

      Related Blogs

      AI Coding Assistants for Mainframe-to-Cloud Integration: What They Can Do, What They Can Not, and What to Expect

      AI coding assistants speed up mainframe to cloud integration, but they cannot replace human expertise. Learn what they can do and what to expect.
      Read More

      Mainframe to AWS Cloud Migration: Tools, Services, and How to Run a Successful Program

      Migrate your mainframe to AWS and cut infrastructure costs by up to 85%. Explore tools, migration phases, security standards, and the AWS MAP program.
      Read More

      Mainframe Modernization to Cloud: What It Means, How It Differs from Migration, and Which Path Is Right for You

      Mainframe modernization to cloud is not just migration. Explore the full spectrum of approaches, ROI ranges of 288 to 362%, and how to avoid the modernization trap.
      Read More