<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Engineered Systems &#8211; Bugra Parlayan | Oracle Database &amp; Exadata Blog</title>
	<atom:link href="https://www.bugraparlayan.com.tr/category/engineered-systems/feed" rel="self" type="application/rss+xml" />
	<link>https://www.bugraparlayan.com.tr</link>
	<description>A technical blog</description>
	<lastBuildDate>Sat, 19 Jul 2025 19:07:30 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.8.2</generator>

<image>
	<url>https://www.bugraparlayan.com.tr/wp-content/uploads/2020/06/cropped-plsql-32x32.jpg</url>
	<title>Engineered Systems &#8211; Bugra Parlayan | Oracle Database &amp; Exadata Blog</title>
	<link>https://www.bugraparlayan.com.tr</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Creating an ASM Disk on Exadata FlashDisk</title>
		<link>https://www.bugraparlayan.com.tr/creating-an-asm-disk-on-exadata-flashdisk.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Sat, 19 Jul 2025 19:06:23 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[ASM]]></category>
		<category><![CDATA[direct path read]]></category>
		<category><![CDATA[disk drop]]></category>
		<category><![CDATA[exadata]]></category>
		<category><![CDATA[flash cache]]></category>
		<category><![CDATA[FlashDisk]]></category>
		<category><![CDATA[performance tuning]]></category>
		<category><![CDATA[planned maintenance]]></category>
		<category><![CDATA[Temp Tablespace]]></category>
		<category><![CDATA[warm-up effect]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1505</guid>

					<description><![CDATA[<p>Recently, we noticed performance drops at certain times on an Exadata server operating as a data warehouse (DWH). Upon analyzing AWR reports, the following wait event stood out: direct path read/write temp waits We knew this event was caused by temporary (temp) data being written to or read from disk during query execution. Although we &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/creating-an-asm-disk-on-exadata-flashdisk.html">Creating an ASM Disk on Exadata FlashDisk</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Recently, we noticed performance drops at certain times on an Exadata server operating as a data warehouse (DWH). Upon analyzing <strong>AWR reports</strong>, the following wait event stood out:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>direct path read/write temp waits</p>
</blockquote>



<p>We knew this event was caused by temporary (temp) data being written to or read from disk during query execution. Although we had applied various SQL optimizations, they did not significantly reduce the wait times.</p>



<h3 class="wp-block-heading"><strong>Root Cause: Heavy TEMP Tablespace Usage</strong></h3>



<p>Large SQL queries in data warehouses—especially those involving operations like hash joins, sorts, and group by—require extensive use of TEMP space. In our case, the <strong>TEMP tablespace</strong> was located on <em>high-latency</em> disk groups by default, which increased IO wait times.</p>



<h3 class="wp-block-heading"><strong>Solution: Move TEMP Tablespace to FlashDisk on Exadata</strong></h3>



<p>To directly improve performance, we decided to move the TEMP tablespace to the <strong>FlashDisk</strong> layer on Exadata. This layer offers much higher IO throughput compared to traditional disks, making it particularly advantageous for handling temporary data operations.</p>



<p><br>P.S. Since we will drop and recreate the existing flash disks during this operation, there may be minor negative impacts during the process and until the flash cache warms up again. Therefore, this work should be performed during a planned and quiet maintenance window.</p>


<div class="wp-block-syntaxhighlighter-code "><pre class="brush: bash; title: ; notranslate">
dcli –g cell_group –l root cellcli -e &quot;alter flashcache all flush&quot;
dcli -g cell_group -l root cellcli -e &quot;LIST CELLDISK ATTRIBUTES name, flushstatus, flusherror&quot; | grep FD
dcli -g cell_group -l root cellcli -e drop flashcache
dcli -g cell_group -l root cellcli -e create flashcache all size=20.2875976562500T;
dcli -g cell_group -l root cellcli -e CREATE GRIDDISK ALL FLASHDISK PREFIX='FLASHTMP';

</pre></div>

<div class="wp-block-syntaxhighlighter-code "><pre class="brush: sql; title: ; notranslate">
sqlplus / as sysasm

alter system set asm_diskstring='o/*/DATA_*','o/*/RECO_*','o/*/FLASHTMP*';

CREATE diskgroup TEMPDG normal redundancy disk 'o/*/FLASHTMP*' attribute 'compatible.rdbms'='19.0.0.0.0', 'compatible.asm'='19.0.0.0.0', 'cell.smart_scan_capable'='TRUE', 'au_size'='4M';
</pre></div>

<div class="wp-block-syntaxhighlighter-code "><pre class="brush: sql; title: ; notranslate">
sqlplus &amp; as sysdba

CREATE TEMPORARY TABLESPACE TEMP_FLASH TEMPFILE '+ TEMPDG' SIZE 32G EXTENT MANAGEMENT LOCAL UNIFORM SIZE 1M;

ALTER DATABASE DEFAULT TEMPORARY TABLESPACE TEMP_FLASH;

DROP TABLESPACE TEMP INCLUDING CONTENTS AND DATAFILES;

</pre></div><p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/creating-an-asm-disk-on-exadata-flashdisk.html">Creating an ASM Disk on Exadata FlashDisk</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Best Practices for Very Large Database (VLDB) Backup and Recovery:</title>
		<link>https://www.bugraparlayan.com.tr/best-practices-for-very-large-database-vldb-backup-and-recovery.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Sat, 14 Jun 2025 17:44:00 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[archive logs]]></category>
		<category><![CDATA[backup retention]]></category>
		<category><![CDATA[differential backup]]></category>
		<category><![CDATA[full backup]]></category>
		<category><![CDATA[incremental backup]]></category>
		<category><![CDATA[point-in-time recovery]]></category>
		<category><![CDATA[recovery point objective]]></category>
		<category><![CDATA[recovery time objective]]></category>
		<category><![CDATA[restore strategy]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1480</guid>

					<description><![CDATA[<p>1. Executive Summary Backing up and recovering Very Large Databases (VLDBs) presents a critical yet increasingly complex challenge for organizations in today&#8217;s data-driven world. With data volumes growing exponentially, traditional backup methods often fall short in meeting performance targets, Recovery Time Objectives (RTOs), and Recovery Point Objectives (RPOs). This report examines best practices in VLDB &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/best-practices-for-very-large-database-vldb-backup-and-recovery.html">Best Practices for Very Large Database (VLDB) Backup and Recovery:</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading">1. Executive Summary</h2>



<p>Backing up and recovering Very Large Databases (VLDBs) presents a critical yet increasingly complex challenge for organizations in today&#8217;s data-driven world. With data volumes growing exponentially, traditional backup methods often fall short in meeting performance targets, Recovery Time Objectives (RTOs), and Recovery Point Objectives (RPOs). This report examines best practices in VLDB backup and recovery, drawing on insights presented in Oracle MAA (Maximum Availability Architecture) blog posts, with a specific focus on Oracle&#8217;s Zero Data Loss Recovery Appliance (ZDLRA) solution.</p>



<p>The ZDLRA is a purpose-built engineered system designed to address these challenges. Its core strategies include &#8220;Incremental Forever&#8221; backups, which significantly reduce the load on production systems; real-time redo protection for near-zero data loss; and continuous recovery validation, enhancing the reliability of backups. <sup></sup> These features are tailored to meet the unique demands of VLDBs, offering substantial improvements in achieving RTO and RPO targets. Oracle&#8217;s development and promotion of a specialized hardware/software appliance like ZDLRA suggest that traditional, software-only backup methods are increasingly inadequate for the scale and criticality of modern VLDBs. This implies the problem&#8217;s complexity has reached a point where integrated hardware and software solutions offer a more effective approach than generic software tools running on general-purpose hardware. This is a significant paradigm shift in high-end backup and recovery strategies. Consequently, organizations managing VLDBs must assess whether their current backup infrastructure can realistically meet future demands or if a specialized appliance approach is becoming a necessity. &nbsp;</p>



<h2 class="wp-block-heading">2. Introduction to VLDB Backup and Recovery Challenges</h2>



<p>Very Large Databases (VLDBs) typically contain terabytes to petabytes of data and are continuously growing. The sheer size and complexity of these databases introduce unique challenges in backup and recovery processes.</p>



<ul class="wp-block-list">
<li><strong>Defining VLDBs and Their Criticality:</strong> VLDBs are central repositories for businesses&#8217; core operations, customer data, financial records, and other vital information. Therefore, any data loss or prolonged downtime in these systems can lead to severe financial losses, reputational damage, and legal repercussions. Business continuity and regulatory compliance are primary drivers for robust backup and recovery strategies for VLDBs.</li>



<li><strong>Common Pain Points:</strong>
<ul class="wp-block-list">
<li><strong>Backup Windows:</strong> Completing full backups of VLDBs within limited timeframes without an acceptable impact on production performance is extremely difficult. As database size increases, full backup times lengthen, often encroaching on business hours and negatively affecting system performance.</li>



<li><strong>Recovery Time Objectives (RTO):</strong> Restoring and recovering massive databases quickly enough to meet business needs in the event of a disaster or failure is a major hurdle. Long recovery times lead to extended business disruptions and, consequently, increased costs.</li>



<li><strong>Recovery Point Objectives (RPO):</strong> There is always a risk of significant data loss due to the time gap between backups. Even hourly or more frequent backups can lead to unacceptable data loss in high-transaction-volume systems.</li>



<li><strong>Performance Impact:</strong> Backup operations generate significant I/O (Input/Output) and CPU (Central Processing Unit) load on production database servers. This load can degrade application performance, especially during peak hours.</li>



<li><strong>Storage Costs:</strong> Managing and storing large volumes of backup data incurs substantial storage costs. Long-term retention policies and multiple backup copies further escalate these costs.</li>



<li><strong>Complexity:</strong> Managing complex backup scripts, schedules, and recovery procedures creates a significant operational burden and increases the risk of human error.</li>
</ul>
</li>
</ul>



<p>These challenges are not just technical but also economic and operational. The &#8220;pain points&#8221; are interconnected; for example, trying to shrink backup windows with traditional methods can increase the performance impact, and aggressive RPO targets can lead to higher storage costs. Because VLDBs are large, backups are inherently time-consuming. Businesses demand short RTOs and near-zero RPOs. Attempting frequent full backups on VLDBs (for RPO) exacerbates backup window and performance impact issues. Using traditional incremental backups can lead to complex and lengthy recovery processes, jeopardizing RTO. This creates a cycle of trade-offs where optimizing one aspect negatively affects another. This highlights the need for a holistic solution that addresses these interconnected challenges simultaneously, rather than in isolation, which is the rationale for an integrated system like ZDLRA.</p>



<h2 class="wp-block-heading">3. Oracle&#8217;s Zero Data Loss Recovery Appliance (ZDLRA): A Purpose-Built Solution for VLDBs (Based on Oracle MAA Blog Insights)</h2>



<p>Oracle&#8217;s Zero Data Loss Recovery Appliance (ZDLRA) is a purpose-built engineered system developed to address the challenges encountered in backing up and recovering Very Large Databases (VLDBs). This section will examine the core features of ZDLRA and its significance in VLDB protection, based on points highlighted in Oracle MAA blogs.</p>



<h3 class="wp-block-heading">3.1. Overview of the ZDLRA Approach</h3>



<p>The ZDLRA is a purpose-built engineered system developed by Oracle to centralize and optimize database backup and recovery operations, focusing on protection, efficiency, and scalability for Oracle databases. It&#8217;s crucial to emphasize that ZDLRA is not merely a software solution but a comprehensive one where hardware and software are co-engineered for optimal performance and reliability in the demanding context of VLDB protection. As stated in the Oracle MAA blog post, &#8220;ZDLRA is a purpose-built engineered system designed to maximize hardware and software to provide a highly available, zero data loss environment. It notes that software alone cannot achieve this, implying that ZDLRA&#8217;s integrated hardware and software approach is critical for meeting stringent RTO and RPO requirements.&#8221; <sup></sup> This positions ZDLRA not just as software, but as a comprehensive solution where hardware and software are co-designed for optimal performance and reliability under the demanding conditions of VLDB protection. &nbsp;</p>



<h3 class="wp-block-heading">3.2. The &#8220;Incremental Forever&#8221; Strategy for VLDBs</h3>



<p>One of ZDLRA&#8217;s most notable features is its &#8220;Incremental Forever&#8221; or &#8220;virtual full&#8221; backup strategy. This strategy fundamentally changes the backup process for VLDBs.</p>



<ul class="wp-block-list">
<li><strong>Mechanism:</strong> After an initial full (Level 0) backup, only changed data blocks are sent from the production database to the ZDLRA. The ZDLRA then synthesizes full backups (&#8220;virtual fulls&#8221;) from these incremental backups. This eliminates the overhead of taking full backups every day.</li>



<li><strong>Benefits for VLDBs:</strong>
<ul class="wp-block-list">
<li><strong>Reduced Production Impact:</strong> &#8220;This strategy reduces the processing load on production systems by only transmitting changed data during daily incremental backups.&#8221;  This minimizes I/O and CPU load on the source database, which is critical for performance-sensitive VLDBs. Traditional Level 0 + Level 1 backups are problematic for VLDBs: Level 0s are too large and impact performance; recovery from many Level 1s is slow. ZDLRA&#8217;s &#8220;Incremental Forever&#8221; sends only changed blocks <em>after the initial full backup</em>. This dramatically reduces the daily backup workload on the production database.  </li>



<li><strong>Storage Efficiency:</strong> Efficient storage of incremental data on the ZDLRA and pointer-based virtual fulls &#8220;can lead to a 10X decrease in space consumption.&#8221;  This offers a significant cost advantage, especially when dealing with large data volumes.  </li>



<li><strong>Faster Backup Completion:</strong> Daily &#8220;backups&#8221; are essentially small incremental backups, significantly shortening the backup window.</li>



<li><strong>Efficient Recovery:</strong> It &#8220;allows for more efficient recovery compared to traditional RMAN incremental-based recovery.&#8221;  Restoring a virtual full backup is similar to restoring a true full backup, without the need to sequentially apply numerous incrementals. ZDLRA takes on the task of creating &#8220;virtual full&#8221; backups from these incrementals. This means the <em>appliance</em>, not the production server, does the heavy lifting. For recovery, the database can be restored from a virtual full, which is much faster than applying a long chain of traditional incrementals. This directly improves RTO.  </li>



<li><strong>Offloading of Backup Operations:</strong> &#8220;By offloading backup compression, deletion, validation, and maintenance operations to the appliance, production systems can focus on workloads.&#8221;  This further frees up production server resources.  </li>
</ul>
</li>
</ul>



<p>This approach fundamentally changes the backup paradigm for Oracle databases. By shifting intelligence and workload from the source database to a specialized appliance, it allows production systems to dedicate their resources to business operations. It also simplifies recovery processes, reducing the potential for human error.</p>



<h3 class="wp-block-heading">3.3. Achieving Real-Time, Near-Zero Data Loss Protection (Near-Zero RPO)</h3>



<p>ZDLRA offers an innovative approach to minimizing the Recovery Point Objective (RPO).</p>



<ul class="wp-block-list">
<li><strong>Mechanism:</strong> &#8220;The Recovery Appliance uses Oracle&#8217;s real-time redo transport to deliver continuous, real-time data protection. Transactional changes (redo) are transmitted directly to the appliance, where archived redo log backups are created and stored.&#8221;  This mechanism is similar to Data Guard redo transport but specifically designed for backup and recovery assurance.  </li>



<li><strong>Benefits:</strong> &#8220;This provides immediate, zero data loss protection of all changes, and directly addresses the RPO objective of minimizing data loss.&#8221;  This means recovery can be performed up to the last committed transaction received by the ZDLRA, achieving an RPO of seconds or sub-seconds rather than hours. Traditional RPO is often tied to the frequency of archived log backups or discrete incremental data backups. For VLDBs, there can still be gaps between these discrete operations (e.g., every 15 mins, 1 hour). Real-time redo transport sends redo data <em>as it&#8217;s generated</em> (or very close to it) to the ZDLRA. The ZDLRA then archives this redo. This means that even if a failure occurs between discrete incremental data backups, the redo logs up to (or very near) the point of failure are already secured on the ZDLRA. This dramatically improves RPO beyond what traditional scheduled backups can offer, allowing for recovery with minimal or no data loss.  </li>
</ul>



<p>This feature is a game-changer for businesses with extremely low tolerance for data loss. It elevates ZDLRA from merely a backup device to a key component of a high-availability and data protection strategy, approaching disaster recovery capabilities for recent transactions. It also implies a tighter integration with the database&#8217;s transaction processing cycle.</p>



<h3 class="wp-block-heading">3.4. Ensuring Data Integrity with Continuous Recovery Validation</h3>



<p>The reliability of backups is paramount for successful recovery. ZDLRA takes a proactive approach to this.</p>



<ul class="wp-block-list">
<li><strong>Process:</strong> &#8220;The appliance performs corruption detection throughout the backup cycle to validate data consistency and immediately alerts administrators if corruption is detected. It checks all incoming and replicated backups for block-level validity. Any corrupted data is detected, recorded, and alerted, allowing administrators to take action.&#8221;   </li>



<li><strong>Benefits:</strong> &#8220;This assurance of valid data is a key component for a successful recovery, directly impacting RTO by ensuring that restored data is usable and not corrupt.&#8221;  This proactive validation prevents the discovery of corruption only at the critical moment of recovery, which could severely impact RTO and business operations. Data block corruption can occur on the primary database and, if undetected, propagate to backups. Traditional validation might happen during the backup process (e.g., RMAN <code>VALIDATE DATABASE</code>) or as a separate scheduled task, but ZDLRA makes it an intrinsic part of data ingestion. By checking blocks as they arrive and as virtual fulls are created/maintained, ZDLRA provides an early warning system. If corruption is detected in an incoming backup, administrators are alerted immediately. This allows them to address the issue on the primary database or ensure subsequent backups are clean, rather than discovering the problem months later during a critical restore. This ensures that backups stored on ZDLRA are known to be good, which is fundamental for a predictable and successful RTO.  </li>
</ul>



<p>This feature increases confidence in the backup repository. It means that when a recovery is initiated, there is a much higher certainty that the restored data will be valid and uncorrupted, reducing the risk of failed recoveries or recoveries that bring back corrupt data, which can be worse than no recovery at all. This also reduces the need for extensive manual validation efforts.</p>



<h3 class="wp-block-heading">3.5. The Significance of an Engineered System Approach for VLDBs</h3>



<p>The idea that ZDLRA is not just software but an integrated hardware and software solution is fundamental to its effectiveness in VLDB protection. &#8220;The article emphasizes that the ZDLRA is a purpose-built engineered system designed to maximize hardware and software&#8230; It states that software alone cannot achieve this.&#8221; <sup></sup> This co-engineering allows for optimizations in I/O, network traffic, storage management, and processing that would be difficult to achieve with general-purpose components. Protecting VLDBs efficiently requires high throughput for backups, fast access for restores, and robust processing for tasks like validation and virtual full creation. General-purpose hardware and backup software might not be optimally configured to work together for these specific, demanding Oracle database workloads. An engineered system allows the vendor (Oracle) to control and optimize all layers: the database-side agents, the network protocols used, the internal processing within the appliance, and the storage layout. This tight integration can lead to performance, reliability, and manageability benefits that are hard to replicate with a piecemeal approach. &nbsp;</p>



<p>The &#8220;engineered system&#8221; argument positions ZDLRA as a premium, high-performance solution where the whole is greater than the sum of its parts. It implies that Oracle has fine-tuned every component of the stack, from database interaction to storage within the appliance, for the specific task of Oracle database protection. While potentially carrying a higher upfront cost, the engineered system approach aims to deliver a lower TCO (Total Cost of Ownership) through operational efficiencies, reduced risk, and superior performance. It also signifies a single-vendor commitment to supporting the entire solution stack, potentially simplifying troubleshooting and support. This is a strategic choice for organizations where VLDB protection is a top-tier priority.</p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Challenge Area</th><th>Traditional Approach Pain Points</th><th>ZDLRA Solution &amp; Key Features Leveraged</th></tr><tr><td>Backup Window</td><td>Long full backups, performance impact</td><td>Incremental Forever, offloaded processing</td></tr><tr><td>RTO</td><td>Slow recovery from many incrementals, risk of corrupt backup</td><td>Virtual Full Backups, Continuous Recovery Validation</td></tr><tr><td>RPO</td><td>Data loss since last backup (hours)</td><td>Real-Time Redo Transport</td></tr><tr><td>Production Impact</td><td>High CPU/IO during backups</td><td>Incremental Forever (sends only changes), offloaded processing (compression, validation) <sup></sup></td></tr><tr><td>Storage Consumption</td><td>Multiple full backups, large incrementals</td><td>Incremental Forever (stores deltas efficiently), space-efficient virtual fulls <sup></sup></td></tr><tr><td>Backup Integrity</td><td>Corruption detected late (at restore or via periodic checks)</td><td>Continuous Recovery Validation (proactive, during backup cycle <sup></sup>)</td></tr><tr><td>Management Complexity</td><td>Complex scripting, scheduling, manual validation</td><td>Centralized appliance management, automated validation and virtual full creation</td></tr></tbody></table></figure>



<p>This table visually reinforces how ZDLRA directly addresses specific, long-standing pain points in VLDB management, making it easier to quickly grasp the benefits that would justify evaluating such a system.</p>



<h2 class="wp-block-heading">4. Key Considerations and Best Practices in ZDLRA-Centric VLDB Backup and Recovery Implementations</h2>



<p>While ZDLRA offers powerful capabilities that significantly improve VLDB backup and recovery processes, fully leveraging these capabilities requires careful planning, configuration, and adherence to operational best practices. This section will translate ZDLRA&#8217;s features into actionable considerations and best practices. Although the provided Oracle blog post summaries indicate they do not offer <em>additional</em> general best practices beyond ZDLRA itself <sup></sup>, this section will focus on <em>how best to leverage</em> ZDLRA&#8217;s capabilities and what to pay attention to <em>within the ZDLRA context</em>. &nbsp;</p>



<h3 class="wp-block-heading">4.1. Optimizing Recovery Time Objective (RTO) with ZDLRA</h3>



<ul class="wp-block-list">
<li>Leverage ZDLRA&#8217;s virtual full backups for rapid restores. This significantly shortens recovery times.</li>



<li>Ensure ZDLRA sizing is adequate to meet restore performance demands. Insufficient resources can lead to missed RTO targets.</li>



<li>Regularly test recovery scenarios to validate RTOs. ZDLRA&#8217;s &#8220;Continuous Recovery Validation&#8221;  ensures backups are valid, a prerequisite for meeting RTOs, but real-world tests confirm the entire process works as expected.  </li>
</ul>



<h3 class="wp-block-heading">4.2. Minimizing Recovery Point Objective (RPO) with ZDLRA</h3>



<ul class="wp-block-list">
<li>Implement and monitor real-time redo transport diligently. This is the primary mechanism for achieving near-zero RPO according to.  </li>



<li>Understand and meet network requirements to ensure real-time redo transport does not lag. Insufficient network bandwidth or high latency can compromise RPO targets.</li>
</ul>



<h3 class="wp-block-heading">4.3. Managing Production System Performance</h3>



<ul class="wp-block-list">
<li>While ZDLRA&#8217;s &#8220;Incremental Forever&#8221; strategy  significantly reduces production impact, confirm this by monitoring baseline database performance metrics post-implementation.  </li>



<li>Optimize network bandwidth between production databases and the ZDLRA. This is critical for the efficiency of both incremental backups and real-time redo transport.</li>
</ul>



<h3 class="wp-block-heading">4.4. Ensuring Backup Data Integrity and Reliability</h3>



<ul class="wp-block-list">
<li>Rely on ZDLRA&#8217;s &#8220;Continuous Recovery Validation&#8221; , but also understand its alerting mechanisms and integrate them into operational monitoring systems. Early warnings allow for proactive resolution of potential issues.  </li>



<li>Consider ZDLRA replication to a secondary ZDLRA for disaster recovery of the backup data itself. This ensures backups are protected even if the primary ZDLRA fails.</li>
</ul>



<h3 class="wp-block-heading">4.5. Storage Management and Efficiency within ZDLRA</h3>



<ul class="wp-block-list">
<li>Understand ZDLRA&#8217;s internal storage management, space reclamation, and how the &#8220;up to 10X decrease in space consumption&#8221;  is achieved and monitored.  </li>



<li>Plan retention policies carefully to balance recovery needs with storage capacity. Overly long retention periods can lead to unnecessary costs, while too short retention can limit recovery capabilities.</li>
</ul>



<h3 class="wp-block-heading">4.6. Network Configuration and Sizing</h3>



<ul class="wp-block-list">
<li>Emphasize the importance of dedicated, high-bandwidth, low-latency network connectivity between production databases and the ZDLRA, especially for real-time redo transport and large data transfers. The network should not be a bottleneck for backup and recovery performance.</li>
</ul>



<h3 class="wp-block-heading">4.7. Regular Testing and Validation of Recovery Procedures</h3>



<ul class="wp-block-list">
<li>Even with ZDLRA&#8217;s automation and validation, conduct periodic, full recovery drills to test the end-to-end process, human procedures, and infrastructure. This validates the entire recovery plan, not just the technology.</li>
</ul>



<p>Implementing ZDLRA is not a &#8220;set it and forget it&#8221; solution. While it automates and optimizes many aspects, careful planning, configuration, ongoing monitoring, and testing are still critical to realizing its full benefits. The &#8220;best practices&#8221; shift from managing the intricacies of RMAN scripts to managing the ZDLRA ecosystem. ZDLRA offers advanced features like &#8220;Incremental Forever,&#8221; &#8220;Real-Time Redo,&#8221; and &#8220;Continuous Validation.&#8221; These features have prerequisites and operational aspects (e.g., network for redo, monitoring alerts for validation, capacity planning for storage). Simply deploying the appliance does not guarantee optimal RTO/RPO or reliability. Administrators must understand how these features work, configure them correctly, monitor their performance, and integrate ZDLRA into broader DR and operational procedures. Regular testing is essential to confirm the entire system (database, network, ZDLRA, recovery procedures) performs as expected under pressure. The role of the Database Administrator (DBA) also evolves in this context. They may spend less time on low-level backup scripting and more on strategic data protection management for ZDLRA, capacity planning, and ensuring end-to-end recoverability of business services. Expertise specific to ZDLRA itself becomes important.</p>



<h2 class="wp-block-heading">5. Conclusion and Recommendations</h2>



<p>As presented in the Oracle MAA blog posts, the Zero Data Loss Recovery Appliance (ZDLRA) offers significant advantages in the realm of Very Large Database (VLDB) backup and recovery. These benefits include near-zero data loss through a vastly improved Recovery Point Objective (RPO), reliable Recovery Time Objective (RTO) via virtual full backups and continuous validation, reduced impact on production systems, and enhanced data integrity. <sup></sup> &nbsp;</p>



<p>As an engineered system, ZDLRA represents a strategic approach to tackling the complexities of VLDB protection. The co-engineering of hardware and software allows for performance and reliability optimizations that are difficult to achieve with general-purpose solutions. This is a critical differentiator, especially in today&#8217;s environment where data volume and transaction rates challenge traditional methods.</p>



<p>However, it must be emphasized that while ZDLRA offers powerful capabilities, successful implementation requires careful planning, a full understanding of its features, and adherence to operational best practices, particularly concerning network configuration, monitoring, and regular recovery testing. Adopting ZDLRA is not merely a technical decision but signifies a commitment to a high level of data protection and availability, driven by the critical nature of the VLDBs it protects. This is an investment that should align with the value of the data and the cost of downtime/data loss.</p>



<p>It is important to note that this report focuses on ZDLRA-centric best practices highlighted in the provided Oracle blog post summaries. A comprehensive discussion of all VLDB backup and recovery techniques, including non-ZDLRA alternatives or complementary strategies like storage snapshots or Oracle Data Guard for DR beyond backup, would require additional resources beyond the scope of the provided material.</p>



<p>In conclusion, organizations should evaluate ZDLRA as part of their overall IT strategy, considering its integration with other systems, the skills required to manage it, and its alignment with long-term data growth and protection needs. When implemented and managed correctly, ZDLRA can provide unparalleled protection and recovery assurance for VLDBs, helping businesses secure one of their most valuable assets: their data.</p>



<p>Ref:</p>



<p><a href="https://blogs.oracle.com/maa/post/very-large-database-backup-and-recovery-best-practices" target="_blank" rel="noopener">https://blogs.oracle.com/maa/post/very-large-database-backup-and-recovery-best-practices</a></p>



<p><a href="https://blogs.oracle.com/maa/post/very-large-database-backup-and-recovery-best-practices-part-2" target="_blank" rel="noopener">https://blogs.oracle.com/maa/post/very-large-database-backup-and-recovery-best-practices-part-2</a></p>



<p></p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/best-practices-for-very-large-database-vldb-backup-and-recovery.html">Best Practices for Very Large Database (VLDB) Backup and Recovery:</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Analysis of Delta Push and Delta Store Mechanisms within ZDLRA</title>
		<link>https://www.bugraparlayan.com.tr/analysis-of-delta-push-and-delta-store-mechanisms-within-zdlra.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Wed, 04 Jun 2025 19:35:29 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[Backup Optimization]]></category>
		<category><![CDATA[Block Change Tracking]]></category>
		<category><![CDATA[Change Tracking]]></category>
		<category><![CDATA[Delta Compression]]></category>
		<category><![CDATA[Delta Push]]></category>
		<category><![CDATA[Delta Store]]></category>
		<category><![CDATA[Incremental Forever]]></category>
		<category><![CDATA[Oracle ZDLRA]]></category>
		<category><![CDATA[Recovery Appliance]]></category>
		<category><![CDATA[RMAN Delta push]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1475</guid>

					<description><![CDATA[<p>I. Introduction to Oracle Zero Data Loss Recovery Appliance (ZDLRA) Technology The Oracle Zero Data Loss Recovery Appliance (ZDLRA or Recovery Appliance), an engineered system specifically designed for Oracle Databases, was developed to eliminate data loss and significantly reduce the data protection workload on production database servers. Its primary goal is to protect transactions in &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/analysis-of-delta-push-and-delta-store-mechanisms-within-zdlra.html">Analysis of Delta Push and Delta Store Mechanisms within ZDLRA</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading">I. Introduction to Oracle Zero Data Loss Recovery Appliance (ZDLRA) Technology</h2>



<p>The Oracle Zero Data Loss Recovery Appliance (ZDLRA or Recovery Appliance), an engineered system specifically designed for Oracle Databases, was developed to eliminate data loss and significantly reduce the data protection workload on production database servers.<sup></sup> Its primary goal is to protect transactions in real-time, enabling databases to be recovered to within less than a second in the event of an outage or ransomware attack.<sup></sup> This approach fundamentally differs from traditional backup solutions, which often lead to data loss measured in hours or even a day.<sup></sup> ZDLRA works in tight integration with Oracle Database and Recovery Manager (RMAN) <sup></sup> and offers capabilities not possible with general-purpose backup solutions. It is built upon Exadata hardware, from which it inherits performance and scalability features.<sup></sup> Positioned for modern cybersecurity protection, ZDLRA offers features like backup immutability and continuous validation.<sup></sup> &nbsp;</p>



<p>The data protection philosophy underlying this system represents a paradigm shift from reactive backup to proactive, continuous data protection. Traditional backup operations are performed periodically (e.g., nightly), which inherently carries the potential for data loss since the last backup.<sup></sup> ZDLRA, on the other hand, captures changes as they occur through mechanisms like &#8220;real-time protection&#8221; <sup></sup> and &#8220;real-time redo transport&#8221;.<sup></sup> This continuous capture reduces the Recovery Point Objective (RPO) to sub-second levels <sup></sup>, significantly mitigating the business risk associated with data loss and moving beyond mere backup to provide near-continuous data assurance. &nbsp;</p>



<p>The fact that ZDLRA is an &#8220;engineered system&#8221; (built on Exadata <sup></sup>) is critical to its performance and reliability. General-purpose backup solutions run on general-purpose hardware and software, which may not be optimized for the unique I/O patterns and metadata intensity of Oracle Database backups.<sup></sup> As an engineered system like Exadata, ZDLRA&#8217;s hardware (storage, nodes, InfiniBand <sup></sup>) and software are co-engineered and pre-tuned for Oracle Database workloads.<sup></sup> This co-engineering enables high throughput (up to 60 TB per hour per rack <sup></sup>), efficient handling of Oracle-specific block formats, and the scalability required for enterprise-wide database protection.<sup></sup> Thus, ZDLRA&#8217;s effectiveness stems not just from its software features but from its holistic system design, purpose-built for Oracle databases. &nbsp;</p>



<h3 class="wp-block-heading">A. &#8220;Incremental Forever&#8221; Backup Strategy: A Paradigm Shift</h3>



<p>ZDLRA implements an &#8220;incremental-forever&#8221; backup architecture.<sup></sup> After an initial one-time full (level 0) backup, only incremental (level 1) backups are sent from the protected databases to the Recovery Appliance.<sup></sup> This strategy eliminates the need for recurring full backups, which are resource-intensive on production systems and can impact application performance.<sup></sup> The &#8220;incremental forever&#8221; approach significantly reduces backup windows, database server load (CPU, memory, I/O), and network traffic.<sup></sup> &nbsp;</p>



<p>The &#8220;incremental forever&#8221; strategy is more than just sending fewer full backups; it is enabled by ZDLRA&#8217;s sophisticated backend processing (Delta Store) that synthesizes these incremental backups into readily usable &#8220;virtual full backups.&#8221; Simply sending only incremental backups without a mechanism to consolidate them would make restores complex and slow, requiring the sequential application of many incrementals.<sup></sup> ZDLRA overcomes this by using the Delta Store to process incoming incremental backups and create &#8220;virtual full backups&#8221;.<sup></sup> These virtual full backups are representations of a full backup at a specific point in time, constructed from the initial level 0 and subsequent level 1 block changes.<sup></sup> This means that for recovery, RMAN can restore a single virtual full backup without the burden of applying numerous incremental backups on the client side <sup></sup>, making the &#8220;incremental forever&#8221; strategy practical and highly efficient for recovery. &nbsp;</p>



<p>The following table summarizes the key concepts of ZDLRA and the benefits it offers:</p>



<p><strong>Table 1: Key ZDLRA Concepts and Benefits</strong></p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Concept</th><th>Brief Definition in ZDLRA Context</th><th>Key Benefit to the Organization</th></tr><tr><td>Zero Data Loss</td><td>Goal of reducing data loss to virtually zero by protecting database transactions in real-time.</td><td>Minimizes critical data loss risk, enhances business continuity.</td></tr><tr><td>Incremental Forever</td><td>Only incremental backups are taken after the initial full backup, eliminating the need for periodic full backups.</td><td>Shortens backup windows, reduces load on production systems, saves storage.</td></tr><tr><td>Real-Time Redo Transport</td><td>Instantaneous transfer of database redo log changes to ZDLRA.</td><td>Provides sub-second RPO (Recovery Point Objective), minimizing data loss.</td></tr><tr><td>Virtual Full Backup</td><td>A logical backup synthesized on ZDLRA from incremental backups, behaving like a full backup.</td><td>Enables fast restores, uses storage space efficiently.</td></tr><tr><td>Sub-Second RPO</td><td>Reduction of data loss tolerance to below one second.</td><td>Minimizes data loss impact for business-critical applications.</td></tr><tr><td>Continuous Validation</td><td>Backup integrity and recoverability are continuously checked by ZDLRA.</td><td>Ensures reliable restores, reduces risk of corrupt backups.</td></tr></tbody></table></figure>



<h2 class="wp-block-heading">II. Delta Push: The Continuous Data Ingestion Engine</h2>



<h3 class="wp-block-heading">A. Delta Push Concept and Objectives</h3>



<p>&#8220;Delta Push&#8221; is a term Oracle uses to describe the process by which protected databases send only the minimum necessary data (i.e., the &#8220;delta difference&#8221; or changes) to ZDLRA for protection.<sup></sup> This process encompasses both incremental backups of data blocks and real-time transport of redo log changes.<sup></sup> The primary objective is to minimize the impact on production systems by transmitting only unique changed data, thereby reducing CPU, I/O, and network load.<sup></sup> It is a source-side optimization enabled by RMAN block change tracking and tight integration with the Oracle database.<sup></sup> &nbsp;</p>



<p>&#8220;Delta Push&#8221; is more than just an incremental backup; it is a holistic strategy to capture <em>all</em> relevant database changes (data blocks and redo) with minimal production impact, forming the ingestion mechanism for ZDLRA&#8217;s continuous data protection. Traditional incremental backups capture changed data blocks at set intervals.<sup></sup> Real-time redo transport continuously captures transaction log changes, even between incremental backups.<sup></sup> While <sup></sup> explicitly states, &#8220;Changes in the database are sent to ZDLRA using the Delta Push process,&#8221; <sup></sup> clarifies, &#8220;Oracle calls Virtual Backups + Real-Time Redo as Delta Push.&#8221; Therefore, Delta Push is not a single technology but a combination of RMAN incremental backups and real-time redo transport working in concert to ensure comprehensive and timely capture of all database modifications. This dual approach is key to achieving both efficient backups and near-zero RPO. &nbsp;</p>



<h3 class="wp-block-heading">B. Operational Mechanisms</h3>



<h4 class="wp-block-heading">1. Leveraging RMAN Incremental Backups</h4>



<p>ZDLRA utilizes the RMAN &#8220;incremental backup&#8221; API to capture changes from the source database.<sup></sup> After the initial level 0 backup, all subsequent backups are level 1 incremental backups.<sup></sup> These can be cumulative backups that use the latest virtual level 0 as their baseline.<sup></sup> The ZDLRA Backup Module (an SBT library) facilitates the transfer of these incremental backups from the protected database to the Recovery Appliance.<sup></sup> RMAN block change tracking on the source database efficiently identifies changed blocks, so only those blocks are read and sent.<sup></sup> &nbsp;</p>



<p>ZDLRA fundamentally transforms the <em>purpose</em> and <em>outcome</em> of an RMAN incremental backup. While the RMAN command on the client-side might look similar, ZDLRA processes it not just as a standalone incremental backup, but as a component of a virtual full backup. Traditionally, an RMAN incremental backup is a set of blocks changed since the last backup of a certain level.<sup></sup> However, <sup></sup> states, &#8220;When a DELTA PUSH is executed, the results are automatically transformed into a VIRTUAL FULL backup in what is known as the Delta Store inside of ZDLRA.&#8221; <sup></sup> explains that ZDLRA opens the RMAN block, reads the data file blocks within it, and creates a new virtual level 0 using the backup already existing on ZDLRA. This indicates that ZDLRA doesn&#8217;t just store the incremental backup set; it actively parses it and integrates its constituent changed blocks into its versioned block store (Delta Store) to synthesize a new point-in-time full representation. This is a critical difference from how incremental backups are handled by traditional backup software. &nbsp;</p>



<h4 class="wp-block-heading">2. Real-Time Redo Transport: Achieving Sub-Second RPO</h4>



<p>This feature is key to ZDLRA&#8217;s &#8220;zero data loss&#8221; claim, providing a zero to sub-second RPO.<sup></sup> Redo data (records of all database changes) is streamed from the protected database&#8217;s memory buffers (LGWR) directly to the Recovery Appliance, typically asynchronously to minimize performance impact.<sup></sup> This is similar to Data Guard redo transport.<sup></sup> ZDLRA validates the redo and writes it to a staging area. Upon a log switch on the protected database, ZDLRA converts these redo changes into compressed archived redo log file backups and tracks them in its catalog.<sup></sup> If the redo stream terminates unexpectedly (e.g., database crash), ZDLRA can create a partial archived redo log, protecting transactions up to the last change received.<sup></sup> With real-time redo transport enabled, this obviates the need for separate archived log backups from the database host to ZDLRA.<sup></sup> &nbsp;</p>



<p>Real-time redo transport effectively decouples redo protection from the source database&#8217;s archiving process, enabling more resilient and immediate capture of transactions. Traditional redo protection often relies on the database successfully writing to its online redo logs and then archiving them.<sup></sup> ZDLRA&#8217;s real-time redo transport taps into the redo stream <em>before</em> or concurrently with local archiving, sending it directly from memory.<sup></sup> Even if the primary database crashes before successfully archiving a log, ZDLRA can construct a partial archive log from the redo it has already received.<sup></sup> This means ZDLRA acts as an independent, highly available redo log destination, guaranteeing transaction capture even if the source database&#8217;s own archiving mechanism is disrupted, which is critical for sub-second RPO. &nbsp;</p>



<h3 class="wp-block-heading">C. Architectural Integration: Data Flow from Protected Database to ZDLRA</h3>



<p>Protected databases use the RMAN client and the Recovery Appliance Backup Module (SBT library) to communicate with ZDLRA.<sup></sup> For incremental backups, RMAN identifies changed blocks. These blocks are packaged and sent via the backup module over the network to an HTTP Server Application (Servlet) on ZDLRA.<sup></sup> Real-time redo is transported similarly to Data Guard (typically via Oracle Net) to a Remote File Server (RFS) process on ZDLRA.<sup></sup> ZDLRA then validates, processes (compression, indexing), and stores the incoming data/redo blocks in the Delta Store.<sup></sup> Metadata is updated in the Recovery Appliance Catalog.<sup></sup> &nbsp;</p>



<p>The data flow architecture for Delta Push is bifurcated (separate paths for incremental block backups and real-time redo) but converges within ZDLRA to provide a unified data protection state. Incremental data block backups are inherently batch-oriented, typically scheduled RMAN operations, even if frequent.<sup></sup> They are processed via the SBT interface.<sup></sup> Real-time redo transport is a continuous, stream-based process, capturing transactional changes as they occur using Data Guard-like mechanisms.<sup></sup> Both data streams—changed blocks and redo records—arrive at ZDLRA and are processed into and cataloged by the Delta Store.<sup></sup> This dual-path ingestion allows ZDLRA to capture both the state of data blocks at specific points in time (via incrementals) and the continuous flow of transactions (via redo), combining the strengths of snapshot-like backups and continuous data replication to enable recovery to almost any point in time. &nbsp;</p>



<h3 class="wp-block-heading">D. Formation of Virtual Full Backups via Delta Push</h3>



<p>While Delta Push is the <em>mechanism</em> for sending changes, its direct result is to enable ZDLRA&#8217;s Delta Store to create and maintain &#8220;Virtual Full Backups&#8221;.<sup></sup> Each Delta Push (incremental backup) operation results in a new Virtual Full Backup becoming available in the ZDLRA catalog.<sup></sup> This means changes are tracked not just from the last physical full backup, but from the previous Virtual Full backup.<sup></sup> &nbsp;</p>



<p>Delta Push acts as the continuous feed that allows the Delta Store to maintain a constantly up-to-date, yet historically deep, set of recovery points represented as virtual full backups. Delta Push transmits the &#8220;deltas&#8221; – the changed blocks.<sup></sup> The Delta Store receives these deltas and intelligently integrates them with previously stored block versions.<sup></sup> This integration allows ZDLRA to construct a logically complete backup image for a specific point-in-time corresponding to an ingested incremental backup.<sup></sup> Therefore, Delta Push is not just about efficient data transfer; it is the critical data pipeline that fuels the Delta Store&#8217;s ability to offer fast, point-in-time recovery through virtual full backups, effectively creating a &#8220;time machine&#8221; for the database.<sup></sup> &nbsp;</p>



<h2 class="wp-block-heading">III. Delta Store: The Intelligent Repository for Protected Data</h2>



<h3 class="wp-block-heading">A. Delta Store Architectural Overview</h3>



<p>The Delta Store is &#8220;the totality of all protected database backup data in the Recovery Appliance storage location&#8221;.<sup></sup> It resides on a dedicated ASM disk group (typically named DELTA) on ZDLRA.<sup></sup> It is described as the &#8220;brains&#8221; of the Recovery Appliance, responsible for validating, compressing, indexing, and storing incoming backup data.<sup></sup> It is not merely a passive storage area; it actively manages backup data to enable efficient virtual full backups and space optimization.<sup></sup> &nbsp;</p>



<p>The Delta Store is an application-aware storage layer, deeply integrated with Oracle Database block structures and RMAN metadata, which distinguishes it from general-purpose deduplication appliances. General-purpose deduplication appliances typically operate at a generic block level without understanding the internal structure of database files.<sup></sup> According to <sup></sup>, ZDLRA&#8217;s Delta Store captures copies of each Oracle block and organizes them hierarchically. <sup></sup> highlights its &#8220;Oracle context sensitivity,&#8221; where it opens RMAN blocks to inspect their contents and index the data blocks for each data file. This database awareness allows for more intelligent deduplication (block versioning rather than just hash-based deduplication of RMAN backup pieces), validation <sup></sup>, and the creation of consistent virtual full backups. This intelligence is what enables ZDLRA to perform block-correctness and RMAN recoverability validation <sup></sup> directly on the appliance, offloading the production server. &nbsp;</p>



<h3 class="wp-block-heading">B. Internal Structure and Data Organization</h3>



<h4 class="wp-block-heading">1. Delta Pools: Granular Management of Data File Backups</h4>



<p>The Delta Store contains &#8220;delta pools&#8221; for all data files across all protected databases.<sup></sup> A delta pool is the set of data file blocks from which the Recovery Appliance constructs virtual full backups. Each distinct data file whose backups are sent to ZDLRA has its own dedicated delta pool.<sup></sup> For example, <code>datafile 10</code> from database <code>prod1</code> has its own delta pool.<sup></sup> &nbsp;</p>



<p>The concept of a delta pool signifies a highly granular and organized approach to managing backup data, enabling efficient block versioning and retrieval at the individual data file level. Databases consist of multiple data files, each with its own lifecycle of changes. By maintaining a separate delta pool per data file <sup></sup>, ZDLRA can track and version blocks specifically for that file. When an incremental backup for a data file arrives, ZDLRA updates the relevant delta pool with the new block versions. This level of granularity is essential for constructing a virtual full backup, as ZDLRA can quickly locate the correct versions of all blocks for each data file belonging to a specific point-in-time backup by querying these distinct pools. It also likely aids in space management and reclamation, as old blocks can be managed within the context of their specific data file pool. &nbsp;</p>



<h4 class="wp-block-heading">2. Block Versioning and Indexing Mechanisms</h4>



<p>The Delta Store is effectively a database of block versions.<sup></sup> As incremental backups (Delta Pushes) arrive, the changed blocks are indexed into the Delta Store.<sup></sup> The Recovery Appliance receives an incremental backup, validates it, compresses it, and writes it to a delta store. It indexes the backup so that corresponding virtual full backups become available.<sup></sup> The ZDLRA metadata database, which includes the RMAN recovery catalog, manages the metadata about these blocks and their versions.<sup></sup> &nbsp;</p>



<p>The indexing of individual database blocks within the Delta Store, not just backup pieces, is the core enabler of &#8220;virtual full backups&#8221; and efficient space utilization. Traditional backups store entire backup pieces (full or incremental). Restoration requires locating and processing these pieces.<sup></sup> ZDLRA, in contrast, extracts individual data blocks from incoming incremental backups and indexes these blocks.<sup></sup> The Delta Store maintains various versions of these blocks. A &#8220;virtual full backup&#8221; is essentially a metadata construct – a list of pointers to the correct versions of all blocks (from various delta pools) that constitute the database at a specific point in time.<sup></sup> This block-level versioning and indexing mean that unchanged blocks are stored only once, and new &#8220;full&#8221; backups are created logically by updating pointers, rather than physically re-copying all data. This is the essence of space efficiency and rapid virtual full creation. &nbsp;</p>



<h3 class="wp-block-heading">C. Creation and Management of Virtual Full Backups within Delta Store</h3>



<p>The Delta Store uses the ingested incremental backups (via Delta Push) to create virtual full backups.<sup></sup> A virtual full backup is a pointer-based representation of a physical full backup at the time of the incremental backup.<sup></sup> It appears as a standard level 0 backup in the RMAN catalog.<sup></sup> To create a virtual full backup, ZDLRA converts an incoming incremental level 1 backup into a virtual representation of an incremental level 0 backup.<sup></sup> It combines the new changed blocks from the incremental backup with the previous unchanged blocks already present in the Delta Store.<sup></sup> These virtual full backups are typically 10 times more space-efficient than physical full backups.<sup></sup> &nbsp;</p>



<p>The creation of virtual full backups is an ongoing, dynamic process within the Delta Store, triggered by each successful Delta Push, ensuring that the latest recovery points are always full representations. <sup></sup> states, &#8220;Each Delta Push sends the latest version of each changed block. Those changed blocks are indexed into the Delta Store and combined with previous un-changed blocks to form a Virtual Full Backup.&#8221; <sup></sup> notes, &#8220;After the process [backup], the catalog reflects all the new virtual full backups that are available.&#8221; This implies a continuous synthesis. As new incremental data arrives, the Delta Store doesn&#8217;t just store the incremental; it actively processes it to update its pointers and metadata, making a new, comprehensive virtual full backup immediately available. This proactive synthesis is why ZDLRA can offer fast restores to recent points in time without the delay of manually applying many incrementals during the restore operation itself. The &#8220;merge&#8221; or &#8220;synthesis&#8221; happens upfront on ZDLRA. &nbsp;</p>



<h3 class="wp-block-heading">D. Storage Optimization and Efficiency</h3>



<h4 class="wp-block-heading">1. Advanced Compression Techniques (including <code>RA_FORMAT</code>)</h4>



<p>ZDLRA employs specialized block-level compression algorithms.<sup></sup> A newer client-side library feature, <code>RA_FORMAT=TRUE</code> (introduced around ZDLRA 23.1), allows for compression of the data <em>within</em> blocks before sending to ZDLRA. This is compatible with ZDLRA&#8217;s ability to create virtual full backups and validate stored backup sets.<sup></sup> This client-side compression can compress the contents of TDE encrypted blocks as well as non-TDE blocks. If RMAN encryption is also on, non-TDE blocks are compressed then encrypted.<sup></sup> This compression reduces network bandwidth for backups and replication, and storage space on ZDLRA.<sup></sup> Archive log compression (BASIC, LOW, MEDIUM, HIGH) can also be configured; LOW, MEDIUM, and HIGH do not require ACO on the protected database when ZDLRA is used.<sup></sup> &nbsp;</p>



<p>The <code>RA_FORMAT</code> feature represents a significant evolution in ZDLRA&#8217;s compression strategy, moving some intelligence to the client to optimize data <em>before</em> transmission and storage, and enabling effective compression even for TDE-encrypted data. Previously, RMAN compression would compress the entire backup set. If the data was TDE encrypted, this compressed backup set was unreadable by ZDLRA for its block-level operations.<sup></sup> <code>RA_FORMAT=TRUE</code> compresses the <em>contents</em> of each block, leaving the block headers intact for ZDLRA to read.<sup></sup> This allows ZDLRA to perform its virtual full backup creation and validation even on backups originating from TDE tablespaces, because the data within the blocks is compressed, but the block structure ZDLRA needs is preserved. This overcomes a major challenge in backup efficiency for encrypted databases, offering both security (TDE) and storage/network efficiency (compression), which were often mutually exclusive or suboptimal with older methods. &nbsp;</p>



<h4 class="wp-block-heading">2. Automated Space Management and Delta Pool Optimization</h4>



<p>The Recovery Appliance performs automated delta pool space management.<sup></sup> This includes deleting old or expired backups (on disk and on tape/cloud) based on recovery window goals and retention policies.<sup></sup> ZDLRA periodically reorganizes delta pools to improve restore performance by maintaining contiguity of blocks (delta pool optimization) as old blocks are deleted and new ones arrive.<sup></sup> &nbsp;</p>



<p>Automated space management and delta pool optimization are critical for sustaining the long-term performance and efficiency of the &#8220;incremental forever&#8221; strategy. An unmanaged &#8220;incremental forever&#8221; system could lead to highly fragmented storage over time as myriad small changes accumulate and old data becomes obsolete. The deletion of old blocks <sup></sup> reclaims space, vital for cost-effectiveness. The reorganization of delta pools <sup></sup> addresses the potential performance degradation from fragmentation that could arise from frequent updates and deletions, ensuring restore operations remain fast by optimizing read access. These automated background tasks are therefore essential for the sustainability of the ZDLRA model, preventing it from becoming unwieldy or slow over long periods of operation. &nbsp;</p>



<p>The following table summarizes the internal components of the Delta Store and their roles:</p>



<p><strong>Table 2: Delta Store Internal Components and Roles</strong></p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Component</th><th>Description/Structure</th><th>Primary Function within Delta Store</th><th>Contribution to ZDLRA Efficiency/Recovery</th></tr><tr><td>Delta Pool</td><td>A logical unit for each data file, containing all backed-up block versions for that specific data file.</td><td>Organizing and managing blocks belonging to a specific data file.</td><td>Granular block management, efficient versioning, and rapid construction of virtual full backups.</td></tr><tr><td>Block Version</td><td>A copy of a data block at a specific point in time.</td><td>Tracking data changes over time.</td><td>Space efficiency (only changed blocks stored), ability to restore to any point in time.</td></tr><tr><td>Index</td><td>Metadata structure tracking the locations and versions of blocks within the Delta Store.</td><td>Enabling rapid location of correct block versions when constructing virtual full backups.</td><td>Fast virtual full backup creation, efficient restore operations.</td></tr><tr><td>Virtual Full Backup Metadata</td><td>Set of pointers to the block versions that constitute a full backup at a specific point in time.</td><td>Providing a logical representation of a physical full backup.</td><td>Storage efficiency (pointers instead of physical full backups), appears as a standard level 0 backup to RMAN, fast recovery.</td></tr></tbody></table></figure>



<h2 class="wp-block-heading">IV. Synergistic Architecture: Operation of Delta Push and Delta Store within ZDLRA</h2>



<h3 class="wp-block-heading">A. End-to-End Data Protection Workflow: From Transaction to Recoverable Backup</h3>



<p>The data protection process begins when a transaction occurs on the protected database. These changes are transmitted to ZDLRA almost instantaneously as part of the Delta Push mechanism.</p>



<ol class="wp-block-list">
<li><strong>Transaction Occurs:</strong> Changes are made in the protected database.</li>



<li><strong>Real-Time Redo Push:</strong> LGWR (or an asynchronous process) sends redo data from memory buffers to ZDLRA. ZDLRA stages and validates this redo.  </li>



<li><strong>Incremental Backup (Delta Push):</strong> Periodically , RMAN performs an incremental level 1 backup. Changed blocks are identified via block change tracking.  </li>



<li><strong>Data Transfer:</strong> The Recovery Appliance Backup Module sends these changed blocks to ZDLRA.  </li>



<li><strong>ZDLRA Ingestion and Processing (Delta Store):</strong>
<ul class="wp-block-list">
<li>Incoming incremental blocks are validated, compressed (if <code>RA_FORMAT=TRUE</code> or with ZDLRA-side compression), and indexed into their respective delta pools within the Delta Store.  </li>



<li>The Delta Store synthesizes a new virtual full backup using these new blocks and existing unchanged blocks.  </li>



<li>Redo logs are converted by ZDLRA into archived log backups upon log switch.  </li>
</ul>
</li>



<li><strong>Catalog Update:</strong> ZDLRA&#8217;s internal RMAN catalog is updated to reflect the new virtual full backup and archived redo logs, making them available for recovery.  </li>



<li><strong>Continuous Validation:</strong> ZDLRA continuously validates backups for recoverability.  </li>



<li><strong>Lifecycle Management:</strong> Policies for retention, replication to another ZDLRA, or archival to tape/cloud are applied.  </li>
</ol>



<p>The synergy between Delta Push (ingestion) and Delta Store (processing and storage engine) creates a closed-loop system for continuous data protection and recovery readiness. Delta Push continuously feeds changed data (blocks and redo) to ZDLRA.<sup></sup> The Delta Store immediately processes this data, integrates it into its versioned block repository, and creates virtual full backups.<sup></sup> The updated catalog then makes these new recovery points instantly available.<sup></sup> This tight, automated loop ensures ZDLRA is always as up-to-date as possible with the state of protected databases, minimizing data loss risk and guaranteeing that recovery assets are constantly refreshed and validated. &nbsp;</p>



<h3 class="wp-block-heading">B. Role of Key ZDLRA Components</h3>



<h4 class="wp-block-heading">1. Recovery Appliance Backup Module (libra.so / SBT Library)</h4>



<p>This Oracle-supplied SBT library is installed on protected database hosts and is used by RMAN to transfer backup data to ZDLRA.<sup></sup> It manages communication for backup and restore operations between RMAN and ZDLRA.<sup></sup> With newer versions (e.g., ZDLRA 23.1), this library can perform client-side compression and formatting (<code>RA_FORMAT=TRUE</code>).<sup></sup> &nbsp;</p>



<p>The Recovery Appliance Backup Module is more than a simple data pipe; it&#8217;s an intelligent client-side agent that actively participates in optimizing the backup stream. Traditionally, SBT libraries are primarily interfaces for RMAN to write to third-party media managers. The ZDLRA backup module, especially with features like <code>RA_FORMAT</code>, performs pre-processing (compression, ZDLRA-specific formatting <sup></sup>) on the client-side. This client-side intelligence reduces load on ZDLRA for certain tasks, optimizes network traffic, and enables advanced features like effectively compressing TDE data before it even reaches the appliance. It acts as an essential, integrated part of the ZDLRA solution, not just a generic connector. &nbsp;</p>



<h4 class="wp-block-heading">2. Recovery Appliance Metadata Database and Catalog</h4>



<p>Residing on each Recovery Appliance, it manages metadata for all backups and contains the RMAN recovery catalog for all protected databases.<sup></sup> This catalog is mandatory and is automatically updated by ZDLRA as backups are processed.<sup></sup> It stores information about backup pieces, archived logs, virtual full backups, delta pools, and block versions, which is essential for orchestrating restores and managing space.<sup></sup> ZDLRA uses two main disk groups: DELTA for backups and CATALOG for RMAN catalog tables.<sup></sup> &nbsp;</p>



<p>ZDLRA&#8217;s centralized and self-managing RMAN catalog serves as the &#8220;single source of truth&#8221; for all protected databases, enabling simplified management and consistent recovery across the enterprise. In traditional environments, RMAN catalogs might be separate or controlfile-based, leading to management complexity for many databases. ZDLRA mandates and manages a central catalog within its own embedded RAC database.<sup></sup> This catalog automatically reflects all virtual full backups and other recovery assets created by ZDLRA.<sup></sup> Database administrators (DBAs) interact with this catalog via standard RMAN commands for restores, without needing to know the internal complexities of virtual backups or delta pools.<sup></sup> This centralization and automation significantly simplify backup administration, especially in large environments. &nbsp;</p>



<h3 class="wp-block-heading">C. Control Flow and Policy Enforcement</h3>



<p>Protection policies are defined on ZDLRA to manage recovery window goals, data retention periods on disk and tape/cloud, replication, and other backup lifecycle aspects.<sup></sup> These policies are applied to protected databases. ZDLRA&#8217;s automated space management tasks (deletion of old backups, delta pool optimization) are driven by these policies.<sup></sup> Enterprise Manager Cloud Control is typically used to manage and monitor ZDLRA and its policies.<sup></sup> &nbsp;</p>



<p>ZDLRA&#8217;s policy-based management automates much of the backup lifecycle, abstracting complexity and ensuring adherence to defined service levels. Manual management of backup retention, replication, and tiering for hundreds of databases is error-prone and labor-intensive. ZDLRA allows administrators to define high-level protection policies (e.g., Gold, Silver, Bronze with different RPOs/retention <sup></sup>). The appliance then automatically enforces these policies, managing space, creating virtual full backups, replicating data, and archiving to secondary storage.<sup></sup> This automation ensures consistency, reduces administrative burden, and helps organizations meet their data protection SLAs reliably. &nbsp;</p>



<p>The following table illustrates the interactive workflow between Delta Push and Delta Store step-by-step:</p>



<p><strong>Table 3: Delta Push and Delta Store Interaction Workflow</strong></p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Step No.</th><th>Action/Process</th><th>Responsible Component(s)</th><th>Key Outcome of the Step</th></tr><tr><td>1</td><td>Database Change</td><td>Protected Database</td><td>Data is modified.</td></tr><tr><td>2</td><td>Redo Sent</td><td>Protected Database (LGWR), ZDLRA (Delta Push Receiver)</td><td>Real-time redo data is transferred to and staged on ZDLRA.</td></tr><tr><td>3</td><td>Incremental Backup Initiated</td><td>Protected Database (RMAN)</td><td>Periodic incremental backup process is triggered.</td></tr><tr><td>4</td><td>Blocks Sent to ZDLRA</td><td>RA Backup Module, ZDLRA (Delta Push Receiver)</td><td>Changed data blocks are transmitted to ZDLRA.</td></tr><tr><td>5</td><td>ZDLRA Validates and Compresses</td><td>ZDLRA (Delta Store)</td><td>Incoming blocks are validated and compressed.</td></tr><tr><td>6</td><td>Blocks Indexed in Delta Pool</td><td>ZDLRA (Delta Store)</td><td>Changed blocks are added to the relevant delta pool and indexed.</td></tr><tr><td>7</td><td>Virtual Full Backup Created</td><td>ZDLRA (Delta Store)</td><td>A new virtual full backup is synthesized using new and existing blocks.</td></tr><tr><td>8</td><td>Catalog Updated</td><td>ZDLRA (Catalog)</td><td>New virtual full backup and archived logs become available in the RMAN catalog for recovery.</td></tr></tbody></table></figure>



<h2 class="wp-block-heading">V. Conclusion: Technical Significance of Delta Push and Delta Store in ZDLRA</h2>



<h3 class="wp-block-heading">A. Summary of Core Architectural and Operational Principles</h3>



<p>Delta Push is an efficient, dual-pronged mechanism (RMAN incrementals + real-time redo) for transferring only necessary changes from protected Oracle databases to ZDLRA. Delta Store is the intelligent, Oracle-aware repository that ingests these changes, versions data blocks at a granular level within delta pools, and synthesizes space-efficient virtual full backups. This &#8220;incremental forever&#8221; approach, powered by Delta Push and Delta Store, minimizes production impact, dramatically reduces RPO, and simplifies recovery.</p>



<h3 class="wp-block-heading">B. Impact on Data Protection, Recovery Speed, and Efficiency</h3>



<p>The combined effect of Delta Push and Delta Store fundamentally redefines Oracle database backup and recovery from a periodic, resource-intensive chore to a continuous, low-impact, and highly reliable data assurance service.</p>



<ul class="wp-block-list">
<li><strong>Data Protection:</strong> Near-zero data loss (sub-second RPO) is achieved thanks to Delta Push&#8217;s real-time redo transport component. Enhanced resilience against ransomware is offered through immutable backups and rapid recovery capabilities. Continuous validation guarantees backup integrity.  </li>



<li><strong>Recovery Speed:</strong> Fast restores from virtual full backups are possible without the need to apply numerous incrementals on the production server. The &#8220;time machine&#8221; feature enables rapid rollback. The Dialog Semiconductor case study showed approximately 4x faster restores.  </li>



<li><strong>Efficiency:</strong> Significant reductions in backup windows, production server load (CPU, I/O), network traffic, and storage consumption are achieved through the incremental forever strategy, Delta Push, virtual full backups, and advanced compression. Backup operations are offloaded from database servers.  </li>
</ul>



<p>Traditional backups are disruptive events.<sup></sup> Delta Push makes data ingestion minimally impactful.<sup></sup> Delta Store optimizes backup data for space and makes it immediately ready in a &#8220;full&#8221; format for rapid recovery.<sup></sup> Automation and continuous validation <sup></sup> add layers of reliability. This transforms the entire data protection posture from a necessary evil to an integrated, efficient, and highly effective component of Oracle database operations, as evidenced by benefits like those at Dialog Semiconductor <sup></sup> and reported overall TCO reductions.<sup></sup> &nbsp;</p>



<p>Furthermore, the Delta Push and Delta Store architecture offers a robust defense mechanism against modern cyber threats like ransomware, not just by enabling fast recovery, but by ensuring the integrity and availability of recovery points up to the last moments before an attack. Ransomware attacks aim to encrypt data and backups, making recovery difficult or impossible.<sup></sup> ZDLRA&#8217;s real-time redo capture via Delta Push allows recovery to within seconds of an attack.<sup></sup> Delta Store&#8217;s continuous validation <sup></sup> helps detect corruption early. Features like backup immutability <sup></sup> protect the backups themselves. The ability to rapidly restore a clean, virtual full backup to a secure location <sup></sup> means organizations can avoid paying ransoms. Thus, the technical design directly translates into enhanced cyber resilience, a critical requirement in today&#8217;s threat landscape</p>



<p></p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/analysis-of-delta-push-and-delta-store-mechanisms-within-zdlra.html">Analysis of Delta Push and Delta Store Mechanisms within ZDLRA</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Exadata Update Utilities: patchmgr and dbnodeupdate.sh</title>
		<link>https://www.bugraparlayan.com.tr/exadata-update-utilities-patchmgr-and-dbnodeupdate-sh.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Sat, 26 Apr 2025 14:56:09 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[cell node patching]]></category>
		<category><![CDATA[database node updates]]></category>
		<category><![CDATA[dbnodeupdate.sh]]></category>
		<category><![CDATA[dbnodeupdate.sh usage]]></category>
		<category><![CDATA[Exadata automation tools]]></category>
		<category><![CDATA[Oracle Exadata patching]]></category>
		<category><![CDATA[patchmgr]]></category>
		<category><![CDATA[patchmgr best practices]]></category>
		<category><![CDATA[rolling patch updates]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1459</guid>

					<description><![CDATA[<p>1. Introduction Oracle Exadata Database Machine is a high-performance, optimized platform for Oracle Database workloads. Regularly updating the software components of this platform—including the operating system, Exadata system software, device drivers, and firmware—is crucial for addressing security vulnerabilities, fixing bugs, and leveraging new features. Oracle provides specialized utilities to manage these update processes. Two of &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/exadata-update-utilities-patchmgr-and-dbnodeupdate-sh.html">Exadata Update Utilities: patchmgr and dbnodeupdate.sh</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading">1. Introduction</h2>



<p>Oracle Exadata Database Machine is a high-performance, optimized platform for Oracle Database workloads. Regularly updating the software components of this platform—including the operating system, Exadata system software, device drivers, and firmware—is crucial for addressing security vulnerabilities, fixing bugs, and leveraging new features. Oracle provides specialized utilities to manage these update processes. Two of the most commonly used tools are <code>patchmgr</code> and <code>dbnodeupdate.sh</code>. This document aims to provide a detailed technical comparison of these two utilities, explaining their functions, key differences, use cases, and parameters for effective <strong>Exadata patching</strong>.</p>



<h2 class="wp-block-heading">2. The <code>patchmgr</code> Utility</h2>



<h3 class="wp-block-heading">2.1. Definition and Purpose</h3>



<p><code>patchmgr</code> is a centralized utility designed to <strong>orchestrate and simplify</strong> software updates for Oracle Exadata infrastructure components.<sup></sup> It allows administrators to update multiple components—database servers, storage servers, and network switches—using a single command structure, streamlining the <strong>Exadata update process</strong>.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">2.2. Scope and Capabilities</h3>



<ul class="wp-block-list">
<li><strong>Broad Component Coverage:</strong><code>patchmgr</code> can update various Exadata components :
<ul class="wp-block-list">
<li>Oracle Exadata Storage Servers (Cells)</li>



<li>Oracle Exadata Database Servers (Compute Nodes)</li>



<li>RDMA Network Fabric Switches (RoCE Switches)</li>



<li>InfiniBand Network Fabric Switches</li>



<li>Management Network Switch (specific models)   </li>
</ul>
</li>



<li><strong>Orchestration:</strong> It manages the update sequence across multiple targets, supporting both rolling and non-rolling updates.
<ul class="wp-block-list">
<li><strong>Rolling Update:</strong> Updates components sequentially (one by one) to maintain overall system availability, ideal for RAC clusters or storage server grids.  </li>



<li><strong>Non-Rolling Update:</strong> Updates all specified components concurrently, which is faster but requires a complete system outage.  </li>
</ul>
</li>



<li><strong>Automation:</strong> For database server updates, <code>patchmgr</code> automates numerous steps, including stopping/starting databases and Grid Infrastructure, managing VMs, handling Oracle Enterprise Manager agents, taking OS backups, relinking Oracle Homes, and applying best practice configurations.  </li>



<li><strong>Centralized Execution:</strong> <code>patchmgr</code> can be executed from a database server within the Exadata system being patched or from a separate, central server (the &#8220;driving system&#8221;) running Oracle Linux or Oracle Solaris. This facilitates managing multiple Exadata systems from one location.  </li>



<li><strong>User and Concurrency:</strong> Can be run by <code>root</code> or a non-root user (requires <code>-log_dir</code>). Multiple <code>patchmgr</code> instances can run concurrently from the same software directory (using distinct <code>-log_dir</code> values) to patch different systems simultaneously.  </li>
</ul>



<h3 class="wp-block-heading">2.3. Platform Support</h3>



<ul class="wp-block-list">
<li><strong>Target Systems:</strong> The Exadata components updated by <code>patchmgr</code> (database servers, storage servers) typically run <strong>Oracle Linux</strong>.</li>



<li><strong>Driving System:</strong> The <code>patchmgr</code> utility itself can be executed from a server running either <strong>Oracle Linux</strong> or <strong>Oracle Solaris</strong>. This means you can initiate and manage the patching of Linux-based Exadata components from a Solaris management server.  </li>
</ul>



<h3 class="wp-block-heading">2.4. Key Parameters and Usage</h3>



<p>The general syntax for <code>patchmgr</code> is: <code>./patchmgr -&lt;component&gt; &lt;component_list_file&gt; -&lt;action&gt; &lt;required_arguments&gt; [optional_arguments]</code> <sup></sup> &nbsp;</p>



<ul class="wp-block-list">
<li><strong><code>-&lt;component></code>:</strong> Specifies the component type:
<ul class="wp-block-list">
<li><code>-cells</code>: For Storage Servers.</li>



<li><code>-dbnodes</code>: For Database Servers.</li>



<li><code>-ibswitches</code>: For InfiniBand Switches.</li>



<li><code>-roceswitches</code>: For RoCE Switches.</li>
</ul>
</li>



<li><strong><code>&lt;component_list_file></code>:</strong> A text file listing the hostnames of the components to be updated.  </li>



<li><strong><code>-&lt;action></code>:</strong> Specifies the operation:
<ul class="wp-block-list">
<li><code>-upgrade</code>: Performs a software upgrade to a specified version.  </li>



<li><code>-rollback</code>: Rolls back to the previous software version.</li>



<li><code>-precheck</code>: Runs prerequisite checks before an upgrade or rollback.  </li>



<li><code>-backup</code>: Performs a backup (typically for DB nodes).</li>
</ul>
</li>



<li><strong><code>[optional_arguments]</code>:</strong> Modifies the action or behavior:
<ul class="wp-block-list">
<li><code>--rolling</code> / <code>-rolling</code>: Perform the action in a rolling fashion.  </li>



<li><code>--iso_repo &lt;path_to_iso></code> / <code>-iso_repo &lt;path_to_iso></code>: Specifies the path to the patch ISO file.  </li>



<li><code>--target_version &lt;version></code> / <code>-target_version &lt;version></code>: Specifies the target software version.  </li>



<li><code>--modify_at_prereq</code>: Allows removal of conflicting RPMs during precheck to resolve dependencies (for DB nodes).  </li>



<li><code>--force_remove_custom_rpms</code>: Forces removal of custom (non-Exadata) RPMs during an OS upgrade (for DB nodes).</li>



<li><code>--log_dir &lt;directory|auto></code>: Specifies a directory for log files or uses automatic naming. Required for concurrent execution or non-root users.  </li>



<li><code>--allow_active_network_mounts</code>: Allows patching to proceed even if active network mounts (like NFS) are detected.  </li>



<li><code>--dbnode_patch_base &lt;path></code>: Specifies the directory on target DB nodes where patch files will be extracted.  </li>



<li><code>--ignore_alerts</code>: Proceeds with patching despite active hardware alerts.  </li>



<li><code>--sequential_backup</code>: Backs up each node immediately before updating it during a rolling update (default is to back up all nodes first).  </li>



<li><code>--update_type &lt;type></code>: Selects security-only (<code>allcvss</code>) or full (<code>full</code>) update.  </li>



<li><code>--live-update-target</code>: Utilizes Exadata Live Update (on supported versions).  </li>
</ul>
</li>
</ul>



<h2 class="wp-block-heading">3. The <code>dbnodeupdate.sh</code> Utility</h2>



<h3 class="wp-block-heading">3.1. Definition and Purpose</h3>



<p><code>dbnodeupdate.sh</code> is a shell script specifically used to update, roll back, or back up the software on a <strong>single Oracle Exadata database server (compute node)</strong>.<sup></sup> Before <code>patchmgr</code> provided orchestration for database servers, updates often involved manually running <code>dbnodeupdate.sh</code> on each node sequentially.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">3.2. Scope and Capabilities</h3>



<ul class="wp-block-list">
<li><strong>Single Database Server Focus:</strong> <code>dbnodeupdate.sh</code> operates only on the database server where it is executed.  </li>



<li><strong>Core Update Engine:</strong> When <code>patchmgr</code> updates database servers, it essentially invokes <code>dbnodeupdate.sh</code> on each target node to perform the actual update, rollback, or backup tasks.  </li>



<li><strong>Operating System Updates:</strong> It handles updates for the Oracle Linux OS, device drivers, and firmware included in the Exadata system software patch. It supports major OS upgrades (e.g., OL6 to OL7, OL7 to OL8), although older <code>dbnodeupdate.sh</code> versions might be needed for older OS transitions.  </li>



<li><strong>Backup and Rollback:</strong> Automatically creates a backup of the root filesystem before starting an update. This backup enables rollback to the previous state using the <code>-r</code> option if the update fails or needs to be reverted.  </li>



<li><strong>Dependency Management:</strong> Includes the <code>-M</code> option to allow the removal of conflicting RPMs during the prerequisite check phase to resolve dependency issues.  </li>
</ul>



<h3 class="wp-block-heading">3.3. Platform Support</h3>



<ul class="wp-block-list">
<li><code>dbnodeupdate.sh</code> runs <em>on</em> the Exadata database server being updated. Modern Exadata database servers exclusively use <strong>Oracle Linux</strong>. While Solaris was supported on older Exadata compute nodes , <code>dbnodeupdate.sh</code> in current contexts targets Linux.  </li>
</ul>



<h3 class="wp-block-heading">3.4. Key Parameters and Usage</h3>



<p>The basic usage is: <code>./dbnodeupdate.sh &lt;options&gt;</code></p>



<ul class="wp-block-list">
<li><strong><code>-u</code>:</strong> Initiates the update process.  </li>



<li><strong><code>-r</code>:</strong> Initiates a rollback from the pre-update backup.  </li>



<li><strong><code>-b</code>:</strong> Performs only the backup step.</li>



<li><strong><code>-c</code>:</strong> Runs the post-reboot completion steps after an update or rollback.  </li>



<li><strong><code>-l &lt;ISO_or_Repo_URL></code>:</strong> Specifies the path to the update ISO file or the YUM repository URL.  </li>



<li><strong><code>-s</code>:</strong> Stops Cluster Ready Services (CRS) before the update/rollback and restarts it afterward.  </li>



<li><strong><code>-p</code>:</strong> Runs the bootstrap phase (typically for upgrades from older versions).  </li>



<li><strong><code>-x &lt;helper_script_dir></code>:</strong> Specifies the directory containing helper scripts (often used with <code>-p</code>).  </li>



<li><strong><code>-M</code>:</strong> Allows removal of conflicting RPMs during prerequisite checks.  </li>



<li><strong><code>-v</code>:</strong> Performs only the prerequisite check.  </li>
</ul>



<h2 class="wp-block-heading">4. <code>patchmgr</code> vs. <code>dbnodeupdate.sh</code>: Key Differences</h2>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Feature</th><th><code>patchmgr</code></th><th><code>dbnodeupdate.sh</code></th></tr><tr><td><strong>Primary Purpose</strong></td><td><strong>Orchestration</strong> for Exadata infrastructure components</td><td><strong>Single</strong> Exadata DB server update/rollback/backup</td></tr><tr><td><strong>Scope</strong></td><td>DB Servers, Storage Servers, Network Switches</td><td>DB Servers Only</td></tr><tr><td><strong>Execution Mode</strong></td><td>Manages multiple targets; invokes <code>dbnodeupdate.sh</code> for DB nodes</td><td>Runs on a single target node</td></tr><tr><td><strong>Update Mode</strong></td><td>Rolling or Non-Rolling</td><td>Affects only the single node it runs on</td></tr><tr><td><strong>Run Location</strong></td><td>DB Node or separate Linux/Solaris server</td><td>Only on the target DB Node being updated</td></tr><tr><td><strong>Typical Use Case</strong></td><td>Standard multi-component/multi-node patching</td><td>Single node update/rollback; invoked by <code>patchmgr</code>; patching last node in rolling update; recovery scenarios</td></tr><tr><td><strong>Driving Platform</strong></td><td>Linux, <strong>Solaris</strong> <sup></sup></td><td>Linux (on the target DB Node)</td></tr></tbody></table></figure>



<h2 class="wp-block-heading">5. Which Tool Should Be Used When?</h2>



<ul class="wp-block-list">
<li><strong>Use <code>patchmgr</code> when:</strong>
<ul class="wp-block-list">
<li>Updating multiple storage servers, database servers, or network switches (<strong>standard and preferred method</strong>).</li>



<li>Choosing between rolling or non-rolling update strategies.</li>



<li>Managing the update process from a central server.</li>



<li>Leveraging automated orchestration steps (service stop/start, backup, relink).</li>
</ul>
</li>



<li><strong>Use <code>dbnodeupdate.sh</code> when:</strong>
<ul class="wp-block-list">
<li>Updating or rolling back <strong>only one</strong> specific database server (e.g., testing on a single node).</li>



<li>Performing a rolling update with <code>patchmgr</code> and needing to patch the <strong>initial driving node last</strong> (by running <code>dbnodeupdate.sh</code> locally on it after other nodes are done).  </li>



<li>Recovering a system after a failed update/rollback by booting from a <strong>diagnostic ISO</strong>.  </li>



<li>As a manual alternative if <code>patchmgr</code> itself encounters issues or is unavailable.</li>
</ul>
</li>
</ul>



<p>As a general rule, <strong><code>patchmgr</code> is the standard utility</strong> for routine Exadata infrastructure patching. <code>dbnodeupdate.sh</code> should be considered an underlying component used by <code>patchmgr</code> for database nodes or a tool for specific single-node scenarios.</p>



<h2 class="wp-block-heading">6. Conclusion</h2>



<p><code>patchmgr</code> and <code>dbnodeupdate.sh</code> are essential tools for maintaining the currency and security of Oracle Exadata platforms. <code>patchmgr</code> serves as the primary orchestration utility, simplifying the update process across multiple Exadata components (database servers, storage servers, switches) and supporting both rolling and non-rolling strategies. It can be driven from Linux or Solaris systems. <code>dbnodeupdate.sh</code> is the core script that performs the actual update, rollback, or backup on individual Linux-based Exadata database servers, often invoked by <code>patchmgr</code> but also usable standalone for specific single-node tasks or recovery situations. Understanding the distinct roles and capabilities of each tool allows administrators to choose the appropriate method for their specific Exadata maintenance requirements, with <code>patchmgr</code> being the standard choice for most patching operations.</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/exadata-update-utilities-patchmgr-and-dbnodeupdate-sh.html">Exadata Update Utilities: patchmgr and dbnodeupdate.sh</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Comprehensive Guide to Oracle Exadata Automatic Hard Disk Scrubbing</title>
		<link>https://www.bugraparlayan.com.tr/comprehensive-guide-to-oracle-exadata-automatic-hard-disk-scrubbing.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Fri, 25 Apr 2025 16:48:27 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[DB_BLOCK_CHECKING Exadata]]></category>
		<category><![CDATA[Exadata cell offload]]></category>
		<category><![CDATA[Exadata data protection]]></category>
		<category><![CDATA[Exadata data validation]]></category>
		<category><![CDATA[Exadata disk scrubbing]]></category>
		<category><![CDATA[Exadata scrubbing]]></category>
		<category><![CDATA[Exadata silent corruption]]></category>
		<category><![CDATA[Oracle HCC scrubbing]]></category>
		<category><![CDATA[smart scan scrubbing]]></category>
		<category><![CDATA[storage level data integrity]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1455</guid>

					<description><![CDATA[<p>I. Introduction: Overview of the Exadata Hard Disk Scrubbing Process Data integrity is a cornerstone of modern computing systems. Errors that may occur during the storage, reading, transmission, and processing of data can have devastating effects on business processes. Various error detection and correction mechanisms have been developed to mitigate these risks. One such mechanism &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/comprehensive-guide-to-oracle-exadata-automatic-hard-disk-scrubbing.html">Comprehensive Guide to Oracle Exadata Automatic Hard Disk Scrubbing</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading">I. Introduction: Overview of the Exadata Hard Disk Scrubbing Process</h2>



<p>Data integrity is a cornerstone of modern computing systems. Errors that may occur during the storage, reading, transmission, and processing of data can have devastating effects on business processes. Various error detection and correction mechanisms have been developed to mitigate these risks. One such mechanism is the &#8220;data scrubbing&#8221; process.</p>



<h3 class="wp-block-heading">A. Data Scrubbing: General Concept</h3>



<p>Data scrubbing is an error correction technique that periodically inspects storage devices or main memory for errors and corrects detected errors using redundant data, such as checksums or backup copies of the data.<sup></sup> Its primary purpose is to reduce the likelihood that single, correctable errors will accumulate over time and lead to uncorrectable errors.<sup></sup> This ensures data integrity and minimizes the risk of data loss.<sup></sup> &nbsp;</p>



<p>This technique is a widely used error detection and correction mechanism in memory modules (with ECC memory), RAID arrays, modern file systems like ZFS and Btrfs, and FPGAs.<sup></sup> For example, a RAID controller can periodically read all hard disks in a RAID array to detect and repair bad blocks before applications access them, thereby reducing the probability of silent data corruption caused by bit-level errors.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">B. Exadata Automatic Hard Disk Scrubbing: Definition and Scope</h3>



<p>The Oracle Exadata platform employs a multi-layered approach to ensure data integrity. One of these layers is the <strong>Exadata Automatic Hard Disk Scrub and Repair</strong> feature. As part of the Exadata System Software (Cell Software), this feature automatically and periodically inspects the <strong>hard disk drives (HDDs)</strong> within the Storage Servers (Cells) when the disks are idle.<sup></sup> &nbsp;</p>



<p>The primary goal of this process is to proactively detect and facilitate the repair of bad sectors or other physical/logical defects on the disks <em>before</em> applications attempt to access the affected data.<sup></sup> This prevents &#8220;latent&#8221; or silent data corruption. &nbsp;</p>



<p>The scope of Exadata scrubbing is important. This feature primarily targets <strong>physical</strong> bad sectors on hard disks.<sup></sup> It focuses on detecting physical media errors that might be missed by standard drive Error Correcting Code (ECC) mechanisms or operating system checks.<sup></sup> This complements, but does not replace, higher-level logical consistency checks performed by the database (e.g., via the <code>DB_BLOCK_CHECKING</code> parameter <sup></sup>) or the manually executable ASM disk scrubbing process.<sup></sup> Furthermore, this automatic scrubbing process does not apply to Flash drives in Exadata; these drives are protected by different mechanisms.<sup></sup> &nbsp;</p>



<p>A distinctive aspect of Exadata scrubbing is its proactive nature. While database block checks typically occur during I/O operations <sup></sup>, Exadata scrubbing specifically targets data that has <strong>not been accessed for a long time</strong>, especially when disks are idle.<sup></sup> This approach ensures that corruption in rarely used data is detected and repaired long before it can cause an access error at a critical moment. &nbsp;</p>



<h3 class="wp-block-heading">C. Differences Between Exadata Hard Disk Scrubbing and ASM Disk Scrubbing</h3>



<p>The term &#8220;scrubbing&#8221; can be used in different contexts within the Oracle ecosystem, so it&#8217;s crucial to distinguish Exadata&#8217;s automatic hard disk scrubbing from the disk scrubbing feature offered by Oracle Automatic Storage Management (ASM).<sup></sup> &nbsp;</p>



<ul class="wp-block-list">
<li><strong>Exadata Automatic Hard Disk Scrubbing:</strong>
<ul class="wp-block-list">
<li><strong>Scope:</strong> Operates at the Exadata Storage Server (Cell) level, managed by the Cell Software.  </li>



<li><strong>Focus:</strong> Checks the integrity of <strong>physical</strong> sectors on hard disks.  </li>



<li><strong>Operation:</strong> Runs automatically based on a schedule configured in CellCLI.  </li>



<li><strong>Resource Usage:</strong> The checking process is local to the storage cell, consuming no CPU on database servers and generating no unnecessary network traffic during the check.  </li>



<li><strong>Monitoring:</strong> Monitored via CellCLI metrics and Cell alert logs.  </li>
</ul>
</li>



<li><strong>ASM Disk Scrubbing:</strong>
<ul class="wp-block-list">
<li><strong>Scope:</strong> Operates at the ASM disk group or file level, managed by ASM.  </li>



<li><strong>Focus:</strong> Searches for <strong>logical</strong> corruptions within ASM blocks/extents.  </li>



<li><strong>Operation:</strong> Typically triggered manually (via SQL*Plus or asmcmd) or through a script (e.g., a cron job ).  </li>



<li><strong>Resource Usage:</strong> The process occurs at the ASM layer and can potentially consume database server resources and inter-cell network traffic.</li>



<li><strong>Monitoring:</strong> Monitored via the <code>V$ASM_OPERATION</code> view and ASM alert logs (<code>alert_+ASM.log</code>).  </li>
</ul>
</li>
</ul>



<p>These two mechanisms are complementary. Exadata scrubbing finds physical errors, potentially preventing them from causing logical corruptions later, while ASM scrubbing can find logical inconsistencies that might arise from sources other than physical media errors (e.g., software bugs). Oracle documentation suggests that due to the presence of automatic Exadata scrubbing in Exadata 11.2.3.3 and later, periodic ASM disk scrubbing becomes less critical for the specific purpose of <em>proactive physical/latent error checking</em>.<sup></sup> However, manual ASM scrubbing retains its value for on-demand logical validation of specific files or disk groups. &nbsp;</p>



<h2 class="wp-block-heading">II. Internal Mechanism of the Exadata Scrubbing Process</h2>



<p>The effectiveness of the Exadata Automatic Hard Disk Scrubbing process relies on the tight integration between the core components of the Exadata architecture: the Storage Servers (Cells) and Oracle Automatic Storage Management (ASM).</p>



<h3 class="wp-block-heading">A. Role of Exadata Storage Servers (Cells)</h3>



<p>The scrubbing process is executed by the Exadata System Software (specifically, the Cell Services &#8211; CELLSRV process) running on each Exadata Storage Server (Cell).<sup></sup> The inspection is local to the cell where the scanned disk resides; data is not sent outside the cell during the sector check phase. This minimizes inter-cell network traffic for the inspection stage.<sup></sup> &nbsp;</p>



<p>The Cell Software continuously monitors disk health and I/O utilization to determine when to start, pause, or throttle the scrubbing process.<sup></sup> Typically, scrubbing begins or resumes when the average disk I/O utilization drops below a certain threshold (often cited as 25%).<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">B. Interaction with Oracle ASM for Detection and Repair</h3>



<p>When the Exadata scrubbing process detects a bad sector on a hard disk, the procedure unfolds as follows <sup></sup>: &nbsp;</p>



<ol class="wp-block-list">
<li><strong>Detection:</strong> The Cell Software identifies a physical read error or inconsistency during its periodic scan.</li>



<li><strong>Request Submission:</strong> The Cell Software that detected the faulty sector automatically sends a repair request to the Oracle ASM instance managing the disk group containing that disk.  </li>



<li><strong>Repair by ASM:</strong> Upon receiving the request, ASM orchestrates the repair by reading a healthy copy of the data block (extent) containing the bad sector from another storage server where a mirrored copy resides.  </li>
</ol>



<p>This interaction exemplifies Exadata&#8217;s &#8220;Intelligent Storage&#8221; philosophy; low-level physical error detection happens within the cell, while ASM, which understands the database structure and data placement, coordinates the logical repair.</p>



<h3 class="wp-block-heading">C. Leveraging ASM Mirroring for Data Recovery</h3>



<p>Oracle ASM mirroring (Normal or High Redundancy) is fundamental to Exadata&#8217;s data protection strategy, and the repair capability of the scrubbing process is entirely dependent on this mechanism.<sup></sup> &nbsp;</p>



<p>ASM distributes redundant copies (extents) of data blocks across different failure groups (which in Exadata are typically the Storage Servers).<sup></sup> This ensures data accessibility even if an entire cell becomes unavailable, as data can be accessed from other copies.<sup></sup> &nbsp;</p>



<p>When ASM receives a repair request triggered by scrubbing, it follows these steps:</p>



<ol class="wp-block-list">
<li><strong>Locate Healthy Copy:</strong> ASM identifies a disk on a different storage cell that holds a valid copy of the affected data block. ASM knows which disks are &#8220;partners&#8221; and where mirrored copies are stored.  </li>



<li><strong>Read Data:</strong> ASM reads the correct data from the disk containing the healthy copy.</li>



<li><strong>Write Over Bad Sector:</strong> ASM uses the correct data read to overwrite the bad sector on the original disk, thus correcting the error.  </li>
</ol>



<p>The success of this repair mechanism hinges entirely on the existence of valid and accessible ASM mirrors. If a second disk failure occurs in a Normal Redundancy (2 copies) disk group before a rebalance completes, or if all three copies become inaccessible simultaneously in a High Redundancy (3 copies) group, scrubbing can <em>detect</em> the error, but ASM <em>cannot repair</em> it.<sup></sup> This underscores why <strong>High Redundancy</strong> is strongly recommended for critical systems <sup></sup>, as the extra copy significantly reduces the probability of losing all copies concurrently. &nbsp;</p>



<p>Furthermore, the scrubbing process not only repairs isolated bad sectors but can also serve as an early indicator of more severe disk problems. If numerous or persistent errors are detected during scrubbing, it can lead ASM to take the corresponding grid disk offline and initiate a rebalance operation to redistribute data onto the remaining healthy disks.<sup></sup> In this context, scrubbing also acts as an early warning system that triggers ASM&#8217;s existing high availability (HA) mechanisms. Monitoring the <code>V$ASM_OPERATION</code> view during or after scrub periods is important for tracking such ASM recovery actions.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">D. Types of Errors Detected</h3>



<p>Exadata Automatic Hard Disk Scrubbing primarily focuses on detecting <strong>physical bad sectors</strong> and <strong>latent media errors</strong> on hard disk drives that might not be caught by standard drive ECC or operating system checks.<sup></sup> Damaged or worn-out sectors or other physical defects fall under this scope.<sup></sup> &nbsp;</p>



<p>The &#8220;logical defects&#8221; mentioned <sup></sup> typically refer to low-level media inconsistencies rather than logical corruptions at the ASM or database level (which is the domain of ASM scrubbing <sup></sup>). The main goal is to find such issues before they impact data access or lead to silent data corruption.<sup></sup> &nbsp;</p>



<h2 class="wp-block-heading">III. Managing and Monitoring the Exadata Scrubbing Process</h2>



<p>Effectively utilizing the Exadata Automatic Hard Disk Scrubbing feature requires proper configuration and continuous monitoring. The primary tool for these tasks is the CellCLI (Cell Command Line Interface) utility.</p>



<h3 class="wp-block-heading">A. CellCLI Commands for Configuration</h3>



<p>CellCLI is the main command-line interface for managing Exadata storage server features.<sup></sup> Scrubbing-related configuration is done using the <code>ALTER CELL</code> command and specific attributes <sup></sup>: &nbsp;</p>



<ul class="wp-block-list">
<li><strong><code>hardDiskScrubInterval</code></strong>: Determines how often the automatic scrubbing process runs. Valid options are:
<ul class="wp-block-list">
<li><code>daily</code>: Every day</li>



<li><code>weekly</code>: Every week</li>



<li><code>biweekly</code>: Every two weeks (default)</li>



<li><code>none</code>: Disables automatic scrubbing and stops any running process.</li>



<li><em>Example:</em> To set weekly scrubbing: <code>CellCLI> ALTER CELL hardDiskScrubInterval=weekly</code>.  </li>
</ul>
</li>



<li><strong><code>hardDiskScrubStartTime</code></strong>: Sets when the next scheduled scrubbing process will start. Valid options are:
<ul class="wp-block-list">
<li>A specific date and time (e.g., in &#8216;YYYY-MM-DDTHH:MI:SS-TZ&#8217; format).</li>



<li><code>now</code>: Triggers the next scrubbing cycle to start immediately (after the current cycle finishes, or for the first run).</li>



<li><em>Example:</em> To start at a specific time: <code>CellCLI> ALTER CELL hardDiskScrubStartTime='2024-10-26T02:00:00-07:00'</code>.  </li>
</ul>
</li>
</ul>



<p>To view the current scrubbing settings, use the command: <code>CellCLI&gt; LIST CELL ATTRIBUTES hardDiskScrubInterval, hardDiskScrubStartTime</code></p>



<p>Note that configuration is done on a per-cell basis, meaning these settings apply to all hard disks within a specific storage server. However, the &#8220;Adaptive Scrubbing Schedule&#8221; feature <sup></sup> can automatically adjust the <em>effective</em> run frequency for specific disks identified as problematic, although the base schedule is configured cell-wide. &nbsp;</p>



<h3 class="wp-block-heading">B. Monitoring Scrubbing Activity</h3>



<p>Several methods are available to understand the status and impact of the scrubbing process:</p>



<ul class="wp-block-list">
<li><strong>CellCLI Metrics:</strong>
<ul class="wp-block-list">
<li>The most direct way to see real-time scrubbing activity is using the <code>LIST METRICCURRENT</code> command. Specifically, the <code>CD_IO_BY_R_SCRUB_SEC</code> metric shows the read I/O generated by scrubbing in MB/second for each hard disk (CD). Non-zero values indicate active scrubbing on that disk.  </li>



<li><em>Example Command:</em> <code>CellCLI> LIST METRICCURRENT WHERE name = 'CD_IO_BY_R_SCRUB_SEC'</code></li>



<li>Other related metrics (discoverable with <code>LIST METRICDEFINITION WHERE name like '%SCRUB%'</code>) might provide additional information about scrubbing wait times or resource usage.</li>
</ul>
</li>



<li><strong>Cell Alert Logs:</strong>
<ul class="wp-block-list">
<li>Informational messages indicating the start (<code>Begin scrubbing celldisk</code>) and finish (<code>Finished scrubbing celldisk</code>) of scrubbing operations are logged in the cell alert logs. These logs can be examined using ADRCI (Automatic Diagnostic Repository Command Interpreter) or directly from files under the <code>$CELLTRACE</code> directory. Messages related to errors encountered during scrubbing or disk issues will also appear in these logs.  </li>



<li><em>Example Command:</em> <code>CellCLI> LIST ALERTHISTORY WHERE message LIKE '%scrubbing%'</code></li>
</ul>
</li>



<li><strong>AWR Reports (Automatic Workload Repository):</strong>
<ul class="wp-block-list">
<li>AWR reports, particularly in their Exadata-specific sections, provide aggregated information about scrubbing I/O activity that occurred during a specific snapshot period. Look for metrics labeled &#8216;scrub I/O&#8217; in the report.  </li>



<li>Seeing high &#8216;scrub I/O&#8217; in AWR during periods of low application I/O is normal and expected. However, understanding whether high scrub I/O correlates with performance degradation requires analyzing the overall system load, IORM configuration, and other sections in AWR like &#8216;Exadata OS I/O Stats&#8217;. AWR provides historical context for evaluating impact over time, while CellCLI metrics offer a real-time view.  </li>
</ul>
</li>



<li><strong>Real-Time Insight:</strong>
<ul class="wp-block-list">
<li>If configured, scrubbing metrics like <code>CD_IO_BY_R_SCRUB_SEC</code> can be sent to a preferred dashboard for visual monitoring of scrubbing activity across all Exadata cells.  </li>
</ul>
</li>



<li><strong>ASM Views:</strong>
<ul class="wp-block-list">
<li>While Exadata scrubbing doesn&#8217;t directly log to <code>V$ASM_OPERATION</code>, if scrubbing triggers an ASM repair or a subsequent rebalance, those operations <em>can</em> be monitored in <code>V$ASM_OPERATION</code>. The <code>V$ASM_DISK_STAT</code> view might also reflect I/O patterns related to scrubbing or repair.  </li>
</ul>
</li>
</ul>



<h3 class="wp-block-heading">C. Starting, Stopping, and Checking Status</h3>



<ul class="wp-block-list">
<li><strong>Starting:</strong> Scrubbing starts automatically based on the <code>hardDiskScrubInterval</code> and <code>hardDiskScrubStartTime</code> settings. The <code>hardDiskScrubStartTime=now</code> setting can be used to trigger the next cycle immediately. There isn&#8217;t a direct command like &#8220;start scrubbing now.&#8221;  </li>



<li><strong>Stopping:</strong> To stop and disable automatic scrubbing, use the <code>hardDiskScrubInterval=none</code> command. This will also stop any currently running scrubbing process.  </li>



<li><strong>Status Check:</strong> There is no single &#8220;scrubbing status&#8221; command. The status is inferred through the monitoring methods described above (CellCLI metrics, logs, AWR) by looking at active I/O rates and log messages.</li>
</ul>



<h3 class="wp-block-heading">D. Table 1: Essential CellCLI Commands for Exadata Hard Disk Scrubbing</h3>



<p>The following table summarizes the key CellCLI commands used to manage and monitor the Exadata hard disk scrubbing process:</p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Command</th><th>Purpose</th><th>Example</th><th>Sources</th></tr><tr><td>`ALTER CELL hardDiskScrubInterval = [daily\</td><td>weekly\</td><td>biweekly\</td><td></td></tr><tr><td>`ALTER CELL hardDiskScrubStartTime = [&#8216;&lt;timestamp&gt;&#8217;\</td><td>now]`</td><td>Sets the start time for the next scheduled scrubbing operation.</td><td><code>ALTER CELL hardDiskScrubStartTime=now</code></td></tr><tr><td><code>LIST CELL ATTRIBUTES hardDiskScrubInterval, hardDiskScrubStartTime</code></td><td>Displays the current scrubbing schedule configuration.</td><td><code>LIST CELL ATTRIBUTES hardDiskScrubInterval, hardDiskScrubStartTime</code></td><td></td></tr><tr><td><code>LIST METRICCURRENT WHERE name = 'CD_IO_BY_R_SCRUB_SEC'</code></td><td>Monitors the real-time scrubbing I/O rate for each hard disk.</td><td><code>LIST METRICCURRENT WHERE name = 'CD_IO_BY_R_SCRUB_SEC'</code></td><td></td></tr><tr><td><code>LIST ALERTHISTORY WHERE message LIKE '%scrubbing%'</code></td><td>Checks logs for scrubbing start/finish/error messages.</td><td><code>LIST ALERTHISTORY WHERE message LIKE '%scrubbing%'</code></td><td><sup></sup></td></tr></tbody></table></figure>



<h2 class="wp-block-heading">IV. Performance Impacts of the Scrubbing Process</h2>



<p>While designed to proactively protect data integrity, Exadata Automatic Hard Disk Scrubbing does have an impact on system resources, particularly the I/O subsystem. Understanding and managing this impact is crucial.</p>



<h3 class="wp-block-heading">A. Resource Consumption (CPU, I/O)</h3>



<p>The primary resource consumed by the scrubbing process is <strong>Disk I/O</strong>.<sup></sup> The operation involves reading sectors from the hard disks. On an otherwise idle system or disk, the scrubbing process can significantly increase disk utilization, potentially reaching close to 100% for the disk being scanned.<sup></sup> &nbsp;</p>



<p><strong>CPU consumption</strong> on the storage server (Cell) for the scrubbing check itself is generally low, as it&#8217;s largely an I/O-bound operation. However, if scrubbing detects an error and triggers a repair via ASM, that repair process (reading the good copy and writing it to the bad location) can consume additional resources (CPU and network) across cells and potentially database nodes, although the Exadata architecture aims to minimize this impact.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">B. Designed Operating Window (Low I/O Utilization)</h3>



<p>A key design principle to minimize the performance impact of Exadata scrubbing is that the process only runs when the storage server detects low average I/O utilization.<sup></sup> This threshold is commonly cited as 25%.<sup></sup> &nbsp;</p>



<p>The system automatically pauses or throttles scrubbing activity when I/O demand from the database workload exceeds this threshold.<sup></sup> This mechanism aims to prevent scrubbing from significantly impacting production workloads. &nbsp;</p>



<p>However, there&#8217;s a nuance to the &#8220;25% utilization&#8221; threshold. It may not mean absolute idleness. There could be a persistent background I/O load running just below this threshold (e.g., 20-24%). Adding the scrubbing I/O on top of this existing load will increase the total I/O. While Exadata I/O Resource Management (IORM) prioritizes user I/O <sup></sup>, even the minimal added load from scrubbing could potentially have a noticeable effect, especially for applications highly sensitive to very low latency. Therefore, while &#8220;low impact&#8221; is the goal, &#8220;zero impact&#8221; is not guaranteed. &nbsp;</p>



<h3 class="wp-block-heading">C. Interaction with I/O Resource Management (IORM)</h3>



<p>Exadata I/O Resource Management (IORM) plays a critical role in managing the performance impact of background tasks like scrubbing.<sup></sup> IORM prioritizes and schedules I/O requests within the storage server based on configured resource plans.<sup></sup> &nbsp;</p>



<p>IORM automatically prioritizes database workload I/O (e.g., user queries, OLTP transactions) over background I/O processes like scrubbing.<sup></sup> This ensures minimal impact on application performance from scrubbing activity. IORM plans can be configured to manage resources among different databases or workloads, indirectly affecting the amount of resources available for background tasks like scrubbing.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">D. Potential Performance Impact and Mitigation Methods</h3>



<p>Despite being designed for low impact, it should be acknowledged that scrubbing can cause spikes in disk utilization and potentially increase latency, especially in situations where the system isn&#8217;t completely idle even when the &#8220;idle&#8221; threshold is met.<sup></sup> The concern about performance impact, though often associated with general ASM scrubbing, can also apply to Exadata scrubbing.<sup></sup> &nbsp;</p>



<p>To mitigate this potential impact, consider these strategies:</p>



<ul class="wp-block-list">
<li><strong>Scheduling:</strong> The most effective mitigation is to schedule the scrubbing process using <code>hardDiskScrubStartTime</code> and <code>hardDiskScrubInterval</code> during periods of genuinely low system activity (e.g., midnight, weekends).  </li>



<li><strong>Monitoring:</strong> Regularly assess when scrubbing runs and its actual impact in your specific environment using AWR  and CellCLI metrics.  </li>



<li><strong>IORM Settings:</strong> Ensure IORM is configured appropriately for your workload priorities.  </li>



<li><strong>Adaptive Scheduling:</strong> Leverage Exadata&#8217;s adaptive scheduling feature. This automatically adjusts the frequency based on need, potentially reducing unnecessary runs on healthy disks.  </li>
</ul>



<h3 class="wp-block-heading">E. Factors Affecting Scrubbing Duration</h3>



<p>The time required to complete a scrubbing cycle depends on several factors:</p>



<ul class="wp-block-list">
<li><strong>Disk Size and Type:</strong> Larger capacity hard disks naturally take longer to scan. Estimates like 8-12 hours for a 4TB disk  or 1-2 hours per terabyte when idle  have been mentioned. Modern High Capacity (HC) drives are much larger (18TB in X9M , 22TB in X10M ), implying potentially much longer scrub times.  </li>



<li><strong>System Load:</strong> Since scrubbing pauses when user workload increases , the busier the system, the longer the <strong>total wall-clock time</strong> required to complete a scrub cycle. On a busy system, completing a cycle could take days.  </li>



<li><strong>Number of Errors Found:</strong> If many bad sectors are found, the time spent coordinating repairs with ASM can increase the total duration.  </li>



<li><strong>ASM Rebalance Activity:</strong> If scrubbing triggers a larger ASM rebalance operation, that separate process will consume its own resources and take time.  </li>



<li><strong>Configured Interval:</strong> While not affecting a single run&#8217;s duration, the interval determines how frequently the process starts.</li>
</ul>



<p>It&#8217;s noteworthy that duration estimates in documentation (S2 vs S9) vary significantly. This highlights that estimates heavily depend on the Exadata generation (disk sizes/speeds), software version (potential efficiency improvements), and most importantly, the actual workload pattern and resulting &#8220;idle&#8221; time on the specific system. Relying on monitoring in your own environment is more accurate than general estimates. For instance, one observation noted a scrubbing rate of approximately 115MB/s per disk.<sup></sup> At this rate, <em>continuously</em> scanning a 22TB disk (X10M <sup></sup>) would take roughly 54 hours. Given that scrubbing runs intermittently based on load <sup></sup>, the actual completion time could be considerably longer. &nbsp;</p>



<h2 class="wp-block-heading">V. Key Benefits of the Exadata Scrubbing Process</h2>



<p>Exadata Automatic Hard Disk Scrubbing is a valuable feature that significantly contributes to the data integrity and high availability capabilities of the Exadata platform.</p>



<h3 class="wp-block-heading">A. Proactive Detection of Latent Errors and Silent Data Corruption</h3>



<p>Its most fundamental benefit is the proactive discovery of physical media errors <em>before</em> they are encountered during normal database operations.<sup></sup> This prevents &#8220;silent&#8221; data corruption, where errors occur on disk but remain undetected until the data is read (which could be much later).<sup></sup> By checking data blocks that haven&#8217;t been accessed recently <sup></sup>, it ensures such hidden threats are uncovered. &nbsp;</p>



<h3 class="wp-block-heading">B. Enhanced Data Integrity and Reliability</h3>



<p>By detecting physical errors and enabling their repair, the scrubbing process directly contributes to the overall data integrity and reliability of the Exadata platform.<sup></sup> This feature complements other protection layers like Oracle HARD (Hardware Assisted Resilient Data) checks <sup></sup>, ASM mirroring <sup></sup>, and database-level checks <sup></sup>, providing robust defense against data corruption. &nbsp;</p>



<h3 class="wp-block-heading">C. Automatic Repair Mechanism</h3>



<p>A significant advantage is that the feature automates not just detection but also the initiation of the repair process. In typical bad sector scenarios, both error detection <em>and</em> the triggering of repair via ASM happen automatically, requiring no manual intervention.<sup></sup> This reduces administrative overhead and ensures timely correction of detected issues. &nbsp;</p>



<h3 class="wp-block-heading">D. Complements Other Exadata High Availability Features</h3>



<p>Scrubbing is part of Exadata&#8217;s comprehensive Maximum Availability Architecture (MAA) strategy.<sup></sup> It works alongside features like redundant hardware components <sup></sup>, Oracle RAC for instance continuity <sup></sup>, ASM for storage virtualization and redundancy <sup></sup>, HARD for I/O path validation <sup></sup>, and potentially Data Guard for disaster recovery.<sup></sup> &nbsp;</p>



<p>This reinforces Exadata&#8217;s &#8220;defense in depth&#8221; approach to data protection. HARD checks the I/O path during writes <sup></sup>; database checks can verify logical structure <sup></sup>; ASM provides redundant copies of data <sup></sup>; and scrubbing proactively inspects the physical media at rest.<sup></sup> No single feature covers all possible scenarios, but working together, they provide robust protection. Scrubbing forms a critical layer in this strategy, specifically targeting latent physical errors that might be missed by other mechanisms. &nbsp;</p>



<h2 class="wp-block-heading">VI. Evolution of the Scrubbing Feature Across Exadata Versions</h2>



<p>The Exadata Automatic Hard Disk Scrubbing feature has evolved along with the platform itself.</p>



<h3 class="wp-block-heading">A. Feature Introduction</h3>



<p>The Automatic Hard Disk Scrub and Repair feature was first introduced with <strong>Oracle Exadata System Software version 11.2.3.3.0</strong>.<sup></sup> At that time, specific minimum database/Grid Infrastructure versions like 11.2.0.4 or 12.1.0.2 were required for the feature to function.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">B. Adaptive Scrubbing Schedule</h3>



<p>A significant enhancement arrived with <strong>Exadata System Software version 12.1.2.3.0</strong>: the Adaptive Scrubbing Schedule.<sup></sup> With this feature, if the scrubbing process finds a bad sector on a disk, the Cell Software automatically schedules the <em>next</em> scrubbing job for <em>that specific disk</em> to run more frequently (typically weekly).<sup></sup> This temporarily overrides the cell-wide <code>hardDiskScrubInterval</code> setting for that disk. If the subsequent, more frequent run finds no errors, the disk&#8217;s schedule reverts to the global <code>hardDiskScrubInterval</code> setting.<sup></sup> This feature also requires specific minimum Grid Infrastructure versions to operate.<sup></sup> &nbsp;</p>



<p>This adaptive approach makes scrubbing more efficient. Instead of frequently scanning all disks, it focuses more attention only on disks showing potential issues. This conserves I/O resources on healthy disks while providing quicker follow-up checks on suspect ones.</p>



<h3 class="wp-block-heading">C. Other Related Developments (Post-12.1.2.3.0)</h3>



<p>Available documentation primarily focuses on the introduction of the scrubbing feature and the adaptive scheduling enhancement. Detailed information about significant changes to algorithms, performance tuning (beyond IORM interaction), or reporting in later versions (e.g., post-12.x, 18.x, 19.x, 20.x, 21.x, 22.x, 23.x <sup></sup>) is not provided in the reviewed sources.<sup></sup> Consulting the release notes for specific Exadata System Software versions might be necessary for details on newer developments. &nbsp;</p>



<h3 class="wp-block-heading">D. Table 2: Evolution of Key Exadata Scrubbing Features</h3>



<p>The following table summarizes the key milestones in the development of the Exadata scrubbing feature:</p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Exadata Software Version</th><th>Key Feature/Enhancement</th><th>Description</th><th>Sources</th></tr><tr><td>11.2.3.3.0</td><td>Automatic Hard Disk Scrub and Repair (Introduction)</td><td>Introduced the core feature for automatic, periodic inspection and initiation of repair via ASM.</td><td></td></tr><tr><td>12.1.2.3.0</td><td>Adaptive Scrubbing Schedule</td><td>Automatically increases scrubbing frequency (e.g., to weekly) for disks where bad sectors were recently detected.</td><td><sup></sup></td></tr><tr><td>Post-12.1.2.3.0</td><td>(Other Enhancements Unspecified)</td><td>(Specific major enhancements for later versions are not detailed in the provided documentation)</td><td></td></tr></tbody></table></figure>



<h2 class="wp-block-heading">VII. Configuration and Best Practices</h2>



<p>To maximize the benefits of the Exadata Automatic Hard Disk Scrubbing feature, proper configuration and adherence to Oracle&#8217;s Maximum Availability Architecture (MAA) principles are important.</p>



<h3 class="wp-block-heading">A. Default Settings and Configuration Options</h3>



<ul class="wp-block-list">
<li><strong>Default Schedule:</strong> By default, the scrubbing process is configured to run <strong>every two weeks (<code>biweekly</code>)</strong>.  </li>



<li><strong>Configuration Options:</strong> The <code>hardDiskScrubInterval</code> (<code>daily</code>, <code>weekly</code>, <code>biweekly</code>, <code>none</code>) and <code>hardDiskScrubStartTime</code> (<code>&lt;timestamp></code>, <code>now</code>) attributes can be set via CellCLI.  </li>



<li><strong>No Intensity/Priority Setting:</strong> There is no direct CellCLI setting to control the &#8220;intensity&#8221; or &#8220;priority&#8221; of the scrubbing process itself. Its impact is primarily managed by the idle-time logic and IORM.  </li>
</ul>



<h3 class="wp-block-heading">B. Recommended Scheduling Strategies for Production Environments</h3>



<ul class="wp-block-list">
<li><strong>Use Defaults:</strong> For many environments, the default bi-weekly schedule and the automatic execution during low I/O periods are sufficient.  </li>



<li><strong>Customize Start Time:</strong> Rather than relying solely on <code>now</code> or random times, explicitly setting <code>hardDiskScrubStartTime</code> to known low-load periods (e.g., 2 AM Sunday morning) offers a more controlled approach.  </li>



<li><strong>Assess Workload:</strong> On very busy, 24/7 systems, evaluate if the <code>biweekly</code> interval allows enough time for the process to complete. If not, consider <code>weekly</code>, but closely monitor the performance impact. Disabling scrubbing (<code>none</code>) is generally not recommended unless there&#8217;s a specific, temporary reason, as it forfeits the proactive detection benefit.  </li>



<li><strong>Align with Maintenance Windows:</strong> Coordinate scrubbing schedules with other planned maintenance windows if possible, although the automatic throttling mechanism should prevent major conflicts.  </li>



<li><strong>Monitor Completion:</strong> Check logs to ensure scrubbing cycles complete successfully within the planned interval. If cycles consistently fail to complete due to high load, the scheduling strategy needs review.  </li>
</ul>



<h3 class="wp-block-heading">C. Importance of ASM Redundancy</h3>



<ul class="wp-block-list">
<li><strong>High Redundancy Recommendation:</strong> Using <strong>High Redundancy (3 copies)</strong> for ASM disk groups on Exadata is strongly recommended, especially for production databases.  </li>



<li><strong>Rationale:</strong> While scrubbing works with Normal Redundancy (2 copies), High Redundancy provides significantly better protection against data loss during the repair window (especially if an unrelated second failure occurs). Scrubbing&#8217;s repair capability depends on having a healthy mirror copy available.  </li>



<li><strong>Requirements:</strong> Properly implementing High Redundancy typically requires at least 5 failure groups (often 3 storage cells + 2 quorum disks on database servers for Quarter/Eighth Rack configurations).  </li>
</ul>



<h3 class="wp-block-heading">D. Integration with Overall MAA Strategy</h3>



<p>Scrubbing is just one part of the MAA best practices recommended by Oracle for Exadata <sup></sup>: &nbsp;</p>



<ul class="wp-block-list">
<li><strong>Regular Health Checks:</strong> Run the <code>exachk</code> utility regularly (e.g., monthly) or rely on AHF (Autonomous Health Framework) to run it automatically to validate configuration against best practices, including storage and ASM settings.  </li>



<li><strong>Use Standby Database:</strong> While Exadata scrubbing and HARD checks protect against many issues, a physical standby database (Data Guard) on a separate Exadata system is critical for comprehensive protection against site failures, certain logical corruptions, and as a secondary validation source.  </li>



<li><strong>Monitoring:</strong> Implement comprehensive monitoring (OEM, AWR, CellCLI metrics, logs, Real-Time Insight) to track system health, performance, and background activities like scrubbing.  </li>



<li><strong>Testing:</strong> Validate recovery procedures and understand the behavior of features like scrubbing and ASM rebalance in your test environment.  </li>
</ul>



<h3 class="wp-block-heading">E. Table 3: Exadata Scrubbing Configuration Attributes and Best Practices</h3>



<p>This table consolidates key configuration parameters and actionable recommendations:</p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Parameter/Area</th><th>Configuration/Setting</th><th>Default</th><th>Recommendation</th><th>Sources</th></tr><tr><td><code>hardDiskScrubInterval</code></td><td><code>daily</code>, <code>weekly</code>, <code>biweekly</code>, <code>none</code></td><td><code>biweekly</code></td><td>Start with default. Consider <code>weekly</code> for busy systems if needed, monitoring impact. Avoid <code>none</code>.</td><td></td></tr><tr><td><code>hardDiskScrubStartTime</code></td><td><code>&lt;timestamp&gt;</code>, <code>now</code></td><td>None</td><td>Explicitly set to a known low-load window (e.g., weekend night).</td><td></td></tr><tr><td>ASM Redundancy</td><td>Normal (2 copies), High (3 copies)</td><td>Normal</td><td>Use <strong>High Redundancy</strong> for production disk groups to maximize repair success probability.</td><td></td></tr><tr><td>Monitoring</td><td>CellCLI Metrics, Cell Logs, AWR, ASM Views, <code>exachk</code></td><td>None</td><td>Regularly monitor scrubbing activity, completion status, performance impact, and overall system health (<code>exachk</code>).</td><td><sup></sup></td></tr><tr><td>Scheduling Strategy</td><td>Workload-dependent</td><td>Idle-based</td><td>Schedule during predictably low-load times; ensure cycles complete.</td><td><sup></sup></td></tr><tr><td>MAA Integration</td><td>Part of overall HA</td><td>None</td><td>Integrate with Data Guard, regular health checks, and robust monitoring per MAA guidelines.</td><td><sup></sup></td></tr></tbody></table></figure>



<h2 class="wp-block-heading">VIII. Conclusion</h2>



<p>Oracle Exadata Automatic Hard Disk Scrub and Repair is a proactive defense mechanism crucial for maintaining data integrity and high availability on the Exadata platform. By periodically scanning hard disks on storage servers for physical errors, this feature detects latent corruptions, especially in infrequently accessed data, before they can impact applications.</p>



<p>The core strength of the scrubbing process lies in the integration between Exadata System Software and Oracle ASM. While the Cell Software detects the error, ASM manages the automatic repair process using mirrored copies. The effectiveness of this repair capability is directly tied to the correctly configured redundancy of ASM disk groups, particularly High Redundancy, which is strongly recommended for production environments.</p>



<p>From a performance perspective, the scrubbing process is designed to run during periods of low I/O utilization detected by the system and is managed by IORM. This aims to minimize the impact on production workloads. However, it remains important for administrators to monitor scrubbing activity via CellCLI metrics, alert logs, and AWR reports, and potentially adjust the schedule based on their environment&#8217;s specific workload patterns.</p>



<p>Introduced in Exadata 11.2.3.3.0 and enhanced with Adaptive Scheduling in 12.1.2.3.0, this feature is an integral part of Exadata&#8217;s multi-layered data protection strategy (including HARD checks, ASM mirroring, RAC, Data Guard, etc.). Properly configuring and operating Exadata Automatic Hard Disk Scrubbing is critical for preserving data integrity, preventing unexpected outages, and maximizing the value of the Exadata investment. For best results, scrubbing configuration and operation should be considered within the framework of Oracle MAA best practices, supported by regular system health checks (<code>exachk</code>) and comprehensive monitoring.</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/comprehensive-guide-to-oracle-exadata-automatic-hard-disk-scrubbing.html">Comprehensive Guide to Oracle Exadata Automatic Hard Disk Scrubbing</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Oracle Exadata X11M Platform: Comprehensive Technical Analysis and Comparison with Previous Generations</title>
		<link>https://www.bugraparlayan.com.tr/oracle-exadata-x11m-platform.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Sun, 06 Apr 2025 14:05:00 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[data center solutions]]></category>
		<category><![CDATA[database infrastructure]]></category>
		<category><![CDATA[enterprise database systems]]></category>
		<category><![CDATA[Exadata X11M features]]></category>
		<category><![CDATA[Exadata X11M performance]]></category>
		<category><![CDATA[high performance database solutions]]></category>
		<category><![CDATA[Oracle database platform]]></category>
		<category><![CDATA[Oracle Exadata benefits]]></category>
		<category><![CDATA[Oracle Exadata X11M]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1397</guid>

					<description><![CDATA[<p>Oracle Exadata X11M: Hardware Architecture Overview The core architecture of Oracle Exadata is based on a scale-out design featuring database servers, intelligent storage servers, and a high-speed, low-latency network fabric connecting these components. Exadata X11M combines this architecture with the latest hardware technologies and optimized software, creating a unique platform for Oracle Database workloads. The &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/oracle-exadata-x11m-platform.html">Oracle Exadata X11M Platform: Comprehensive Technical Analysis and Comparison with Previous Generations</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p></p>



<p></p>



<h2 class="wp-block-heading">Oracle Exadata X11M: Hardware Architecture Overview</h2>



<p>The core architecture of Oracle Exadata is based on a scale-out design featuring database servers, intelligent storage servers, and a high-speed, low-latency network fabric connecting these components. Exadata X11M combines this architecture with the latest hardware technologies and optimized software, creating a unique platform for Oracle Database workloads. The co-engineered nature of hardware and software aims to maximize performance at every layer of the system.</p>



<h3 class="wp-block-heading">Database Server (Standard X11M)</h3>



<p>Exadata X11M database servers are designed for compute-intensive database operations. Key hardware specifications include:</p>



<ul class="wp-block-list">
<li><strong>CPU:</strong> Equipped with two <strong>AMD EPYC&#x2122; 9J25 processors</strong>, each featuring 96 cores per socket (2.6 GHz base, up to 4.5 GHz boost), totaling <strong>192 physical cores</strong> per server. These cores are stated to be up to 25% faster than their X10M counterparts.</li>



<li><strong>RAM:</strong> Utilizes high-performance <strong>6400MT/s DDR5</strong> DIMMs. Memory capacity options per server are <strong>512 GB</strong> (16x32GB), <strong>1.5 TB</strong> (24x64GB), <strong>2.25 TB</strong> (24x96GB), or <strong>3 TB</strong> (24x128GB). This provides up to 33% more DRAM bandwidth compared to X10M.</li>



<li><strong>System Storage:</strong> Standard configuration includes 2 x 3.84 TB NVMe devices, expandable to 4. Typically used for the OS and system software.</li>



<li><strong>RDMA Network Fabric:</strong> Features one dual-port ConnectX-7 (CX7) RDMA Network Fabric adapter (PCIe 5.0) providing <strong>2 x 100 Gb/s RoCE</strong> (RDMA over Converged Ethernet) ports in an active-active configuration for a total bandwidth of 200 Gb/s.</li>



<li><strong>Client/Management Network:</strong> Includes dedicated 1GbE Base-T ports for Admin and ILOM. Two dual-port 10/25 GbE adapters (CX6-LX) are factory-installed for client connectivity. Optional adapters include additional 10/25 GbE, 100 GbE (CX6-DX), or Quad 10GBASE-T.</li>
</ul>



<h3 class="wp-block-heading">Storage Servers (HC and EF)</h3>



<p>Exadata storage servers not only store data but also intelligently offload significant portions of SQL query processing from the database servers, reducing CPU load and network traffic. The X11M generation offers High Capacity (HC) and Extreme Flash (EF) storage server types:</p>



<ul class="wp-block-list">
<li><strong>CPU:</strong> Both HC and EF servers feature two <strong>AMD EPYC&#x2122; 9J15 processors</strong> (32 cores per socket, 2.95 GHz base, up to 4.4 GHz boost), providing <strong>64 cores per server</strong> dedicated to SQL offload and system operations. These are up to 11% faster than X10M storage server CPUs.</li>



<li><strong>RAM:</strong> Each server contains <strong>1.5 TB of 6400MT/s DDR5 DRAM</strong>, offering 33% more bandwidth than X10M.</li>



<li><strong>Exadata RDMA Memory (XRMEM):</strong> A significant portion of the RAM, <strong>1.25 TB</strong>, is configured as XRMEM, an ultra-low latency cache tier. Accessible directly via RDMA, XRMEM sits between the database buffer cache and Flash Cache, enabling OLTP read latencies as low as <strong>14µs</strong>. RDMA bypasses much of the OS and network stack overhead, accessing DRAM (XRMEM) which is inherently faster than flash, to achieve this low latency.</li>



<li><strong>Exadata Smart Flash Cache:</strong> Utilizes performance-optimized <strong>PCIe 5.0 NVMe flash cards</strong> (F680 v2).
<ul class="wp-block-list">
<li><strong>HC Server:</strong> 4 x 6.8 TB cards (Total 27.2 TB cache).</li>



<li><strong>EF Server:</strong> 4 x 6.8 TB cards (Total 27.2 TB cache).</li>



<li>This new generation flash provides up to <strong>2.2X faster</strong> analytical I/O (scan throughput) compared to X10M. SQL Scan throughput from flash reaches <strong>100 GB/s per server</strong>.</li>
</ul>
</li>



<li><strong>Persistent Storage:</strong>
<ul class="wp-block-list">
<li><strong>HC Server:</strong> 12 x 22 TB 7,200 RPM SAS HDDs (Total 264 TB raw capacity).</li>



<li><strong>EF Server (All-Flash):</strong> 4 x 30.72 TB capacity-optimized NVMe PCIe Flash drives (Total 122.88 TB raw capacity).</li>
</ul>
</li>



<li><strong>System Storage:</strong> 2 x 480 GB NVMe devices (likely for OS/Exadata software).</li>



<li><strong>RDMA Network Fabric:</strong> Same as database server: 1 x dual-port CX7 adapter (PCIe 5.0), 2&#215;100 Gb/s active-active RoCE (Total 200 Gb/s).</li>
</ul>



<p>The synergy between faster AMD EPYC cores, higher-bandwidth DDR5 RAM, significantly faster PCIe 5.0 flash scans, and the large XRMEM cache accessed via the 200 Gb/s RoCE network forms the foundation for X11M&#8217;s overall workload acceleration. Performance gains stem from the interaction of these components (e.g., faster CPUs fed by faster memory/flash over a faster network) rather than any single element.</p>



<p><strong>Table 1: Exadata X11M Server Component Specifications</strong></p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Component</th><th>Database Server (X11M)</th><th>Storage Server (HC)</th><th>Storage Server (EF)</th></tr><tr><td><strong>CPU</strong></td><td>2 x AMD EPYC&#x2122; 9J25 (96 Cores/socket, 192 Total)</td><td>2 x AMD EPYC&#x2122; 9J15 (32 Cores/socket, 64 Total)</td><td>2 x AMD EPYC&#x2122; 9J15 (32 Cores/socket, 64 Total)</td></tr><tr><td><strong>RAM (Total Size)</strong></td><td>512GB / 1.5TB / 2.25TB / 3TB</td><td>1.5 TB</td><td>1.5 TB</td></tr><tr><td><strong>RAM (Type/Speed)</strong></td><td>DDR5 / 6400MT/s</td><td>DDR5 / 6400MT/s</td><td>DDR5 / 6400MT/s</td></tr><tr><td><strong>XRMEM Size</strong></td><td>N/A</td><td>1.25 TB</td><td>1.25 TB</td></tr><tr><td><strong>Flash Cache (Type)</strong></td><td>N/A</td><td>Perf. Opt. F680 v2 NVMe PCIe 5.0</td><td>Perf. Opt. F680 v2 NVMe PCIe 5.0</td></tr><tr><td><strong>Flash Cache (Size)</strong></td><td>N/A</td><td>4 x 6.8 TB (Total 27.2 TB)</td><td>4 x 6.8 TB (Total 27.2 TB)</td></tr><tr><td><strong>Persistent Storage (Type)</strong></td><td>System NVMe (2&#215;3.84TB, exp. to 4)</td><td>SAS HDD (7200 RPM)</td><td>Cap. Opt. NVMe PCIe Flash</td></tr><tr><td><strong>Persistent Storage (Size)</strong></td><td>&#8211;</td><td>12 x 22 TB (Total 264 TB Raw)</td><td>4 x 30.72 TB (Total 122.88 TB Raw)</td></tr><tr><td><strong>RDMA Network</strong></td><td>1x Dual Port CX7 (PCIe 5.0), 2x100Gb/s Active-Active RoCE</td><td>1x Dual Port CX7 (PCIe 5.0), 2x100Gb/s Active-Active RoCE</td><td>1x Dual Port CX7 (PCIe 5.0), 2x100Gb/s Active-Active RoCE</td></tr></tbody></table></figure>



<p></p>



<p></p>



<h2 class="wp-block-heading">Introducing the Exadata X11M-Z Variant</h2>



<p>Alongside the standard high-performance configurations, the Exadata X11M family introduces a more economical entry-level option designated <strong>&#8220;X11M-Z&#8221;</strong>. This variant is designed for customers with smaller workloads or those seeking access to core Exadata capabilities at a lower cost point. The X11M-Z replaces the fixed &#8220;Eighth Rack&#8221; concept of previous generations, offering a more flexible starting configuration. Customers can start with X11M-Z servers and scale out by adding more Z servers or standard X11M servers as needs grow. This approach potentially broadens Exadata&#8217;s market reach by attracting customers with smaller initial needs or budgets who might otherwise opt for less optimized platforms or cloud services lacking Exadata features.</p>



<h3 class="wp-block-heading">X11M-Z Database Server Specifications</h3>



<ul class="wp-block-list">
<li><strong>CPU:</strong> Utilizes a single-socket configuration with <strong>1 x 32-core AMD EPYC&#x2122; 9J15</strong> processor (2.95 GHz base, up to 4.4 GHz boost).</li>



<li><strong>RAM:</strong> Uses 6400MT/s DDR5 DIMMs. Memory options are <strong>768 GB</strong> (12x64GB) or <strong>1.125 TB</strong> (12x96GB).</li>



<li><strong>System Storage:</strong> Standard 2 x 3.84 TB NVMe devices, expandable to 4.</li>



<li><strong>RDMA Network Fabric:</strong> Same as standard X11M DB server: 1 x dual-port CX7 adapter (PCIe 5.0), 2&#215;100 Gb/s active-active RoCE (Total 200 Gb/s).</li>



<li><strong>Client/Management Network:</strong> Similar options to standard X11M DB server, but with only one optional field-installable adapter slot.</li>
</ul>



<h3 class="wp-block-heading">X11M-Z High Capacity (HC-Z) Storage Server Specifications</h3>



<ul class="wp-block-list">
<li><strong>CPU:</strong> Single-socket configuration with <strong>1 x 32-core AMD EPYC&#x2122; 9J15</strong> processor.</li>



<li><strong>RAM:</strong> <strong>768 GB</strong> 6400MT/s DDR5 DRAM.</li>



<li><strong>XRMEM:</strong> <strong>576 GB</strong> of the RAM is allocated as XRMEM.</li>



<li><strong>Flash Cache:</strong> Two 6.8 TB performance-optimized F680 v2 NVMe PCIe 5.0 flash cards.</li>



<li><strong>Persistent Storage:</strong> Six 22 TB 7,200 RPM SAS HDDs.</li>



<li><strong>System Storage:</strong> 2 x 480 GB NVMe devices.</li>



<li><strong>RDMA Network Fabric:</strong> Same as standard HC/EF storage servers: 1 x dual-port CX7 adapter (PCIe 5.0), 2&#215;100 Gb/s active-active RoCE (Total 200 Gb/s).</li>
</ul>



<h3 class="wp-block-heading">Positioning and Use Cases</h3>



<p>The <strong>X11M-Z</strong> variant is ideal for smaller databases, development/test environments, or departmental applications that require core Exadata software features (Smart Scan, Storage Indexes, HCC, IORM) and performance characteristics but do not need the full scale of standard X11M. It offers a lower entry cost and a reduced power consumption footprint. The single-socket design inherently consumes less power than dual-socket standard servers, aligning with the overall theme of improved energy efficiency and sustainability in X11M. Offering a lower-power hardware option alongside software-based power management strengthens Exadata&#8217;s efficiency narrative.</p>



<p><strong>Table 2: Exadata X11M vs. X11M-Z Key Hardware Differences</strong></p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Component</th><th>Standard X11M Server</th><th>X11M-Z Server</th></tr><tr><td><strong>Server Type</strong></td><td>Database / Storage (HC/EF)</td><td>Database / Storage (HC-Z)</td></tr><tr><td><strong>CPU Sockets</strong></td><td>2</td><td>1</td></tr><tr><td><strong>CPU Cores/Server</strong></td><td>DB: 192 / Storage: 64</td><td>DB: 32 / Storage: 32</td></tr><tr><td><strong>Max RAM/Server</strong></td><td>DB: 3 TB / Storage: 1.5 TB</td><td>DB: 1.125 TB / Storage: 768 GB</td></tr><tr><td><strong>XRMEM Size (Storage)</strong></td><td>1.25 TB</td><td>576 GB</td></tr><tr><td><strong>Flash Cache (Storage)</strong></td><td>4 x 6.8 TB Cards (HC/EF)</td><td>2 x 6.8 TB Cards (HC-Z)</td></tr><tr><td><strong>Disk Drives (HC)</strong></td><td>12 x 22 TB</td><td>6 x 22 TB (HC-Z)</td></tr></tbody></table></figure>



<p></p>



<h2 class="wp-block-heading">Generational Performance Leap: Exadata X11M vs. X10M</h2>



<p>Exadata X11M delivers significant performance increases across all major workload types compared to the previous X10M generation. These gains stem from the combination of updated hardware components (CPU, memory, flash, network) and continuous software optimizations. Oracle offering these improvements at the <strong>same starting price</strong> as X10M enhances the platform&#8217;s value proposition.</p>



<h3 class="wp-block-heading">AI Vector Search Acceleration</h3>



<p>With the rise of AI-powered applications, <strong>AI Vector Search</strong> for semantic similarity search, RAG applications, recommendation systems, and anomaly detection has become critical. Exadata X11M is explicitly optimized for these workloads.</p>



<p>Exadata&#8217;s unique <strong>AI Smart Scan</strong> feature plays a central role by offloading vector distance calculations and top-K filtering to storage servers. Data is processed in XRMEM or Flash Cache on the storage servers, minimizing network traffic and database server load.</p>



<p>X11M provides the following concrete gains over X10M for AI Vector Search:</p>



<ul class="wp-block-list">
<li><strong>Persistent Vector Index (IVF) Searches:</strong> Up to <strong>55% faster</strong> due to hardware acceleration and transparent storage offload.</li>



<li><strong>In-Memory Vector Index (HNSW) Queries:</strong> Up to <strong>43% faster</strong>, benefiting from faster database server CPUs and memory.</li>



<li><strong>Software Optimizations (All Exadata Platforms):</strong>
<ul class="wp-block-list">
<li><strong>4.7X more data filtering</strong> on storage servers (due to improved top-K efficiency).</li>



<li>Up to <strong>32X faster queries</strong> when searching BINARY vector dimensions. This offers significant speedup with minimal accuracy impact for some use cases.</li>



<li>Offload of vector distance projection to storage servers.</li>
</ul>
</li>
</ul>



<h3 class="wp-block-heading">OLTP Performance Enhancements</h3>



<p>Low latency and high IOPS are crucial for financial transactions, e-commerce, and other critical OLTP systems. Exadata X11M delivers significant improvements:</p>



<ul class="wp-block-list">
<li><strong>IOPS:</strong> Delivers up to <strong>25.2 Million</strong> 8K SQL read IOPS in a single rack. Competitive claims mention significantly higher IOPS compared to rivals like Pure Storage (33x) and Dell PowerMax (3.3x).</li>



<li><strong>Latency:</strong> SQL 8K read latency is reduced to as low as <strong>14 microseconds (14μs)</strong> thanks to XRMEM and RDMA. This is a 21% improvement over X10M&#8217;s 17µs. Cloud comparisons claim up to 70x lower latency than AWS RDS and Azure SQL.</li>



<li><strong>Throughput:</strong> Up to <strong>25% faster</strong> serial transaction processing and up to <strong>25% higher</strong> concurrent transaction throughput compared to X10M, driven by faster cores.</li>



<li><strong>Flash Performance:</strong> Single block reads from flash are up to <strong>43% faster</strong>.</li>
</ul>



<h3 class="wp-block-heading">Analytics Workload Acceleration</h3>



<p>Data warehouses, reporting, and large-scale analytics demand high scan throughput. Exadata X11M shows significant progress here:</p>



<ul class="wp-block-list">
<li><strong>Scan Throughput:</strong>
<ul class="wp-block-list">
<li>Analytical I/O (scan throughput) on storage servers is up to <strong>2.2X faster</strong> than X10M. This is enabled by faster PCIe 5.0 flash and potentially the ability to scan from both flash and XRMEM concurrently.</li>



<li>Flash scan throughput reaches <strong>100 GB/s per server</strong>.</li>



<li>Scanning columnar data cached in XRMEM reaches <strong>500 GB/s per server</strong>.</li>



<li>Total scan throughput per rack can reach up to <strong>8.5 TB/s</strong> (likely from XRMEM). This is a substantial increase from the 1 TB/s per rack claimed for X10M.</li>
</ul>
</li>



<li><strong>Query Processing:</strong> Up to <strong>25% faster</strong> analytics query processing compared to X10M. This results from faster database server cores and faster storage offload processing.</li>



<li><strong>Database In-Memory:</strong> Database In-Memory scans increase up to <strong>500 GB/s</strong>. Exadata extends In-Memory columnar formats into Flash Cache and XRMEM, leveraging ultra-fast SIMD Vector instructions.</li>
</ul>



<p>This balanced improvement across workloads reinforces Exadata&#8217;s suitability for diverse and consolidated environments. The significant gains allow more demanding workloads to coexist efficiently on the same platform. While hardware upgrades provide the raw potential, software optimizations like AI Smart Scan enhancements, RDMA protocols, and Exadata System Software unlock this potential. Some software optimizations, like the 32x faster binary vector queries, are available across Exadata platforms (with appropriate software), demonstrating software value independent of hardware, but achieving peak potential on the latest hardware like X11M.</p>



<h2 class="wp-block-heading">Key Software and Efficiency Improvements</h2>



<p>Beyond raw performance, Exadata X11M introduces significant software and hardware features focused on improving operational efficiency and resource utilization, helping to lower Total Cost of Ownership (TCO) and address modern data center concerns like sustainability and management overhead.</p>



<h3 class="wp-block-heading">Advanced Power Management</h3>



<p>Introduced with Exadata System Software 25.1 and specifically designed for X11M database servers, these capabilities help organizations meet energy efficiency goals and reduce operational costs. Key features include:</p>



<ul class="wp-block-list">
<li><strong>Core Disablement:</strong> When the active core count (<code>pendingCoreCount</code>) is set to 128 or lower, the system automatically powers off 64 unused CPU cores (32 per socket), saving approximately 80 Watts per server without impacting performance for the licensed cores. Cores can be re-enabled if needed.</li>



<li><strong>Power Capping:</strong> Allows setting an overall power consumption target for the database server, useful for regulatory compliance or cooling management. Performance scales linearly with the power cap (e.g., a 10% power reduction results in roughly a 10% peak processing capacity reduction).</li>



<li><strong>Low Power Mode:</strong> Enables scheduling automatic transitions to a low power mode during anticipated low-usage periods (e.g., nights, weekends). The system automatically exits low power mode if demand unexpectedly increases (based on CPU, I/O, or network thresholds) to maintain performance.</li>
</ul>



<p>Combined with the ability to consolidate more workloads onto fewer systems due to X11M&#8217;s higher performance, these power management features can lead to significant savings in infrastructure, power, cooling, and data center space costs.</p>



<h3 class="wp-block-heading">Storage Efficiency</h3>



<ul class="wp-block-list">
<li><strong>Exascale Free Space Management (Exadata SW 25.1):</strong> This enhancement significantly reduces the amount of free space Exascale storage pools require compared to traditional ASM disk groups to successfully complete an automatic data rebalance after a storage device failure. For example, with 9+ HC servers, Exascale requires only 3% free space versus 9% for ASM. This increases usable storage capacity. Such improvements indicate the maturation of the Exascale architecture beyond basic pooling.</li>



<li><strong>Hybrid Columnar Compression (HCC):</strong> A standard Exadata feature that significantly compresses data (often 5x-20x), especially for analytics, saving storage costs and improving performance by reducing I/O.</li>



<li><strong>Exascale Thin Cloning:</strong> Redirect-on-write technology enables space-efficient clones, particularly useful for dev/test environments.</li>
</ul>



<h3 class="wp-block-heading">Other Notable Software Features (Exadata SW 25.1)</h3>



<p>Exadata System Software 25.1 introduces other improvements relevant to the X11M platform:</p>



<ul class="wp-block-list">
<li><strong>Automatic Tuning of ASM Rebalance:</strong> Dynamically adjusts the <code>asm_power_limit</code> based on available I/O bandwidth and client database workload presence. It speeds up rebalance when resources are free and slows it down to prioritize user workloads, minimizing performance impact and reducing manual tuning needs. This contrasts with the traditional static power limit approach.</li>



<li><strong>Simpler Package Management:</strong> Streamlines management of additional non-Exadata software packages during database server updates (that don&#8217;t change the major OS version), reducing maintenance window duration.</li>



<li><strong>Exascale Volume Cloning:</strong> Ability to create clones directly from existing Exascale volumes.</li>



<li><strong>Secure Fabric Default:</strong> Secure internal communication layer is now recommended and enabled by default.</li>



<li><strong>Cache Observability Enhancements:</strong> Improved monitoring of Exadata cache performance via <code>ecstat</code> utility enhancements.</li>



<li><strong>Faster Cisco Switch Upgrades:</strong> Reduces downtime associated with network switch software updates.</li>
</ul>



<p>These efficiency-focused features demonstrate that X11M offers a holistic approach, addressing not just raw speed but also operational costs, management complexity, and environmental impact.</p>



<h2 class="wp-block-heading">Comparative Summary: Exadata X11M vs. X10M</h2>



<p>Exadata X11M represents a significant step up from its predecessor, X10M, offering substantial hardware and software enhancements that translate into tangible performance and efficiency gains. The table below summarizes key technical and performance differences:</p>



<p><strong>Table 3: Feature and Performance Comparison (Exadata X11M vs. X10M)</strong></p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Feature/Metric</th><th>Exadata X10M</th><th>Exadata X11M</th><th>Improvement</th></tr><tr><td><strong>Hardware</strong></td><td></td><td></td><td></td></tr><tr><td>DB Server CPU</td><td>2x AMD EPYC&#x2122; 9J14 (96 Cores/socket)</td><td>2x AMD EPYC&#x2122; 9J25 (96 Cores/socket)</td><td>Up to 25% faster core performance</td></tr><tr><td>Storage Server CPU</td><td>2x AMD EPYC&#x2122; (32 Cores/socket)</td><td>2x AMD EPYC&#x2122; 9J15 (32 Cores/socket)</td><td>Up to 11% faster core performance</td></tr><tr><td>Memory Type/Speed</td><td>DDR5 / 4800MT/s</td><td>DDR5 / 6400MT/s</td><td>33% more bandwidth</td></tr><tr><td>Flash Technology</td><td>PCIe 4.0 NVMe</td><td>PCIe 5.0 NVMe</td><td>Faster (Up to 2.2x for Analytics I/O)</td></tr><tr><td>Network Fabric Speed</td><td>2x100Gb/s RoCE (Active-Active)</td><td>2x100Gb/s RoCE (Active-Active)</td><td>Same nominal speed</td></tr><tr><td>XRMEM Latency</td><td>&lt; 17µs</td><td>&lt; 14µs</td><td>Up to 21% lower</td></tr><tr><td><strong>AI Vector Search</strong></td><td></td><td></td><td></td></tr><tr><td>IVF Search Speedup</td><td>Baseline</td><td>Up to 55% faster</td><td>Up to 55%</td></tr><tr><td>HNSW Search Speedup</td><td>Baseline</td><td>Up to 43% faster</td><td>Up to 43%</td></tr><tr><td>Binary Vector Query Speedup</td><td>Baseline (Software dependent)</td><td>Up to 32x faster (Software Opt.)</td><td>Up to 32x (Software Opt.)</td></tr><tr><td><strong>OLTP</strong></td><td></td><td></td><td></td></tr><tr><td>Max Read IOPS (per rack)</td><td>25.2 Million</td><td>25.2 Million</td><td>Equal (Likely network/protocol limited)</td></tr><tr><td>SQL Read Latency</td><td>&lt; 17µs</td><td>&lt; 14µs</td><td>Up to 21% lower</td></tr><tr><td>Serial Transaction Speedup</td><td>Baseline</td><td>Up to 25% faster</td><td>Up to 25%</td></tr><tr><td>Concurrent Throughput Increase</td><td>Baseline</td><td>Up to 25% more</td><td>Up to 25%</td></tr><tr><td><strong>Analytics</strong></td><td></td><td></td><td></td></tr><tr><td>Analytics I/O Speedup (Storage)</td><td>Baseline</td><td>Up to 2.2x faster</td><td>Up to 2.2x</td></tr><tr><td>Analytics Query Processing Speedup</td><td>Baseline</td><td>Up to 25% faster</td><td>Up to 25%</td></tr><tr><td>Max In-Memory Scan Speed (Server)</td><td>~227 GB/s (Implied/Est. for X10M)</td><td>500 GB/s (from XRMEM)</td><td>&gt;2x increase (with XRMEM scan)</td></tr><tr><td><strong>Efficiency</strong></td><td></td><td></td><td></td></tr><tr><td>Advanced Power Management Features</td><td>No</td><td>Yes (Core disable, Capping, Low Power Mode)</td><td>New</td></tr><tr><td>X11M-Z Option</td><td>No (Eighth Rack existed)</td><td>Yes</td><td>New (More flexible entry-level)</td></tr></tbody></table></figure>



<p>]



<h2 class="wp-block-heading">Conclusion</h2>



<p>Oracle Exadata X11M is a powerful engineered system representing a significant advancement over the previous X10M generation in performance, efficiency, and flexibility. The integration of the latest AMD EPYC processors, faster DDR5 memory, PCIe 5.0 flash, and Exadata RDMA Memory (XRMEM) provides a substantial hardware uplift.</p>



<p>These hardware improvements, combined with intelligent software optimizations like AI Smart Scan and features introduced in Exadata System Software 25.1 (Advanced Power Management, Automatic ASM Rebalance Tuning, enhanced Exascale storage management), further amplify X11M&#8217;s capabilities. The platform delivers demonstrable performance gains across critical workloads, including AI Vector Search (up to 55% faster IVF, 32x faster binary queries), OLTP (up to 25% higher throughput, latency down to 14µs), and Analytics (2.2x faster storage I/O, up to 500 GB/s In-Memory scans).</p>



<p>Crucially, these advancements are offered at the same price point as the previous generation, making X11M a compelling value proposition. Enhanced power management and consolidation potential contribute to lower TCO and improved sustainability, while the introduction of the X11M-Z variant makes the platform accessible to a broader range of customers.</p>



<p>With deployment flexibility across on-premises, OCI, Cloud@Customer, and major multicloud environments (Azure, AWS, Google Cloud), Exadata X11M allows organizations to run their critical Oracle Database workloads wherever needed without application changes.</p>



<p>In summary, Oracle Exadata X11M stands as one of the most advanced and strategic platforms available for organizations seeking peak performance, scalability, and efficiency for their Oracle Database workloads. Its strong emphasis on AI capabilities signals its readiness for future enterprise computing trends.</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/oracle-exadata-x11m-platform.html">Oracle Exadata X11M Platform: Comprehensive Technical Analysis and Comparison with Previous Generations</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Exadata 25ai Future : Automatic Tuning of ASM Rebalance Operations</title>
		<link>https://www.bugraparlayan.com.tr/exadata-25ai-future-automatic-tuning-of-asm-rebalance-operations.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Tue, 18 Mar 2025 13:12:00 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[ASM rebalance automation]]></category>
		<category><![CDATA[ASM rebalance operations]]></category>
		<category><![CDATA[automatic tuning ASM rebalance]]></category>
		<category><![CDATA[dynamic asm_power_limit]]></category>
		<category><![CDATA[Exadata I/O management]]></category>
		<category><![CDATA[Exadata storage optimization]]></category>
		<category><![CDATA[Exadata system software 25.1]]></category>
		<category><![CDATA[Oracle ASM performance]]></category>
		<category><![CDATA[Oracle database automation]]></category>
		<category><![CDATA[Oracle Exadata 25ai]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1389</guid>

					<description><![CDATA[<p>The Oracle Exadata platform continuously evolves, delivering industry-leading performance, availability, and scalability for Oracle Database workloads. A key part of this evolution is the increasing role of automation in database infrastructure management. Oracle Automatic Storage Management (ASM), a volume manager and file system specifically designed for Oracle databases, plays a critical role in simplifying storage &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/exadata-25ai-future-automatic-tuning-of-asm-rebalance-operations.html">Exadata 25ai Future : Automatic Tuning of ASM Rebalance Operations</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p></p>



<p>The Oracle Exadata platform continuously evolves, delivering industry-leading performance, availability, and scalability for Oracle Database workloads. A key part of this evolution is the increasing role of automation in database infrastructure management. <strong>Oracle Automatic Storage Management (ASM)</strong>, a volume manager and file system specifically designed for Oracle databases, plays a critical role in simplifying storage administration.<sup></sup> One of ASM&#8217;s core functions is ensuring balanced data distribution across disks after storage configuration changes (like adding/removing disks) or hardware failures – a process known as <strong>ASM rebalance</strong>.<sup></sup> Rebalancing is vital for maintaining data integrity and optimizing <strong>database performance</strong>. &nbsp;</p>



<p>Oracle Exadata System Software 25ai (specifically release 25.1.0) introduces several innovations.<sup></sup> Among these is the <strong>&#8220;Automatic Tuning of ASM Rebalance Operations&#8221;</strong> feature.<sup></sup> This document provides an in-depth look at this new capability, explaining its mechanism, comparing it to traditional methods, and highlighting its key benefits for <strong>Exadata performance</strong> and <strong>storage management</strong>. &nbsp;</p>



<p><strong>1. Defining &#8220;Automatic Tuning of ASM Rebalance Operations&#8221; in Exadata 25ai</strong></p>



<ul class="wp-block-list">
<li><strong>Definition:</strong> Introduced with Oracle Exadata System Software 25.1.0, &#8220;Automatic Tuning of ASM Rebalance Operations&#8221; is a capability that dynamically adjusts the speed or power of <strong>ASM rebalance</strong> operations based on the system&#8217;s real-time I/O (Input/Output) status and current database workloads. This feature allows ASM to intelligently manage the intensity of its data movement operations, considering the overall health and <strong>performance</strong> of the Exadata system.  </li>



<li><strong>Core Purpose:</strong> The primary goal is twofold:<ol><li>Ensure <strong>ASM rebalance</strong> operations (e.g., restoring data redundancy after disk failure, redistributing data after adding storage, resynchronizing data post-rolling update) complete as quickly as possible.  </li><li>Minimize the negative performance impact on critical database workloads (like OLTP transactions or analytical queries) running concurrently on the system while the rebalance is in progress.  </li></ol>Essentially, this feature aims for rebalance operations that are both efficient (fast) and &#8220;neighbor-friendly,&#8221; sharing system resources harmoniously with other essential processes.</li>
</ul>



<p><strong>2. How Automatic ASM Rebalance Tuning Works</strong></p>



<p>The core principle involves the Exadata system software automatically and dynamically managing the <code>ASM_POWER_LIMIT</code> initialization parameter, which traditionally required manual tuning by a Database Administrator (DBA).<sup></sup> &nbsp;</p>



<ul class="wp-block-list">
<li><strong>Dynamic <code>asm_power_limit</code> Adjustment:</strong>
<ul class="wp-block-list">
<li>The <code>ASM_POWER_LIMIT</code> parameter dictates how many resources (primarily I/O bandwidth and CPU) a rebalance operation consumes, thus determining its speed. Traditionally, DBAs set a static value. A high value (e.g., 11, or up to 1024 in newer versions ) speeds up rebalancing but can negatively impact other workloads. A low value is less intrusive but slows down the rebalance.  </li>



<li>The automatic tuning feature in <strong>Exadata 25ai</strong> replaces this static approach. The system software continuously adjusts the <code>asm_power_limit</code> based on real-time system conditions.  </li>
</ul>
</li>



<li><strong>Monitored System Conditions:</strong> The tuning mechanism primarily monitors:
<ul class="wp-block-list">
<li><strong>I/O Performance and Available Bandwidth:</strong> Exadata software constantly observes the available I/O bandwidth and overall I/O performance of the storage subsystem. If significant I/O capacity is free (low I/O contention), the software automatically increases <code>asm_power_limit</code> to accelerate the rebalance.  </li>



<li><strong>Client Database Workloads:</strong> The system detects active client database workloads requiring I/O resources. When such workloads are present, the software automatically lowers <code>asm_power_limit</code> to protect their performance. This allows the rebalance to run less aggressively in the background, freeing up resources for priority tasks.  </li>
</ul>
</li>



<li><strong>Underlying Algorithms/Enhancements:</strong> Oracle documentation states this feature is enabled by specific &#8220;enhancements and algorithms&#8221; added in Exadata System Software 25.1.0. However, the precise technical details of these algorithms (specific metrics used, decision thresholds, adjustment frequency) are not publicly disclosed, considered part of Exadata&#8217;s proprietary optimizations.  </li>
</ul>



<p>This feature represents a shift from reactive manual tuning (adjusting <code>asm_power_limit</code> after observing issues <sup></sup>) to proactive, automated optimization.<sup></sup> It continuously balances the critical resource of I/O bandwidth between the need for rebalancing and the demands of ongoing database operations. This aligns with Oracle&#8217;s broader automation and autonomous database strategy <sup></sup>, reducing manual intervention and increasing the self-managing capabilities of the Exadata platform, particularly in <strong>storage management</strong>. &nbsp;</p>



<p><strong>3. Solving Traditional ASM Rebalance Challenges</strong></p>



<p>While effective, the traditional manual approach to ASM rebalancing presented several challenges:</p>



<ul class="wp-block-list">
<li><strong>Challenges of Manual Methods:</strong>
<ul class="wp-block-list">
<li><strong>The <code>asm_power_limit</code> Dilemma:</strong> Administrators faced a constant trade-off: complete the rebalance quickly (especially after critical events like disk failure) or slow it down to protect production workload performance. Finding the optimal static value often required trial-and-error and continuous monitoring. Incorrect settings could lead to unnecessarily long rebalances (increasing risk exposure) or unacceptable application slowdowns.  </li>



<li><strong>Potential Performance Impact:</strong> High <code>asm_power_limit</code> values generate intense I/O activity , potentially starving normal database operations of required bandwidth and causing slowdowns.  </li>



<li><strong>&#8220;Overbalancing&#8221; Inefficiency:</strong> Performing storage changes like adding <em>then</em> dropping disks in separate steps could trigger two full rebalance operations, nearly doubling the I/O load and duration compared to a single combined operation. This was a common inefficiency in manual procedures.  </li>



<li><strong>Difficulty Adapting to Variable Workloads:</strong> A static <code>asm_power_limit</code> couldn&#8217;t dynamically adapt to changing workload profiles (e.g., batch processing vs. OLTP). Rebalance might run slower than possible during idle periods or cause contention during peak hours.</li>



<li><strong>Unawareness of Higher Power Limits:</strong> Not all administrators might realize that <code>ASM_POWER_LIMIT</code> can go up to 1024 (not just 11) in newer ASM versions, potentially running rebalances slower than necessary.  </li>
</ul>
</li>



<li><strong>Solutions Provided by Automatic Tuning:</strong> The <strong>Exadata 25ai</strong> automatic tuning feature addresses these issues:
<ul class="wp-block-list">
<li><strong>Eliminates Manual Tuning:</strong> Dynamically adjusts <code>asm_power_limit</code>, removing the need for administrators to manually find and manage this complex balance.  </li>



<li><strong>Dynamic Balancing Act:</strong> Automatically balances rebalance speed (increasing when resources are free) against workload priority (slowing rebalance when critical workloads are detected).  </li>



<li><strong>Optimal Resource Utilization:</strong> Aims for the most efficient use of system resources (especially I/O) based on the current situation, speeding up rebalance during quiet times and protecting workloads during busy periods.</li>



<li><strong>Consistency:</strong> Strives to minimize the impact of rebalancing on other operations, contributing to more predictable and consistent system <strong>performance</strong> regardless of workload.</li>
</ul>
</li>
</ul>



<p>Automatic tuning also reduces operational risks associated with manual configuration errors (incorrect <code>asm_power_limit</code>, inefficient &#8220;overbalancing&#8221; steps).<sup></sup> By automating these decisions <sup></sup>, the likelihood of such errors and their potential negative impact on performance or rebalance duration is significantly lowered. &nbsp;</p>



<p><strong>4. Key Benefits of Automatic ASM Rebalance Tuning</strong></p>



<p>The &#8220;Automatic Tuning of ASM Rebalance Operations&#8221; feature offers tangible advantages for Exadata administrators and users:</p>



<ul class="wp-block-list">
<li><strong>Minimized Performance Impact:</strong> This is the primary benefit. The system automatically throttles rebalance activity when it detects critical database workloads, reducing I/O and CPU consumption. This ensures minimal degradation in application and end-user performance, crucial for latency-sensitive OLTP systems.  </li>



<li><strong>Enhanced Data Resilience and Distribution Health:</strong>
<ul class="wp-block-list">
<li><strong>Faster Rebalancing:</strong> When system resources (especially I/O bandwidth) are available, the feature accelerates the rebalance by increasing <code>asm_power_limit</code>. This significantly shortens the time needed to restore data redundancy after a disk or storage server failure, reducing the window of vulnerability to subsequent failures and improving overall data resilience.  </li>



<li><strong>Efficient Data Distribution:</strong> Speeds up the process of evenly distributing data across disk groups after configuration changes (like adding/removing storage). This promotes efficient use of storage resources and prevents potential performance bottlenecks.  </li>
</ul>
</li>



<li><strong>Improved Manageability:</strong>
<ul class="wp-block-list">
<li><strong>Reduced Administrative Burden:</strong> Eliminates the need for DBAs and system administrators to constantly monitor, evaluate, and manually adjust <code>asm_power_limit</code> during rebalance operations. This boosts <strong>operational efficiency</strong> and frees up administrators for more strategic tasks.  </li>



<li><strong>Simplified Operations:</strong> Greatly simplifies rebalance management, especially in dynamic environments with frequent storage changes or highly variable workload profiles.</li>
</ul>
</li>



<li><strong>Synergy with Exadata Exascale:</strong>
<ul class="wp-block-list">
<li><strong>Ideal for Dynamic, Large-Scale Environments:</strong> The <strong>Exadata Exascale</strong> architecture introduces more flexible, shared, and dynamic management of storage and compute resources. Automatic rebalance tuning helps manage the inherent complexity of storage operations in such large-scale, potentially multi-tenant environments. It works synergistically with other Exascale storage efficiency features like &#8220;Improved Free Space Management&#8221;  to create a more autonomous and efficient storage layer.  </li>



<li><strong>Complements Exascale Efficiency Goals:</strong> A core goal of Exascale is to improve storage resource utilization and reduce costs. Automatic rebalance tuning directly supports this by ensuring rebalance operations themselves use resources efficiently. By preventing unnecessary performance dips and reducing operational overhead, it reinforces Exascale&#8217;s overall value proposition of performance, efficiency, and agility. The dynamism of Exascale might lead to more frequent or complex rebalance scenarios, making manual management even more challenging. Automatic tuning naturally adapts to this dynamism, aiding Exascale in achieving its goals.  </li>
</ul>
</li>
</ul>



<p><strong>5. Configuration and Monitoring</strong></p>



<ul class="wp-block-list">
<li><strong>Configuration:</strong> Current Oracle Exadata System Software 25.1.0 documentation does not provide specific commands or interface options to enable, disable, or fine-tune the &#8220;Automatic Tuning of ASM Rebalance Operations&#8221; feature. This strongly suggests the feature is <strong>enabled by default</strong> on compatible Exadata systems and likely requires no direct user configuration, unlike other offload controls like <code>CELL_OFFLOAD_PROCESSING</code>.  </li>



<li><strong>Monitoring:</strong>
<ul class="wp-block-list">
<li><strong>No Direct Monitoring Mechanism:</strong> The documentation does not describe specific V$ views or metrics to directly monitor the <em>internal workings</em> of the automatic tuning mechanism itself (e.g., the dynamically calculated <code>asm_power_limit</code> value or influencing factors).  </li>



<li><strong>Indirect Monitoring (Observing Effects):</strong> The feature&#8217;s presence and effectiveness can be observed indirectly using standard ASM and system monitoring tools:
<ul class="wp-block-list">
<li><strong><code>V$ASM_OPERATION</code> / <code>GV$ASM_OPERATION</code>:</strong> These core ASM views show the status (<code>STATE</code>), type (<code>OPERA</code>), current power level (<code>POWER</code>), progress (<code>SOFAR</code>, <code>EST_WORK</code>), and estimated time remaining (<code>EST_MINUTES</code>) for ongoing rebalance operations. Observing fluctuations in <code>POWER</code> (if reflected dynamically), <code>EST_RATE</code>, or <code>EST_MINUTES</code> during a rebalance, especially correlating with changes in system load, could indicate automatic tuning at work. For instance, <code>EST_MINUTES</code> might increase when workload increases (rebalance slows) and decrease when workload drops.  </li>



<li><strong>I/O Performance Metrics:</strong> Exadata-specific tools like <code>ecstat</code>  or standard OS (<code>iostat</code>) and database (AWR, ASH) monitoring tools can track overall I/O performance (IOPS, throughput MB/s, latency). OCI Database Management service also offers ASM performance monitoring. Observing changes in these metrics during rebalance (e.g., increased rebalance I/O) while ensuring the I/O performance (especially latency) of critical workloads remains stable would demonstrate the effectiveness of automatic tuning.  </li>



<li><strong>Workload Performance:</strong> The most crucial indicator is the performance of critical applications and database workloads during rebalance operations. Key Performance Indicators (KPIs) like end-user response times and transaction throughput should not degrade significantly, proving the feature is achieving its primary goal.</li>
</ul>
</li>
</ul>
</li>
</ul>



<p>While Oracle likely enables this feature by default due to confidence in its stability, standard monitoring practices remain essential. Administrators should continue to use tools like <code>V$ASM_OPERATION</code> and monitor overall system/workload I/O performance during rebalances to verify expected behavior and detect any potential anomalies.</p>



<p><strong>Manual vs. Automatic ASM Rebalance Tuning Comparison</strong></p>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Feature</th><th>Traditional Manual Tuning</th><th>Exadata 25ai Automatic Tuning</th></tr><tr><td><strong><code>asm_power_limit</code> Setting</strong></td><td>Static, manually set by DBA <sup></sup></td><td>Dynamic, automatically adjusted by Exadata software <sup></sup></td></tr><tr><td><strong>Core Mechanism</strong></td><td>Operates at a fixed parallelism/resource limit</td><td>Continuously adjusts limit based on system I/O &amp; workload <sup></sup></td></tr><tr><td><strong>Factors Considered</strong></td><td>DBA experience, general system expectations</td><td>Real-time I/O bandwidth, active client workloads <sup></sup></td></tr><tr><td><strong>Primary Goal</strong></td><td>Manually balance speed vs. low impact <sup></sup></td><td>Maximize speed AND minimize impact automatically <sup></sup></td></tr><tr><td><strong>Workload Impact</strong></td><td>Can be significant at high limits <sup></sup></td><td>Automatically reduced when workload detected <sup></sup></td></tr><tr><td><strong>Rebalance Speed</strong></td><td>Tied to static limit, potentially suboptimal</td><td>Automatically increased when resources available <sup></sup></td></tr><tr><td><strong>Administrative Effort</strong></td><td>Requires ongoing monitoring &amp; potential tuning <sup></sup></td><td>Minimal/None <sup></sup></td></tr><tr><td><strong>Adaptability</strong></td><td>Cannot adapt to variable workloads without intervention</td><td>Automatically adapts to changing I/O &amp; workload conditions <sup></sup></td></tr></tbody></table></figure>



<p><strong>Conclusion</strong></p>



<p>The <strong>&#8220;Automatic Tuning of ASM Rebalance Operations&#8221;</strong> feature in Exadata System Software 25ai marks a significant advancement for Oracle&#8217;s flagship database machine platform. It makes the critical, yet potentially disruptive, <strong>ASM rebalance</strong> process smarter, more efficient, and considerably more sensitive to concurrent system workloads.</p>



<p>By dynamically adjusting the <code>ASM_POWER_LIMIT</code> based on real-time I/O conditions and database workload demands <sup></sup>, the feature intelligently accelerates rebalancing when resources permit and throttles it during peak usage times, minimizing performance impact. &nbsp;</p>



<p>This automatic tuning directly enhances Exadata&#8217;s overall <strong>performance</strong>, availability, and manageability. It improves system resilience by speeding up redundancy restoration <sup></sup>, reduces administrative overhead by eliminating manual tuning <sup></sup>, and lowers operational risks associated with manual configuration errors. &nbsp;</p>



<p>&#8220;Automatic Tuning of ASM Rebalance Operations&#8221; aligns perfectly with Oracle&#8217;s broader strategy towards greater automation and autonomous capabilities within the Exadata and database ecosystem. By automating this crucial aspect of <strong>storage management</strong>, it contributes to making Exadata a more efficient, self-managing platform, particularly beneficial in dynamic, large-scale <strong>Exadata Exascale</strong> environments.<sup></sup> This feature is a valuable innovation reinforcing Exadata&#8217;s leadership in <strong>database performance</strong> and ease of management.</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/exadata-25ai-future-automatic-tuning-of-asm-rebalance-operations.html">Exadata 25ai Future : Automatic Tuning of ASM Rebalance Operations</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>AI Smart Scan on Oracle Exadata: Accelerating AI Vector Search for RAG and Similarity Search</title>
		<link>https://www.bugraparlayan.com.tr/ai-smart-scan-on-oracle-exadata-accelerating-ai-vector-search-for-rag-and-similarity-search.html</link>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Wed, 12 Feb 2025 12:50:00 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<guid isPermaLink="false">https://www.bugraparlayan.com.tr/?p=1385</guid>

					<description><![CDATA[<p>1. Introduction: Exadata and the Rise of In-Database AI 1.1. Oracle Exadata: High-Performance Database Platform Oracle Exadata stands as a premier enterprise database platform, engineered to run Oracle Database workloads with exceptional performance, availability, and security across all scales and criticalities. Its architecture integrates high-performance database servers, intelligent storage servers, and an ultra-fast, low-latency internal &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/ai-smart-scan-on-oracle-exadata-accelerating-ai-vector-search-for-rag-and-similarity-search.html">AI Smart Scan on Oracle Exadata: Accelerating AI Vector Search for RAG and Similarity Search</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p></p>



<h2 class="wp-block-heading">1. Introduction: Exadata and the Rise of In-Database AI</h2>



<h3 class="wp-block-heading">1.1. Oracle Exadata: High-Performance Database Platform</h3>



<p><strong>Oracle Exadata</strong> stands as a premier enterprise database platform, engineered to run <strong>Oracle Database workloads</strong> with exceptional performance, availability, and security across all scales and criticalities.<sup></sup> Its architecture integrates high-performance database servers, <strong>intelligent storage servers</strong>, and an ultra-fast, low-latency internal network fabric (typically RDMA over InfiniBand).<sup></sup> The core philosophy is the co-design of hardware and software, enabling unique optimizations for database operations at both compute and storage layers.<sup></sup> &nbsp;</p>



<p>Exadata is optimized for both Online Transaction Processing (OLTP) and Data Warehousing (DW)/Analytics. It has evolved to support modern workloads like in-memory analytics, <strong>Artificial Intelligence (AI)</strong>, and Machine Learning (ML), facilitating efficient mixed-workload consolidation.<sup></sup> Its scale-out design allows balanced expansion of compute, storage, and network resources to meet growing demands.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">1.2. The Trend Towards In-Database AI</h3>



<p>Integrating <strong>AI and ML capabilities</strong> directly into database platforms is a significant trend. This approach minimizes or eliminates the need to move data to separate systems for analysis or model training, reducing complexity, latency, and security risks associated with data movement.<sup></sup> Processing data in place allows AI/ML algorithms to leverage the database&#8217;s transactional capabilities, security models, and consistency guarantees.<sup></sup> &nbsp;</p>



<p>Oracle addresses this through its <strong>Converged Database</strong> strategy, managing diverse data types (relational, JSON, graph, spatial, and now vector) and workloads within a single database engine.<sup></sup> This aims to eliminate data silos and management complexity associated with specialized databases. &nbsp;</p>



<h3 class="wp-block-heading">1.3. Report Focus: AI Smart Scan for Vector Processing</h3>



<p>This technical report provides an in-depth analysis of the <strong>AI Smart Scan</strong> feature on the <strong>Oracle Exadata</strong> platform, specifically its <strong>vector processing</strong> capabilities. It defines AI Smart Scan within the Exadata context, explains how it accelerates <strong>Oracle AI Vector Search</strong>, details the <strong>offloading mechanism</strong> for vector tasks to Exadata storage servers (especially within the <strong>Exascale architecture</strong>), and documents claimed performance gains.</p>



<p>The report highlights the importance of this feature for modern <strong>database AI applications</strong>, particularly <strong>Similarity Search</strong> and <strong>Retrieval-Augmented Generation (RAG)</strong> for Generative AI.</p>



<h2 class="wp-block-heading">2. Understanding AI Smart Scan in the Exadata Context</h2>



<h3 class="wp-block-heading">2.1. Evolution from Traditional Smart Scan (SQL Offload)</h3>



<p>Understanding <strong>AI Smart Scan</strong> requires knowledge of its predecessor, the traditional <strong>Smart Scan</strong> (or SQL Offload), a cornerstone of Exadata.<sup></sup> This technology pushes data-intensive SQL processing from database servers to <strong>intelligent storage servers</strong>. &nbsp;</p>



<p>In conventional architectures, full table scans move all data blocks across the network to the database server for filtering (<code>WHERE</code> clauses) and projection (<code>SELECT</code> list), consuming network bandwidth and database server CPU.<sup></sup> &nbsp;</p>



<p>Exadata Smart Scan optimizes this by sending SQL filters and column projections to the storage servers.<sup></sup> Storage servers apply these filters <em>as data is read</em>, returning only relevant rows and columns to the database server.<sup></sup> This dramatically reduces data transfer and database server CPU usage, boosting <strong>database performance</strong>.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">2.2. Defining AI Smart Scan</h3>



<p><strong>AI Smart Scan</strong> is a set of Exadata-specific optimizations designed to accelerate the <strong>AI Vector Search</strong> capabilities introduced in Oracle Database 23ai.<sup></sup> It extends the Smart Scan philosophy to AI vector operations, specifically offloading compute-intensive <strong>vector distance calculations</strong> and <strong>Top-K nearest neighbor filtering</strong> to the storage servers.<sup></sup> &nbsp;</p>



<p>Introduced with Oracle Exadata System Software 24.1 and requiring Oracle Database 23ai or later <sup></sup>, AI Smart Scan leverages the processing power of Exadata storage servers and the high-speed internal network to optimize vector-based queries. &nbsp;</p>



<h3 class="wp-block-heading">2.3. Core Objectives: Performance and Efficiency</h3>



<p>The primary goal of <strong>AI Smart Scan</strong> is to achieve <strong>orders-of-magnitude performance improvements</strong> for <strong>AI Vector Search</strong> queries, especially on large datasets.<sup></sup> This is accomplished by addressing the computational intensity of vector processing. &nbsp;</p>



<p>By offloading distance calculations and Top-K filtering to the <strong>Exadata storage servers</strong> where data resides <sup></sup>, AI Smart Scan achieves: &nbsp;</p>



<ol class="wp-block-list">
<li><strong>Reduced Database Server Load:</strong> Frees up database server CPU resources for other tasks, improving overall system throughput. &nbsp;</li>



<li><strong>Minimized Network Traffic:</strong> Only filtered results (e.g., top K vectors) are sent back, significantly reducing data movement, crucial for high-dimensional vectors. &nbsp;</li>



<li><strong>Low Latency and High Throughput:</strong> Processing data closer to the source, combined with Exadata&#8217;s low-latency RDMA network, results in faster query responses. &nbsp;</li>
</ol>



<p>AI Smart Scan represents a logical extension of Exadata&#8217;s core principle: moving processing closer to the data. It adapts the proven SQL offload architecture to the demanding requirements of modern <strong>AI workloads</strong>, particularly the compute-heavy nature of <strong>vector processing</strong>. The claim of &#8220;orders of magnitude&#8221; performance gains <sup></sup> signifies a fundamental architectural advantage, positioning Exadata as a high-performance platform for vector search, competitive with specialized vector databases. &nbsp;</p>



<h2 class="wp-block-heading">3. Accelerating AI Vector Search with AI Smart Scan</h2>



<h3 class="wp-block-heading">3.1. Oracle AI Vector Search Fundamentals</h3>



<p><strong>Oracle AI Vector Search</strong> is an integrated database capability enabling semantic search based on meaning, not just keywords.<sup></sup> It uses <strong>vector embeddings</strong> – multi-dimensional numerical representations – to capture the semantic meaning of structured and unstructured data (text, images, audio).<sup></sup> Semantically similar items have vectors closer in the vector space.<sup></sup> &nbsp;</p>



<p>Key use cases include:</p>



<ul class="wp-block-list">
<li><strong>Semantic Search:</strong> Searching documents, products by meaning. &nbsp;</li>



<li><strong>Recommendation Systems:</strong> Suggesting similar items based on user preferences. &nbsp;</li>



<li><strong>Anomaly Detection:</strong> Identifying outliers. &nbsp;</li>



<li><strong>Image/Video Search:</strong> Finding visually similar content. &nbsp;</li>



<li><strong>Retrieval-Augmented Generation (RAG):</strong> Enhancing Large Language Model (LLM) accuracy with relevant enterprise data. &nbsp;</li>
</ul>



<p>Oracle Database 23ai provides:</p>



<ul class="wp-block-list">
<li><strong><code>VECTOR</code> Data Type:</strong> Native storage for vector embeddings. &nbsp;</li>



<li><strong>Vector Indexes:</strong> Optimized indexes (HNSW for in-memory, IVF for disk-based) to accelerate similarity searches. &nbsp;</li>



<li><strong>SQL Operators/Functions:</strong> New SQL capabilities (<code>VECTOR_DISTANCE</code>) for performing similarity searches and combining them with other data types. &nbsp;</li>
</ul>



<h3 class="wp-block-heading">3.2. AI Smart Scan&#8217;s Role in Acceleration</h3>



<p><strong>AI Smart Scan</strong> is central to optimizing <strong>AI Vector Search</strong> query performance, especially for large-scale vector data scans.<sup></sup> It addresses the compute-intensive nature of finding nearest neighbors in vast vector datasets. &nbsp;</p>



<p>Acceleration mechanisms include:</p>



<ul class="wp-block-list">
<li><strong>Compute Offload:</strong> Intensive vector distance calculations and Top-K filtering are executed on storage servers, not the database server.</li>



<li><strong>Parallel Processing:</strong> Exadata&#8217;s scale-out architecture allows these offloaded operations to run in parallel across multiple storage servers.</li>



<li><strong>Data Reduction:</strong> Only filtered results (top K vectors) are sent back to the database server, minimizing network traffic.</li>



<li><strong>Hardware Optimization:</strong> AI Smart Scan leverages Exadata&#8217;s ultra-fast storage tiers (XRMEM, Smart Flash Cache) and low-latency RDMA network. &nbsp;</li>
</ul>



<p>These combined mechanisms deliver <strong>low-latency</strong> responses and <strong>high-throughput</strong> processing for AI Vector Search queries.<sup></sup> Oracle&#8217;s strategy of integrating AI Vector Search into the database and accelerating it with <strong>AI Smart Scan</strong> embodies the &#8220;bring AI to the data&#8221; approach <sup></sup>, simplifying architectures compared to using separate vector databases and offering significant efficiency gains, especially for existing Oracle users. The focus on handling &#8220;massive volumes&#8221; and &#8220;high concurrency&#8221; <sup></sup> positions this technology for demanding, mission-critical enterprise AI workloads. &nbsp;</p>



<h2 class="wp-block-heading">4. Offloading Vector Processing to Exascale Storage Servers</h2>



<h3 class="wp-block-heading">4.1. The Offload Mechanism Explained</h3>



<p><strong>AI Smart Scan</strong> pushes the most CPU-intensive vector search steps—<strong>vector distance calculations</strong> and <strong>Top-K filtering</strong>—down to the <strong>Exadata storage servers</strong>.<sup></sup> &nbsp;</p>



<ol class="wp-block-list">
<li><strong>Vector Distance Calculation Offload:</strong> Computations using metrics like Cosine Similarity or Euclidean Distance are performed directly on storage servers, distributing the CPU load. &nbsp;</li>



<li><strong>Top-K Filtering Offload:</strong> Identifying the &#8216;K&#8217; nearest neighbors to a query vector also happens at the storage layer. Only these top candidates (or potential improvements) are returned, preventing unnecessary network transfer of irrelevant vectors. &nbsp;</li>
</ol>



<p>This <strong>storage offload</strong> dramatically reduces network traffic between database and storage servers, conserving bandwidth and lightening the load on the database server, especially critical for high-dimensional vectors.<sup></sup> &nbsp;</p>



<h3 class="wp-block-heading">4.2. Leveraging Exadata Hardware for Vector Processing</h3>



<p>AI Smart Scan&#8217;s effectiveness is tightly coupled with Exadata&#8217;s hardware:</p>



<ul class="wp-block-list">
<li><strong>Exadata RDMA Memory (XRMEM) &amp; Smart Flash Cache:</strong> AI Smart Scan processes vector data at &#8220;memory speed&#8221; using these ultra-fast storage tiers, offering much lower latency than traditional storage. <strong>RDMA (Remote Direct Memory Access)</strong> allows direct data transfer to database server memory, bypassing network stacks for further latency reduction and throughput gains. &nbsp;</li>



<li><strong>Scale-out Architecture:</strong> Offloaded vector operations are parallelized across all available storage servers in Exadata&#8217;s scale-out design, leveraging numerous CPU cores for rapid processing of large datasets. &nbsp;</li>
</ul>



<h3 class="wp-block-heading">4.3. Enhancements in Exadata System Software 25.1</h3>



<p>Oracle continuously refines AI Smart Scan. Release 25.1.0 introduced key improvements <sup></sup>: &nbsp;</p>



<ul class="wp-block-list">
<li><strong>Enhanced Top-K Filtering:</strong> Storage servers maintain a running Top-K set locally, only sending results that improve the current best set back to the database server. This significantly reduces network traffic and improves performance, especially for large K values. &nbsp;</li>



<li><strong>INT8 and BINARY Vector Support:</strong> AI Smart Scan now supports these compact, efficient formats alongside high-precision FLOAT types. BINARY vectors offer up to <strong>32x smaller size</strong> and <strong>40x faster distance computation</strong> compared to FLOAT32, with minimal impact on search quality in some tests. INT8 offers 4x compression with negligible quality difference in evaluations. This caters to diverse accuracy vs. performance needs. &nbsp;</li>



<li><strong>Vector Distance Projection Offload:</strong> When a query selects the vector distance (<code>SELECT VECTOR_DISTANCE(...)</code>), this calculation is now also offloaded to storage servers, further reducing network traffic by avoiding the transfer of large vectors just to compute the distance on the database server. &nbsp;</li>
</ul>



<p>Support for INT8/BINARY formats <sup></sup> broadens AI Smart Scan&#8217;s applicability beyond high-accuracy scenarios to use cases prioritizing performance and efficiency. Offloading distance projection <sup></sup> demonstrates Oracle&#8217;s commitment to refining the offload mechanism for maximum network optimization. &nbsp;</p>



<h3 class="wp-block-heading">4.4. Exascale Architecture and Vector Offload</h3>



<p><strong>Oracle Exadata Exascale</strong> is a next-generation architecture merging Exadata&#8217;s performance with cloud elasticity and cost-effectiveness.<sup></sup> Its loosely-coupled design, separating compute and storage into shared resource pools <sup></sup>, provides an ideal foundation for <strong>AI Smart Scan offload</strong>. &nbsp;</p>



<p>The <strong>Exascale intelligent storage cloud</strong> hosts the storage servers targeted by AI Smart Scan.<sup></sup> Shared storage pools and the RDMA fabric <sup></sup> enable efficient distribution and parallel execution of offloaded vector tasks. Exascale combines the elasticity needed for AI workloads with the raw performance delivered by AI Smart Scan, making Exadata attractive for dynamic, cloud-native AI applications. &nbsp;</p>



<h2 class="wp-block-heading">5. Measured Performance Gains for AI Vector Search</h2>



<p>Oracle reports significant performance improvements for <strong>AI Vector Search</strong> on Exadata using <strong>AI Smart Scan</strong>.</p>



<h3 class="wp-block-heading">5.1. General Acceleration Claims</h3>



<p>Marketing and technical documents often cite acceleration of <strong>up to 30X</strong> <sup></sup> or <strong>up to 32X</strong> <sup></sup> compared to non-offloaded architectures or potentially older Exadata versions. These figures highlight the fundamental benefit of the offload mechanism. &nbsp;</p>



<h3 class="wp-block-heading">5.2. Exadata X11M Platform Gains</h3>



<p>Compared to the previous generation X10M, the latest <strong>Exadata X11M</strong> platform delivers specific gains <sup></sup>: &nbsp;</p>



<ul class="wp-block-list">
<li><strong>Persistent Vector Index (IVF) Searches:</strong> Up to <strong>55% faster</strong> due to intelligent storage offload.</li>



<li><strong>In-Memory Vector Index (HNSW) Queries:</strong> Up to <strong>43% faster</strong>.</li>
</ul>



<p>These improvements reflect the combined effect of newer hardware (AMD EPYC processors) and software optimizations on X11M.</p>



<h3 class="wp-block-heading">5.3. Software Optimizations (All Platforms)</h3>



<p>Optimizations in <strong>Exadata System Software 25.1</strong> benefit all modern Exadata platforms:</p>



<ul class="wp-block-list">
<li><strong>Data Filtering:</strong> <strong>4.7X more</strong> data filtering capacity in storage servers. &nbsp;</li>



<li><strong>BINARY Vectors:</strong> Queries using the newly supported BINARY format can run up to <strong>32X faster</strong> than with FLOAT32 vectors, due to smaller size and faster distance computation. &nbsp;</li>



<li><strong>BINARY Distance Computation:</strong> The distance calculation itself for BINARY vectors can be up to <strong>40X faster</strong> than for FLOAT32. &nbsp;</li>
</ul>



<h3 class="wp-block-heading">5.4. Summary Table of Performance Claims</h3>



<figure class="wp-block-table"><table class="has-fixed-layout"><tbody><tr><th>Performance Claim</th><th>Comparison Point / Context</th><th>Relevant Hardware / Software</th><th>Source Snippet IDs</th></tr><tr><td>Up to 30X faster AI Vector Search</td><td>Traditional architecture / Non-offload</td><td>Exadata (General) / ESS 24ai+</td><td><sup></sup></td></tr><tr><td>Up to 32X faster AI Vector Search</td><td>Traditional architecture / Previous X10M</td><td>Exadata Exascale / ESS 24ai+</td><td><sup></sup></td></tr><tr><td>Up to 55% faster IVF searches</td><td>Exadata X10M platform</td><td>Exadata X11M / ESS 25.1+</td><td><sup></sup></td></tr><tr><td>Up to 43% faster HNSW queries</td><td>Exadata X10M platform</td><td>Exadata X11M / ESS 25.1+</td><td><sup></sup></td></tr><tr><td>4.7X more data filtering</td><td>Previous software versions</td><td>All Exadata Platforms / ESS 25.1+</td><td><sup></sup></td></tr><tr><td>Up to 32X faster BINARY query</td><td>Queries with FLOAT32 vectors</td><td>All Exadata Platforms / ESS 25.1+</td><td><sup></sup></td></tr><tr><td>Up to 40X faster BINARY distance calc</td><td>Distance calc with FLOAT32 vectors</td><td>All Exadata Platforms / ESS 25.1+</td><td><sup></sup></td></tr></tbody></table></figure>



<p></p>



<p>The variety in performance claims highlights that actual gains depend on multiple factors: Exadata hardware generation, software version, vector type (FLOAT, INT8, BINARY), dimensionality, index type (IVF, HNSW), dataset size, and query complexity. Users should evaluate these nuances for their specific scenarios. The strong focus on <strong>AI Vector Search performance</strong> in recent Exadata X11M <sup></sup> and ESS 25.1 <sup></sup> announcements underscores its strategic importance for Oracle, positioning Exadata as a key platform for the growing AI/ML market. &nbsp;</p>



<h2 class="wp-block-heading">6. Importance for Modern Database Applications and AI</h2>



<p>Exadata&#8217;s accelerated <strong>AI Vector Search</strong> via <strong>AI Smart Scan</strong> has significant implications for modern applications, especially those driven by AI.</p>



<h3 class="wp-block-heading">6.1. Enabling In-Database AI Applications</h3>



<p>AI Smart Scan facilitates running AI logic directly within the Oracle database where the data resides, offering key advantages <sup></sup>: &nbsp;</p>



<ul class="wp-block-list">
<li><strong>Reduced Data Movement:</strong> Minimizes costly, complex, and potentially insecure data transfers to external AI platforms. &nbsp;</li>



<li><strong>Enhanced Performance:</strong> Processing data locally with offload capabilities significantly improves query latency for AI algorithms.</li>



<li><strong>Improved Data Security:</strong> Sensitive data remains within the database boundary, protected by Oracle&#8217;s robust security features. &nbsp;</li>



<li><strong>Simplified Architecture:</strong> Reduces the need for complex integrations between disparate data stores and AI tools.</li>



<li><strong>Consistency:</strong> AI operations can benefit from database ACID guarantees.</li>
</ul>



<h3 class="wp-block-heading">6.2. Powering RAG (Retrieval-Augmented Generation) Architectures</h3>



<p><strong>Retrieval-Augmented Generation (RAG)</strong> enhances LLM responses by grounding them in external, often private or real-time, data.<sup></sup> RAG mitigates LLM limitations like knowledge cut-offs and potential inaccuracies (&#8220;hallucinations&#8221;).<sup></sup> &nbsp;</p>



<p><strong>AI Smart Scan-accelerated AI Vector Search</strong> is ideal for the crucial <em>retrieval</em> step in RAG.<sup></sup> It efficiently finds semantically relevant information (documents, records) within the Oracle database based on a user&#8217;s query (converted to a vector). This retrieved context is then fed to the LLM, enabling it to generate more accurate, relevant, and trustworthy responses based on specific enterprise data.<sup></sup> This is vital for building reliable generative AI applications like chatbots and internal knowledge systems. &nbsp;</p>



<h3 class="wp-block-heading">6.3. Driving Similarity Search Use Cases</h3>



<p>AI Vector Search excels at finding semantically similar items beyond simple keyword matching.<sup></sup> <strong>AI Smart Scan</strong> makes large-scale similarity searches efficient, enabling applications like: &nbsp;</p>



<ul class="wp-block-list">
<li><strong>Product/Content Recommendations:</strong> Suggesting items similar to user preferences. &nbsp;</li>



<li><strong>Visual Search:</strong> Finding similar images/videos. &nbsp;</li>



<li><strong>Document Similarity:</strong> Locating related documents in large corpora. &nbsp;</li>



<li><strong>Fraud/Anomaly Detection:</strong> Identifying patterns similar to known fraud or deviations from normal behavior. &nbsp;</li>



<li><strong>Customer Support:</strong> Matching queries to relevant knowledge base articles. &nbsp;</li>



<li><strong>Bioinformatics/Medicine:</strong> Comparing medical images or symptoms to known cases. &nbsp;</li>
</ul>



<p>In all these scenarios, AI Smart Scan&#8217;s offload capabilities ensure efficient execution on large datasets. This integration of high-performance vector processing within the <strong>Converged Database</strong> <sup></sup> is a key differentiator for Oracle, simplifying architectures and potentially lowering TCO compared to using separate specialized databases.<sup></sup> For RAG, performing the retrieval step securely within the database <sup></sup> before potentially sending filtered context to an LLM is a critical advantage for protecting sensitive enterprise data. &nbsp;</p>



<h2 class="wp-block-heading">7. Conclusion and Summary</h2>



<h3 class="wp-block-heading">7.1. Key Findings Summarized</h3>



<ul class="wp-block-list">
<li><strong>AI Smart Scan Defined:</strong> A critical, Exadata-specific optimization accelerating <strong>AI Vector Search</strong> (introduced in Oracle DB 23ai) by offloading compute-intensive vector distance calculations and Top-K filtering to intelligent storage servers.</li>



<li><strong>Offload Mechanism:</strong> Leverages Exadata hardware (RDMA, XRMEM, Flash Cache) to perform vector operations at memory speed near the data, drastically reducing network traffic and database server load.</li>



<li><strong>Continuous Improvement (ESS 25.1):</strong> Enhancements include more efficient Top-K filtering, support for performant INT8/BINARY vector formats, and offloading of vector distance projection.</li>



<li><strong>Performance Gains:</strong> Oracle claims significant speedups (up to 30X/32X) over non-offloaded methods. Exadata X11M offers further gains (up to 55% faster IVF, 43% faster HNSW vs. X10M), and software optimizations provide boosts like 32X faster queries with BINARY vectors. Actual results vary based on configuration and workload.</li>



<li><strong>Importance for Modern AI:</strong> Enables efficient <strong>in-database AI</strong>, simplifies architectures, enhances security, and is crucial for high-performance <strong>RAG</strong> retrieval and various large-scale <strong>similarity search</strong> applications.</li>
</ul>



<h3 class="wp-block-heading">7.2. Future Outlook and Strategic Significance</h3>



<p>Oracle&#8217;s ongoing investment signals the strategic importance of <strong>database AI</strong>. Future enhancements to <strong>AI Smart Scan</strong> might include broader operation offload and deeper integration with cloud architectures like <strong>Exascale</strong>.</p>



<p>The shift towards <strong>in-database AI</strong> promises more agile, efficient, and secure solutions by processing data where it resides. <strong>Exadata AI Smart Scan</strong> is central to Oracle&#8217;s <strong>Converged Database</strong> strategy, offering a compelling alternative to specialized vector databases by combining Exadata&#8217;s proven enterprise capabilities with cutting-edge AI workload acceleration.</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/ai-smart-scan-on-oracle-exadata-accelerating-ai-vector-search-for-rag-and-similarity-search.html">AI Smart Scan on Oracle Exadata: Accelerating AI Vector Search for RAG and Similarity Search</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Oracle Exadata’yı Hızlı Yapan Nedir? Bölüm 2</title>
		<link>https://www.bugraparlayan.com.tr/oracle-exadatayi-hizli-yapan-nedir-bolum-2.html</link>
					<comments>https://www.bugraparlayan.com.tr/oracle-exadatayi-hizli-yapan-nedir-bolum-2.html?noamp=mobile#respond</comments>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Mon, 22 Mar 2021 14:05:00 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[exadata]]></category>
		<category><![CDATA[exadata x8]]></category>
		<category><![CDATA[oracle exadata]]></category>
		<guid isPermaLink="false">http://www.bugraparlayan.com.tr/?p=823</guid>

					<description><![CDATA[<p>Active Storage – Cell OffloadOracle Exadata Node üzerinde storage yada Cell üzerine Cell Offload Processing adını verdiği veri aktarma teknolojisine sahiptir. Bu teknoloji mevcut süreçlerin Storage üzerinde çalıştığı anlamını taşır. Piyasadaki diğer ürünler veri işlemeyi sunucu üzerinde gerçekleştirdiği için bu teknolojiye göre daha az performans gösterir. Ayrıca Oracle Exadata diğer veri tabanı sunucularına nazaran yüksek &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/oracle-exadatayi-hizli-yapan-nedir-bolum-2.html">Oracle Exadata’yı Hızlı Yapan Nedir? Bölüm 2</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p><strong>Active Storage – Cell Offload</strong><br>Oracle Exadata Node üzerinde storage yada Cell üzerine Cell Offload Processing adını verdiği veri aktarma teknolojisine sahiptir. Bu teknoloji mevcut süreçlerin Storage üzerinde çalıştığı anlamını taşır. Piyasadaki diğer ürünler veri işlemeyi sunucu üzerinde gerçekleştirdiği için bu teknolojiye göre daha az performans gösterir. Ayrıca Oracle Exadata diğer veri tabanı sunucularına nazaran yüksek performanslar elde edebilmek için bir çok iş yükünü Cell üzerinde çalıştrmaktadır. Bu sayede diğer All-Flash diskler Exadata performansına erişemez.<br></p>



<p>Varsayılan olarak açık gelen bu özellik CELL_OFFLOAD_PROCESSING parametresi ile düzenlenmektedir. Cell Offload teknolojisinin bazı özellikleri aşağıdaki gibidir.</p>



<ul class="wp-block-list"><li>SQL Offload</li><li>XML &amp; JSON Offload</li><li>RMAN Backup (BCT) Filtering</li><li>Data file vs. REDO I/O Segregation</li><li>Encryption/Decryption Offload</li><li>Fast Data File Creation</li></ul>



<p><strong>Massively Parallel Processing (MPP)</strong><br>MPP teknolojisi çok sayıda işlemicininin yada birden fazla sunucunun aynı amaca hizmeti için tasarlanmış bir teknolojidir. Exadata üzerinde shared memory özelliğini kullandığınızda iş yükleri için tüm kaynaklar seferber edilebilir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/02/image023.gif" alt="" class="wp-image-557759"/></figure>



<p>MPP teknolojisi yatay olarak büyüyebilir, Shared everything yapısına göre kayıp oluşmamaktadır. Shared nothing yapısına göre tek bir veri tabanı kavramı kullanıldığı için avantaj sağlar.</p>



<p><strong>Bloom Filters</strong><br>Exadata, Storage üzerindeki JOIN işlemlerini doğrudan işlemek yerine bunları kolaylaştıran Bloom filtrelerini kullanır. Filtre ölçüleri, ilgili verileri filtrelemek için Exadata Storage katmanına aktarılır. Sürecin son adımı RAC üzerinde yapılırken, asıl iş gücü Bloom Filtreleme mimarisine göre Cell üzerinde gerçekleşir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/02/AsnPV.png" alt="" class="wp-image-557768"/></figure>



<p><strong>Fast Node &amp; Cell Death Detection</strong><br>Tipik bir Cluster yapısı, ethernet üzerinden standart mimarisini kullanır. Bu tüm platformlarda desteklenmesi ve yüksek performans sağlamasına rağmen bazı dezavantajlar içerir. Exadata ilk sürümlerinden bügüne Cluster içeirisndeki beklemeleri engellemek ve Node’lar arasındaki hataları ekarte etmek için Maximum Availability Architecture ( MMA ) temel edinmiştir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/02/oracle-database-12c-with-real-application-clusters-rac-high-availability-ha-best-practices-6-638.jpg" alt="" class="wp-image-557769"/></figure>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/02/screen-shot-2018-02-09-at-12-56-35-pm.png" alt="" class="wp-image-557763"/></figure>



<p>Yukarıdaki grafikte RAC üzerinde kopan bir node için bekleme süresin göstermektedir. MMA teknolojisine göre bunun belirlenmesi 30 saniye ve altında olurken, exadata dışındaki klasik sistemlerde 120 saniyeye kadar çıkabilmektedir.</p>



<p>Yukarıda da belirttiğimiz gibi Exadata entegre bir çözümdür. RAC üzerinde düşen node’ları belirlemek için Infiniband hızını kullanır. Aşağıdaki örnek bir grafik verilmiştir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/02/screen-shot-2018-02-09-at-1-07-51-pm.png" alt="" class="wp-image-557764"/></figure>



<p><strong>Large Write Caching &amp; TEMP Performance</strong></p>



<p>Exadata 128kb üstü verileri Flash Cache üzerinde saklayarak bu teknolojiden tam olarak yayarlanır. Bu mimari herzaman üretilen kapasite ile paralele yürümüştür. Bu sayede yeni çıkan her Exadata makinasında Flash Cache boyutuda aynı ölçüde artar.</p>



<p><strong>Adaptive SQL Optimization</strong><br>Bu özellikle sadece Exadata makinasına has olmamasına rağmen , Exadata için kritik bir öneme sahiptir.DWH ortamlarının çoğunun Exadata üzerine taşındığı günümüzde SQL ifadelerinin optimizasyonu son derece önemlidir. Exadata SQL çalışma planını her zaman üst düzey verimlilikte optimize eder. Planda bir değişiklik algılanırsa bunu hızlı bir şekilde düzenleme yeteneğine sahiptir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/02/tgsql_vm_069.png" alt="" class="wp-image-557770"/></figure>



<p><strong>Optimizing storage use and I/O through compression</strong><br>Oracle Exadata Hybrid Columd adını verdiğini benzersiz bir sıkıştırma teknolojisi sunar. HCC teknolojisi ile büyük veri tabanlarında yüksek miktarda kapasite tasarrufu sağlanır. Bu teknoloji veri sıkıştırmada yenilikçi bir yaklaşımdır.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/02/hcc.jpg" alt="" class="wp-image-557771"/></figure>



<p><strong>Mission Critical HA</strong><br>Oracle Exadata yüksek erişilebilirlik mimarisine tam uyumlu olarak tasarlanmıştır.Her türlü arızlara, insan hatalarına ve kesintilere karşı yedekli olarak tasarlanmıştır.Bununla birlikte Dataguard mimarisine entegre olarak herhangi bir durumda oluşabilecek felaketlere karşı yüksek koruma sağlanmaktadır.</p>



<p>2 bölümde Exadata’yı hızlı yapan belirgin özellikleri dilimiz döndüğünce anlatmaya çalıştık. Diğer yazılarımızda görüşmek üzere.</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/oracle-exadatayi-hizli-yapan-nedir-bolum-2.html">Oracle Exadata’yı Hızlı Yapan Nedir? Bölüm 2</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.bugraparlayan.com.tr/oracle-exadatayi-hizli-yapan-nedir-bolum-2.html/feed</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Oracle Exadata&#8217; yı Hızlı Yapan Nedir? Bölüm 1</title>
		<link>https://www.bugraparlayan.com.tr/oracle-exadata-yi-hizli-yapan-nedir-bolum-1.html</link>
					<comments>https://www.bugraparlayan.com.tr/oracle-exadata-yi-hizli-yapan-nedir-bolum-1.html?noamp=mobile#respond</comments>
		
		<dc:creator><![CDATA[Bugra Parlayan]]></dc:creator>
		<pubDate>Mon, 22 Mar 2021 14:04:00 +0000</pubDate>
				<category><![CDATA[Engineered Systems]]></category>
		<category><![CDATA[exadata]]></category>
		<guid isPermaLink="false">http://www.bugraparlayan.com.tr/?p=821</guid>

					<description><![CDATA[<p>Değerli Dostlar, Bildiğiniz üzere Oracle Exadata makinasının X8M sürümü duyuruldu. Bir çok çevrede bu ürünün neden bu kadar pahalı yada verdiğimiz paranın hakkını veriyor mu? sorularını sık sık duyuyoruz. 2020 yılına girişimizle beraber , Ocak ayının 3. haftasında Oracle duayenlerinden Cem zorba özel bir etkinlikte ürün hakkında nefis bilgiler sundu. Bende yaptığım araştırmaları ve tecrübeleri &#8230;</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/oracle-exadata-yi-hizli-yapan-nedir-bolum-1.html">Oracle Exadata&#8217; yı Hızlı Yapan Nedir? Bölüm 1</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Değerli Dostlar,</p>



<p>Bildiğiniz üzere Oracle Exadata makinasının X8M sürümü duyuruldu. Bir çok çevrede bu ürünün neden bu kadar pahalı yada verdiğimiz paranın hakkını veriyor mu? sorularını sık sık duyuyoruz. 2020 yılına girişimizle beraber , Ocak ayının 3. haftasında Oracle duayenlerinden Cem zorba özel bir etkinlikte ürün hakkında nefis bilgiler sundu. Bende yaptığım araştırmaları ve tecrübeleri dilim döndüğünce sizlerle paylaşmak istedim. Bu makalede Oracle Exadata’yı bu kadar iyi ve rakipsiz yapan şey nedir onu irdeleyeceğiz.</p>



<p>Oracle Exadata’nın günümüzde en çok kıyaslamayı IBM Power serisi ile yaşadığını görüyoruz. Bununla beraber Exadata Dell / EMC yada Purge Storage gibi All-Flash diskleri bulunan ürünlerlede karşılaştırıldığını görebilirsiniz.Bir çok karşılaştırma testinde Exadata makinasının ( özellikle makina olarak belirtiyorum ) bu ürünlere açık ara fark yarattığını fark edeceksiniz. Peki bu nasıl oluyor ?</p>



<p><strong>Bütünleşik Donanım ve Yazılım</strong><br>Öncelikle belirtmeliyim ki ilk fark aslında mimarinin temelinde ortaya çıkıyor. Exadata gerek yazılımı gerekse donanımı ile bütünleşik bir ürün olarak ortaya çıkarken rakipleri bir bileşen olarak kendini göstermektedir. Exadata makinası ilk duyurulduğu dönemlerde HP ürünlerini kabin içerisinde kullanırken, Sun Microsystem satın alması ile artık kendi stratejisi ile ürettiği ürünleri kullanmaktadır ve bileşenleri aşağıdaki gibidir.<br><br></p>



<ul class="wp-block-list"><li>Clustered Servers</li><li>Cluster Interconnect Network</li><li>Storage Network</li><li>Active Storage</li><li>Software</li></ul>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/01/sagug002.gif" alt="" class="wp-image-556292"/></figure>



<p>Yapının bütünleşik olması bize elma ile elmayı, armut ile armutu karşılaştırmamız gerektiğini gösterir. Örneğin Exadata makinasını bir IBM sunucu ile karşılaştırırsanız Exadata Cell yani storage’leri yok sayarsınız. Yine Exadata makinasını Full-Flash storage ürünleri ile karşılaştırırsanız bu seferde üzerinde bulunan sunucuları yok sayarsanız. Exadata makinası bir bütündür ve bütünü oluşturan parçalar sinerji oluşturarak bize fazlasını sağlar. Bu sebeple bir karşılaştırma yaparken aynı çizgide bütünleşik bir mimari ile karşılaştırılması gerekir.</p>



<p><strong>Oracle Database ile tam entegrasyon</strong><br>Oracle Database tartışmasız olarak dünyanın en iyi veri tabanı olduğu ortadadır ve neredeyse bütün platformlarda performanslı olarak çalışabilir. Özellikle açık kaynak sistemleri desteklemesi ile kazandığı işletim sistemi tecrübesi ve uyguladığı innavasyon günümüz sistemlerinin ilerisinde bir noktaya taşıyınca bütünleşik bir mimarinin gerekliliği ortaya çıktı. Exadata’nın ortaya çıkışı Oracle Database ve Oracle Linux ile birlikte tamamen bütünleşik bir yapı oluşturdu. Bu sayede birbiri ile konuşan tam stabil bir ürün ortaya çıktı.</p>



<p><strong>Yüksek Kullanılabilirlik, Yedeklilik ve Ölçeklendirme</strong><br>Exadata’nın yüksek kullanılabilirlik ve ölçeklenebilir yapısı ayrıca tamamen yedekli olarak çalışmaktadır. Sistem mimarisindeki bu gerçek, Exadata makinasındaki tasarımın temelini oluşturur. Örneğin Exadata üzerinde en az üç adet Exadata Cell bulunur. Ürünün üç adet olması bize öncelikli olarak yedekliliği ve kesinti yapmadan yama geçme yeteneği katar. Oracle Exadata üzerinde kesintisiz storage patch geçebilir yada üzerindeki mevcut storage’leri dönüşümlü olarak bakıma alabilirsiniz. Aynı zamanda üzerinde buluan 2 adet Node sayesinde Cluster yapısı gelmektedir.</p>



<p><strong>Büyük Bellek Özellikleri ( DRAM )</strong><br>Exadata büyük ve hız gerektiren veritabanı sunucuları arasında farklı olarak DRAM özelliğine sahiptir. X7-2 modeli ile beraber Nod başına 1.5TB DRAM ile birlikte gelir. Bununla her Nod yatay olarak 12TB kapasiteye kadar genişleyebilir. DRAM sayesinde veritabanı işlemleri fiziksel diskten önce bellek üzerinde işenir ve bu performans sağlar.</p>



<p><strong>Exadata’yı özel kılan nedir ?<br></strong>Yukarıda Oracle Exadata makinasındaki en belirgin özellikleri en basit hali ile aktarmaya çalıştım. Bununla birlikte alt özelliklerede değinmek gerekir.</p>



<ul class="wp-block-list"><li>SQL Offload</li><li>Active Storage – Cell Offload</li><li>Massively Parallel Processing (MPP) Design</li><li>Bloom Filters</li><li>Storage Indexes</li><li>High Bandwidth, Low Latency Storage Network</li><li>High Bandwidth, Low Latency Cluster Interconnect</li><li>Fast Node Death Detection</li><li>Smart Flash Cache</li><li>Write-Back Flash Cache</li><li>Large Write Caching &amp; Temp Performance</li><li>Smart Flash Logging</li><li>Smart Fusion Block Transfer</li><li>NVMe Flash Hardware</li><li>Columnar Flash Cache</li><li>In-Memory Fault Tolerance</li><li>Adaptive SQL Optimization</li><li>Exadata Aware Optimizer Statistics</li><li>DRAM Cache in Storage</li></ul>



<p><strong>SQL Offload</strong><br>Exadata ilk çıkışı ile beraber yaptığı yenilik SQL Offload olarak görücüye çıktı. Yukarıda bahsettiğimiz gibi Exatada’nın bütünleşik bir mimari olması ve Cell adını verdiğini storage yapısı rutin sorgulardaki işleyişi ters düz etti. Normal şartlarda veri tabanında istenilen veriler sunucu üzerinde işlenirken Exadata üzerinden çağırılan veriler cell üzerinden döner ve bu performans üzerinde ciddi kazanımlar sağlar.</p>



<p>Farklı bir değişle çağırdığınız sorgular Node üzerindeki değil, storage üzerindeki işlemcide işlenir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/01/screen-shot-2018-02-09-at-2-20-45-pm.png" alt="" class="wp-image-556289"/></figure>



<p><strong>High Bandwidth, Low Latency Storage Network</strong><br>Exadata üzerinde barındırdığı ürünler arasında Infiniband teknolojisi ile haberleşir.Mevcut ağ yapısı Exadata makinasının konumlandığı ağ üzerinden izole şekilde çalıştığı için public olarak tanımlanmaktadır. Bu sayede güvenli bir iletişim sağlanır. Oracle tavsiyelerinde bu ağın genişletilmesini önermez.<br><br>Exadata içerisindeki Infiniband ağı saniyede 40 Gigabit olarak çalışır.DWH ortamlarında ve Büyük veri çalışmalarında bant genişliği önemlidir. Infiniband sayesinde Exadata makinası yüke binmeden veri trafiğinin üstesinden gelebilir.<br><br>Bununla birlikte Oracle veritabanı için REDO LOG kavramı hayati bir önem taşımaktadır. Redo Log yazma işlemleri I/O Wait ve commit işlemlerine duyarlı olduğu için storage üzerindeki işlemlerin maksimum seviyede olması istenir. Infiniband sayesinde bu süreç problemsiz giderilir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/01/A2848-IB_cnnct_Extdta-full_rck-1024x768.jpg" alt="" class="wp-image-556301"/></figure>



<p><strong>DRAM Cache in Storage</strong></p>



<p>Oracle Exadata veri erişiminde öncelikle DRAM teknolojisini kullanır. Mevcut veriler bu sayede ilk olarak Disk yerine DRAM üzerindeki önbellekten okunmaktadır. Exadata Smart Flash Cache hızına ek olarak 2.5 kat daha fazla verim elde edilir.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/01/screen-shot-2018-02-10-at-11-49-55-am.png" alt="" class="wp-image-556303"/></figure>



<p>Yukarıdaki görsel örnekte görüldüğü üzere Exadata bellek yapısında minumum I/O ile yüksek performans sağlamaktadır. DRAM cache özelliğini daha geniş incelemek için aşağıdaki video takip edilebilir.</p>



<figure><iframe width="730" height="411" src="https://www.youtube.com/embed/jsEwiYdU2gE?feature=oembed" allowfullscreen=""></iframe></figure>



<p><strong>In-Memory Fault Tolerance</strong><br>Oracle In-Memory özelliği duyurulduğu 12C sürümü ile birlikte bütün platformlarda desteklenmektedir. Bu özelliği Oracle Exadata makinasında çalıştırmanız bazı avantajları bulunur. In-Memory özelliğini Exadata üzerinde çalıştırmanız en önemli avantajı yine Exadata’ya özgü bellek içi hata töleransıdır.</p>



<figure class="wp-block-image"><img decoding="async" src="https://www.cozumpark.com/wp-content/uploads/2020/01/screen-shot-2018-02-09-at-4-28-41-pm.png" alt="" class="wp-image-556320"/></figure>



<p>In-Memory Fault Tolerance verilerin Exadata üzerinde koşan Oracle veri tabanı node’ları arasında etkili bir şekilde mirror edilebileceği anlamına gelir. Bu özellik geliştiricilere DRAM ön belleğini daha agrasif kullanma avantajlarını sağlatabilir.</p>



<p>Oracle Exadata makinasının fiziksel ve yapısal özelliklerini tanıttığımız bu yazı dizimizin sonuna geldik. Bölüm 2 de görüşmek üzere.</p>
<p>The post <a rel="nofollow" href="https://www.bugraparlayan.com.tr/oracle-exadata-yi-hizli-yapan-nedir-bolum-1.html">Oracle Exadata&#8217; yı Hızlı Yapan Nedir? Bölüm 1</a> appeared first on <a rel="nofollow" href="https://www.bugraparlayan.com.tr">Bugra Parlayan | Oracle Database &amp; Exadata Blog</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.bugraparlayan.com.tr/oracle-exadata-yi-hizli-yapan-nedir-bolum-1.html/feed</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
