Internal ORA-00600 error code, parameter [kdsgrp1]

Source: Internet
Author: User
Physical standby database error: Internal ORA-00600 error code, parameter [kdsgrp1], [], [], [], [], [], [], [], [], [], [], [] the SQL statement that produces an error is one second.

Physical standby database error: Internal ORA-00600 error code, parameter [kdsgrp1], [], [], [], [], [], [], [], [], [], [], [] the SQL statement that produces an error is one second.

Physical standby database error: Internal ORA-00600 error code, parameter [kdsgrp1], [], [], [], [], [], [], [], [], [], [], []

The SQL statement that produces the error is a select statement.

Query this error:

Applies:
Oracle Server-Enterprise Edition-Version 10.2.0.4 and later
Information in this document applies to any platform.
* ** Checked for relevance on 12-Dec-2012 ***


Purpose
This document discusses the ora-600 [kdsgrp1] error, its possible causes and the work around solutions that can be tried.

Troubleshooting Steps
The ora-600 [kdsgrp1] error is thrown when a fetch operation fails to find the expected row. the error is hit in memory and so may be a memory only error or an error that results from specified uption on disk.

This error may indicate (but is not restricted to) any of the following conditions:

Lost writes
Parallel DML issues
Index upload uption
Data block upload uption
Consistent read [CR] issues
Buffer cache uption

Common Work Around Solutions

If the issue is in memory only we can try to immediately resolve the issue by flushing the buffer cache but remember to consider the performance impact on production systems:

Alter system flush buffer_cache;
If we have an intermittent consistent read issue we can try disabling rowCR which is an optimization to reduce consistent-read rollbacks during queries by setting _ row_cr = FALSE in the initialization files. however, this cocould lead to performance degradation of queries. please check the ratio of the two statistics "RowCR hits"/"RowCR attempts" to determine whether the workaround is to be used.

If this is a result of index upload uption then we can drop and rebuild the index. Note that this will require a maintenance window on production systems.

Root Cause Determination
Now lets look at how we discover the root cause of the problem: the first step in finding the root cause of this issue is to inspect the generated trace file. the ora-600 will generate both a trace file in the trace directory and an incident file under the incident id within the incident directory.
The top part of the trace file tells us the SQL that was being run when the error was hit:

----- Current SQL Statement for this session (SQL _id = 9mamr7xn4wg7x )-----

This immediately shows us the data objects that were accessed. searching the trace file for the text string 'Plan table' will locate the SQL execution Plan that is dumped within this trace file. for a persistent issue this allows us to determine which indexes have been accessed and so identify indexes that shocould be validated to check for block upload uption:

SQL> analyze index scott. pk_dept validate structure online;

Index analyzed.

An other approach we can take is to use the file and block information contained in the trace file. At the top of the trace file we will find information on the block where the uption was found:

* ** Session id: (3202.5644) 04:12:16. 910
Row 07c7c8c7. a continuation
File #31 block #510151 slot 11 not found

This information can be used to identify the object details in dba_extents:

Select owner, segment_name, segment_type, partition_name, tablespace_name
From dba_extents
Where relative_fno =
And Between block_id and (block_id + blocks-1 );

We can then validate this object, for example a table and all it's indexes:

Analyze table scott. dept validate structure cascade online;
Remember that we may be dealing with a permanent upload uption that is not located in the object blocks themselves. Examples of this include:

Dictionary upload uption issue from transportable tablespace operations: check dba_tablespaces to see if the tablespace has been plugged in.
Lost writes in ASM diskgroup mirrors-most likely to be seen when there is heavy IO and disk resync activity. To check this run dbms_diskgroup.checkfile to detect mirror discrepancies
If analyze reports no longer uption then check if there are any chained rows on the table. if these exist then we may have an undetected reply uption and the issue shocould reproduce whenever the SQL is run. exporting the table will also detect this issue.

If analyze and exporting the table (in the presence of chained rows) both report no errors then this shoshould be considered a consistent read issue.

Once you understand the nature of the problem you can review the list of known bugs and determine which one matches your condition. if you cannot determine which issue is affecting you then open a service request with Oracle Support and upload the RDBMS and ASM (if applicable) instance alert logs for all nodes, any trace and incident files generated and a full description of the nature of the problem.

Our database is not damaged. If you refresh the buffer cache, no error is reported.

Related reading:

GoldenGate performs two-way Oracle-Oracle replication without using a Data Pump

One-way Oracle-Oracle Replication Using the GoldenGate Data Pump

How to debug Oracle Data Pump (expdp/impdp)

Oracle Database Export data pump (EXPDP) file storage location

Export of Oracle 10g Data Pump Partition Table

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.