dimesnion no present value in the data file

Hi all

I have a data file that does not contain the value of an dimesnion. It is a constant value. In the cartographic section, I tried to write this constant value so that dimesnion. But thiis gives an error. How we specify the value of this dimension?

Thank you

If you try to predict a constant value of type char or varchar, mapping of Stagging and provide a value like this: '12' (or) 'value '.
In case another number or date you can provide it to the level of the target or release. If you use the target, then I$ do not use this column it will be directly inserted into the target Table.

I hope this helps. Can you please provide the error message if the above does not help you.

Tags: Business Intelligence

Similar Questions

  • Enter the value for the data files:

    Hi all
    I use 10.2.0.1.0
    I have a tablespace "hg" with two data files.
    whenever I use after an order to drop this tablespace I hv asked values for data files
    Welcome is:

    SQL > drop tablespace hg, including content and data files;

    SQL > drop tablespace hg, including content and data files;

    I have two data files to the tablespace named as hg.dbf, hg1.dbf

    now what I hv value to enter? or how to calculate the value.


    Thank you

    Honey says:
    Yes I know its 'AND' but when I using the '&' then oracle asked me of ' enter the values of the data files: "k".
    Please try first, then the calculation of the value.
    It can be a variable binding value.
    If you find the answer please suggest steps

    Simple words, the "DROP TABLESPACE" command does not support '&' character. & is a substitution of placeholder for SQL variable * more.

    Even if you provide the value for the "data files" prompt, the command does not work as it would lead to "drop tablespace including contents tsname *"details of your file *.

  • Defining the new path for the data files for restoring using the VALUE of NEWNAME FOR DATABASE

    Version: 11.2.0.3 Linux

    Today, I had to do a restore RMAN to a new server and I came across the post following RTO on the VALUE of NEWNAME FOR DATABASE

    ALTER database open resetlogs upgraded;         error to throw

    So, I thought to use it to indicate the new location of the data files to restore.

    That's what I did
    ===================

    Restore the control file and catalog items to backup using the command of CATALOGUE START WITH. Then I started the restoration
    $ rman target / cmdfile=restore.txt
    
    Recovery Manager: Release 11.2.0.3.0 - Production on Thu Jul 26 04:40:41 2012
    
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    
    connected to target database: SPIKEY (DBID=2576163333, not open)
    
    RMAN> run
    2>  {
    3>  SET NEWNAME FOR DATABASE TO '/fnup/hwrc/oradata/spikey';
    4>  restore database  ;
    5>  }
    6>
    7>
    8>
    executing command: SET NEWNAME
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of set command at 07/26/2012 04:40:43
    RMAN-06970: NEWNAME '/fnup/hwrc/oradata/spikey' for database must include %f or %U format
    
    Recovery Manager complete.
    Don't know how it worked for Levi without %f or %U. So, I added %f
     $ vi restore.txt
     $ cat restore.txt
    run
     {
     SET NEWNAME FOR DATABASE TO '/fnup/hwrc/oradata/spikey/%f';
     restore database  ;
     }
    
    
     $ rman target / cmdfile=restore.txt
    
    Recovery Manager: Release 11.2.0.3.0 - Production on Thu Jul 26 04:45:45 2012
    
    Copyright (c) 1982, 2011, Oracle and/or its affiliates.  All rights reserved.
    
    connected to target database: SPIKEY (DBID=2576163333, not open)
    
    RMAN> run
    2>  {
    3>  SET NEWNAME FOR DATABASE TO '/fnup/hwrc/oradata/spikey/%f';
    4>  restore database  ;
    5>  }
    6>
    7>
    8>
    executing command: SET NEWNAME
    
    Starting restore at 26-JUL-12
    using target database control file instead of recovery catalog
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=19 device type=DISK
    
    channel ORA_DISK_1: starting datafile backup set restore
    channel ORA_DISK_1: specifying datafile(s) to restore from backup set
    channel ORA_DISK_1: restoring datafile 00001 to /fnup/hwrc/oradata/spikey/1
    channel ORA_DISK_1: restoring datafile 00002 to /fnup/hwrc/oradata/spikey/2
    channel ORA_DISK_1: restoring datafile 00003 to /fnup/hwrc/oradata/spikey/3
    channel ORA_DISK_1: restoring datafile 00004 to /fnup/hwrc/oradata/spikey/4
    channel ORA_DISK_1: restoring datafile 00005 to /fnup/hwrc/oradata/spikey/5
    channel ORA_DISK_1: restoring datafile 00006 to /fnup/hwrc/oradata/spikey/6
    channel ORA_DISK_1: restoring datafile 00007 to /fnup/hwrc/oradata/spikey/7
    channel ORA_DISK_1: restoring datafile 00008 to /fnup/hwrc/oradata/spikey/8
    channel ORA_DISK_1: restoring datafile 00009 to /fnup/hwrc/oradata/spikey/9
    channel ORA_DISK_1: reading from backup piece /u07/bkpfolder/SPIKEY_full_01nh0028_1_1_20120725.rmbk
    channel ORA_DISK_1: errors found reading piece handle=/u07/bkpfolder/SPIKEY_full_01nh0028_1_1_20120725.rmbk
    channel ORA_DISK_1: failover to piece handle=/u07/dump/bkpfolder/SPIKEY_full_01nh0028_1_1_20120725.rmbk tag=SPIKEY_FULL
    channel ORA_DISK_1: restored backup piece 1
    channel ORA_DISK_1: restore complete, elapsed time: 00:01:56
    Finished restore at 26-JUL-12
    
    Recovery Manager complete.
    As you can see, RMAN restore data files to the desired location. But the data file names ended up as
    1
    2
    3
    .
    .      
    .
    9
    !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! -----------| Holy Cow |-----------------------------


    So I had to rename each file as below
    $ mv 1 /fnup/hwrc/oradata/spikey/system01.dbf
    $ mv 2 /fnup/hwrc/oradata/spikey/sysaux01.dbf
    $ mv 3 /fnup/hwrc/oradata/spikey/undotbs01.dbf
    I would have been better in execution of the order for each data below file
    SET NEWNAME FOR DATAFILE
    Now, I think, there is no advantage in using NEWNAME SET of DATABASE to. Only the disadvantages. I did anything wrong above?

    Martin;

    On the issue of the VALUE of NEWNAME FOR DATABASE, you must specify at least one of the first three of the following substitution variables to avoid collisions of names: %b f % U. see semantic entry for TO 'filename' for a description of the possible substitution variables.

    You use %f

    %b
    
    Specifies the filename without the fully qualified directory path. For example, the datafile name /oradata/prod/financial.dbf is transformed to financial.dbf. This variable enables you to preserve the names of the datafiles while you move them to different directory. During backup, it can be used for the creation of image copies. The variable cannot be used for OMF datafiles or backup sets.
    
    %f
    
    Specifies the absolute file number of the datafile for which the new name is generated. For example, if datafile 2 is duplicated, then %f generates the value 2.
    
    %U
    
    Specifies a system-generated unique filename. The name is in the following format: data-D-%d_id-%I_TS-%N_FNO-%f. The %d variable specifies the database name. For example, a possible name might be data-D-prod_id-22398754_TS-users_FNO-7.
    

    Source - E10643-01

    Backup and recovery reference

    http://docs.Oracle.com/CD/E14072_01/backup.112/e10643/rcmsynta2014.htm

    I see CKPT and I agree on that!

    Best regards

    mseberg

  • Noise DC value of the ini file values

    Hello

    I am trying to build a generator of signals (sine, triangle, sawtooth, square random noise, DC). I managed to get the job of waveform using the signal generator of NOR built based on VI. but I want to add the DC case that I managed to do. but I want the values (magnitude and frequency) to be in an ini file. My idea is manual sets the values in the file, once I run the vi and select the case of DC noise, it should read the file values and the data output and once I have stop, it records the values of dc before returning.

    So, I'm stuck trying to figure how to do this. I hope that I was clear enough. I don't have a code to view since I haven't started yet coding on this part. Help, please.

    Clearly enough explained.

    Take a look at the screw configuration file, found in e/s from file > Config File live. There also many examples n the community, as well as in the Finder of the example, illustrating the use of these. A good starting point is the sample Configuration settings file write example Finder, and reading Configuration settings file.

    Your .ini file will look something like this, once created:

    [DC noise settings]

    Amplitude = 1.234567

    Frequency = 8.901234

    This is equivalent to:

    [Name of section]

    Key 1 = value

    Button 2 = value

    When asked to read a specific key in a certain section, VI only through the file for the named section, and then search for they key the name you provided it. Remember, you will need to know what type of data is when you read back.

    You can use VI writing to modify specific keys in specific sections as soon as you close the file.

    This is one of a multitude of methods, you can use to do this, but are probably the simplest.

  • DAQmx PDM, add the external value to the log file

    Hello

    I would like to know if it is possible to permanently add an external double value to the TDMS file that is created by the DAQmx Configure logging (PDM) VI?

    Thus, for example, I collect double data of 8 channels, DAQmx, speed of 1 Hz (analog double, 1 d). I'll get these 8 values with corresponding time stamps in the TDMS file created. Is there a way to "tunnel" a double value external to the same TDMS file?

    If there is no such option, I will configure just the TDMS logging explicitly not through the DAQmx driver.

    Thank you!

    with any possibility of foor loop iteration time zero will give you a default value of reference if your reference will not be valid in this case.

  • Generate the analog waveform based on the data file

    I want to create an analog voltage output that follows I have a data file (excel, csv, text (which is easy)).  The data file creates a waveform with equal time between steps (dT =.0034 sec).  After the output through all the data points, I want it repeat indefinitely.

    What is the best way to create the waveform of a data file?

    To create a type of waveform data, calculate the dt by subtracting two values in column 1 and get the array of Y from column 2. If you save the file as a comma separated or tab text file, you can then use the spreadsheet file read. After obtaining a 2D array, you would use the index table and the subset of table functions.

    Assuming you use a capture card data OR for the output signal, you can pass a type of waveform data to a writing DAQmx and set for the generation of types.

  • Open the data file Outlook - access denied

    I have an Outlook data file created and used when I was running Outlook 2003 on Windows XP - Pro. I have a new computer with Windows 7 Home Premium. I'm still on Outlook 2003. When I try to open the data file, I get - file access is denied. You don't have permission to access the file C:\Users\Valued Customer\Outlook_xxx.pstr

    I am logged in as administrator.
    The permissions for all folders in the path and the file itself grant full control.

    I changed the owner to my new login name.

    NG wrong - try: http://social.answers.microsoft.com/Forums/en-US/category/officeoutlook TaurArian [MVP] 2005-2010-implementation to date of Services

  • When OMF add the data file in the tablespace

    Hi friends,

    We use oracle 11.2 with CMS (oracle managed file) for the database on the linux platform. the parameter - DB_CREATE_FILE_DEST had been set. at the present time, we received a message from global warming this tablespace criterion is 85% full.

    According to the document of the oracle, OMF auto will add a data file in the tablespace tast. more than a week, I do not see any data file has been added by OMF.

    I want to know when OMF adds the data file in the tablespace? 85% or 95% or some (parameter) setting this action?

    Thank you

    newdba

    OMF does not automatically add a new data file.  You must explicitly add a new data file with the ALTER TABLESPACE ADD DATAFILE command tbsname.

    What is OMF is to provide a Unique name and a default size (100 MB) for the data file.  That is why the ALTER TABLESPACE... Command ADD DATAFILE that you run didn't need to specify the file size or file name.

    Hemant K Collette

  • run the Idoc function in the data file returned by the service of GET_FILE

    Hello

    I'm new to this forum, so thank you in advance for any help and forgive me of any error with the post.

    I'm trying to force the execution of a custom Idoc function in a data file Complutense University of MADRID, when this data file is requested from the University Complutense of MADRID through service GET_FILE.

    The custom Idoc function is implemented as a filter of the computeFunction type. One of the datafile has appealed to my custom Idoc function:
    * < name wcm:element = "MainText" > [! - $myIdocFunction ()-] < / wcm:element > *.

    The data file is then downloaded with CRMI via service GET_FILE, but the Idoc function is not called.

    I tried to implement another filter Idoc type sendDataForServerResponse or sendDataForServerResponseBytes, that store objects cached responseString and responseBytes, personalized in order to look for any call to my function in the response object Idoc, eventually run the Idoc function and replace the output of the Idoc in the response. But this kind of filter will never run.

    The Idoc function myIdocFunction is executed correctly when I use WCM_PLACEHOLDER service to get a RegionTemplate (file .hcsp) associated with the data file. In this case, the fact RegionTemplate refers to the element of "MainText" data file with <!-$wcmElement ("MainText")->. But I need to make it work also with service GET_FILE.

    I use version 11.1.1.3.0 UCM.

    Any suggestion?
    Thank you very much
    Francesco

    Hello

    Thank you very much for your help and sorry for this late reply.

    Your trick to activate the complete detailed follow-up was helpful, because I found out I could somehow use the filter prepareForFileResponse for my purpose and I could also have related to the implementation of the native filter pdfwatermark. PdfwFileFilter .

    I managed to set up a filter whose purpose is to force the Idoc assessment of a predefined list of functions Idoc on the output returned by the service GET_FILE. Then I paste the code I have written, in which case it may be useful for other people. In any case, know that this filter can cause performance problems, which must be considered carefully in your own use cases.

    First set the filter in the set of filters in file .hda from your device:

    Filters @ResultSet

    4

    type

    location

    parameter

    loadOrder

    prepareForFileResponse

    mysamplecomponent. ForceIdocEvaluationFilter

    null

    1

    @end

    Here is a simplified version of the implementation of the filter:

    / public class ForceIdocEvaluationFilter implements FilterImplementor {}

    public int doFilter (workspace ws, linking DataBinder, ExecutionContext ctx) survey DataException, ServiceException {}

    Service string = binder.getLocal ("IdcService");

    String dDocName = binder.getLocal ("dDocName");

    Boolean isInternalCall = Boolean.parseBoolean (binder.getLocal ("isInternalCall"));

    If ((ctx instanceof FileService) & service.equals ("GET_FILE") &! isInternalCall) {}

    FileService fileService = ctx (FileService);

    checkToForceIdocEvaluation (dDocName, fileService);

    }

    continue with other filters

    Back to CONTINUE;

    }

    ' Private Sub checkToForceIdocEvaluation (String dDocName, FileService fileService) throws DataException, ServiceException {}

    PrimaryFile file = IOUtils.getContentPrimaryFile (dDocName);

    Ext = FileUtils.getExtension (primaryFile.getPath ());

    If (ext.equalsIgnoreCase ("xml")) {}

    forceIdocEvaluation (primaryFile, fileService);

    }

    }

    forceIdocEvaluation Private Sub (file primaryFile FileService fileService) throws ServiceException {}

    String multiplesContent = IOUtils.readStringFromFile (primaryFile);

    Replacement ForceIdocEvaluationPatternReplacer = new ForceIdocEvaluationPatternReplacer (fileService);

    String replacedContent = replacer.replace (fileContent);

    If (replacer.isMatchFound ()) {}

    setNewOutputOfService (fileService, replacedContent);

    }

    }

    ' Private Sub setNewOutputOfService (FileService fileService, String newOutput) throws ServiceException {}

    File newOutputFile = IOUtils.createTemporaryFile ("xml");

    IOUtils.saveFile (newOutput, newOutputFile);

    fileService.setFile (newOutputFile.getPath ());

    }

    }

    public class IOUtils {}

    public static getContentPrimaryFile (String dDocName) survey DataException, ServiceException {queue

    DataBinder serviceBinder = new DataBinder();

    serviceBinder.m_isExternalRequest = false;

    serviceBinder.putLocal ("IdcService", "GET_FILE");

    serviceBinder.putLocal ("dDocName", dDocName);

    serviceBinder.putLocal ("RevisionSelectionMethod", "Latest");

    serviceBinder.putLocal ("isInternalCall", "true");

    ServiceUtils.executeService (serviceBinder);

    String vaultFileName = DirectoryLocator.computeVaultFileName (serviceBinder);

    String vaultFilePath = DirectoryLocator.computeVaultPath (vaultFileName, serviceBinder);

    return new File (vaultFilePath);

    }

    public static String readStringFromFile (File sourceFile) throws ServiceException {}

    try {}

    return FileUtils.loadFile (sourceFile.getPath (), null, new String [] {"UTF - 8"});

    } catch (IOException e) {}

    throw new ServiceException (e);

    }

    }

    Public Shared Sub saveFile (String source, destination of the file) throws ServiceException {}

    FileUtils.writeFile (source, destination, "UTF - 8", 0, "is not save file" + destination);

    }

    public static getTemporaryFilesDir() leader throws ServiceException {}

    String idcDir = SharedObjects.getEnvironmentValue ("IntradocDir");

    String tmpDir = idcDir + "custom/MySampleComponent";

    FileUtils.checkOrCreateDirectory (tmpDir, 1);

    return new File (tmpDir);

    }

    public static createTemporaryFile (String fileExtension) leader throws ServiceException {}

    try {}

    The file TmpFile = File.createTempFile ("tmp", "." + fileExtension, IOUtils.getTemporaryFilesDir ());

    tmpFile.deleteOnExit ();

    return tmpFile;

    } catch (IOException e) {}

    throw new ServiceException (e);

    }

    }

    }

    Public MustInherit class PatternReplacer {}

    Private boolean matchFound = false;

    public string replace (CharSequence sourceString) throws ServiceException {}

    Matcher m = expand () .matcher (sourceString);

    StringBuffer sb = new StringBuffer (sourceString.length ());

    matchFound = false;

    While (m.find ()) {}

    matchFound = true;

    String matchedText = m.group (0);

    String replacement = doReplace (matchedText);

    m.appendReplacement (sb, Matcher.quoteReplacement (replacement));

    }

    m.appendTail (sb);

    Return sb.toString ();

    }

    protected abstract String doReplace(String textToReplace) throws ServiceException;

    public abstract Pattern getPattern() throws ServiceException;

    public boolean isMatchFound() {}

    Return matchFound;

    }

    }

    SerializableAttribute public class ForceIdocEvaluationPatternReplacer extends PatternReplacer {}

    private ExecutionContext ctx;

    idocPattern private model;

    public ForceIdocEvaluationPatternReplacer (ExecutionContext ctx) {}

    This.ctx = ctx;

    }

    @Override

    public getPattern() model throws ServiceException {}

    If (idocPattern == null) {}

    List of the functions = SharedObjects.getEnvValueAsList ("forceidocevaluation.functionlist");

    idocPattern = IdocUtils.createIdocPattern (functions);

    }

    Return idocPattern;

    }

    @Override

    protected String doReplace(String idocFunction) throws ServiceException {}

    Return IdocUtils.executeIdocFunction (ctx, idocFunction);

    }

    }

    public class IdocUtils {}

    public static String executeIdocFunction (ExecutionContext ctx, String idocFunction) throws ServiceException {}

    idocFunction = convertIdocStyle (idocFunction, IdocStyle.ANGULAR_BRACKETS);

    PageMerger activeMerger = (PageMerger) ctx.getCachedObject("PageMerger");

    try {}

    String output = activeMerger.evaluateScript (idocFunction);

    return output;

    } catch (Exception e) {}

    throw the new ServiceException ("cannot run the Idoc function" + idocFunction, e);

    }

    }

    public enum IdocStyle {}

    ANGULAR_BRACKETS,

    SQUARE_BRACKETS

    }

    public static String convertIdocStyle (String idocFunction, IdocStyle destinationStyle) {}

    String result = null;

    Switch (destinationStyle) {}

    case ANGULAR_BRACKETS:

    result = idocFunction.replace ("[!-$","<$").replace("--]", "$="">" "]");

    break;

    case SQUARE_BRACKETS:

    result = idocFunction.replace ("<$", "[!--$").replace("$="">", "-] '");

    break;

    }

    return the result;

    }

    public static model createIdocPattern ( list idocFunctions) throws ServiceException {}

    If (idocFunctions.isEmpty ()) throw new ServiceException ("list of Idoc functions to create a template for is empty");

    StringBuffer patternBuffer = new StringBuffer();

    model prefix

    patternBuffer.append ("(\\ [\\!--|)")<>

    Features GOLD - ed list

    for (int i = 0; i)

    patternBuffer.append (idocFunctions.get (i));

    If (i

    }

    model suffix

    patternBuffer.append ("") (. +?) (--\\]|\\$>)");

    String pattern = patternBuffer.toString ();

    log.trace ("Functions return Idoc model", model);

    Return Pattern.compile (pattern);

    }

    }

    public class ServiceUtils {}

    Private Shared Workspace getSystemWorkspace()}

    Workspace workspace = null;

    WsProvider provider = Providers.getProvider ("SystemDatabase");

    If (null! = wsProvider) {}

    workspace = wsProvider.getProvider ((workspace));

    }

    Returns the workspace;

    }

    getFullUserData private static UserData (String userName, cxt ExecutionContext, workspace ws) throws DataException, ServiceException {}

    If (null == ws) {}

    WS = getSystemWorkspace();

    }

    UserData userData is UserStorage.retrieveUserDatabaseProfileDataFull (name of user, ws, null, cxt, true, true);.

    ws.releaseConnection ();

    return userData;

    }

    public static executeService (DataBinder binder) Sub survey DataException, ServiceException {}

    get a connection to the database

    Workspace workspace = getSystemWorkspace();

    Look for a value of IdcService

    String cmd = binder.getLocal ("IdcService");

    If (null == cmd) {}

    throw new DataException("!csIdcServiceMissing");

    }

    get the service definition

    ServiceData serviceData = ServiceManager.getFullService (cmd);

    If (null == serviceData) {}

    throw new DataException (LocaleUtils.encodeMessage ("!")) csNoServiceDefined", null, cmd));

    }

    create the object for this service

    The service = ServiceManager.createService (serviceData.m_classID, workspace, null, Binder, serviceData);

    String userName = 'sysadmin ';

    UserData fullUserData = getFullUserData (username, service, workspace);

    service.setUserData (fullUserData);

    Binder.m_environment.put ("REMOTE_USER", username);

    try {}

    init service do not return HTML

    service.setSendFlags (true, true);

    create the ServiceHandlers and producers

    service.initDelegatedObjects ();

    do a safety check

    service.globalSecurityCheck ();

    prepare for service

    service.preActions ();

    run the service

    service.doActions ();

    } catch (ServiceException e) {}

    } {Finally

    service.cleanUp (true);

    If (null! = workspace) {}

    workspace.releaseConnection ();

    }

    }

    }

    }

  • Move the data files on the standby server without moving primary Oracle 11 g r2

    Hi all


    Oracle 11g 2 with EBS 11.5.0.2 version.

    The size of one of the mounting points on the standby is full and I'm going through one of the data files to another mount point without making any changes to the primary level.

    I tried Google and came across this link:
    http://oraclesea.blogspot.in/2011/12/move-datafiles-on-standby-server.html
    But it did not work... I had to start the database with the spfile to work.

    The steps mentioned in the blog:
    Include below parameter in standby parameter file
    DB_FILE_NAME_CONVERT = '/primary_location/xyz.dbf','/standby_location/xyz.dbf'
    ALTER DATABASE RECOVER MANAGED STANDBY DATABASE cancel;
    shut immediate
    startup nomount pfile=initSCSL.ora
    alter database mount standby database ;
    alter system set standby_file_management='MANUAL' SCOPE=MEMORY ; 
    ! cp /primary_location/xyz.dbf'  /standby_location/xyz.dbf
    alter database rename file  '/primary_location/xyz.dbf' to '/standby_location/xyz.dbf';
    alter system set standby_file_management='AUTO' SCOPE=MEMORY ; 
    alter database recover managed standby database parallel 4 disconnect from session;
    Can you please help me with the right steps...


    Concerning
    KK

    Edited by: 903150 Sep 26, 2012 22:41

    Here is an example for you.

    Database pending:

    SQL> select status,instance_name,database_role from v$database,v$instance;
    
    STATUS       INSTANCE_NAME    DATABASE_ROLE
    ------------ ---------------- ----------------
    OPEN         srprim           PHYSICAL STANDBY
    
    SQL> select file_name from dba_data_files;
    
    FILE_NAME
    --------------------------------------------------------------------------------
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\USERS01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\UNDOTBS01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\SYSAUX01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\SYSTEM01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\USERS02.DBF
    
    SQL> select process,status,sequence# from v$managed_standby;
    
    PROCESS   STATUS        SEQUENCE#
    --------- ------------ ----------
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    RFS       IDLE                  0
    RFS       IDLE                154
    MRP0      WAIT_FOR_LOG        154
    
    7 rows selected.
    
    SQL> show parameter name_convert
    
    NAME                                 TYPE        VALUE
    ------------------------------------ ----------- ------------------------------
    db_file_name_convert                 string
    log_file_name_convert                string
    SQL> alter database recover managed standby database cancel;
    
    Database altered.
    
    SQL> shut immediate
    Database closed.
    Database dismounted.
    ORACLE instance shut down.
    SQL> exit
    Disconnected from Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    
    M:\>copy C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\USERS02.DBF C:\APP\SHIVANANDA.RAO\ORADATA\DBTEST\USERS02.DBF
    1 file(s) copied.
    
    M:\>sqlplus sys/oracle@srprim as sysdba
    
    SQL*Plus: Release 11.2.0.1.0 Production on Thu Sep 27 14:57:16 2012
    
    Copyright (c) 1982, 2010, Oracle.  All rights reserved.
    
    Connected to an idle instance.
    
    SQL> startup mount
    ORACLE instance started.
    
    Total System Global Area  778387456 bytes
    Fixed Size                  1374808 bytes
    Variable Size             494929320 bytes
    Database Buffers          276824064 bytes
    Redo Buffers                5259264 bytes
    Database mounted.
    SQL> alter system set standby_file_management=manual;
    
    System altered.
    
    SQL> alter database rename file 'C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\USERS02.DBF' to 'C:\APP\SHIVANANDA.RAO\ORADA
    A\DBTEST\USERS02.DBF';
    
    Database altered.
    
    SQL> alter database recover managed standby database disconnect from session;
    
    Database altered.
    
    SQL> select process,status,sequence# from v$managed_standby;
    
    PROCESS   STATUS        SEQUENCE#
    --------- ------------ ----------
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    ARCH      CONNECTED             0
    RFS       IDLE                  0
    RFS       IDLE                155
    MRP0      WAIT_FOR_LOG        155
    
    7 rows selected.
    
    SQL> select name from v$datafile;
    
    NAME
    --------------------------------------------------
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\SYSTEM01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\SYSAUX01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\UNDOTBS01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\SRPRIM\USERS01.DBF
    C:\APP\SHIVANANDA.RAO\ORADATA\DBTEST\USERS02.DBF
    

    1. you must close the database pending.
    2 copy the file that you want to move to the different mount point using the OS commands.
    3. Mount the standby database.
    4 rename the data through database level. Make sure that the standby_file_management is set to MANUAL
    5. start the MRP on the standby database.

    Please do not use more than one response to the thread ID. Because you created this thread with ID 903150, I propose to answer with the same ID, not the ID of the other.

  • How to check the data file is set to unlimilted autoextend

    S/N,

    Oracle version: 10.2.0.4
    OS: linux

    could someone tell me how can I check the data file are set to unlimited autoextend?

    Thank you
    Bahadur.

    Salvation;

    Pelase see:
    How to check the CanGrow data files the value of Maxsize unlimited [ID 468096.1]

    Respect of
    HELIOS

  • Restore the data file

    Hello

    Initial situation and config:

    Server: Win2k3 Sp1 32 bit
    Database: 10.2.0.4 Standard noarchivelog mode
    RMAN catalog on different server.
    Cold backup


    Therefore, I was invited to do a restore from a backup of database. What he needs, it's to restore the controlfile, spfile and a data from a backup file complete performed via RMAN. He does NOT INTEND to use them, they just want to see their restored from the backup.

    Restore the spfile and the controlfile is not a problem, and I was able to restore them to a Test directory. But since I can't restore datafile 1 to 'C:\dest', I googled how to make these.

    I came up with these, which I think who should restore a data file to a different location, but I don't know if it's okay. And since I started on the database server, I prefer to ask first.

    The value of newname to 'C:\datafiles\datafile1.dbf' to 'C:\Test\datafile1.dbf ';
    restore the data file 'C:\Test\datafile1.dbf '.

    Would this work? It must be run with the open database altering State and work and will not somehow it. If not, can someone help me here? The only thing needed, he did restore a data file just to get...

    Thanks for the tips!

    Add this

    RMAN > run {}
    value of newname for datafile 'C:\datafiles\datafile1.dbf' to 'C:\Test\datafile1.dbf ';
    restore the data file ' C:\Test\datafile1.dbf';l
    switch datafile;
    }

    It will update the control file with the new location of this data file

  • Data in the data file header

    RDBMS version: 10.2.0.4.0, 11.2.0.2

    During the backup hot (RMAN and user managed backup), data can be written and read from data files. But its headers are frozen. Then, where will be data for the headers of data stored up to that hot backup is made?

    Why must 'means for the headers of data data' be 'stored' somewhere?

    The header is not updated. Header updates are Checkpoint SCN, LogSequence (which is not visible to us) etc. These updates are applied to the data file when the database tablespace backup is complete. Intermediate of the SCN values need not be "stored". When the ends of backup and a control point is completed, Oracle updates the header with the current SNA Checkpoint.

    Hemant K Collette

  • Fractured block found during the backup of the data file

    Hello

    Version: 10.2.0.3

    I take backups RMAN and export of db every night. The following error is recorded in the file alertlog while taking rman full backup. Also, the backup is very slow and it worked at 18hrs now (200 GB db size). Could someone please explain this to me that this error on and the action to take to solve this problem? Thank you

    Kill Mar 3 20:21:52 2009
    A hexadecimal dump (file 16, block 490151) in file /opt/oracle/admin/hol/udump/holiday_ora_11507.trc trace
    Corrupt block relative s/n: 0x04077aa7 (16, block 490151 file)
    Fractured block found during the backup of the data file
    Bad block data:
    type: 2 format: 2 rdba: 0x04077aa7
    last change the SNA: following 0x0013.830b4809: 0 x 4 flg: 0x04
    spare1: 0 x 0 spare2: 0 x 0 spare3: 0x0
    consistency in tail value: 0 x 10340201
    check the value in the block header: 0x7e2e
    checksum calculated block: 0 x 361
    Review of blocknum = 490151, file=/opt/oracle/oradata/hol/undotbs03.dbf. found valid data

    The problem may occur when the data being backed up file is simultaneously and widely used. It does not necessarily mean there is really block corruption. And in your case it seems that all is well:

    Review of blocknum = 490151, file=/opt/oracle/oradata/hol/undotbs03.dbf. found valid data

    On metalink:

    A broken block Messages in the Alert.log during RMAN of Datafile backup
    DOC - ID: 199933.1

    Werner

  • What is the minimum value for the data type double?

    ... a more academic question: what is the minimum value of the data type double? I thought 'double', it is the 8-byte IEEE standard and what I remember, it's + - 1.7E + - 308.  The more different number of 0 should be 1.797E - 308.

    This is why I was a little surprised when I looked at the data output of my calculation of CVI, where I found the numbers with exponents of E-319 or E-324 as the smallest value...

    But I'm sure there's an explanation :-)

    Thank you for sharing with me,

    Wolfgang

    The largest number is indeed the order of + 1.7E + 308; There are a smaller number of - 1.7E + 308. Note that this is in fact the most negative number, which is the smallest by some reckoning. For the small beach of the Exhibitor, things get a little more complicated. 1E-308 can be represented with complete accuracy and resolution, but if you're willing to lose a few bits of precision, you can go down even further. For example:

    Double x, y, z:

    x = 1E-300;

    y = 1E7;

    z = x / y;               Give 1.000000000000000E - 307

    y = 1E8;

    z = x / y;               Give 9.999999999999999E - 309

    y = 1E18;

    z = x / y;               Give 9.999987484955998E - 319

    y = 1E23;

    z = x / y;               Give 9.881312916824931E - 324

    So, as you can see, values less than 1E-308 are possible, but they become increasingly more inaccurate and should not be relied upon.

    JR

Maybe you are looking for