Validation test for the relationship between network_name, services and relaxation

11g R2 running on RHEL 5 U4
Primary (NO CARS) with a physical Standby (NO CARS)
I'm trying to set up a basic configuration so that the client can ONLY connect to the primary database.

Here's what I think it should all look like

-entry on the primary and Standby physical tnsnames
DMS =
(DESCRIPTION =
(ADDRESS_LIST =
(ADDRESS = (PROTOCOL = TCP) (HOST = PRIMAIRE1)(PORT = 1521))
(ADDRESS = (PROTOCOL = TCP) (HOST = STBY_DR01)(PORT = 1521))
)
(CONNECT_DATA =
(SERVICE_NAME = DG_DMS)
)
)

-the service created on the primary
exec DBMS_SERVICE. () CREATE_SERVICE
network_name = > "DMS",.
service_name = > 'DG_DMS '.
);

-the service created on the primary
CREATE OR REPLACE TRIGGER manage_service after starting on the database
DECLARE
role varchar (30);
BEGIN
SELECT DATABASE_ROLE IN the role OF V$ DATABASE;
IF role = "PRIMARY" THEN
DBMS_SERVICE. START_SERVICE ('DG_DMS');
END IF;
END;


Sounds good?
If this is not the case, can you please explain how to set up.

Yes, if you plan to have a service being started on your physical waiting while it is in standby mode, you don't need to create the service with DBMS_SERVICE previously. CREATE_SERVICE. When the waiting becomes the main, Oracle Restart starts the appropriate services as defined by you using the command "srvctl add service d DMSPRMY s DG_DMS-l primary SCHOOL.

If you need a container of service started on a standby physical while it is in standby mode, then you would have still to do the creation on the primary service so that the service is available on the day before.

But in both cases you don't have to create the trigger.

Larry

Tags: Database

Similar Questions

  • Explain the relationship between libraries, albums and photos in pictures.  Where are the images stored?

    Looking for info on the relationship between libraries, albums and photos in the Photos app.  I can create several libraries containing photos and unique albums.  Can I put the same photo in libraries and the different albums?  By making a copy it creates a link to the original or a new photo?  Where are located the photo files?  When scanning photo to my Mac Pro files are placed in a folder structure, copy these photos in an album Photos using the original file or he places a copy in a database?

    Bottomline, I'm frustrated by the method of collection, the albums, the place and date.  I want store related photos in a separate album.   Example: A library for a family, containing albums of members of their family. Maybe the photos have duplicate in different albums.

    Some of your questions are answered by Leonie contributor in this thread:

    Re: Too many libraries...

    More answers can be found by using help in Photos app, or in general to help the Viewer.

    Apple Support website has links to information in articles; a search on the web is sometimes useful

    to locate the Apple pages faster than the company's site. Others in line include:

    Photos for OS X FAQ:

    http://www.IMore.com/photos-OS-x-FAQ

    How to use the Photos for OS x:

    http://www.IMore.com/how-use-photos-OS-x-ultimate-guide

    Good luck!

  • I'm looking for the relationship between the unit ORGANIZATION and self-service workflow?

    Hi all:
    I have 3 UO and 3 files of wft different workflow self-service in the my HR system. How can I know the relationship between the unit ORGANIZATION and self-service workflow? I mean what wft file corresponds to an ORGANIZATIONAL unit?

    my environment is oracle 11.5.9. self-service 4.1 | workflow Builder 2.6

    for example:
    OR 1 = > HRSSA1.wft

    OR 2 = > HRSSA3.wft

    OR 3 = > HRSSA2.wft

    I'm looking for this relationship?


    Concerning
    Terry

    Terry,

    I mean what wft file corresponds to an ORGANIZATIONAL unit?

    Open the workflow designer using wft files and see if you can find any reference to these OUs.

    Thank you
    Hussein

  • for the relationship between the total points for the individual subject

    Hi, I have the STUDENT_MARKS table with the following fields

    name varchar2 (20), varchar2 (10) standard, topic varchar2 (20), marksnumber (3), number of ratio (5.2).

    The following values are entered in the table

    Name Standard Topic brands of ratio
    A1firstS130
    A1firstS220
    A1firstS310
    B1secondS120
    B1secondS220
    C1thirdS110

    Now I need a report of field is filled with the sum of all marjks of the given topic studednt according to standard divided by individual notes:

    For A1 and the topic should be value s1 (30 + 20 + 10) / 30

    A1 of the enemy and subject s2, it should be (30 + 20 + 10) / 20

    .....

    For B1 and subject S1, it should be (20 + 20) / 20

    can someone please help me with the query for the same number of subjects are subject to change in the future for the student of gicven with the standard.

    Not code anything hard, just use my only query with your table... my with clause is just for the representation of the data.

    You can directly use the query on your table:

    SELECT a.*, SUM (brands) (PARTITION BY NAME, STANDARD) / ratio of brands

    From your_table;   -Replace to appropriate table your_table

    See you soon,.

    Manik

  • The relationship between C-200 and 6120XP

    Hi all

    I set up a lab of the UCS enviroment fot our society. I have 6 servers of C-200 M2 (only with the ports of 2x1Gig LOM) couple of blades and 2 x 6120XP x. I have no problems on the connection of chassis to 6120.But, I have a few questions about the connection of the 6120 C-200 servers.

    I want to use 1Gig LOM ports for traffic data so I need to use some type of (or type in) SFP + for the creation of links between c - 200-6120.

    or I can do something else.

    for now, I have no FEX or other solutions. We ordered and waiting for delivery.

    Hey Burak,

    For the integration of series C, you must have the FEX to which LOM plug ports for server management and then ports 10 gig (to an ANC) on the 6120 for the data path.

    Without the FEX you will not be able to connect them to the 6120.

    . / Afonso

  • Relationship between a Tablespace and the BLOB type columns

    Hello
    When I import using cmd imp, all tables were imported except the table that has columns of type BLOB. After some pain in my head, I noticed that it was because of tablespace (I don't have the source table space in my destination).

    Can someone explain the reason for the relationship between the tables BLOB and tablespaces (because all other imported tables)?

    Oracle Version: 10.2.0

    Check this thread
    problem importing table with the blob data type
    http://asktom.Oracle.com/pls/asktom/f?p=100:11:0:::P11_QUESTION_ID:378418239571}

  • What is the difference/relationship between a subject and a page?

    Hi all

    I'm not sure of the relationship between a topic and a new page: it seems that a subject can be made up of pages, but when I look at the HTML of the subject, there is no new section (DIV) for the addition of the page.

    Also, not sure about the relationship between the book and the pages/topics.

    You would be grateful if someone could either point of reference material or provide a few clarifying words on that.

    Thank you, Donna

    In regards to the benefits of the pages pointing to topics:

    Placeholders (table of contents Pages) allow a single topic to appear in several places in the table of contents. If the table of contents used directly the subject, each topic could appear only in one place. For example, maybe help is divided into sections, user and administrator, and administrators and regular users can add coordinates to a person in the application. Using a placeholder allows a single procedure to be created (as the task is the same for both users) and then referenced in the two sections of the table of contents.  If the subjects were used directly, two different subjects will have to be created. You can also organize your subjects in a folder structure that suits you and use a different structure in the table of contents.

    The placeholder allows also a display name different be used, for example, if you have a long topic name but want a shorter name to appear in the table of contents, or if the topic appears in several places in your table of contents and you want to focus slightly different in each place, say for different groups of users.

  • What relationships between flash cs4 and Pixel Bender Toolkit 2?

    I want to know the relationship between flash cs4 and Pixel Bender Toolkit 2. Now I use Pixel Bender Toolkit 2 to make effects, use file-> export filter for flash player, then it will create a .pbj file. I don't know how flash cs4 to use the .pbj file? I use google for search, but I can't find answer. Could someone give me example of detail on flash cs4 use the .pbj file?

    Thank you

    http://www.flepstudio.org/Forum/Flash-CS4-tutorials/4168-using-pixel-Bender-Flash-CS4.html

  • Problem of password of account, without that we can connect to the network on this computer. The error indicates, "the trust relationship between this device and the network is no longer valid.

    Hi, I hope you can help with this problem. A friend has 6 computers networked to a Tower Server and have a problem. A worker student put computers into hibernation, and since then, nobody can connect to the network on this computer. The error indicates, "the trust relationship between this device and the network is no longer valid. I tried to get into the pc to try removing network and replace it, but I can't get on the Administrators account because it has been disabled. I'm at a loss on how to proceed. My friend cannot afford a service call, so I'll try to help him.  Thank you

    If the computer is on a domain then there are at least two accounts that can be used to remove the computer from the domain:

    • A domain administrator account.
    • A local administrator account.

    When your friend has launched Windows for the first time after buying the machine, he was prompted to enter the name of an account. This account is created automatically with administrator privileges. He should use it now.

  • How to establish the relationship between the tables contact and Global (entity) in Oracle Service Cloud to preconfigure data

    Hello

    We strive to integrate the OPA attributes to the Cloud Service Oracle tables to store session data for the customer portal users.

    We tried mapping the attributes of the OPA with relevant tables of cloud Service Oracle? Is there any installation of additional data to create to retrieve the session data for the particular user / contact?

    We managed to save tha data in arrays of cloud Service for users of ananymous and also well regarding customers portal users but to preconfigure data do we need to do any additional configuration / mapping to the contact table?

    Please help on the establishment of the relationship between the contact and Global (new table created in cloud Service) tables.

    Thank you

    Vivek

    Hi Vivek,

    To load data from the Contact in the the OPA policy template, you must configure the widget of the OPA. For instructions, see the following articles:

    - Incorporate an interview that uses data portals in Oracle Service Cloud Customer Portal

    - Deploy and configure the sample OPA widget

    - Insert the widget to sample in a customer portal page

    Obviously, you need to do the mapping in OPM as well, but even if the mapping is correct, you will not be able to load the Contact data unless the widget of the OPA is implemented.

    See you soon,.

    Jasmine

  • What are the relationships between the logging and IKM?

    What is the best method to use in the following scenario:
    I have about 20 tables with the large amount of data sources.
    I need to create interfaces that join the source tables in target tables.
    The tables are inserted every few seconds with about hundreds of thousands lines.
    There may be a gap of a few seconds between the insertion of different tables that could be attached.
    The source and target tables are on the same Oracle instance and schema.
    I want to understand the role of: 'Logging CDC' and "IKM - incremental" and
    How can I use it in my script?
    In general, what are the relationships between "Logging" and 'IKM '?
    Use both? Or maybe it's better deelte and insert the target tables?

    I want to understand what is the role of "Logging CDC"?
    Can 'IKM - incremental' work without "logging"?
    Must 'Logging' have PK on the tables?
    What should I do if I can't say PK (there may be several identical lines)?

    Yael thanks in advance

    user604062 wrote:
    Hello
    Thanks for your quick response!

    No probs - its still fresh in memory I did a major project on this topic last year (400 tables, millions of lines per day (inserts, updates, deletes), sup-5 minute latency). The problem is it isn't that well written on the web, that you have read the blog of the example I linked to in my first answer? See also here: http://odiexperts.com/changed-data-capture-cdc/

    Always on logging:
    My source table is inserted all the time.
    The interface to join the source table in the target table.

    In ODI, the correct term would be your source table "fits" in the table target, unless you mean literally that want to join the the source with the taget table table? My question if you want to do with the result of the join?

    What exactly the "journaling" CDC updates?
    It updates the model of ODI? interfaces? The source of data in the model of ODI? The target table?

    Logging CDC configures and deploys the data capture mechanism (Triggers or log based capture, IE Logminer/streams/Goldengate) - it is not updated the model as such, she pointed out the metadata of the model of ODI repositoty as a CDC data store, allowing you, the developer say ODI to use log data if you wish (reported in the interface) There is no change in the target table, you get an indicator of metadata (IND_UPD) against a line during the integration (in C$ and I have tables$) that tells you if its insertion (I) and update (U) or deletion (D). It had ' lines allow you to synchronize the deletions, but yoy say its inserts only then you probably used use this option. "
    So the only changes are the source table to your interface, another diary data (if you use logging) or the table of the actual source (if not using the logging).

    This is the main thing that I don't understand!

    I hope I made a little clearer.

    Try the following as a quick test:

    Reverse a source table an engineer and the target (at least) table.
    Import the update incremental LKM and IKM.
    Import of the JKM you want to use.

    Create an interface between the source and the target without any deployed JKM.
    Configure the options of JKM on the model, the "Start log" to start the process of capture - this is quite a complex stage and a lot of things to understand what is happening in the source database, better to check code ODI sends to the database and to review the documentation of Oracle database for a description of what his weight (instantiate Tables (, sets of creating change, creation of subscribers etc. establishment of newspaper groups, creating views Journalising etc.) -you will need to consult your Source DBA database initially as ODI wants to make many changes to the source DB (in mode Archivelog process max, parallelism, size, Java etc.)

    Now, edit your interface and mark the table source for use "Journalized data bank.
    Restart your interface
    Compare the difference in the generated code in the journal of the operator, see the differences of the operator.

    >

    Thank you, Yael

  • What gets the relationship between the number of blocks and coherent?

    SQL> CREATE TABLE TEST(ID INT ,NAME VARCHAR2(10));
    
    SQL> CREATE INDEX IND_IDN ON TEST(ID);
    
    SQL> BEGIN
      2  FOR I IN 1 .. 1000
      3  LOOP
      4  EXECUTE IMMEDIATE 'INSERT INTO TEST VALUES('||I||',''LONION'')';
      5  END LOOP;
      6  COMMIT;
      7  END;
      8  /
    
    SQL> EXEC DBMS_STATS.GATHER_TABLE_STATS(USER,'TEST',CASCADE=>TRUE);
    
    SQL> SELECT DISTINCT DBMS_ROWID.rowid_block_number(ROWID) BLOCKS FROM TEST;
    BLOCKS
    -----------
    61762
    61764
    61763         
    
               >> above , there have 3 blocks in table TEST .
    
    SQL> SET AUTOTRACE TRACEONLY;                                     
    
    SQL> SELECT * FROM TEST;
    
    Execution Plan
    ----------------------------------------------------------                      
    Plan hash value: 1357081020                                                     
                                                                                    
    --------------------------------------------------------------------------      
    | Id  | Operation         | Name | Rows  | Bytes | Cost (%CPU)| Time     |      
    --------------------------------------------------------------------------      
    |   0 | SELECT STATEMENT  |      |  1000 | 10000 |     2   (0)| 00:00:01 |      
    |   1 |  TABLE ACCESS FULL| TEST |  1000 | 10000 |     2   (0)| 00:00:01 |      
    --------------------------------------------------------------------------      
    
    
    Statistics information
    ----------------------------------------------------------                      
              0  recursive calls                                                    
              0  db block gets                                                      
             72  consistent gets                                    >>  there have 72  consistent gets                  
              0  physical reads                                                     
              0  redo size                                                          
          24957  bytes sent via SQL*Net to client                                   
           1111  bytes received via SQL*Net from client                             
             68  SQL*Net roundtrips to/from client                                  
              0  sorts (memory)                                                     
              0  sorts (disk)                                                       
           1000  rows processed                                                     
    
    SQL> SELECT /*+ INDEX_FFS(TEST IND_IDN)*/ * FROM TEST WHERE ID IS NOT NULL;
    
    Execution Plan
    ----------------------------------------------------------                      
    Plan hash value: 1357081020                                                     
                                                                                    
    --------------------------------------------------------------------------      
    | Id  | Operation         | Name | Rows  | Bytes | Cost (%CPU)| Time     |      
    --------------------------------------------------------------------------      
    |   0 | SELECT STATEMENT  |      |  1000 | 10000 |     2   (0)| 00:00:01 |      
    |*  1 |  TABLE ACCESS FULL| TEST |  1000 | 10000 |     2   (0)| 00:00:01 |      
    --------------------------------------------------------------------------      
                                                                                    
    Predicate Information (identified by operation id):                             
    ---------------------------------------------------                             
                                                                                    
       1 - filter("ID" IS NOT NULL)                                                 
    
    
    Statistics information
    ----------------------------------------------------------                      
              1  recursive calls                                                    
              0  db block gets                                                      
             72  consistent gets                                   >>  there have 72  consistent gets                  
              0  physical reads                                                     
              0  redo size                                                          
          17759  bytes sent via SQL*Net to client                                   
           1111  bytes received via SQL*Net from client                             
             68  SQL*Net roundtrips to/from client                                  
              0  sorts (memory)                                                     
              0  sorts (disk)                                                       
           1000  rows processed                                                     
    
                                             
    
    SQL> SELECT COUNT(*) FROM TEST;
    
    Execution Plan
    ----------------------------------------------------------                      
    Plan hash value: 1950795681                                                     
                                                                                    
    -------------------------------------------------------------------             
    | Id  | Operation          | Name | Rows  | Cost (%CPU)| Time     |             
    -------------------------------------------------------------------             
    |   0 | SELECT STATEMENT   |      |     1 |     2   (0)| 00:00:01 |             
    |   1 |  SORT AGGREGATE    |      |     1 |            |          |             
    |   2 |   TABLE ACCESS FULL| TEST |  1000 |     2   (0)| 00:00:01 |             
    -------------------------------------------------------------------             
    
    Statistics information
    ----------------------------------------------------------                      
              0  recursive calls                                                    
              0  db block gets                                                      
              5  consistent gets                                      >>  there have 5  consistent gets               
              0  physical reads                                                     
              0  redo size                                                          
            408  bytes sent via SQL*Net to client                                   
            385  bytes received via SQL*Net from client                             
              2  SQL*Net roundtrips to/from client                                  
              0  sorts (memory)                                                     
              0  sorts (disk)                                                       
              1  rows processed                                                     
    
                                                  
    
    SQL> SELECT COUNT(*) FROM TEST WHERE ID IS NOT NULL;
    
    Execution Plan
    ----------------------------------------------------------                      
    Plan hash value: 735384656                                                      
                                                                                    
    --------------------------------------------------------------------------------
    -                                                                               
                                                                                    
    | Id  | Operation             | Name    | Rows  | Bytes | Cost (%CPU)| Time     
    |                                                                               
                                                                                    
    --------------------------------------------------------------------------------
    -                                                                               
                                                                                    
    |   0 | SELECT STATEMENT      |         |     1 |     4 |     2   (0)| 00:00:01 
    |                                                                               
                                                                                    
    |   1 |  SORT AGGREGATE       |         |     1 |     4 |            |          
    |                                                                               
                                                                                    
    |*  2 |   INDEX FAST FULL SCAN| IND_IDN |  1000 |  4000 |     2   (0)| 00:00:01 
    |                                                                               
                                                                                    
    --------------------------------------------------------------------------------
    -                                                                               
                                                                                                                                                                
    Predicate Information (identified by operation id):                             
    ---------------------------------------------------                             
                                                                                    
       2 - filter("ID" IS NOT NULL)                                                 
    
    
    Statistics information
    ----------------------------------------------------------                      
              0  recursive calls                                                    
              0  db block gets                                                      
              5  consistent gets                                                >>  there have 5 consistent gets     
              0  physical reads                                                     
              0  redo size                                                          
            408  bytes sent via SQL*Net to client                                   
            385  bytes received via SQL*Net from client                             
              2  SQL*Net roundtrips to/from client                                  
              0  sorts (memory)                                                     
              0  sorts (disk)                                                       
              1  rows processed                                                     
    
                                              
    
    SQL> SELECT COUNT(ID) FROM TEST WHERE ID IS NOT NULL;
    
    Execution Plan
    ----------------------------------------------------------                      
    Plan hash value: 735384656                                                      
                                                                                    
    --------------------------------------------------------------------------------
    -                                                                               
                                                                                    
    | Id  | Operation             | Name    | Rows  | Bytes | Cost (%CPU)| Time     
    |                                                                               
                                                                                    
    --------------------------------------------------------------------------------
    -                                                                               
                                                                                    
    |   0 | SELECT STATEMENT      |         |     1 |     4 |     2   (0)| 00:00:01 
    |                                                                               
                                                                                    
    |   1 |  SORT AGGREGATE       |         |     1 |     4 |            |          
    |                                                                               
                                                                                    
    |*  2 |   INDEX FAST FULL SCAN| IND_IDN |  1000 |  4000 |     2   (0)| 00:00:01 
    |                                                                               
                                                                                    
    --------------------------------------------------------------------------------
    -                                                                               
                                                                                                                                                              
    Predicate Information (identified by operation id):                             
    ---------------------------------------------------                             
                                                                                    
       2 - filter("ID" IS NOT NULL)                                                 
    
    
    Statistics information
    ----------------------------------------------------------                      
              0  recursive calls                                                    
              0  db block gets                                                      
              5  consistent gets                                        >>  there have 5  consistent gets             
              0  physical reads                                                     
              0  redo size                                                          
            409  bytes sent via SQL*Net to client                                   
            385  bytes received via SQL*Net from client                             
              2  SQL*Net roundtrips to/from client                                  
              0  sorts (memory)                                                     
              0  sorts (disk)                                                       
              1  rows processed                                                     
    QUESTION:

    What gets the relationship between the number of blocks and coherent? How to calculate become consistent?

    You can see that your uniform is getting down to 6 to 12, is it not? Reading of the below thread Asktom.
    http://asktom.Oracle.com/pls/Apex/f?p=100:11:0:P11_QUESTION_ID:880343948514

    Aman...

  • I keep getting "the trust relationship between this workstation and the primary domain failed" error windows 2008R2

    Hi all

    I appreciate your help. And I hope that this issue has been addressed previously, although I couldn't find any solution there.

    I manage a domain windows 2008 with 3 domain controllers. Recently my workstations continue to fall out of the field. I get "the trust relationship between this workstation and the primary domain failed" to many workstations.

    I know how to fix properly, use netdom.exe. But the mistakes keep coming back. I don't know where to look for the source of this error. A possible problem might be the time in any field is out of sync. But all my workstations synchronize with the domain controller with the primary domain controller role and seems to work correctly.

    Are there other sources, that I need to check? This is getting frustrating.

    Thank you much for the help.

    Sincerely,

    Hello

    Post your question in the TechNet Server Forums, as your question kindly is beyond the scope of these Forums.

    http://social.technet.Microsoft.com/forums/WindowsServer/en-us/home?category=WindowsServer

    See you soon.

  • What is the relationship between the source code (LKS) file and the process file (L4P)?

    I know that Lookout produces a file of source code with an LKS extension when you save a file to process.  Can someone explain the relationship between the two files, especially while the (L4P) process is running?

    1. Is this just a backup file can be recompiled in a process file?
    2. A file corrupt LKS cause strange problems with operation process file?

    I currently have a very strange intermittent behavior with a process file run.  I first thought that the problem was associated with my Fieldpoint and/or their configuration modules.  Since then, I found that my process file has a file corrupt LKS.  I repaired and recompiled my LKS file to a new file to process.  I still don't know if I have solved the problem or not.  So, the problem is intermittent, I did that about 20-30 SECONDS to resolve the problems there before he goes.  Then he can not show up again for another 2-3 days.

    The .lks file is just the source code of your process. It can be opened by a different version of lookout, but does not have the .l4p file.

    The .lks file is not be used while a process is running. Lookout does not read the file more after his execution. So it should not affect the running process.

    What kind of problem, is it?

  • the trust relationship between this machine and the primary domain failed

    Hi I have a virtual server (Server 03) with a vista machine, I can log on to the vista computer, but it happens with the trust relationship between this machine and primary domain failed. I tried to delete the domain vista machine, but he says you need to keep a main area

    Any suggestions gratefully received

    Hello

    Your question is more complex than what is generally answered in the Microsoft Answers forums. It is better suited for Windows Server on Technet. Please post your question in the Technet forums. You can follow the link to your question:

Maybe you are looking for