NLS parameters in a connection pool environment

I have limited experience of localization/globalization and I am building an application, using Oracle APEX 4.2/Oracle 11 g db, which takes care of several users of different nationalities.  I have a specific question, but first... can anyone recommend good books, blogs and websites to learn more about best practices for the many things you need to think in this area?

Second, my precise question is... Since my application will use a connection pool, how it works as far as setting the NLS parameters for a given client session?  In the typical client/server architecture that I know that you would simply change the NLS parameter when the user connects, and this change would govern just the session of this user database.  It's a different approach when we are in an environment of connection pool?  If this is not the case, how Oracle now manages the parameters of the client session this session of customer may engage in more than one session of db in its lifetime?

You can set the runtime: http://docs.oracle.com/cd/E16655_01/appdev.121/e17961/global_primary_lang.htm#HTMDB14002

Thank you

Sergiusz

Tags: Database

Similar Questions

  • For the other sessions to find the NLS parameters.

    Hi all
    Database server Environment Details:
    2 Node Oracle 10.2.0.4 RAC on Solaris Operating System
    Parameter in Spfile:
    nls_sort = BINARY_CI
    nls_comp = LINGUISTIC
    nls_language = AMERICAN
    Parameters in database:
    SQL> select * from nls_database_parameters;
    
    PARAMETER                      VALUE
    ------------------------------ ------------------------------------------------------------
    NLS_LANGUAGE                   AMERICAN
    NLS_NCHAR_CHARACTERSET         AL16UTF16
    NLS_TERRITORY                  AMERICA
    NLS_CURRENCY                   $
    NLS_ISO_CURRENCY               AMERICA
    NLS_NUMERIC_CHARACTERS         .,
    NLS_CHARACTERSET               AL32UTF8
    NLS_CALENDAR                   GREGORIAN
    NLS_DATE_FORMAT                DD-MON-RR
    NLS_DATE_LANGUAGE              AMERICAN
    NLS_SORT                       BINARY
    NLS_TIME_FORMAT                HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT           DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT             HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT        DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY              $
    NLS_COMP                       BINARY
    NLS_LENGTH_SEMANTICS           BYTE
    NLS_NCHAR_CONV_EXCP            FALSE
    NLS_RDBMS_VERSION              10.2.0.4.0
    
    20 rows selected.
    Parameters at client's environment variable(Its Application sever on Windows):
    Oracle Client 10.2.0.4
    nls_sort = BINARY_CI
    nls_comp = LINGUISTIC
    Now, each index on columns of type of data characters in this database are created as function of the Index of base to support Case Insensitive search using nls_sort to BINARY_CI, as shown below.
    CREATE UNIQUE INDEX UX_NAME_BR ON ONS (NLSSORT("NAME",'nls_sort=''BINARY_CI'''),"TYPE_ID", "UNIT_NUMBER", "DOMICILE", "IS_ACTIVE", "ID")
    What worries me started when I created by mistake an index btree normal on a character column. After this error just by curiosity, I enabled index followed up on this clue, when I checked in v$ object_usage it got USED!

    So I strongly suspect that there are a few sessions connected to this database with different values of NLS instead of nls_sort = BINARY and nls_comp = LINGUISTIC...

    Sort of, I did a little test in production to confirm this, as shown below:
    PARAMETER                                          VALUE
    -------------------------------------------------- ------------------------------------------------------------
    NLS_LANGUAGE                                       AMERICAN
    NLS_TERRITORY                                      AMERICA
    NLS_CURRENCY                                       $
    NLS_ISO_CURRENCY                                   AMERICA
    NLS_NUMERIC_CHARACTERS                             .,
    NLS_CALENDAR                                       GREGORIAN
    NLS_DATE_FORMAT                                    DD-MON-RR
    NLS_DATE_LANGUAGE                                  AMERICAN
    NLS_SORT                                           BINARY
    NLS_TIME_FORMAT                                    HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT                               DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT                                 HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT                            DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY                                  $
    NLS_COMP                                           BINARY
    NLS_LENGTH_SEMANTICS                               CHAR
    NLS_NCHAR_CONV_EXCP                                FALSE
    
    17 rows selected.
    
    SQL> set autotrace traceonly exp
    SQL> select id,is_active from ons where domicile = 'US' and id = 440 and name = 'AMERICAN COMPANY';
    
    Execution Plan
    ----------------------------------------------------------
    Plan hash value: 1171456783
    
    ---------------------------------------------------------------------------------------------------------------------
    | Id  | Operation                          | Name           | Rows  | Bytes | Cost (%CPU)| Time     | Pstart| Pstop |
    ---------------------------------------------------------------------------------------------------------------------
    |   0 | SELECT STATEMENT                   |                |     1 |    49 |     4   (0)| 00:00:01 |       |       |
    |   1 |  PARTITION RANGE ALL               |                |     1 |    49 |     4   (0)| 00:00:01 |     1 |     3 |
    |*  2 |   TABLE ACCESS BY LOCAL INDEX ROWID| ONS            |     1 |    49 |     4   (0)| 00:00:01 |     1 |     3 |
    |*  3 |    INDEX RANGE SCAN                | IDX_COD        |     1 |       |     4   (0)| 00:00:01 |     1 |     3 |
    ---------------------------------------------------------------------------------------------------------------------
    
    Predicate Information (identified by operation id):
    ---------------------------------------------------
    
       2 - filter("ID"=440)
       3 - access("DOMICILE"='US' AND "NAME"='AMERICAN COMPANY')
    
    SQL> alter session set nls_sort=BINARY_CI;
    
    Session altered.
    
    SQL> alter session set nls_comp=LINGUISTIC;
    
    Session altered.
    
    SQL> select * from nls_session_parameters;
    
    PARAMETER                                          VALUE
    -------------------------------------------------- ------------------------------------------------------------
    NLS_LANGUAGE                                       AMERICAN
    NLS_TERRITORY                                      AMERICA
    NLS_CURRENCY                                       $
    NLS_ISO_CURRENCY                                   AMERICA
    NLS_NUMERIC_CHARACTERS                             .,
    NLS_CALENDAR                                       GREGORIAN
    NLS_DATE_FORMAT                                    DD-MON-RR
    NLS_DATE_LANGUAGE                                  AMERICAN
    NLS_SORT                                           BINARY_CI
    NLS_TIME_FORMAT                                    HH.MI.SSXFF AM
    NLS_TIMESTAMP_FORMAT                               DD-MON-RR HH.MI.SSXFF AM
    NLS_TIME_TZ_FORMAT                                 HH.MI.SSXFF AM TZR
    NLS_TIMESTAMP_TZ_FORMAT                            DD-MON-RR HH.MI.SSXFF AM TZR
    NLS_DUAL_CURRENCY                                  $
    NLS_COMP                                           LINGUISTIC
    NLS_LENGTH_SEMANTICS                               CHAR
    NLS_NCHAR_CONV_EXCP                                FALSE
    
    17 rows selected.
    
    
    SQL> select id,is_active from ons where domicile = 'US' and id = 440 and name = 'AMERICAN COMPANY';
    
    Execution Plan
    ----------------------------------------------------------
    Plan hash value: 270874147
    
    ------------------------------------------------------------------------------------------------------------------------------
    | Id  | Operation                          | Name                    | Rows  | Bytes | Cost (%CPU)| Time     | Pstart| Pstop |
    ------------------------------------------------------------------------------------------------------------------------------
    |   0 | SELECT STATEMENT                   |                         |     1 |    49 |     2   (0)| 00:00:01 |       |       |
    |   1 |  TABLE ACCESS BY GLOBAL INDEX ROWID| ONS                     |     1 |    49 |     2   (0)| 00:00:01 | ROWID | ROWID |
    |*  2 |   INDEX RANGE SCAN                 | UX_NAME_BR              |     1 |       |     2   (0)| 00:00:01 |       |       |
    ------------------------------------------------------------------------------------------------------------------------------
    
    Predicate Information (identified by operation id):
    ---------------------------------------------------
    
       2 - access(NLSSORT("NAME",'nls_sort=''BINARY_CI''')=HEXTORAW('669696E6720636F6D70
                  616E7900')  AND "ID"=440)
           filter("ID"=440 AND NLSSORT(INTERNAL_FUNCTION("DOMICILE"),'nls_sort=''BINARY_CI
                  ''')=HEXTORAW('7588700') )
    IDX_COD index is a regular btree index...

    So, how can I find which sessions have access to these indexes and find the session parameters nls level connected to this database.

    I know there is no way to find settings nls for the other sessions without tracing it...

    Kickbacks will not work me
    NLS_DATABASE_PARAMETERS
    NLS_SESSION_PARAMETERS


    Could someone help me please to find these sessions these sessions that are not level session as BINARY_CI and LANGUAGE settings

    Trigger may fail if the storage space for the trace table is full. This trigger does not use dynamic SQL code: in this perspective, it is a bit more secure.

    I don't know any other way supported to retrieve another session NLS parameters. But he can there have no documented similar to this one http://dioncho.wordpress.com/2009/07/18/spying-on-the-other-session/.

  • Missing for DBAdapter cluster Weblogic Congfiguration outbound connection Pool entry

    Hello

    When I created an entry for outbound connection pool in DBAdapter weblogic console in a clustered (1-Admin server and servers managed by 2) environment and updated adapter DB, I have identified only input connection is updated in that plan a managed server and other server managed, it is not updated. For this reason, we are unable to access the data sources for some requests for the application.

    As a work-around if we plan. XML file to another server managed manually with update and DB adapter connection entry, we are able to find the entrance of connection for managed servers.

    Outside work around there at - it another solution to this problem so that I can add a connection adapter DB Console entry that will update the two plan files.

    Thank you and best regards,

    Vincent

    Can I propose to maintain the file plan of shared drive accessible from all nodes in the server, this will reduce the manual work.

    Concerning

    Albin I

    http://www.albinsblog.com/

  • vRO (vCO) HTTP-REST Workflow - connection pool closing

    Hi all - I've recently upgraded to vRealize Automation 6.2 and works collaboratively with vRealize Orchestrator to create workflows.  I created HTTP and REST workflows for request/release an IP address from our system IPAM (VitalQIP) and it seems to be failing with the description / the following exception:

    Connection pool stop (workflow: demand-IP-du-QIP / Scripting (item3) #14)

    vRO-1.pngvRO-2.png

    Looking for the schema > script line #14 is in red below:

    prepare the application

    Do not edit

    var inParamtersValues = [subnet, hostname, ddns, mac, type, w2kdom, comment];

    var request = restOperation.createRequest (inParamtersValues, null);

    Set the type of content query

    request.contentType = "";

    System.log ("request:" + request);

    System.log ("the request URL:" + request.fullUrl);

    Customize the request here

    request.setHeader ("HeaderName", "headerValue");

    run the query

    Do not edit

    var response = request.execute ();

    prepare the output parameters

    System.log ("response:" + response);

    statusCode = response.statusCode;

    statusCodeAttribute = statusCode;

    System.log ("status code:"+ statusCode ');

    contentLength = response.contentLength;

    headers = response.getAllHeaders ();

    contentAsString = response.contentAsString;

    System.log ("content in the string:" + contentAsString);

    Grep for IPaddress

    Patt var = ('forward IP");

    IPAddress = contentAsString.match (patt) m:System.NET.SocketAddress.ToString () var;

    IPAddress = ipaddress.split (/ [\s,] + /);

    IPAddress = ipaddress [2];

    System.log ("IP address: =" + ipaddress);

    vRO-3.png

    I'm not sure how to work around this problem and I'm looking for assistance because I'm fairly new to vRO.  Thank you very much!




    There is now an official fix for this problem: Technical preview of REST plugin version

    Please provide your comments.

  • API - POST - REST VCO connection pool closing

    Hello

    I'm working on an automation scenario where I need a VCO A workflow to start another workflow VCO B using the VCO REST Plugin (1.0.2) on VCO 5.1.1.

    I did it with success while creating a multitude of REST and surgery REMAINS by using the appropriate workflow. However, my approach is a lot of static and I need to create REST operations on the fly. I have had a look at workflows Plugin and try to do the same thing in my workflow.

    However I am stuck with the error 'Connection Pool close' when you try to run the POST request.

    My approach is to store the host REMAINS as a static attribute with VCO configuration object and retrieve it from there. I can run GET operations without any problem with the owner, i.e.


    restRequest = restHost.createRequest (restType, restUrl, restContent);

    restResponse = restRequest.execute ();

    My problem is actually running a POST operation. I prepared the content in XML format, which is

    " <-xmlns = execution context" http://www.VMware.com/VCO "> "
    < Parameters >

    < name parameter = "param1" type = "string" >

    < string > 1 < / string >

    < / parameter >

    < name parameter = "param2" type = "string" >

    < string > value2 < / string >

    < / parameter >

    < / Parameter >

    < / execution context >

    and use the content type "application/xml". Earlier in my code, I run a query GET to retrieve my ID for workflow target to create a Url like this:

    flow of work/98c3b8ee-9569-4940-acc1-b8fbc2e64649 / executions.

    This is code that I use to make the POST while creating an operation of REST when needed and delete it later.


    var restOperation = new RESTOperation (System.nextUUID ());

    restOperation.method = restType;

    restOperation.urlTemplate = restUrl;

    restOperation.defaultContentType = restContentType;

    restHost.addOperation (restOperation);

    RESTHostManager.updateHost (restHost);

    inputParameters var = [];

    var restRequest = restOperation.createRequest (inputParameters, restContent);

    restRequest.contentType = restContentType;

    var restResponse = restRequest.execute ();

    restHost.removeOperation (restOperation.id);

    RESTHostManager.updateHost (restHost);

    My code does not work when calling for the lifting of restRequest.execute: ' InternalError: connection pool arrested "." If anyone else has tried a similar scenario or has any idea what I may be missing?

    Thank you

    Bernd

    Thanks chap - update the objects did the trick.

    In my case, that initially caused the problem travel of COULD happen. Of course store the restHost object in the configuration of VCO leads to an out-of-date version at run time. I'm now store the ID of restHost only. Then, copy the following code works:

    var restHostId is System.getModule("my.library").myGetConfigAttr ("Folder", "VCO", "restHostId");.

    var restHost = RESTHostManager.getHost (restHostId);

    If (restType == 'GET') {}
    try {}
    restRequest = restHost.createRequest (restType, restUrl, restContent);
    restResponse = restRequest.execute ();
    }
    catch (exp) {throw ("request REST to run: cannot run '" + restType + "' '" + restUrl + "' request-reason = '" + exp + "'") ;}}
    }
    else {}
    try {}
    var restOperation = new RESTOperation (System.nextUUID ());
    restOperation.method = restType;
    restOperation.urlTemplate = restUrl;
    restOperation.defaultContentType = restContentType;
     
    restHost.addOperation (restOperation);
    RESTHostManager.updateHost (restHost);
     
    restHost = RESTHostManager.getHost (restHost.id);
    restOperation = restHost.getOperation (restOperation.id);

    inputParameters var = [];
    var restRequest = restOperation.createRequest (inputParameters, restContent);
    restRequest.contentType = restContentType;

    var restResponse = restRequest.execute ();

    restHost.removeOperation (restOperation.id);
    RESTHostManager.updateHost (restHost);
    }
    catch (exp) {throw ("request REST to run: cannot run '" + restType + "' '" + restUrl + "' request-reason = '" + exp + "'") ;}}
    }

  • the connection of WebLogic connection pool timeout issue too soon

    We are having a problem with the connection to weblogic pool. Our web application is based on the Spring Framework, deploy to Weblogic using the data source to connect to Oracle. We use a duration of 1 min and max setting of connection pool 15 on weblogic. But the problem that we face is that the pool does not manage any connection as a pool. Each request makes a new connection to the Oracle server and right closed within a few seconds of connection. The number of new connections could be easily up to more than 100 within 10 minutes.

    Is there a timeout parameter to check the living connection? Could someone please share your thoughts how to solve the problem? Your help is really appreciated.

    Framework Spring (2.5.3), Weblogic (10.3), Oracle (11 g 2)

    This is the setting on our data source:

    -----------------------
    Driver class: oracle.jdbc.OracleDriver

    Initial capacity: 30
    Maximum capacity: 30
    Ability to increment: 1
    Type of instruction Cache: LRU
    Statement Cache Size: 10

    Test the connections on the reserve: checked
    Test frequency: 120
    Name of the test Table: SQL SELECT 1 FROM DUAL
    Seconds to trust an idle connection pool: 10
    Reduce the frequency: 900
    The login retry frequency: 0
    Connect time: 0
    Timeout:0 connection inactive
    Maximum pending connection: 2147483647
    Timeout:10 store login
    Reporting deadline:-1
    Ignore the connections in use: checked
    Pins-wire: unchecked
    Remove Infected connections Enabled: checked
    ---------------------
    Of the tracking tool, we have seen the following statistics:

    Average number of active connections: 0
    Current number of active connections: 0
    High number of active connections: 1
    Total number of connections: 1523
    Current capacity: 1
    Doesn't have a number of requests for reserve: 46
    More high Num available: 30
    Number of connections leak: 0
    Wait for the high number of seconds: 0
    Number of requests for reserve: 1531

    Other parameters are all 0.

    --------------------------
    For the adjustment of the spring:

    < bean id = "ourdatasource" class = "org.springframework.jndi.JndiObjectFactoryBean" >
    < property name = "jndiName" > < value > weblogic.datasource < / value > < / property >

    <!-adds the following to the question if debugging caused by cluster or not->
    < property name = "cache" value = "false" / >
    < property name = "lookupOnStartup" value = "false" / >
    < property name = "proxyInterface" value="javax.sql.DataSource"/ >
    < / bean >

    Try to uncheck ' delete infected enabled connections "and let me know...

  • Data DRCP connection pooling resident

    Hello

    I have a quesiton form SR

    Please confirm do you use DRCP "database resident connection pool?


    How can I determine if DRCP is enabled in my environment

    Could someone help. It's urgent

    Thank you
    Prasad

    Salvation;

    It is better to ask the Oracle support?

    Please see:
    How the program installation and database resident connection Trace pooling (DRCP) [ID 567854.1]
    Example of grouping (DRCP) usage [ID 577865.1] resident of database connection

    Respect of
    HELIOS

  • How to compile the connection pool java example code

    I recently bought David Knox of 'Effective Oracle Database 10' g security by Design and worked through his examples with identifiers of customers and how to exploit the database security with anonymous connections pools.

    The side of the things database I totally understand and have implemented, but I now want to compile / the java run the examples of code, but do not know what is required to compile this.

    I am running Oracle 10.2.0.4 (64-bit) Enterprise Edition on a Linux (RHEL 5) PC. Version Java is 1.6.0_20. .Bash_profile relevant environment variables are:
    export LD_LIBRARY_PATH = $ORACLE_HOME/lib: / lib: / usr/lib
    Export CLASSPATH = $ORACLE_HOME/jre: $ORACLE_HOME/jlib: $ORACLE_HOME/rdbms/jlib
    export PATH = $ORACLE_HOME/bin: $ORACLE_HOME/OPatch: / usr/kerberos/bin: / usr/local/bin: / bin: / usr/bin

    When I try to compile, I get:
    Oracle Java: DB10204$ FastConnect.java
    Exception in thread "main" java.lang.NoClassDefFoundError: java/FastConnect
    Caused by: java.lang.ClassNotFoundException: FastConnect.java
    in java.net.URLClassLoader$ 1.run(URLClassLoader.java:202)
    at java.security.AccessController.doPrivileged (Native Method)
    at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
    to Sun.misc.Launcher$appclassloader$ AppClassLoader.loadClass (Launcher.java:301)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:248)
    The main class is not found: FastConnect.java. Program ends.

    Below is an example java source code. Is anyone able to point me in the right direction as to how to get this to compile? I have just a configuration syntax change and/or the environment to do, or there it more for this example doesn't work? Any help much appreciated.
    oracle:DB10204$   cat FastConnect.java 
    package OSBD;
    import java.sql.*;
    import oracle.jdbc.pool.OracleDataSource;
    
    public class FastConnect
    {
      public static void main(String[] args)
      {
        long connectTime=0, connectionStart=0, connectionStop=0;
        long connectTime2=0, connectionStart2=0, connectionStop2=0;
        ConnMgr cm = new ConnMgr();
        // time first connection. This connection initializes pool.
        connectionStart = System.currentTimeMillis();
        Connection conn = cm.getConnection("SCOTT");
        connectionStop = System.currentTimeMillis();
        String query = "select ename, job, sal from person_view";
        try {
          // show security by querying from View
          Statement stmt = conn.createStatement();
          ResultSet rset = stmt.executeQuery(query);
          while (rset.next()) {
            System.out.println("Name:   " + rset.getString(1));
            System.out.println("Job:    " + rset.getString(2));
            System.out.println("Salary: " + rset.getString(3));
        }
          stmt.close();
          rset.close();
          // close the connection which resets the database session
          cm.closeConnection(conn);
          // time subsequent connection as different user
          connectionStart2 = System.currentTimeMillis();
          conn = cm.getConnection("KING");
          connectionStop2 = System.currentTimeMillis();
          // ensure database can distinguish this new user
          stmt = conn.createStatement();
          rset = stmt.executeQuery(query);
          while (rset.next()) {
            System.out.println("Name:   " + rset.getString(1));
            System.out.println("Job:    " + rset.getString(2));
            System.out.println("Salary: " + rset.getString(3));
          }
          stmt.close();
          rset.close();
          cm.closeConnection(conn);
        } catch (Exception e)    { System.out.println(e.toString()); }
        // print timing results
        connectTime = (connectionStop - connectionStart);
        System.out.println("Connection time for Pool: " + connectTime + " ms.");
        connectTime2 = (connectionStop2 - connectionStart2);
        System.out.println("Subsequent connection time: " +
                            connectTime2 + " ms.");
      }
    }
    Code download is: http://www.mhprofessional.com/getpage.php?c=oraclepress_downloads.php & cat = 4222
    I'm looking at Chapter 6.

    I extracted the .jar (jar - xvf) file

    The ODBC .jar file? Don't, don't. Add it to your classpath as it is, or paste it into the WEB-INF/lib directory, or what you have to do.

  • Download weekdaynum, independent of the nls parameters.

    What is the best way to get Weekdaynum for a date?
    The algorithm must be independent of the nls parameters.

    This solution depends on nls_teritory and works differently when
    nls_territory = "AMERICA."
    and when it is ESTONIA:
    select decode (
       to_char(sysdate,'DY','nls_date_language=english'), 
       'MON', 1,
       'TUE', 2,
       'WED', 3,
       'THU', 4,
       'FRI', 5,
       'SAT', 6,
       'SUN', 7
    ) WeekdayNum
    from dual;

    Something like this should also work, it just takes mod 7 the Julian date and add the value 1.

    with data as
    (select sysdate + rownum - 1 as dt from dual connect by rownum < 10)
    
    select dt, decode(mod(to_char(dt,'J'),7) + 1,
                 '1', 'Monday',
                 '2', 'Tuesday',
                 '3', 'Wednesday',
                 '4', 'Thursday',
                 '5', 'Friday',
                 '6', 'Saturday',
                 'Sunday') as result
                  from data;
    
  • connection pool and inefficient query plan

    There is a single query that covers almost 90% of cpu DB.

    Select * from employee e, Department d where e.departement_id = d.department_id

    Us will gather statistics for this table in two, and the issues is resolved.

    This occurs every 2 weeks. That said, the query runs fine for 2 weeks and then we have questions... we will bring together the statistics of these two tables... and things are good for another 2 weeks.

    This query is a proc to store oracle pl - sql, which is called by the JDBC code. I introduced 2 months behind connection pooling, and today, we are facing this problem.

    The query has been accounting for 5 years with no problems.

    Before that I presented the connection pool, the jdbc code created a new connection before calling the store proc.

    Do you think that my connection to the connection pool has introduced this problem.

    The DBA tell me that the query runs a bad plan. Oracle recovers not the more effective plan (and that leads to this high CPU utilization).

    I guess that after 2 weeks Oracle begins to pick up plans that are effective in.

    Do you think all that this question never has nothing to do with my connection pooling code.

    I use Oracle 10 G.

    Hi Mike and ground beach thanks for your response.

    It was just a doubt... that you guys allowed.

    Thank you

    m

  • connection pool

    HI the OBIEE Experts

    How many pools of connections you used in your project

    To get data from another schema, then what you must activate in the connection pool

    . Can we create multiple pools of connection under the object of a database


    Thanks in advance

    You seriously to come on the forum and posting simply questions of an interview questionnaire or a certification test?

    You have 4 threads right now with 2 different user accounts. I suggest that you stop this behavior. Start posting questions with a specific context and meaning. We only are here to get a job, that you should better start for your own good if you do not know the answers and even less here to help you cheat your way to certification.

  • OBIEE 11 g RPD migration - how to automate the Connection Pool passwords changes in environments

    Hello gurus

    We have a situation in our client, where I am responsible for automating RPD Migrations in environments. In our current scenarios, (SID database) data source names and user names (schema name) are same in environments. The current process deploy us the RPD by EM then SPR online by using the administration tool and manually open update passwords in each connection pool (we have 9 pools of connections). There has been a lot of problems with this approach you can imagine (especial when do things the manual way).

    Can you please provide a better approach to automate this process in order to avoid human error.

    Thank you very much.

    Take a look at BIServerT2PProvisioner.jar http://docs.oracle.com/cd/E29542_01/bi.1111/e10540/conn_pool.htm#BHBBDDFF

  • SQL Plugin - connection pool &amp; retries

    People-

    We use vCO workflows which a lot of CRUD operations on the database. In light of these are long-running workflows with a decent amount of load, we want to make sure that this solution we have in place is efficient and reliable. To this end:

    1. we want to ensure that we are able to reuse the database connections - y at - it the notion of Pools of connections or equivalent.

    2 How about reconnecting to the db when it falls down? Is that something it will take the code in the workflow through say managers of exceptions etc..

    I'm sure that others have encountered similar problems - would appreciate all comments that you may have.

    See you soon.

    Hello

    Unfortunately does not support the plugin SQL connection pooling. We have never considered it. We will discuss it and may include in the next version of the plugin update.

    Regarding your second question - we don't keep connection open for a long time and call evry DB is atomic. When you run workflows CRUD a new conncetion is created, the corresponding SQL query is executed / clerk and the connection is closed. If there is a problem, as the disconnection of the server Oracle, the right way to ensure that your operation has passed is to write a custom error handling by using the Scripting of the Orchestrator features. In pseudo-code:

    var hasPassed = false;
    for (i = 0; i< maxretries="" -1="" &&="" !haspassed="" ;="" i++)="">

    hasPassed = false;

    try {}

    Execute the statement using the JDBC API exposed or by using CRUD actions

    hasPassed = true;

    } catch (error) {}

    Save the problem

    }

    }

    Hope this helps you. Any additional comments will be great.

    Best regards

    Boyan

  • Connection pool is already checked or temporarily locked

    After the reboot of the servers BI showing this error

    And deployments BI Publisher Application is not comeing up, showing bellow error.

    [nQSError: 36004] Object real-time OLTP connection pool is already checked or temporarily locked.

    Pls me how reslove

    Thank you

    Make sure you close the tool of administration or the RPD (when changes in the RPD online) before starting the services

  • Pool table of event in the connection pool of 11g OIC for obiee 11g

    Hi all, I tried to use the event pool table using the OIC 11 g connection pool. A deleted event, but cache pool table records did not get purged after 60 minutes interval. Should we do addition together upwards to event pool table when you use oci 11g other that makes it active? It seems that it works only with odbc to connect. http://gerardnico.com/wiki/dat/obiee/event_table thanks, Sushil

    Import the table with ODBC and change the interface call OCI connection pool.

Maybe you are looking for