There will be review certification essentials (OCS) for the GREAT DATA Oracle?

Hello

I often hear about this subject called BIG DATA much in the arena of the database

and even at Oracle.

Will there be a Big data Oracle Certification?

Roger

For now, I have no information on this subject.

Kind regards
Brandye Barrington

Oracle Certification program.

Tags: Oracle

Similar Questions

  • Is there an option upgrade version of PC for the Mac version Acrobat v10?

    Is there an option upgrade version of PC for the Mac version Acrobat v10?

    Hi seanl70961755,

    If you wish, can upgrade to Acrobat Pro DC for Mac. Please find more info here: Plans and prices | Adobe Acrobat DC

    Thank you

    Abhishek

  • Hey there, I have a problem, I paid for the creative cloud each month. 30 days ago, I downloaded the new update only in the trial version of photoshop and today the trial version has expired and I could not work with my old adobe program, although I pay f

    Hey there, I have a problem, I paid for the creative cloud each month. 30 days ago, I downloaded the new update only in the trial version of photoshop and today the trial version has expired and I could not work with my old adobe program, although I pay for it.

    Aileen,

    Please see suggestions here: FAQ: what should I do if I have a subscription, but my application acts as if I had a trial?

    Kind regards

    Guinot

  • certification of operating systems for the installation of oracle 10G database

    Can you get it someone please let me know which are the operating systems certified for the installation of oracle 10G (10.1 and 10.2) database
    Refers to the operating system officially supported by oracle.

    Published by: arnaud on December 17, 2009 07:08

    Arnaud says:
    Can you get it someone please let me know which are the operating systems certified for the installation of oracle 10G (10.1 and 10.2) database
    Refers to the operating system officially supported by oracle.

    The Installation Guide provides the official list. Go to http://tahiti.oracle.com, select your version and find the Setup instructions for your OS family.

    http://www.Oracle.com/pls/db102/homepage, you will have to the docs of 10gr 2. ON this page, to the right, you will see a region speak installation docs. Chapter 2 (http://download.oracle.com/docs/cd/B19306_01/install.102/b15660/pre_install.htm#sthref85) tells you the requirements before you begin. It tells you the operating system.

    Operating system      One of the following operating system versions:
    
        *      Red Hat Enterprise Linux AS/ES 3.0 (Update 4 or later)
        *      Red Hat Linux 4.0
        *      SUSE Linux Enterprise Server 9.0 with SP 2 or later
        *      Asianux 1.0
        *      Asianux 2.0
    
  • Functions in pipeline for the csv data analysis?

    Hi all

    I currently have a pl/sql procedure that is used to load and parse a CSV file in a table of database within the Apex.

    Downloading csv files are quite large (nearly 1 million rows or more) and there is a time of significant waiting for the course ends. I tried both Wizard 4.2 data that was very slow loading and the apex plugin excel2collection who timed out/never finished.

    I heard functions in pipeline and how they can offer great time savings for insert instructions where the database lines have no interconnect/dependencies to each other.

    My question is, would the data through pipes to offer me a gain with my time insert statements, and if so someone could help me to implement? The current procedure is listed below, less any code validation etc. for readability. The CSV is first uploaded to a table in a BLOB file before be analyzed by the procedure.

    -- Chunk up the CSV file and split into a line at a time
      rawChunk := dbms_lob.substr(bloContent, numChunkLength, numPosition + numExtra);
      strConversion := strConversion || utl_raw.cast_to_varchar2(rawChunk);
    
      numLineEnd := instr(strConversion,chr(10),1);  --This will return 0 if there is no chr(10) in the String
    
    
      strColumns := replace(substr(strConversion,1,numLineEnd -numTrailChar),CHR(numSpacer),',');
    
      strLine := substr(strConversion,1,numLineEnd);
      strLine := substr(strLine,1,length(strLine) - numTrailChar);
       
      -- Break each line into columns using the delimeter
      arrData := wwv_flow_utilities.string_to_table (strLine, '|');
    
        FOR i in 1..arrData.count
        LOOP
      
         --Now we concatenate the Column Values with a Comma
          strValues := strValues || arrData(i) || ','; 
    
        END LOOP;
    
         --Remove the trailing comma
          strValues := rtrim(strValues,',');
    
         -- Insert the values into target table, one row at a time
        BEGIN
          EXECUTE IMMEDIATE 'INSERT INTO ' || strTableName || ' (' || strColumns || ')
                             VALUES (' || strValues ||  ')';
        END;
      
        numRow := numRow + 1; --Keeps track of what row is being converted
    
        
       -- We set/reset the values for the next LOOP cycle
        strLine := NULL;
        strConversion := null;
        strValues := NULL;
        numPosition := numPosition + numLineEnd;
        numExtra := 0;
        numLineEnd := 0;
      END IF;
    END LOOP;
    
    

    Apex-user wrote:

    Hi Chris,

    I'm trying to expand your code to use more tou both current columns, but having trouble with the format here...

    1. While (l_clob) dbms_lob.getlength > l_off and l_off > 0 loop
    2. l_off_new: = instr (l_clob, c_sep, l_off, c_numsep);
    3. line (csv_split_type)
    4. substr (l_clob, l_off, instr (l_clob, c_sep, l_off)-l_off)
    5. , substr (l_clob, instr (l_clob, c_sep, l_off) + 1, l_off_new - instr (l_clob, c_sep, l_off) - 1)
    6. ));
    7. l_off: = l_off_new + 2; -to switch c_sep and line (10 sep

    How can I add more columns to this code? I'm mixed with all segments of substr and instr.

    I've done a rewrite on it (12 sec for 50,000 lines, 4 columns ~ 7 MB, 2.2 sec for 10,000 lines)

    create or replace function get_csv_split_cr (blob p_blob)

    return csv_table_split_type

    pipelined

    as

    c_sep constant varchar2 (2): = "";

    c_line_end constant varchar2 (1): = Chr (10);

    l_row varchar2 (32767).

    number of l_len_clob;

    number of l_off: = 1;

    CLOB l_clob;

    -below is used only for the call of dbms_lob.converttoclob

    l_src_off pls_integer: = 1;

    l_dst_off pls_integer: = 1;

    number of l_ctx: = dbms_lob. DEFAULT_LANG_CTX;

    number of l_warn: = dbms_lob. WARN_INCONVERTIBLE_CHAR;

    Start

    DBMS_LOB.CREATETEMPORARY (l_clob, true);

    DBMS_LOB.converttoclob (l_clob, p_blob, dbms_lob.lobmaxsize, l_src_off, l_dst_off, dbms_lob. DEFAULT_CSID, l_ctx, l_warn);

    -Attention: hypothesis that there is at least a 'correct' csv-line

    -should perhaps find a better guard condition

    -Hypothesis: last column ends with the separator

    l_len_clob: = length (l_clob);

    While l_len_clob > l_off and l_off > 0 loop

    l_row: = substr (l_clob, l_off, instr (l_clob, c_line_end, l_off)-l_off);

    line (csv_split_type)

    -start of the first; occurrence - 1

    substr (l_row, 1, instr (l_row, c_sep) - 1)

    -first; second occurrence; accident - first; occurrence

    , substr (l_row, instr (l_row, c_sep, 1, 1) + 1, instr (l_row, c_sep, 1, 2) - instr (l_row, c_sep, 1, 1) - 1)

    -second; third occurrence; occurrence - second; occurrence

    , substr (l_row, instr (l_row, c_sep, 1, 2) + 1, instr (l_row, c_sep, 1, 3) - instr (l_row, c_sep, 1, 2) - 1)

    - and so on

    , substr (l_row, instr (l_row, c_sep, 1, 3) + 1, instr (l_row, c_sep, 1, 4) - instr (l_row, c_sep, 1, 3) - 1)

    ));

    l_off: = l_off + length (l_row) + 1; -to switch c_sep and line (10 sep

    end loop;

    return;

    end;

    You must change the csv_split_type also.

    Update: I had to correct, combined version of two upward.

  • The number of heartbeat for the host data warehouses is 0, which is less than required: 2

    Hello

    I have trouble creating my DRS cluster + storage of DRS, I have 3 hosts esxi 5.1 for the task

    First, I created the cluster, no problem with that, so the DRS storage was created and now I can see in the Summary tab

    "The number of heartbeat for the host data warehouses is 0, which is less than required: 2".

    I search the Web and there are similar problems when people has only a single data store (the one that came with ESXi) and need to add another but in my case... vcenter detects any...

    In the views of storage I see the store of data (VMFS) but for some strange reason the cluster not

    In order to achieve data warehouses minimum (2) can I create an NFS and map it in THE 3 esxi hosts? Vcenter which consider a play config?

    Thank you

    You probably only have local data warehouses, which are not that HA would require for this feature (pulsations datastore) to work properly.

    You will need either 2 iSCSI, FC 2 or 2 NFS volumes... Or a combination of the any of them, for this feature to work. If you don't want to use this feature, you can also turn it off:

    http://www.yellow-bricks.com/2012/04/05/the-number-of-vSphere-HA-heartbeat-datastores-for-this-host-is-1-which-is-less-than-required-2/

  • What is the proposal for substitution for the AMPERSAND (&) in oracle?

    Hi all

    It's my request...

    Select * from
    (
    Select 'abc' a Union double
    Select 'def' a Union double
    Select 'abc and xyz' double
    ) where a = 'abc and xyz '.

    Output must be "abc and xyz."

    Since I use '&' it prompts you for a value between my keyboard.

    Yes, I know, I'll get the output using set scan off the coast, together set off, set escape on.

    What is the proposal for substitution for the AMPERSAND (&) in oracle?

    "Abc and xyz" value will be passed forward and will be passed to the select...

    Kind regards
    Varma Koffi N

    nkvkashyap wrote:
    Hi all

    It's my request...

    Select * from
    (
    Select 'abc' a Union double
    Select 'def' a Union double
    Select 'abc & xyz' double
    ) where a ='abc & xyz';

    Output must be 'abc & xyz '.

    Since I use '&' it prompts you for a value between my keyboard.

    Yes, I know, I'll get the output using set scan off the coast, together set off, set escape on.

    What is the alternate to sign & (and) in oracle?

    The value 'abc & xyz' will be passed forward and will be passed to the select statement.

    Kind regards
    Varma Koffi N

    I think you're confused. Oracle (i.e. DB) doesn't care &. However the frontends like SQL * more: and request you substitution (unless the whole set off etc.).

    If you use Java, and pass channels with & without problem.

  • Recommended value for the BAM data expiration time

    Hello

    Can someone tell me what is the recommended value for the BAM data expiration time?

    Enterprise Server default is 24 hours, but I would like to be able to raise the average runtime instance after several months. Is it reasonable to the value of the time-out value a high value? Or it will have an impact on the performance of BPM/BAM?

    Thanks in advance.

    Best regards
    CA

    Normally, we keep the BAM data expiration time at halfway with 24 to 72 hours. For historical reports that you are looking for the Data Mart / Data Warehouse DB are more logical. This database stores the data forever and takes pictures at longer intervals, normally 24 hours. These data are not in time real normally then because a capture instant is only taken once per day but will give you historical reports that you are looking for. The data from this database structure is almost identical to the BAM DB.

  • Need to create a structure for the target data store?

    Hi Experts,

    If I create a structure for the target data store, and then load the data from source to target works fine. If I make mistakes.

    Is necessary to create a structure for target?

    Please help me...

    Thanks in advance.

    A.Kavya.

    I found the answer. No need to create the structure for target temporary data store. and we need to create the structure for the permanent target data store.

  • BPM deployment error: java.lang.IllegalArgumentException: Argument 0 for the @NotNull of oracle/bpm/project/interfaces/ConversationHelper.getReferenceNameFrom parameter must not be null

    Hey guys,.

    I get this error when I tried to deploy the process. Din could not do anything on the net

    java.lang.IllegalArgumentException: Argument 0 for the @NotNull of oracle/bpm/project/interfaces/ConversationHelper.getReferenceNameFrom parameter must not be null

    My composite application contains a certain number of processes. I added a BPEL process and then removed, but since then, appears it problems and could not deploy the composite application more. Any hint is appreciated.

    the entire stack trace is:

    Error when bean calling "domain manager": error deployment suitcase BPMN.
    error when trying to deploy the component BPMN "D:\oracle\domains\BPMSOA_DOMAIN\servers\BPM_1\dc\soa_6b5472e6-f4dc-4cd1-a7b5-c7b81f5e24ab" file the reported exception is: java.lang.IllegalArgumentException: Argument 0 for the @NotNull of oracle/bpm/project/interfaces/ConversationHelper.getReferenceNameFrom parameter must not be null

    This error was an exception thrown by the underlying deployment module.
    Check the trace for the exception in the log (the connection value level debug mode).
    ORABPEL-05250

    Suitcase BPMN deployment by mistake.
    error when trying to deploy the component BPMN "D:\oracle\domains\BPMSOA_DOMAIN\servers\BPM_1\dc\soa_6b5472e6-f4dc-4cd1-a7b5-c7b81f5e24ab" file the reported exception is: java.lang.IllegalArgumentException: Argument 0 for the @NotNull of oracle/bpm/project/interfaces/ConversationHelper.getReferenceNameFrom parameter must not be null

    This error was an exception thrown by the underlying deployment module.
    Check the trace for the exception in the log (the connection value level debug mode).

    at com.collaxa.cube.engine.deployment.DeploymentManager.deployComponent(DeploymentManager.java:202)
    at com.collaxa.cube.ejb.impl.CubeServerManagerBean._deployOrLoadComponent(CubeServerManagerBean.java:949)
    at com.collaxa.cube.ejb.impl.CubeServerManagerBean.deployComponent(CubeServerManagerBean.java:128)
    at sun.reflect.GeneratedMethodAccessor1654.invoke (unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:606)
    at com.bea.core.repackaged.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:310)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:182)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:149)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.doProceed(DelegatingIntroductionInterceptor.java:131)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.invoke(DelegatingIntroductionInterceptor.java:119)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.oracle.pitchfork.spi.MethodInvocationVisitorImpl.visit(MethodInvocationVisitorImpl.java:34)
    at weblogic.ejb.container.injection.EnvironmentInterceptorCallbackImpl.callback(EnvironmentInterceptorCallbackImpl.java:54)
    at com.oracle.pitchfork.spi.EnvironmentInterceptor.invoke(EnvironmentInterceptor.java:42)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.interceptor.ExposeInvocationInterceptor.invoke(ExposeInvocationInterceptor.java:89)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)


    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.doProceed(DelegatingIntroductionInterceptor.java:131)
    at com.bea.core.repackaged.springframework.aop.support.DelegatingIntroductionInterceptor.invoke(DelegatingIntroductionInterceptor.java:119)
    at com.bea.core.repackaged.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at com.bea.core.repackaged.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:204)
    to com.sun.proxy. $Proxy345.deployComponent (unknown Source)
    at oracle.bpm.bpmn.engine.ejb.impl.BPMNServerManagerBean_6gbx7k_IBPMNCubeServerManagerLocalBeanImpl.__WL_invoke (unknown Source)
    at weblogic.ejb.container.internal.SessionLocalMethodInvoker.invoke(SessionLocalMethodInvoker.java:39)
    at oracle.bpm.bpmn.engine.ejb.impl.BPMNServerManagerBean_6gbx7k_IBPMNCubeServerManagerLocalBeanImpl.deployComponent (unknown Source)
    at oracle.fabric.CubeServiceEngine.load(CubeServiceEngine.java:923)
    at oracle.bpm.bpmn.engine.service.BPMNServiceEngine.load(BPMNServiceEngine.java:703)
    at oracle.bpm.bpmn.engine.service.BPMNServiceEngine.load(BPMNServiceEngine.java:157)
    at oracle.integration.platform.blocks.deploy.CompositeDeploymentConnection.deployComponents(CompositeDeploymentConnection.java:243)
    at oracle.integration.platform.blocks.deploy.CompositeDeploymentConnection.deploy(CompositeDeploymentConnection.java:94)
    at oracle.integration.platform.blocks.deploy.CompositeDeploymentManagerImpl.initDeployment(CompositeDeploymentManagerImpl.java:185)
    at oracle.integration.platform.blocks.deploy.CompositeDeploymentManagerImpl.load(CompositeDeploymentManagerImpl.java:63)
    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:606)
    at org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:307)
    at org.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:182)
    at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:149)
    at oracle.integration.platform.blocks.deploy.DeploymentEventPublisher.invoke(DeploymentEventPublisher.java:86)
    at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:171)
    at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:204)
    to com.sun.proxy. $Proxy438.load (unknown Source)
    at oracle.integration.platform.blocks.deploy.StandaloneCompositeDeploymentCoordinatorImpl.coordinateCompositeDeployment(StandaloneCompositeDeploymentCoordinatorImpl.java:67)
    at oracle.integration.platform.blocks.deploy.servlet.BaseDeployProcessor.deployNewComposite(BaseDeployProcessor.java:467)
    at oracle.integration.platform.blocks.deploy.servlet.BaseDeployProcessor.deploySARs(BaseDeployProcessor.java:268)
    at oracle.integration.platform.blocks.deploy.servlet.DeployProcessor.doDeployWork(DeployProcessor.java:203)
    at oracle.integration.platform.blocks.deploy.servlet.DeployProcessor.doDeployWork(DeployProcessor.java:147)
    at oracle.integration.platform.blocks.deploy.servlet.DeployProcessor.doDeploy(DeployProcessor.java:134)
    at oracle.integration.platform.blocks.deploy.servlet.DeployProcessor.process(DeployProcessor.java:100)
    at oracle.integration.platform.blocks.deploy.servlet.CompositeDeployerServlet.doPostInsideLoggingSession(CompositeDeployerServlet.java:221)
    at oracle.integration.platform.blocks.deploy.servlet.CompositeDeployerServlet.doPost(CompositeDeployerServlet.java:130)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
    to weblogic.servlet.internal.StubSecurityHelper$ ServletServiceAction.run (StubSecurityHelper.java:227)
    at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
    at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:301)
    at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    to oracle.security.jps.ee.http.JpsAbsFilter$ 1.run(JpsAbsFilter.java:119)
    at java.security.AccessController.doPrivileged (Native Method)
    at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:324)
    at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:460)
    at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
    at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    to oracle.security.jps.ee.http.JpsAbsFilter$ 1.run(JpsAbsFilter.java:119)
    at java.security.AccessController.doPrivileged (Native Method)
    at oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:324)
    at oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:460)
    at oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)
    at oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)
    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    at oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:163)
    at weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)
    to weblogic.servlet.internal.WebAppServletContext$ ServletInvocationAction.wrapRun (WebAppServletContext.java:3730)
    to weblogic.servlet.internal.WebAppServletContext$ ServletInvocationAction.run (WebAppServletContext.java:3696)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
    at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)
    at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2273)
    at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2179)
    at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1490)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)

    The answer late, sorry it was a bug in the product, was fixed after you apply the hotfix.

    Cheers, nasser

  • Need to retrieve the data for the current date.

    Hello

    I have a table which then retrieves information when using this command.

    Select ta_acct, shift, created_on track_alerts;

    Technicolor A 24 March 14

    Technicolor A 24 March 14

    Technicolor A 24 March 14

    Technicolor A 24 March 14

    Manitoba telecom a 24 March 14 system

    Technicolor A 24 March 14

    I used this statement to retrieve the data for the given date.

    Select ta_acct, shift, created_on track_alerts where created_on = 24 March 14 ';

    Its not data recovery.

    Need help.

    Kind regards

    Prasad K T,.

    984002170

    Prasad K T wrote:

    Yes the created data type is date.

    CREATED_ON DATE

    Partha thanks it works now.

    Select ta_acct, shift, created_on in track_alerts where to move is: Shift and TRUNC (created_on) = TO_DATE('24-MAR-2014','DD-MON-YYYY');

    Still, I made a small change to my querry.

    Select ta_acct, shift, created_on track_alerts where to move is: shft and TRUNC (created_on) = TO_DATE (select double sysdate,'MON-DD-YYYY "");

    For this statement, it does not work.

    of course not...

    first: sysdate returns a date so no need of conversion here

    and

    second SYSDATE includes time, so your application should look like this:

    Select ta_acct, shift, created_on in track_alerts where to move is: Shift and TRUNC (created_on) = trunc (sysdate)

    or

    Select ta_acct, shift, created_on in track_alerts where to move is: shft and created_on > = trunc (sysdate) and created_on<>

    HTH

  • Preparations for the launch of Oracle Universal Installer from/tmp/OraInstall2014-02-25_01-35-08 PM. Please wait... Error in CreateOUIProcess(): 13: permission denied

    Hi all

    Im getting below error when installing on RHEL 6 12

    Preparations for the launch of Oracle Universal Installer from/tmp/OraInstall2014-02-25_01-35-08 PM. Please wait... Error in CreateOUIProcess(): 13
    : Permission denied

    I got permission to

    mkdir Pei /u01/app/oracle/product/12.1.0/db_1

    chown-r oracle: oinstall/U01

    chmod-r 775/U01


    permission on temp

    drwxrwxrwt. 9 root root 4096/tmp


    Bash profile


    export TMP = / tmp
    export TMPDIR = $TMP
    export ORACLE_HOME=/u01/app/oracle/product/12.1.0/db_1
    export ORACLE_BASE = / u01/app/oracle
    export ORACLE_SID = BDEV
    export PATH = $ORACLE_HOME/bin: $PATH
    #export PATH = $ORACLE_HOME/OPatch: $PATH
    export LD_LIBRARY_PATH = $ORACLE_HOME/lib: / lib: / usr/lib
    Export CLASSPATH = $ORACLE_HOME/jlib: $ORACLE_HOME/rdbms/jlib

    The link refers to the TEMP variable as so defined.

    I think you would have to remove and back up/tmp at all the less for the new permissions to take effect.  I would not have thought that it is a dynamic change that takes effect by simply changing the file.

  • Difference in the number of records for the same date - 11 GR 2

    Guy - 11 GR on Windows2005 2, 64-bit.

    BILLING_RECORD_KPN_ESP - is a monthly partitioned table.
    BILLING_RECORD_IDX #DATE - is a local index on "charge_date" in the table above.

    SQL > select / * + index (BILLING_RECORD_KPN_ESP BILLING_RECORD_IDX #DATE) * /.
    2 (trunc (CHARGE_DATE)) CHARGE_DATE;
    3 count (1) Record_count
    4. IN "RATOR_CDR". "" BILLING_RECORD_KPN_ESP ".
    where the 5 CHARGE_DATE = January 20, 2013.
    Group 6 by trunc (CHARGE_DATE)
    5 m

    CHARGE_DATE RECORD_COUNT
    ------------------ ------------
    2401 20 January 13-> > some records here.

    -> > Here I can see only '2041' records for Jan/20. But in the query below, it shows "192610" for the same date.

    Why is this difference in the number of records?

    SQL > select / * + index (BILLING_RECORD_KPN_ESP BILLING_RECORD_IDX #DATE) * /.
    (trunc (CHARGE_DATE)) CHARGE_DATE,
    2 count (1) Record_count
    3. FOR "RATOR_CDR." "" BILLING_RECORD_KPN_ESP ".
    "4 where CHARGE_DATE > 20 January 2013."
    Group of 5 by trunc (CHARGE_DATE)
    6 order by trunc (CHARGE_DATE)
    5 m

    CHARGE_DATE RECORD_COUNT
    ------------------ ------------
    192610 20 January 13-> > more records here
    JANUARY 21, 13 463067
    JANUARY 22, 13 520041
    23 JANUARY 13 451212
    JANUARY 24, 13 463273
    JANUARY 25, 13 403276
    JANUARY 26, 13 112077
    27 JANUARY 13 10478
    28 JANUARY 13 39158

    Thank you!

    Because in the second example you also select rows that have a nonzero component.

    The first example selects only rows that are 00:00:00

    (by the way, you should ask questions like this in the forum SQL)

  • ESXi is unable to install 'his place on the disk for the dump data' no ideas?

    It is an older server.

    Data sheet:

    GOING Linux 2200 series

    Intel P3 700 MHz x 2

    768 MB OF RAM

    9.1 GB SCSI x 3 (configured in RAID 5 now)

    I have attached a picture of the error message that is received and typed most of the message below.

    NOT_IMPLEMENTED /build/mts/release/bora-123629/bora/vmkernel/sched/sched.c:5075

    Frame 0x1402ce0 ip = 0x62b084 cr2 = cr3 = 0 x 0 = ox3000 cr4 = 0 x 20

    are is 0xffffffff ds is 0xffffffff fs = 0xffffffff gs = 0xffffffff

    EAX = 0xffffffff = 0xffffffff = 0xffffffff edx ecx ebx = 0xffffffff

    = 0x1402e3c = 0xffffffff edi esi EBP = 0xffffffff err =-1 eflags = 0xffffffff

    * 0:0 / & lt; NULL & gt;  1:0 / & lt; NULL & gt;

    0x1402e3c: battery: 0x830c3f, 0x1402e58, 0x1402e78

    VMK availability: 0:00:00:00.026 TSC: 222483259709

    No space on disk for the dump data

    Waiting for debugger... (World 0)

    Debugger is listening on the serial port...

    Press ESC to enter the local debugger

    This could be a simple problem or not, I'm not sure. I spent several hours already trying to reconfigure the readers to try to get the installation to recognize.

    Any help is greatly appreciated.

    I agree with Matt, the material can be simply too old-

    If you find this or any other answer useful please consider awarding points marking the answer correct or useful

  • Not able to start agent cache for the requested data store

    Hello

    This is my first attempt in TimesTen. I am running TimesTen on the same host Linux (RHES 5.2) running Oracle 11 g R2. TimesTen version is:

    TimesTen Release 11.2.1.4.0


    Trying to create a simple cache.

    The DSN entry section for ttdemo1 to. odbc.ini is as follows:

    + [ttdemo1] +.
    Driver=/home/Oracle/TimesTen/TimesTen/lib/libtten.so
    Data store = / work/oracle/TimesTen_store/ttdemo1
    PermSize = 128
    TempSize = 128
    UID = hr
    OracleId = MYDB
    DatabaseCharacterSet = WE8MSWIN1252
    ConnectionCharacterSet = WE8MSWIN1252

    With the help of ttisql I connect

    Command > Connect "dsn = ttdemo1; pwd = oracle; oraclepwd = oracle;
    Successful login: DSN = ttdemo1; UID = hr; DataStore = / work/oracle/TimesTen_store/ttdemo1; DatabaseCharacterSet = WE8MSWIN1252; ConnectionCharacterSet=WE8MSWIN1252;DRIVER=/home/oracle/TimesTen/timesten/lib/libtten.so;OracleId=MYDB; PermSize = 128; TempSize = 128; TypeMode = 0; OracleNetServiceName = MYDB;
    (Default AutoCommit = 1).
    Command > call ttcacheuidpwdset ('ttsys', 'oracle');
    Command > call ttcachestart;
    * 10024: could not start agent cache for the requested data store. Could not initialize Handle.* Oracle environment
    The command failed.

    The following text appears in the tterrors.log:

    15:41:21.82 Err: ORA: 9143: ora-9143 - 1252549744-xxagent03356: database: TTDEMO1 OCIEnvCreate failed. Return - 1 code
    15:41:21.82 Err: 7140: oraagent says it failed to start: could not initialize manage Oracle environment.
    15:41:22.36 Err: 7140: TT14004: failed to create the demon TimesTen: couldn't reproduce oraagent for "/ work/oracle/TimesTen_store/ttdemo1 ': has not been initialized Handl Oracle environment

    What are the reasons that the demon cannot happen again to another agent? FYI, the environment variables are defined as:

    ORA_NLS33=/U01/app/Oracle/product/11.2.0/Db_1/ocommon/NLS/Admin/data
    ANT_HOME = / home/oracle/TimesTen/ttdemo1/3rdparty/ant
    CLASSPATH=/home/oracle/TimesTen/ttdemo1/lib/ttjdbc5.jar:/home/oracle/TimesTen/ttdemo1/lib/orai18n.jar:/home/oracle/TimesTen/ttdemo1/lib/timestenjmsxla.jar:/home/oracle/TimesTen/ttdemo1/3rdparty/jms1.1/lib/jms.jar:.
    Oracle@rhes5:/Home/Oracle/TimesTen/ttdemo1/info% echo $LD_LIBRARY_PATH
    /home/oracle/TimesTen/ttdemo1/lib:/home/oracle/TimesTen/ttdemo1/ttoracle_home/instantclient_11_1:/u01/app/oracle/product/11.2.0/db_1/lib:/u01/app/oracle/product/11.2.0/db_1/network/lib:/lib:/usr/lib:/usr/ucblib:/usr/local/lib


    See you soon

    I see no problem here. The ENOENTs are superfluous because it locates libtten here:

    23302 open ("/ home/oracle/TimesTen/11.2.1.4.0/TimesTen/ttimdb1/lib/libtten.so", O_RDONLY) = 3

    without doubt, it does the same thing trying to find the libttco.so?

    23302 open ("/ home/oracle/TimesTen/11.2.1.4.0/TimesTen/ttimdb1/ttoracle_home/instantclient_11_1/libttco.so", O_RDONLY) =-1 ENOENT (no such file or directory)

    Thank you for taking the trace. I really want to have a look at the complete file if you can send it to me?

Maybe you are looking for