Loading using SQL rules

Hi - I have a single charge rule dimension loading a hierarchy in a cube ASO. It uses SQL to return the values of parent/child, with the operator of consolidation and several UDA values

When I run the State of charge with one set of SQL (what happens is not to have a prescription), data are retieved and the hierarchy is to build successfully, but with members of Level0 not in the right order, but always under the correct parents.

When I use a different piece of SQL (with an order by clause to correctly classify members), it will fail to load and the MaxL script will fail.

The SQLs were tested in TOAD and the overview of data on Essbase, so they return both the same set of data. Nothing useful is returned to work or log of Essbase applications

Someone met by amythinjg like this before?


What's really weird is that the SQL code that is a failure previously worked. The only thing that changed was that a new set of basic data has been loaded from the database

Any help gratefully received

THX

When you use the "order by" clause, what is the SQL code that fails, or just the rule of the load?
How the 'order by' clause orders the members?

My guess is that it is not for lack of SQL (you have already proved it a TOAD), but that members really are not sorted correctly for loading. In a dimension build parent/child of the higher level in the dimension member must be in the first record as the "Parent"(c'est le membre de dimension dans la plupart deles de cas). "»

Robert

Tags: Business Intelligence

Similar Questions

  • Date fromats loading using SQL * Loader

    Hello

    I have the data to load into an Oracle DB using sqlldr below

    333. 789. 6 ||| 01-08-2013 | 2014-08-01 |

    334. 789. 6 ||| 01-08-2013 | 2014-08-01 |

    335. 789. 6 ||| 01-08-2013 | 2014-08-01 |

    It fails while the failed to load on date formats. How can I fix everything by using load data.

    Thank you

    Sylvie

    Works for me. You must validate the CTL file.

    Here is an example:

    Control line: -.

    DOWNLOAD THE DATA

    INFILE *.

    IN THE TABLE test_table

    REPLACE

    FIELDS TERMINATED BY ', '.

    (

    Col DATE 'yyyy-mm-dd ".

    )

    BEGINDATA

    2013-08-01

    2014-09-01

    2015-10-01

    2016-11-01

    2017-12-01

    -Create table

    SQL > create table test_table (date of the pass);

    Table created.

    -SQL * Loader

    C:\Users\43729434>sqlldr user/password@db_alias name control=C:\fakepath\test_ctl.ctl

    SQL * Loader: release 11.2.0.1.0 - Production Tue Oct 24 10:26:40 2013

    Copyright (c) 1982, 2009, Oracle and/or its affiliates.  All rights reserved.

    Commit the point reached - the number of logical records 4
    Commit the point reached - the number of logical records 5

    -Check the data in the database

    SQL > alter session set nls_date_format = 'DD-Mon-YYYY ';

    Modified session.

    SQL > select col from test_table;

    COL
    -----------
    August 1, 2013
    01-sep-2014
    October 1, 2015
    November 1, 2016
    December 1, 2017

    SQL >

  • How to tune the loading using SQL LOADER

    Hi all

    I use SQLLDR to load the file into the target. All the files I am receving is ' | ' delimited. Some of the files size more than 3 GB. It takes a lot of time during loading.

    Help me to know if it is possible of tuner SQLLDR.

    Here is the example of control file.

    DOWNLOAD THE DATA
    UTF8 CHARACTER
    INFILE ' / var/hulu/ptp/ABC.txt'
    BADFILE ' / var/hulu/ptp/ABC.bad'
    DISCARDFILE ' / var/hulu/ptp/ABC.dsc'

    IN THE TABLE "PTP_ABC".
    TRUNCATE

    FIELDS TERMINATED BY ' |'
    TRAILING NULLCOLS
    (
    CODE,
    DESCRIPTION,
    FILLING,
    COKE
    )

    Change you your current sqlldr using external tables, this make the parallel load automatically, or launch sqlldr processes in the background of a single script.

    We doing something similar to this with the help of a shell like this script:

    nohup sqlldr userid=/ data=data01.dat log=my1.log control=my.ctl bad=my1.bad direct=yes silent=all parallel=true &
    nohup sqlldr userid=/ data=data02.dat log=my2.log control=my.ctl bad=my2.bad direct=yes silent=all parallel=true &
    nohup sqlldr userid=/ data=data03.dat log=my3.log control=my.ctl bad=my3.bad direct=yes silent=all parallel=true &
    nohup sqlldr userid=/ data=data04.dat log=my4.log control=my.ctl bad=my4.bad direct=yes silent=all parallel=true &
    nohup sqlldr userid=/ data=data05.dat log=my5.log control=my.ctl bad=my5.bad direct=yes silent=all parallel=true &
    nohup sqlldr userid=/ data=data06.dat log=my6.log control=my.ctl bad=my6.bad direct=yes silent=all parallel=true &
    nohup sqlldr userid=/ data=data07.dat log=my7.log control=my.ctl bad=my7.bad direct=yes silent=all parallel=true &
    ...
    

    We are currently opening 20 simultaneous processes and we get an overal 400 000 platelets per second speed. You test what is the maximum number of concurrent processes, you can open.

    Course which you need a first step to split the log into smaller chuncks. In my example, the file data.dat got cut into 20 smaller files, and each of them at the same time got loaded in parallel in the database. If you can not split the file, then you should go for external tables.

    You can always test direct = true without any parallel option. This should be more than noticiable.

    LW

  • Question to load data using sql loader in staging table, and then in the main tables!

    Hello

    I'm trying to load data into our main database table using SQL LOADER. data will be provided in separate pipes csv files.

    I have develop a shell script to load the data and it works fine except one thing.

    Here are the details of a data to re-create the problem.

    Staging of the structure of the table in which data will be filled using sql loader

    create table stg_cmts_data (cmts_token varchar2 (30), CMTS_IP varchar2 (20));

    create table stg_link_data (dhcp_token varchar2 (30), cmts_to_add varchar2 (200));

    create table stg_dhcp_data (dhcp_token varchar2 (30), DHCP_IP varchar2 (20));

    DATA in the csv file-

    for stg_cmts_data-

    cmts_map_03092015_1.csv

    WNLB-CMTS-01-1. 10.15.0.1

    WNLB-CMTS-02-2 | 10.15.16.1

    WNLB-CMTS-03-3. 10.15.48.1

    WNLB-CMTS-04-4. 10.15.80.1

    WNLB-CMTS-05-5. 10.15.96.1

    for stg_dhcp_data-

    dhcp_map_03092015_1.csv

    DHCP-1-1-1. 10.25.23.10, 25.26.14.01

    DHCP-1-1-2. 56.25.111.25, 100.25.2.01

    DHCP-1-1-3. 25.255.3.01, 89.20.147.258

    DHCP-1-1-4. 10.25.26.36, 200.32.58.69

    DHCP-1-1-5 | 80.25.47.369, 60.258.14.10

    for stg_link_data

    cmts_dhcp_link_map_0309151623_1.csv

    DHCP-1-1-1. WNLB-CMTS-01-1,WNLB-CMTS-02-2

    DHCP-1-1-2. WNLB-CMTS-03-3,WNLB-CMTS-04-4,WNLB-CMTS-05-5

    DHCP-1-1-3. WNLB-CMTS-01-1

    DHCP-1-1-4. WNLB-CMTS-05-8,WNLB-CMTS-05-6,WNLB-CMTS-05-0,WNLB-CMTS-03-3

    DHCP-1-1-5 | WNLB-CMTS-02-2,WNLB-CMTS-04-4,WNLB-CMTS-05-7

    WNLB-DHCP-1-13 | WNLB-CMTS-02-2

    Now, after loading these data in the staging of table I have to fill the main database table

    create table subntwk (subntwk_nm varchar2 (20), subntwk_ip varchar2 (30));

    create table link (link_nm varchar2 (50));

    SQL scripts that I created to load data is like.

    coil load_cmts.log

    Set serveroutput on

    DECLARE

    CURSOR c_stg_cmts IS SELECT *.

    OF stg_cmts_data;

    TYPE t_stg_cmts IS TABLE OF stg_cmts_data % ROWTYPE INDEX BY pls_integer;

    l_stg_cmts t_stg_cmts;

    l_cmts_cnt NUMBER;

    l_cnt NUMBER;

    NUMBER of l_cnt_1;

    BEGIN

    OPEN c_stg_cmts.

    Get the c_stg_cmts COLLECT in BULK IN l_stg_cmts;

    BECAUSE me IN l_stg_cmts. FIRST... l_stg_cmts. LAST

    LOOP

    SELECT COUNT (1)

    IN l_cmts_cnt

    OF subntwk

    WHERE subntwk_nm = l_stg_cmts (i) .cmts_token;

    IF l_cmts_cnt < 1 THEN

    INSERT

    IN SUBNTWK

    (

    subntwk_nm

    )

    VALUES

    (

    l_stg_cmts (i) .cmts_token

    );

    DBMS_OUTPUT. Put_line ("token has been added: ' |") l_stg_cmts (i) .cmts_token);

    ON THE OTHER

    DBMS_OUTPUT. Put_line ("token is already present'");

    END IF;

    WHEN l_stg_cmts EXIT. COUNT = 0;

    END LOOP;

    commit;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    for dhcp


    coil load_dhcp.log

    Set serveroutput on

    DECLARE

    CURSOR c_stg_dhcp IS SELECT *.

    OF stg_dhcp_data;

    TYPE t_stg_dhcp IS TABLE OF stg_dhcp_data % ROWTYPE INDEX BY pls_integer;

    l_stg_dhcp t_stg_dhcp;

    l_dhcp_cnt NUMBER;

    l_cnt NUMBER;

    NUMBER of l_cnt_1;

    BEGIN

    OPEN c_stg_dhcp.

    Get the c_stg_dhcp COLLECT in BULK IN l_stg_dhcp;

    BECAUSE me IN l_stg_dhcp. FIRST... l_stg_dhcp. LAST

    LOOP

    SELECT COUNT (1)

    IN l_dhcp_cnt

    OF subntwk

    WHERE subntwk_nm = l_stg_dhcp (i) .dhcp_token;

    IF l_dhcp_cnt < 1 THEN

    INSERT

    IN SUBNTWK

    (

    subntwk_nm

    )

    VALUES

    (

    l_stg_dhcp (i) .dhcp_token

    );

    DBMS_OUTPUT. Put_line ("token has been added: ' |") l_stg_dhcp (i) .dhcp_token);

    ON THE OTHER

    DBMS_OUTPUT. Put_line ("token is already present'");

    END IF;

    WHEN l_stg_dhcp EXIT. COUNT = 0;

    END LOOP;

    commit;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    for link -.

    coil load_link.log

    Set serveroutput on

    DECLARE

    l_cmts_1 VARCHAR2 (4000 CHAR);

    l_cmts_add VARCHAR2 (200 CHAR);

    l_dhcp_cnt NUMBER;

    l_cmts_cnt NUMBER;

    l_link_cnt NUMBER;

    l_add_link_nm VARCHAR2 (200 CHAR);

    BEGIN

    FOR (IN) r

    SELECT dhcp_token, cmts_to_add | ',' cmts_add

    OF stg_link_data

    )

    LOOP

    l_cmts_1: = r.cmts_add;

    l_cmts_add: = TRIM (SUBSTR (l_cmts_1, 1, INSTR (l_cmts_1, ',') - 1));

    SELECT COUNT (1)

    IN l_dhcp_cnt

    OF subntwk

    WHERE subntwk_nm = r.dhcp_token;

    IF l_dhcp_cnt = 0 THEN

    DBMS_OUTPUT. Put_line ("device not found: ' |") r.dhcp_token);

    ON THE OTHER

    While l_cmts_add IS NOT NULL

    LOOP

    l_add_link_nm: = r.dhcp_token |' _TO_' | l_cmts_add;

    SELECT COUNT (1)

    IN l_cmts_cnt

    OF subntwk

    WHERE subntwk_nm = TRIM (l_cmts_add);

    SELECT COUNT (1)

    IN l_link_cnt

    LINK

    WHERE link_nm = l_add_link_nm;

    IF l_cmts_cnt > 0 AND l_link_cnt = 0 THEN

    INSERT INTO link (link_nm)

    VALUES (l_add_link_nm);

    DBMS_OUTPUT. Put_line (l_add_link_nm |) » '||' Has been added. ") ;

    ELSIF l_link_cnt > 0 THEN

    DBMS_OUTPUT. Put_line (' link is already present: ' | l_add_link_nm);

    ELSIF l_cmts_cnt = 0 then

    DBMS_OUTPUT. Put_line (' no. CMTS FOUND for device to create the link: ' | l_cmts_add);

    END IF;

    l_cmts_1: = TRIM (SUBSTR (l_cmts_1, INSTR (l_cmts_1, ',') + 1));

    l_cmts_add: = TRIM (SUBSTR (l_cmts_1, 1, INSTR (l_cmts_1, ',') - 1));

    END LOOP;

    END IF;

    END LOOP;

    COMMIT;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    control files -

    DOWNLOAD THE DATA

    INFILE 'cmts_data.csv '.

    ADD

    IN THE STG_CMTS_DATA TABLE

    When (cmts_token! = ") AND (cmts_token! = 'NULL') AND (cmts_token! = 'null')

    and (cmts_ip! = ") AND (cmts_ip! = 'NULL') AND (cmts_ip! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:cmts_token))' cmts_token,

    cmts_ip ' RTRIM (LTRIM(:cmts_ip)) ")". "

    for dhcp.


    DOWNLOAD THE DATA

    INFILE 'dhcp_data.csv '.

    ADD

    IN THE STG_DHCP_DATA TABLE

    When (dhcp_token! = ") AND (dhcp_token! = 'NULL') AND (dhcp_token! = 'null')

    and (dhcp_ip! = ") AND (dhcp_ip! = 'NULL') AND (dhcp_ip! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:dhcp_token))' dhcp_token,

    dhcp_ip ' RTRIM (LTRIM(:dhcp_ip)) ")". "

    for link -.

    DOWNLOAD THE DATA

    INFILE 'link_data.csv '.

    ADD

    IN THE STG_LINK_DATA TABLE

    When (dhcp_token! = ") AND (dhcp_token! = 'NULL') AND (dhcp_token! = 'null')

    and (cmts_to_add! = ") AND (cmts_to_add! = 'NULL') AND (cmts_to_add! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:dhcp_token))' dhcp_token,

    cmts_to_add TANK (4000) RTRIM (LTRIM(:cmts_to_add)) ")" ""

    SHELL SCRIPT-

    If [!-d / log]

    then

    Mkdir log

    FI

    If [!-d / finished]

    then

    mkdir makes

    FI

    If [!-d / bad]

    then

    bad mkdir

    FI

    nohup time sqlldr username/password@SID CONTROL = load_cmts_data.ctl LOG = log/ldr_cmts_data.log = log/ldr_cmts_data.bad DISCARD log/ldr_cmts_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    nohup time username/password@SID @load_cmts.sql

    nohup time sqlldr username/password@SID CONTROL = load_dhcp_data.ctl LOG = log/ldr_dhcp_data.log = log/ldr_dhcp_data.bad DISCARD log/ldr_dhcp_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    time nohup sqlplus username/password@SID @load_dhcp.sql

    nohup time sqlldr username/password@SID CONTROL = load_link_data.ctl LOG = log/ldr_link_data.log = log/ldr_link_data.bad DISCARD log/ldr_link_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    time nohup sqlplus username/password@SID @load_link.sql

    MV *.log. / log

    If the problem I encounter is here for loading data in the link table that I check if DHCP is present in the subntwk table, then continue to another mistake of the newspaper. If CMTS then left create link to another error in the newspaper.

    Now that we can here multiple CMTS are associated with unique DHCP.

    So here in the table links to create the link, but for the last iteration of the loop, where I get separated by commas separate CMTS table stg_link_data it gives me log as not found CMTS.

    for example

    DHCP-1-1-1. WNLB-CMTS-01-1,WNLB-CMTS-02-2

    Here, I guess to link the dhcp-1-1-1 with balancing-CMTS-01-1 and wnlb-CMTS-02-2

    Theses all the data present in the subntwk table, but still it gives me journal wnlb-CMTS-02-2 could not be FOUND, but we have already loaded into the subntwk table.

    same thing is happening with all the CMTS table stg_link_data who are in the last (I think here you got what I'm trying to explain).

    But when I run the SQL scripts in the SQL Developer separately then it inserts all valid links in the table of links.

    Here, she should create 9 lines in the table of links, whereas now he creates only 5 rows.

    I use COMMIT in my script also but it only does not help me.

    Run these scripts in your machine let me know if you also get the same behavior I get.

    and please give me a solution I tried many thing from yesterday, but it's always the same.

    It is the table of link log

    link is already present: dhcp-1-1-1_TO_wnlb-cmts-01-1

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-02-2

    link is already present: dhcp-1-1-2_TO_wnlb-cmts-03-3
    link is already present: dhcp-1-1-2_TO_wnlb-cmts-04-4

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-5

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-01-1

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-8
    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-6
    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-0

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-03-3

    link is already present: dhcp-1-1-5_TO_wnlb-cmts-02-2
    link is already present: dhcp-1-1-5_TO_wnlb-cmts-04-4

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-7

    Device not found: wnlb-dhcp-1-13

    IF NEED MORE INFORMATION PLEASE LET ME KNOW

    Thank you

    I felt later in the night that during the loading in the staging table using UNIX machine he created the new line for each line. That is why the last CMTS is not found, for this I use the UNIX 2 BACK conversion and it starts to work perfectly.

    It was the dos2unix error!

    Thank you all for your interest and I may learn new things, as I have almost 10 months of experience in (PLSQL, SQL)

  • Unable to load data using the rules file

    Hi all

    I am facing a strange problem while loading data by using a rules file:

    1. I don't get any error/kick. It gives fair attention No. data values are modified by this data file loading. But nothing will load inside.

    2. I tried with Maxl and directly from EAS but gettting same warning.

    3 has tried to block sending n to entires in the data file and it does not work (so no Dynamics members etc..)

    4. rule file validates correctly.

    5. I tried with different types of different files (Csv, Xls, Txt, etc.) but no luck.

    Any suggestions / ideas of what could be bad?

    Don't know what exactly was the problem, but I recreated the rule files and it solved the problem... May be a case of corrupted file rule?

  • Error loading data using SQL loader

    I get an error message like "SQL * Loader - 350 combination illegal syntax of non-alphanumeriques characters error during loading of a file using SQL loader in RHEL." The command used to run SQL * Loader is:

    Sqlldr userid = < user name > / < password > control = data.ctl

    The control file is data.ctl:

    DOWNLOAD the data

    INFILE ' / home/oraprod/data.txt'

    Add in the table test

    {

    EmpID completed by «,»,

    fname completed by «,»,

    lname completed by «,»,

    treatment is completed with a white space

    }

    The data.txt file is:

    1, Kaushal, Hamad, 5000

    2, Chetan, Hamad, 1000

    Hopefully, my question is clear.

    Please get back with the answer to my query.

    Concerning

    Replace "{" by "("dans votre fichier de contrôle) "

    DOWNLOAD the data

    INFILE 'c:\data.txt.

    Add the emp_t table

    (

    EmpID completed by «,»,

    fname completed by «,»,

    lname completed by «,»,

    treatment is completed with a white space

    )

    C:\>sqlldr user/pwd@database control = c.ctl

    SQL * Loader: release 10.2.0.3.0 - Production on Wed Nov 13 10:10:24 2013

    Copyright (c) 1982, 2005, Oracle.  All rights reserved.

    Commit the point reached - the number of logical records 1

    Commit the point reached - the number of logical records 2

    SQL > select * from emp_t;

    EMPID, FNAME LNAME SALARY

    ---------- -------------------- -------------------- ----------

    1 kone hamadi 5000

    2 Chetan Hamad 1000

    Best regards

    Mohamed Houri

  • Ask about the load of metadata in Essbase using the rules file

    Hello

    In our application, we load the metadata using the Essbase rules file and here, we are faced with a problem. Suppose we have load a member of level 0, called "xxxx" under two different parent, assume parents are Parent_AAAA and Parent_BBBB. But our demand is the operator of consolidation for the first occurrence (i.e. under the Parent_AAAA) is '+' and for the Member shared under Parent_BBBB is ' - '. But, when we are load using rule file the operator of consolidation for the Member always shared loading as ' + '. Y at - it a trick to change the operator of consolidation for other than manually changing outline shared member?

    Pls help... Thanks in advance.

    I don't think that there is a way to change the default setting.

    However, what you can do is to have a different charge and rule file different text with only the metadata that you want to have an exception for shared members. In this way, you can get that you don't have to change existing rules file.

    Concerning

    ORACLE | Essbase

  • Problem loading xml using sql loader file

    I am trying to load data into the table test_xml (xmldata XMLType)

    I have an xml file and I want any file to load in a single column

    When I use the following control file and run from the command-line as follows
    sqlldr $1@$TWO_TASK direct control=$XXTOP/bin/LOAD_XML.ctl = true; :

    DOWNLOAD THE DATA
    INFILE *.
    TRUNCATE INTO TABLE test_xml
    XmlType (XMLDATA)
    FIELDS
    (
    tank fill ext_fname (100),
    XMLDATA LOBFILE (ext_fname) COMPLETED BY expressions of folklore
    )
    START DATA
    U01/appl/apps/apps_st/appl/XXTop/12.0.0/bin/file. XML

    the file is loaded in the table perfectly.

    Unfortunately I can't hard-code the name of file as file name will be changed dynamically.

    so I removed the block

    START DATA
    U01/appl/apps/apps_st/appl/XXTop/12.0.0/bin/file. XML

    control file and tried to run by giving the following command line path

    sqlldr $1@$TWO_TASK control=$XXTOP/bin/LOAD_XML.ctl direct data=/u01/APPL/apps/apps_st/appl/xxtop/12.0.0/bin/file.xml = true;

    But strangely it attempts to load each line of the xml file in the table instead of the whole file

    Please find the log of the program with the error

    ------------------------------------------------------------------
    Loading XML through SQL * Loader begins
    ------------------------------------------------------------------
    SQL * Loader-502: cannot open the data file ' <? XML version = "1.0"? > ' table field TEST_XML XMLDATA
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file '< root >' XMLDATA field table TEST_XML
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file '< ScriptFileType >' field XMLDATA table TEST_XML
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file ' < Type > forms < / Type > ' table field TEST_XML XMLDATA
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file ' < / ScriptFileType > ' table field TEST_XML XMLDATA
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file '< ScriptFileType >' field XMLDATA table TEST_XML
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file ' < Type > PLL < / Type > ' table field TEST_XML XMLDATA
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file ' < / ScriptFileType > ' table field TEST_XML XMLDATA
    SQL * Loader-553: file not found
    SQL * Loader-509: System error: no such file or directory
    SQL * Loader-502: cannot open the data file '< ScriptFileType >' field XMLDATA table TEST_XML

    Please help me how can I load full xml in a single column using command line without Hardcoding in the control file

    Published by: 907010 on January 10, 2012 02:24

    But strangely it attempts to load each line of the xml file in the table instead of the whole file

    Nothing strange, that the data parameter specifies the file containing the data to load.
    If you use the name of the XML here, the control file will try to interpret each line of XML as being separate ways.

    The traditional approach this is to have the name of the file stored in another file, say filelist.txt and use, for example:

    echo "/u01/APPL/apps/apps_st/appl/xxtop/12.0.0/bin/file.xml" > filelist.txt
    sqlldr $1@$TWO_TASK control=$XXTOP/bin/LOAD_XML.ctl data=filelist.txt direct=true;
    
  • Sign Flip using the rule of load

    Hi Experts,

    I'm doing a sign flip using the rule of load. I know it's possible to do this but we must define a UDA. Is it possible to do it on my data field without using the UDA? Thank you.

    Kind regards
    Gerd

    If you want to return the entire file, you can use the scale factor for the data column and it set to-1. You can also set the file to subtract from existing values, but I prefer the first way

  • set the load of SQL to Essbase Cube

    Hi, I have a cube ASO which I load the data using SQL. The database user id and password changes every quarter. I wanted to make the load demand... so I want to know if theres any way I can set the user name and password for the SQL code in the file.so evry rule don't quarter I have no need to change user id and pswd evry area in all the rule files. something like variable substitution


    If you load planning, I assume you are using MaxL.   You can use parameters to pass the username password in the statement the data import and SQL.  You don't set them up in the rule of the load itself.

    You can't have them set up, encrypt to MaxL (that I know) but if you can get away with unencrypted passwords then it will work very well.

  • CRM using SQL database in an environment of small business often inaccessible with the following error message

    Hello

    This is my first time on this forum so bear with my ignorance!  I'll expose the background of my setup and then explain the error.

    We have a small business with a server running SBS2011 and PC running Windows 8 and 8.1 connected to the server in a domain environment.

    We use a CRM product called Emu Tall that uses SQL Server 2008 R2 on the server for the database and the PC run a local application that connects to this database.

    On a daily basis, but completely randomly and sometimes several times a day, a user will try and open the application on a PC and get the following error message.

    Could not load file or assembly 'Microsoft.SqlServer.SqlEnum, Version = 10.0.0.0, Culture = neutral, PublicKeyToken = 89845dcd8080cc91' or one of its dependencies. The system cannot find the specified file.

    At first, I thought it must be a matter of waiting time as it seemed to happen every morning, but I've proved since that it occurs at any time of the day and is apparently unrelated to amount of users or idle time or something like that.

    Here's another oddity that hopefully sheds some light on it. If I go to the server and open the CRM application on the server, it works very well since it is only the PC that have the connection problem.  In addition, as soon as I connected to the server of any PC can connect immediately (until it happens again later in the day).  I don't need to stay connected to the server, connect one time and close the session and he "fixed" it for a few hours for all PC users.

    I contacted the CRM vendor, and they are completely confused and have no idea what to do.

    We are trolling along at the moment with me opening a session on the server for launch launch whenever a user receives the error message but it is far from satisfactory.

    A final observation is that it does not always.  It used to work fine.  The problem began just two months ago and seems to coincide with the increase in our user from 5 to 12 users license even if the CRM provider swear that it is nothing to do with the license.  They may agree; I don't know enough to argue the point.

    Any help will be appreciated! I'm sure it will be a simple solution and someone there will know why and the answer immediately but is elusive at this time.

    This issue is beyond the scope of this site and must be placed on Technet or MSDN

    http://social.technet.Microsoft.com/forums/en-us/home

    http://social.msdn.Microsoft.com/forums/en-us/home

  • How can I load different CSS rules according to the browser?

    Basically I have background images set in place and when the page is scrolled on a desktop computer, the images are locked in place that their holders scroll upward and down, so I have a very simple parallax effect going on. Problem is that when I view on mobile devices and especially Apple iPhone, iPad, or iPod, the background is fixed yet, but it's huge!

    To after my research, background-attachment: fixed rule that is what creates the effect of simple Parallax seems to destroy the background-size rule: cover, simple force the image to view full size.

    So what I think is, to see what browser uses the user and through PHP or Javascript have a separate css file load with the rule background-attachment: fixed if the user is on a laptop Desk/computer or hide if they are on an apple device. I can't understand how do however. Is this possible?

    I don't want to use media queries as they are about to be unusable for this instance based on the screen size of a part of the device that are now producing people like Apple.

    I just your thoughts and or possible examples that you may have encountered.

    Thank you.

    Mobile Detect - Lightweight PHP class for detecting mobile devices (including tablets)

    Nancy O.

  • MySQL-&gt; table using SQL Developer Oracle

    Greetings,

    Currently, I am a (student) temporary employee who is responsible for the migration of the MySQL database of a company to their new (gr 11, 2) Oracle database.

    So far, I use SQL Developer, because I learned a few minor techniques with this application.

    I have tried different methods for the migration, including those of the Oracle (Migration of MySQL) page

    I however got no positive results with most of the methods.

    -J' have established a connection to the MySQL database Oracle database aswell, within SQL Developer

    Now, so far, I've copied manually (right click on the table, copy to Oracle) 90% of the tables from the MySQL to Oracle database successfully.

    However, these tables contain only about 5% of the data. The problem is in the 2 tables that actually have a decent amount of records.

    FOR INFO:

    File 1: 75MB file .csv, ~150.000 records (individual records contain all select queries, where the size)

    File 02:50 MB of files .csv, ~2.200.000 records

    My previous attempts to use the same copy for Oracle option to right click on one of the two largest tables resulted in the following error:

    The table of queries down. Message: java.io.IOException: IO error: socket read timed out

    My first action to prevent this error has been to add a second file data to the current storage space, giving it 200M maxsize.

    Okay it's coming, my real questions!

    This could be a possible solution to avoid the error mentioned earlier to return?

    What other issues are possible in case of failure?

    Uses an external table to a faster and more effective method?

    (The reason for which I have not yet tried external tables is because I had no SYS (TEM) privileges, so anything that requires privileges, I first have to discuss with my superior).

    Any help is appreciated! Although my next answer is maybe not earlier that come Monday 17.

    Welcome, Brent

    Hello

    I just discovered there is a csv Import Wizard in sqldeveloper also - don't know if you tried?

    tables in the destination database (oracle) - right-click and choose Import data - this will bring you in a Wizard of csv.

    However, I think that due to the volumes of data it may have similar problems.

    sqlldr is probably your best choice - it's pretty simple once you find a decent example - it seems to cover the basics of ok - Oracle SQL Stuff (for example): SQL-Loader: the tutorial step by step - example 1 (CSV file)

    See you soon,.

    Rich

  • Connection to SQL rules

    Hello

    Simple question, I managed to use SQL together with rules in Essbase file, but given that you are forced to

    SELECT < code here > < code here > WHERE < code here >

    Is anyway to order data, as this can be critical with metadata?

    Is anyway use UNION / UNION ALL / UNMIS etc, set operators in short?

    Thanks for your comments,

    Robert.

    Robert,

    You can place the select statement together in the first box (chic), just don't forget to remove the word "select" from your query. You can put in the group, of the order by, union, etc. without problem.  Just make sure you don't have anything in the other two boxes (to where).

  • Using sql developer

    Hello

    I have mysql on a server and oracle on other sever (both under linux). I want to migrate data from mysql to oracle using sql Developer installed on another remote computer. What are the neccessities base that I have to keep in mind in this regard, or should I just use the installed with oracle sql developer.

    I do not have the required display variables defined on the oracle server and also an older version of sqldeveloper is installed with oracle. I think installing the latest version of sql developer on windows platform and then want to access the database on linux machines.
    Please notify.

    Thanx

    11 g DBA wrote:
    Hello

    I have mysql on a server and oracle on other sever (both under linux). I want to migrate data from mysql to oracle using sql Developer installed on another remote computer. What are the neccessities base that I have to keep in mind in this regard, or should I just use the installed with oracle sql developer.

    I do not have the required display variables defined on the oracle server and also an older version of sqldeveloper is installed with oracle. I think installing the latest version of sql developer on windows platform and then want to access the database on linux machines.
    Please notify.

    Thanx

    Download the latest developer SQL on your PC. Enable your firewall rules accept connections from your PC to the database port (usually 1521 - Oracle, 3306 - MySQL).
    Establish a connection with your data (assuming that the privileges are sorted) and start the migration.
    Note: For MySQL, you will need to download third party connector (a quick Google and you'll find the download and installation instructions).

    You already have the URL for the migration.

Maybe you are looking for