Issue while loading data using the file Rules Essbase

Hi all

I am facing problem while loading data using the Rules file. In the rules file, I rejected several members in two areas (two dimensions). Now if I load the data using the rules file I'm getting errors for all members in the dataload.err file. If I reject mutiple members of a single field, the data load without settling errors in the dataload.err file.

I want to know how rmany members of several fields of ejection for loading data using the file Rules Essbase? Is it possible?

Okay, okay... I think that you must assign Global Select / reject Boolean in the parameters of loading data as 'Or':

Tags: Business Intelligence

Similar Questions

  • Unable to load data using the rules file

    Hi all

    I am facing a strange problem while loading data by using a rules file:

    1. I don't get any error/kick. It gives fair attention No. data values are modified by this data file loading. But nothing will load inside.

    2. I tried with Maxl and directly from EAS but gettting same warning.

    3 has tried to block sending n to entires in the data file and it does not work (so no Dynamics members etc..)

    4. rule file validates correctly.

    5. I tried with different types of different files (Csv, Xls, Txt, etc.) but no luck.

    Any suggestions / ideas of what could be bad?

    Don't know what exactly was the problem, but I recreated the rule files and it solved the problem... May be a case of corrupted file rule?

  • Can I use the file rule to load the metadata in the planning

    We are about to load data into the planning and was curious to know if we can create contour and loads the data in planning using Rulefiles (I know that we can in Essbase and want to make sure the planning)

    What could be the best option for creating contour and loading of metadata in the Planning.

    Thanks in advance

    No charge is just for essbase.
    You can use a rule to load to load directly into essbase for a planning application, but you can not for metadata.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Failed to load data using the contour load utility

    I am unable to load data using contour utility charge.
    Assigned data loading dimension such as: account
    Driver-Dimension such as: period
    Member as a pilot: Jan

    Connection file:

    Account, Jan, Point of view, the name of Cube data loading
    Investment, 1234, "India, current, heck, FY13", Plan1


    Utility of contour control
    OutlineLoad A:pract U:admin /I:C:\test1.csv /D:Account /L:C:\lg.log /X:C:\ex.exc


    Exception file

    [Thu Mar 28 18:05:39 GMT + 05:30 2013] Error loading of data record 1: investments, 1234, '' ' India, common, project, FY14' ' ', Plan1
    [Thu Mar 28 18:05:39 GMT + 05:30 2013] com.hyperion.planning.InvalidMemberException: Member India, common, rough, FY14 does not exist or you do not have access to it.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Planning of vector data store processes loaded with exceptions: exceptions have occurred, examine the file exception for more information. 1 data record has been read 1 record of data have been processed, 0 were loaded successfully, 1 was rejected.


    Log file:


    Connected application "Rlap", liberation 11.113, adapter Interface Version 5, supported and not enabled workforce, CapEx no taken care and not active, CSS Version 3
    [Thu Mar 28 18:05:38 GMT + 05:30 2013] Input file located and opened successfully "C:\load.csv".
    [Thu Mar 28 18:05:38 GMT + 05:30 2013] Record header fields: account, Jan, Point of view, the name of Cube data loading
    [Thu Mar 28 18:05:38 GMT + 05:30 2013] Located by using 'Account' dimension and for the loading of data into the application 'Rlap.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Loading dimension 'Account' has been successfully opened.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] A refresh of the cube operation will not be run.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Create filters for safe operation will not be performed.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Look at the files of newspapers of Essbase to status if Essbase data have been loaded.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Planning of vector data store processes loaded with exceptions: exceptions have occurred, examine the file exception for more information. 1 data record has been read 1 record of data have been processed, 0 were loaded successfully, 1 was rejected.



    Infact members exist in outline.
    Any help would be appreciated.

    can you double check your csv file, open it in a text editor, because the error is showing additional citations.

    [Thu Mar 28 18:05:39 GMT + 05:30 2013] Error loading of data record 1: investments, 1234, '' ' India, common, project, FY14' ' ', Plan1

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Time of loading images using the File.applicationDirectory or File.applicationStorageDirectory charger

    While I try to load several images with File.applicationDirectory charger or File.applicationStorageDirectory, I found something strange.

    It took about 2 times as long to load the File.applicationDirectory images of File.applicationStorageDirectory.

    I googled for a long time, but I found out why.

    Anyone know why?

    Thank you!!

    It is on Android by chance? If so, it makes sense. On Android, the assets of applicationDirectory is stored in the original format (zipped) .apk and file IO is abstract automatically for you so that you think that you access files on the file system, but that also do is extract the bytes of the first compressed content files.

    Same effect can be seen when you see the copy operations slower when copying a group of individual files one at a time of File.applicationDirectory instead of copying a file zip of applicationDirectory and expanding the zip in a different directory to the place. Mentioned in this discussion:

    File.copyTo (and copyToAsync) extremely slow on Android. for example 1/2 second by little XML file. Help?

  • Error in reading data using the file adapter... need help.

    Hi all

    We try to read a flat file delimited using file adapter.
    The following data works.
    abc|"sssss"|jjjj
    hhhj|fgfdg|dsfsf
    sfdf|dsf|adfd
    But when a space is present after the closing for the second column, that the record quotation mark is rejected

    Format below does not
    abc|"sssss"   |jjjj
    hhhj|fgfdg|dsfsf
    sfdf|dsf|adfd
    any help how to read such data... help please.

    Kind regards

    Published by: Oraacler on April 27, 2010 21:30
     
    
    
      
        
          
            
              
                
                  
                  
                  
                  
                  
                  
                
              
            
          
        
      
    
    
    
    
  • How to automate the process of loading data using load file & Task Scheduler

    Hello

    I do the automated processes to load the data into Hyperion Planning application using the file data_Load.bat & Scheduler of tasks.

    I created Data_Load.bat file, but the rest of the process, I cannot complete.

    Could help you me, how to automate the process of loading data using the file Data_load.bat & task Scheduler or what are the rest of the file is require it to achieve.

    Thank you

    In response to your question using the maxl for loading scripts?

    If Yes, I've seen and deliver in the batch (ex: load_data.bat) that is you do not have the path of the maxl script complete with a batch when passing through the event the task scheduler will work, but the log file and / or error will not be created. Which means lots claims it linked task scheduler, although he did not do what you need to.

    If you use maxl use this as the batch

    "essmsh C:\data\DataLoad.mxl" or you can also use the full path for the maxl or work elsewhere. The only reason why I think that the maxl can then not work is if you do not have the updated batch updated to call on all LANE changes maxl or if you need to update your environment variables to correct the command essmsh to work in a command prompt.

  • Load SQL data using the Regional service

    Hello

    I've been loading data using text file, now I want to try to load from SQL server directly, but screen load data from EAS, "Data Source" is grayed, as soon as I clicked on SQL as the Data Source.

    What I miss here?, I configure ODBC to the server level.

    First of all, don't you configured the source SQL and ODBC in State of charge? Open the rule of the load and the "select" menu-> open SQL. Set up the sql statement in the form, then click ok/collect to test.

    Once that the rule of the load seems good, so to actually load the data, show the daa load screen as you did. Select SQL. The data source must be a Virgin because it is for the flat file. Select the rule of load (it has the SQL in it). If you look to the right of the screen, and you may need to expand to see, rather than enter the sqlID and password there. Enter those, and click ok. It should load as usual.

    If you want to leave MaxL, there is a syntax similar to tell the import statement is a sql load and give you it the id and the password

  • Automate the process of loading data using Task Scheduler

    Hello

    I do the automated processes to load the data into Hyperion Planning application with the help of the Task Scheduler.

    I created Data_Load.bat file, but the rest of the process, I cannot complete.

    So could you help me, how to automate the loading of the process data using the file Data_load.bat & Task Scheduler.

    Thank you

    Windows 2008? Open the Task Scheduler, Action > creates a fundamental task, give it a name, select the frequency, select start a program, find the batch script, then you're pretty much done for a basic task.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Problems loading data to the Essbase.

    Hi all

    I am trying to load data to essbase and I get the following error. He said from an unknown source. But when I do the reverse, that it works fine without any errors.
    I do a simple load data from the file... Please help me.


    org.apache.bsf.BSFException: exception of Jython:
    Traceback (innermost last):

    "< String >" file, line 23, in there?

    com.hyperion.odi.essbase.ODIEssbaseException: error reached records the maximum error threshold: 1


    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData (unknown Source)


    at sun.reflect.GeneratedMethodAccessor100.invoke (unknown Source)


    at sun.reflect.DelegatingMethodAccessorImpl.invoke (unknown Source)


    at java.lang.reflect.Method.invoke (unknown Source)


    at org.python.core.PyReflectedFunction.__call__ (PyReflectedFunction.java)


    at org.python.core.PyMethod.__call__ (PyMethod.java)


    at org.python.core.PyObject.__call__ (PyObject.java)


    at org.python.core.PyInstance.invoke (PyInstance.java)


    to org.python.pycode._pyx125.f$ 0 (< string >: 23)


    to org.python.pycode._pyx125.call_function (< string >)


    at org.python.core.PyTableCode.call (PyTableCode.java)


    at org.python.core.PyCode.call (PyCode.java)


    at org.python.core.Py.runCode (Py.java)


    at org.python.core.Py.exec (Py.java)


    at org.python.util.PythonInterpreter.exec (PythonInterpreter.java)


    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)


    at com.sunopsis.dwg.codeinterpretor.k.a (k.java)


    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting (SnpSessTaskSql.java)


    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)


    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)


    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt (SnpSessTaskSql.java)


    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt (SnpSessTaskSqlI.java)


    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask (SnpSessTaskSql.java)


    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep (SnpSessStep.java)


    at com.sunopsis.dwg.dbobj.SnpSession.treatSession (SnpSession.java)


    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand (DwgCommandSession.java)


    at com.sunopsis.dwg.cmd.DwgCommandBase.execute (DwgCommandBase.java)


    at com.sunopsis.dwg.cmd.e.j (e.java)


    at com.sunopsis.dwg.cmd.h.z (h.java)


    at com.sunopsis.dwg.cmd.e.run (e.java)


    at java.lang.Thread.run (unknown Source)


    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: error reached records the maximum error threshold: 1


    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.checkMaxError (unknown Source)


    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.sendRecordArrayToEsbase (unknown Source)


    ... 31 more




    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: error reached records the maximum error threshold: 1



    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)

    at com.sunopsis.dwg.codeinterpretor.k.a (k.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt (SnpSessTaskSqlI.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep (SnpSessStep.java)

    at com.sunopsis.dwg.dbobj.SnpSession.treatSession (SnpSession.java)

    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand (DwgCommandSession.java)

    at com.sunopsis.dwg.cmd.DwgCommandBase.execute (DwgCommandBase.java)

    at com.sunopsis.dwg.cmd.e.j (e.java)

    at com.sunopsis.dwg.cmd.h.z (h.java)

    at com.sunopsis.dwg.cmd.e.run (e.java)

    at java.lang.Thread.run (unknown Source)


    Thank you
    K

    Hello

    Well just use the 10.1.3.5_2 version as it looks like this is the right one.

    Ok?

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Error while loading data file with using a file of rules through a MAXL.

    I think that the functionality to generate a. Records error during loading data in the ERR file is supported only if there is a. RUL file being used.

    Is this good?

    I tried to get. ERR files to generate by using the following statement:
    Import database PL_RPT. Reprting data data_file
    "E:\Hyperion\AnalyticServices\APP\PL\PL.txt" to load_buffer with buffer_id 17
    Error writing to "E:\Hyperion\Scripts\Pln\Logs\LoadlData.err."

    When I run this, if there are errors, it is not generated any file errors and stops the load.

    I saw the technical reference Essbase ASO MAXL data loading and code syntax diagram indicates that it is supported.

    Any suggestions will be greatly appreciated.

    Thank you

    Hello

    Here are a few suggestions for trapping errors. I hope that one of them will meet your needs:

    1._____________________________________________

    spool to 'D:\logs\maxlresults.out ';

    function SIERREUR 'WRITE_ERRORS ';

    / * do stuff * /.

    Define the label 'WRITE_ERRORS ';
    spool off;
    spool to 'D:\logs\maxlerrors.out ';
    "exit";

    2._____________________________________________

    coil stdout to "D:\logs\maxlresults.out."
    coil stderr to "D:\logs\maxlerrors.out."

    3._____________________________________________

    essmsh script.msh 2 > D:\logs\maxlresults.out

    Robb

  • How to load data from ms sql to the by the file rules and maxl essbase

    Hi, everybody!
    Pretty Shure the kind of topic exists already on the forum, but unfortunately can't find it.

    I want to understand, how to load data from the database to ms sql for the PB of essbase application.

    (I) so I have:
    1. a request for 'society '.
    2 and its database 'Plan '.
    3. with simple outline, which contains only two dimensions:

    -The time period < 1 > (Label only)
    -Total year (~)
    -Qtr1 (+)
    Jan (+)
    Feb (+)
    Mar (+)
    -Qtr2 (+)
    APR (+)
    May (+)
    Jun (+)

    -Accounts < 1 > (Label only)
    -Lost / based (~) (Label only)
    -Revenues (+)
    L1.1 (+)
    L1.2 (+)
    -Costs (-)
    L2.1 (+)
    L2.2 (+)

    (II) also, I created a rules file called "CO_DWH" and associated with this schema
    It (Rules file) a 3 columns:
    Data of the 'period', 'account '.

    There is also the option checked "load data".

    (III) in MS SQL Server, I have a database of "hypdb" and "essadmin" with "password" login

    (IV) I executed my bat file:
    C:\Hyperion\EssLoad.bat
    There is only one line:

    EssLoad.bat-
    startMaxl.cmd EssLoad.txt
    ----------------------------------------------

    EssLoad.txt-
    login password admin on erpserver;
    Import database data Company.Plan connect in essadmin password using the server rules_file 'CO_DWH' error abort;
    disconnection;
    "exit";
    --------------------------------------------


    All plans that I copied well worked ERP system, but I don't understand, this is exactly table in the MS SQL database, data loading to essbase db?

    TOU have to do a few things
    1. on the server of Essbase, put in place a system odbc for your MySQL database connection
    2. in the State of charge, go to the file menu and select open source SQL database and in the source of data in your SQL statement. A few tips. Where it says select, don't put in the select Word and when he says to not put in of. The system adds for you. TI enter a simple SQL statement is to do this in the box so select if your SQL would be normanlly as select * from myTable just enter the code as * from myTable in the upscale neighborhood

    The ol/recover the click and fill your connection information. Should bring back the data in the State of charge. Save the State of charge and use

  • Question to load data using sql loader in staging table, and then in the main tables!

    Hello

    I'm trying to load data into our main database table using SQL LOADER. data will be provided in separate pipes csv files.

    I have develop a shell script to load the data and it works fine except one thing.

    Here are the details of a data to re-create the problem.

    Staging of the structure of the table in which data will be filled using sql loader

    create table stg_cmts_data (cmts_token varchar2 (30), CMTS_IP varchar2 (20));

    create table stg_link_data (dhcp_token varchar2 (30), cmts_to_add varchar2 (200));

    create table stg_dhcp_data (dhcp_token varchar2 (30), DHCP_IP varchar2 (20));

    DATA in the csv file-

    for stg_cmts_data-

    cmts_map_03092015_1.csv

    WNLB-CMTS-01-1. 10.15.0.1

    WNLB-CMTS-02-2 | 10.15.16.1

    WNLB-CMTS-03-3. 10.15.48.1

    WNLB-CMTS-04-4. 10.15.80.1

    WNLB-CMTS-05-5. 10.15.96.1

    for stg_dhcp_data-

    dhcp_map_03092015_1.csv

    DHCP-1-1-1. 10.25.23.10, 25.26.14.01

    DHCP-1-1-2. 56.25.111.25, 100.25.2.01

    DHCP-1-1-3. 25.255.3.01, 89.20.147.258

    DHCP-1-1-4. 10.25.26.36, 200.32.58.69

    DHCP-1-1-5 | 80.25.47.369, 60.258.14.10

    for stg_link_data

    cmts_dhcp_link_map_0309151623_1.csv

    DHCP-1-1-1. WNLB-CMTS-01-1,WNLB-CMTS-02-2

    DHCP-1-1-2. WNLB-CMTS-03-3,WNLB-CMTS-04-4,WNLB-CMTS-05-5

    DHCP-1-1-3. WNLB-CMTS-01-1

    DHCP-1-1-4. WNLB-CMTS-05-8,WNLB-CMTS-05-6,WNLB-CMTS-05-0,WNLB-CMTS-03-3

    DHCP-1-1-5 | WNLB-CMTS-02-2,WNLB-CMTS-04-4,WNLB-CMTS-05-7

    WNLB-DHCP-1-13 | WNLB-CMTS-02-2

    Now, after loading these data in the staging of table I have to fill the main database table

    create table subntwk (subntwk_nm varchar2 (20), subntwk_ip varchar2 (30));

    create table link (link_nm varchar2 (50));

    SQL scripts that I created to load data is like.

    coil load_cmts.log

    Set serveroutput on

    DECLARE

    CURSOR c_stg_cmts IS SELECT *.

    OF stg_cmts_data;

    TYPE t_stg_cmts IS TABLE OF stg_cmts_data % ROWTYPE INDEX BY pls_integer;

    l_stg_cmts t_stg_cmts;

    l_cmts_cnt NUMBER;

    l_cnt NUMBER;

    NUMBER of l_cnt_1;

    BEGIN

    OPEN c_stg_cmts.

    Get the c_stg_cmts COLLECT in BULK IN l_stg_cmts;

    BECAUSE me IN l_stg_cmts. FIRST... l_stg_cmts. LAST

    LOOP

    SELECT COUNT (1)

    IN l_cmts_cnt

    OF subntwk

    WHERE subntwk_nm = l_stg_cmts (i) .cmts_token;

    IF l_cmts_cnt < 1 THEN

    INSERT

    IN SUBNTWK

    (

    subntwk_nm

    )

    VALUES

    (

    l_stg_cmts (i) .cmts_token

    );

    DBMS_OUTPUT. Put_line ("token has been added: ' |") l_stg_cmts (i) .cmts_token);

    ON THE OTHER

    DBMS_OUTPUT. Put_line ("token is already present'");

    END IF;

    WHEN l_stg_cmts EXIT. COUNT = 0;

    END LOOP;

    commit;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    for dhcp


    coil load_dhcp.log

    Set serveroutput on

    DECLARE

    CURSOR c_stg_dhcp IS SELECT *.

    OF stg_dhcp_data;

    TYPE t_stg_dhcp IS TABLE OF stg_dhcp_data % ROWTYPE INDEX BY pls_integer;

    l_stg_dhcp t_stg_dhcp;

    l_dhcp_cnt NUMBER;

    l_cnt NUMBER;

    NUMBER of l_cnt_1;

    BEGIN

    OPEN c_stg_dhcp.

    Get the c_stg_dhcp COLLECT in BULK IN l_stg_dhcp;

    BECAUSE me IN l_stg_dhcp. FIRST... l_stg_dhcp. LAST

    LOOP

    SELECT COUNT (1)

    IN l_dhcp_cnt

    OF subntwk

    WHERE subntwk_nm = l_stg_dhcp (i) .dhcp_token;

    IF l_dhcp_cnt < 1 THEN

    INSERT

    IN SUBNTWK

    (

    subntwk_nm

    )

    VALUES

    (

    l_stg_dhcp (i) .dhcp_token

    );

    DBMS_OUTPUT. Put_line ("token has been added: ' |") l_stg_dhcp (i) .dhcp_token);

    ON THE OTHER

    DBMS_OUTPUT. Put_line ("token is already present'");

    END IF;

    WHEN l_stg_dhcp EXIT. COUNT = 0;

    END LOOP;

    commit;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    for link -.

    coil load_link.log

    Set serveroutput on

    DECLARE

    l_cmts_1 VARCHAR2 (4000 CHAR);

    l_cmts_add VARCHAR2 (200 CHAR);

    l_dhcp_cnt NUMBER;

    l_cmts_cnt NUMBER;

    l_link_cnt NUMBER;

    l_add_link_nm VARCHAR2 (200 CHAR);

    BEGIN

    FOR (IN) r

    SELECT dhcp_token, cmts_to_add | ',' cmts_add

    OF stg_link_data

    )

    LOOP

    l_cmts_1: = r.cmts_add;

    l_cmts_add: = TRIM (SUBSTR (l_cmts_1, 1, INSTR (l_cmts_1, ',') - 1));

    SELECT COUNT (1)

    IN l_dhcp_cnt

    OF subntwk

    WHERE subntwk_nm = r.dhcp_token;

    IF l_dhcp_cnt = 0 THEN

    DBMS_OUTPUT. Put_line ("device not found: ' |") r.dhcp_token);

    ON THE OTHER

    While l_cmts_add IS NOT NULL

    LOOP

    l_add_link_nm: = r.dhcp_token |' _TO_' | l_cmts_add;

    SELECT COUNT (1)

    IN l_cmts_cnt

    OF subntwk

    WHERE subntwk_nm = TRIM (l_cmts_add);

    SELECT COUNT (1)

    IN l_link_cnt

    LINK

    WHERE link_nm = l_add_link_nm;

    IF l_cmts_cnt > 0 AND l_link_cnt = 0 THEN

    INSERT INTO link (link_nm)

    VALUES (l_add_link_nm);

    DBMS_OUTPUT. Put_line (l_add_link_nm |) » '||' Has been added. ") ;

    ELSIF l_link_cnt > 0 THEN

    DBMS_OUTPUT. Put_line (' link is already present: ' | l_add_link_nm);

    ELSIF l_cmts_cnt = 0 then

    DBMS_OUTPUT. Put_line (' no. CMTS FOUND for device to create the link: ' | l_cmts_add);

    END IF;

    l_cmts_1: = TRIM (SUBSTR (l_cmts_1, INSTR (l_cmts_1, ',') + 1));

    l_cmts_add: = TRIM (SUBSTR (l_cmts_1, 1, INSTR (l_cmts_1, ',') - 1));

    END LOOP;

    END IF;

    END LOOP;

    COMMIT;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    control files -

    DOWNLOAD THE DATA

    INFILE 'cmts_data.csv '.

    ADD

    IN THE STG_CMTS_DATA TABLE

    When (cmts_token! = ") AND (cmts_token! = 'NULL') AND (cmts_token! = 'null')

    and (cmts_ip! = ") AND (cmts_ip! = 'NULL') AND (cmts_ip! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:cmts_token))' cmts_token,

    cmts_ip ' RTRIM (LTRIM(:cmts_ip)) ")". "

    for dhcp.


    DOWNLOAD THE DATA

    INFILE 'dhcp_data.csv '.

    ADD

    IN THE STG_DHCP_DATA TABLE

    When (dhcp_token! = ") AND (dhcp_token! = 'NULL') AND (dhcp_token! = 'null')

    and (dhcp_ip! = ") AND (dhcp_ip! = 'NULL') AND (dhcp_ip! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:dhcp_token))' dhcp_token,

    dhcp_ip ' RTRIM (LTRIM(:dhcp_ip)) ")". "

    for link -.

    DOWNLOAD THE DATA

    INFILE 'link_data.csv '.

    ADD

    IN THE STG_LINK_DATA TABLE

    When (dhcp_token! = ") AND (dhcp_token! = 'NULL') AND (dhcp_token! = 'null')

    and (cmts_to_add! = ") AND (cmts_to_add! = 'NULL') AND (cmts_to_add! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:dhcp_token))' dhcp_token,

    cmts_to_add TANK (4000) RTRIM (LTRIM(:cmts_to_add)) ")" ""

    SHELL SCRIPT-

    If [!-d / log]

    then

    Mkdir log

    FI

    If [!-d / finished]

    then

    mkdir makes

    FI

    If [!-d / bad]

    then

    bad mkdir

    FI

    nohup time sqlldr username/password@SID CONTROL = load_cmts_data.ctl LOG = log/ldr_cmts_data.log = log/ldr_cmts_data.bad DISCARD log/ldr_cmts_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    nohup time username/password@SID @load_cmts.sql

    nohup time sqlldr username/password@SID CONTROL = load_dhcp_data.ctl LOG = log/ldr_dhcp_data.log = log/ldr_dhcp_data.bad DISCARD log/ldr_dhcp_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    time nohup sqlplus username/password@SID @load_dhcp.sql

    nohup time sqlldr username/password@SID CONTROL = load_link_data.ctl LOG = log/ldr_link_data.log = log/ldr_link_data.bad DISCARD log/ldr_link_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    time nohup sqlplus username/password@SID @load_link.sql

    MV *.log. / log

    If the problem I encounter is here for loading data in the link table that I check if DHCP is present in the subntwk table, then continue to another mistake of the newspaper. If CMTS then left create link to another error in the newspaper.

    Now that we can here multiple CMTS are associated with unique DHCP.

    So here in the table links to create the link, but for the last iteration of the loop, where I get separated by commas separate CMTS table stg_link_data it gives me log as not found CMTS.

    for example

    DHCP-1-1-1. WNLB-CMTS-01-1,WNLB-CMTS-02-2

    Here, I guess to link the dhcp-1-1-1 with balancing-CMTS-01-1 and wnlb-CMTS-02-2

    Theses all the data present in the subntwk table, but still it gives me journal wnlb-CMTS-02-2 could not be FOUND, but we have already loaded into the subntwk table.

    same thing is happening with all the CMTS table stg_link_data who are in the last (I think here you got what I'm trying to explain).

    But when I run the SQL scripts in the SQL Developer separately then it inserts all valid links in the table of links.

    Here, she should create 9 lines in the table of links, whereas now he creates only 5 rows.

    I use COMMIT in my script also but it only does not help me.

    Run these scripts in your machine let me know if you also get the same behavior I get.

    and please give me a solution I tried many thing from yesterday, but it's always the same.

    It is the table of link log

    link is already present: dhcp-1-1-1_TO_wnlb-cmts-01-1

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-02-2

    link is already present: dhcp-1-1-2_TO_wnlb-cmts-03-3
    link is already present: dhcp-1-1-2_TO_wnlb-cmts-04-4

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-5

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-01-1

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-8
    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-6
    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-0

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-03-3

    link is already present: dhcp-1-1-5_TO_wnlb-cmts-02-2
    link is already present: dhcp-1-1-5_TO_wnlb-cmts-04-4

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-7

    Device not found: wnlb-dhcp-1-13

    IF NEED MORE INFORMATION PLEASE LET ME KNOW

    Thank you

    I felt later in the night that during the loading in the staging table using UNIX machine he created the new line for each line. That is why the last CMTS is not found, for this I use the UNIX 2 BACK conversion and it starts to work perfectly.

    It was the dos2unix error!

    Thank you all for your interest and I may learn new things, as I have almost 10 months of experience in (PLSQL, SQL)

  • How to use the date of the file, when you download a file?

    Using the basics of 12 c, APEX 5.0 and IE11

    I am creating a file upload feature for my application.  I'm able to get the upload├⌐ file in my table, but one of my requirements is that I fill in a value in the table with the "file date".

    Is there an easy way to get the "file date" in the file being downloaded and fill a field for all records with this date?

    Tillie

    Tillie says:

    Using the basics of 12 c, APEX 5.0 and IE11

    I am creating a file upload feature for my application.  I'm able to get the upload├⌐ file in my table, but one of my requirements is that I fill in a value in the table with the "file date".

    Is there an easy way to get the "file date" in the file being downloaded and fill a field for all records with this date?

    That depends on your definition of 'easy '...

    The HTML5 File interface can be used to access date in browsers supportthe file was last modified.

    1. Add a hidden page that contains the file download point (P6_FILE_DATE in this example).

    2. create a dynamic action of change on the file upload question (P6_FILE):

    Event: Change

    Selection type: Agenda

    Region:

    Real Actions

    Action: Run the JavaScript Code

    Fire on Page load: NO.

    Code

    console.log('Get file modification date:')
    var fileItem = $x('P6_FILE');
    if (window.File) {
      if (fileItem.files.length > 0) {
        var file = fileItem.files[0];
        if (typeof file.lastModifiedDate !== 'undefined') {
          console.log('...lastModifiedDate for file \"' + file.name + '\": ' + file.lastModifiedDate);
          apex.item('P6_FILE_DATE').setValue(file.lastModifiedDate);
        }
        else {
          console.log('...browser does not support the HTML5 File lastModifiedDate property.');
        }
      }
    }
    else {
      console.log('...browser does not support the HTML5 File interface.');
    }
    

    If the browser has support for the File interface and lastModifiedDate property, the date, the file was last modified will be presented in the P6_FILE_DATE article in the form:

    Fri Jul 17 2015 10:43:44 GMT+0100 (GMT Daylight Time)

    You can analyze and convert your data load process to obtain a TIMESTAMP or DATE value according to the needs.

  • An error in loading data from .csv file using sql * loader

    Hello


    I wrote a control file, to load data into the table.
    The .csv (input file) contains a value that contains within it. So that the value is placed in "" (double quotes)

    The record that contains the value that is quoted is not loaded into the file.

    example: -.
    input file (.csv file)
    001, apple, 1, 31 December 00
    002, "abc, def & ltd", December 1, 31 00

    in the example above second row is not inserted into the table.

    How can I insert all the rows in the table.
    any help is appreciated.

    Thanks in advance...
    RXG

    looks like your control file?

    Try this.
    fields ended by ',' optionally surrounded "" "

Maybe you are looking for