Loading data, including data Smart Lists using ODI planning KM

Hi all

a question about the ODI SmartData lists loading...

Can someone confirm if we can load the lists of data from a source file that has the chip labels, instead of the id values of active lists (ex: YES instead of 1)? I guess planning will make the conversion during the loading of essbase

Moreover, load contour utility works the same way?

Thanks in advance

I don't think it's possible with the KM in its current state that you always update driver members in planning or update the HSP_DRIVER_DIMENSION table (can't rememeber if this table is cached and run it all changes require a restart of planning)

See you soon

John
http://John-Goodwin.blogspot.com/

Tags: Business Intelligence

Similar Questions

  • Cannot load data into Essbase using ODI

    Hi guys,.

    Help help. I have problem loading data into essbase using ODI. The error message is
    java.sql.SQLException: unexpected token: ACCOUNT in the statement [select C1_ACCOUNT "" account]

    I have a very simple flat file that are similar to the below:

    Account, resources, time, data
    Active, Na_Resource, Jan, 10
    Active, Na_Resource, 12, February

    With the same flat files, I am able to load data to load rules.


    I use 9.3.1.0 and ODI 10.1.3.4.0 essbase. I use the ODI to load members and data in the planning without any problem.


    Thank you

    Hello

    It seems to generate an extra set of quotation marks around the SQL, in my interface it generates.

    SQL = "" "select C1_ACCOUNT 'Account', C2_PERIOD 'Period', C3_RESOURCE 'Resource', C4_DATA 'Data' of the" C$ _0TestApp_testData "where (1 = 1) «»

    Note the single quotes around the account.

    If you go to the topology Manager, on the tab of the physical architecture, right-click 'Hyperion Planning' > 'change '.
    Select the "Langugage" tab for the "JYTHON" line, make sure that the "Object Delimiter" field has no quotes, if it's remove and apply and save.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Loading data from text using ODI in planning

    Hi John,.

    Hope you do well...

    In your blog, more to life than that...: ODI - get the textual data in the planning of the, you mentioned to restart services to update the textual data, once it is loaded into the Cell_Text table. Is it always applies to the current version of planning. I have a requirement to load text data into Workforce Planning.

    Thank you

    Sravan

    Yes, you put combinations of members in the POV, it's the same concept that when you use the outlineload utility.

    See you soon

    John

    http://John-Goodwin.blogspot.com/

  • ODI 11 g to load data to Hyperion - error ODI-1228

    Hi friends,

    I'm using ODI 11 g to load the data for Hyperion Essbase using the "memory Sunopsis Staging Area table engine', I several created interfaces, but some have failed at the third stage of the process '3 - Carga - SrcSet0 - Load data'.

    Number of interfaces I created worked without problems by using the same configuration, topology, and KM, I thought that the problem was created by the file this extraction I´m. Ce file failed in other interfaces, but if I dupplicate the last line of the file (copy and paste at the end) interfaces for the file is complete.

    After the first appearance of the error, I created other interfaces, most of them works without problem, but a couple of them generates exactly the same error as one explained using different files.

    I use:

    ODI 11 g 11.1.1.5.2.

    IKM: IKM: IKM SQLto Hyperion Essbase (DATA).

    HSQLDB: Version 2.0.0.

    Essbase: 11.1.2.1.

    Source of the cube: BSO.

    The error that generated the ODI is:

    ODI-1228: Fallo en SrcSet0 (Carga) en el destino SUNOPSIS_ENGINE conexion SUNOPSIS_MEMORY_ENGINE task.

    Caused by: java.sql.SQLException: sentencia SQL don't get runs en modo 'batch '.

    at org.hsqldb.jdbc.Util.sqlException (unknown Source)

    at org.hsqldb.jdbc.Util.sqlException (unknown Source)

    at org.hsqldb.jdbc.Util.sqlExceptionSQL (unknown Source)

    at org.hsqldb.jdbc.JDBCPreparedStatement.executeBatch (unknown Source)

    at oracle.odi.runtime.agent.execution.sql.SQLCommand.end(SQLCommand.java:267)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.endExecution(SQLExecutor.java:156)

    at oracle.odi.runtime.agent.execution.sql.SQLExecutor.endExecution(SQLExecutor.java:1)

    at oracle.odi.runtime.agent.execution.DataMovementTaskExecutionHandler.handleTask(DataMovementTaskExecutionHandler.java:113)

    ...

    ...

    ...

    I hope that you can guide me to find a solution.

    Thanks for all, best regards.

    I had the same problem in a seemingly random way. I could not find any resolution so the safest option is to choose a different staging technology.

  • loading data to essbase using EAS against back-end script

    Good afternoon

    We have noticed recently that when loading of our ASO cube (with the help of back-end scripts - esscmd etc.) it seems to load much more overhead then when using EAS and loading files of data individually.  When loading using scripts, the total size of the cube has been 1.2 gig.  When you load the files individually to EAS, the size of the cube has been 800 meg.  Why the difference?  is there anything we can do in scripts supported to reduce this burden?

    Thank you

    You are really using EssCmd to load the ASO cubes?  You should use MAxL with buffers to load. Default EAS uses a buffer to load when you load multiple files. Esscmd (and without the command buffer MAxL) won't. It means loads long and larger files. Loading ASO, it takes the existing .dat file and adds the new data. When you are not using a buffer load, it takes the .dat file and the .tmp file and merges together, then when you do the second file, it takes the files .dat (which includes the first load data) and repeats the process. Whenever he does that he has to move together (twice) .dat file and there is growth of the .dat files. If nothing else I'll call it fragmentation, but I don't think it's as simple as that. I think it's just the way the data is stored. When you use a buffer and no slices, he need only do this once.

  • Loading data to Essbase with ODI rules file

    Hi all

    I need to load the flat file DATA (no metadata) or an Oracle table in to ESSBASE using ODI.

    Where can I find an example or a tutorial on the loading procedure.
    I have not have how to use ODI with a rules file:

    (1) in the standard rule file I can insert a transformation (join, split ecc.) can be used
    a rule like this in ODI file?

    (2) if I put the file of IKM SQL for Hyperion Essbase (DATA) rules I have to manually set the mapping
    between the source and target? the mapping is not set in the rule file?

    (3) can I use both the type of rules (flat file or SQL) file?


    Thanks a lot for your help!

    Edited by: 882454 on 19-game-2011 2.57

    Edited by: 882454 on 2.58 19-game-2011

    If you go to the next page in my blog, you should see a link to all the different posts around ODI and hyperion knowledge modules - http://john-goodwin.blogspot.com/2011/08/odi-series-summary.html

    In ODI, you can do all the transformations, the rule of the load is really used to optimize the process of loading the data.
    You must configure your interface with source and target mapping, and then create a rule to load that corresponds to the mapping of the target.
    You can use any type of source.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Error while loading data file with using a file of rules through a MAXL.

    I think that the functionality to generate a. Records error during loading data in the ERR file is supported only if there is a. RUL file being used.

    Is this good?

    I tried to get. ERR files to generate by using the following statement:
    Import database PL_RPT. Reprting data data_file
    "E:\Hyperion\AnalyticServices\APP\PL\PL.txt" to load_buffer with buffer_id 17
    Error writing to "E:\Hyperion\Scripts\Pln\Logs\LoadlData.err."

    When I run this, if there are errors, it is not generated any file errors and stops the load.

    I saw the technical reference Essbase ASO MAXL data loading and code syntax diagram indicates that it is supported.

    Any suggestions will be greatly appreciated.

    Thank you

    Hello

    Here are a few suggestions for trapping errors. I hope that one of them will meet your needs:

    1._____________________________________________

    spool to 'D:\logs\maxlresults.out ';

    function SIERREUR 'WRITE_ERRORS ';

    / * do stuff * /.

    Define the label 'WRITE_ERRORS ';
    spool off;
    spool to 'D:\logs\maxlerrors.out ';
    "exit";

    2._____________________________________________

    coil stdout to "D:\logs\maxlresults.out."
    coil stderr to "D:\logs\maxlerrors.out."

    3._____________________________________________

    essmsh script.msh 2 > D:\logs\maxlresults.out

    Robb

  • Load data into Essbase using the rules file

    Hello

    IAM trying to create a rule to load for the fresh Essbase cube associated with the Hyperion Planning Application.We having 14 Dimensions (including the HSP_Rates Dimension).

    The falt file used to generate the rule file has columns related to the 13 dimensions as well as the data as the 14th column column. When I validate the rules file, he said "the rules file is correct for the loading of the data.

    My concern here is iam not taking care of Member Dimnesion related to the Dimension of "HSP_Rates" (not even as the Difinition of the header). I'm doing something wrong here? How validation can MOVE independently without taking into account one dimension?

    If I insert the data, while recovering the data in Excel which dimension member HSP_Rates should I use?

    Any response is greatly appreciated.

    Thank you
    RNG

    Hello

    You should charge against 'HSP_InputValue' in the HSP_Rates dimension, i.e. where the input values must be loaded.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • EAS load data and contour review

    Dear all,

    I am currently using Hyperion planning 11.1.2.2, I created an application with appropriate (see below) outline

    HSP_Rates/account/period/year/scenario/Version/currency/entity/brand/product/customer/measure

    and created the corresponding tables in essbase I want to load the data via Eas, I created a file to load data (a table) with exactly the same column names as my plan dimensions names EXCEPT HSP_Rates which is absent in my data file. I'm sure all the dimension members are correct / exist and there is no empty fields.

    When loading the file, the following error message appears:
    Value data [2056] met before that all selected Dimensions, complete Records [2]
    Unexpected error Essbase 1003007

    Question: I need to add a column named HSP_Rates in my data file? should what value I put under this column?


    PS: workload data file is an Excel file

    Thanks in advance for your help.
    Best regards.

    the problem is resolved:

    Load data into Essbase using the rules file

    the answer is Yes, create a column named HSP_Rates with 'HSP_InputValue' as values in the data file.

    Thank you
    you

  • Loading data into Essbase directly

    Hi all

    I have a question about loading data into Essbase, rather than planning.
    Essbase I found another dimension called HSP_Rates. How can I adjust this dimension when downloading to essbase directly?
    Or the question from another perspective: when writing data to a form in the planning, to which Member of the HSP_Rates is she responsible?

    I hope you got the point that I'm ugly questions :D
    Your help is very appreciated.

    BR
    ioGGo.

    Hello

    This link should provide a little more information:-http://download.oracle.com/docs/cd/E12825_01/epm.111/hp_admin/ch10s04s19s02.html

    What data loading loading essentially it against Hsp_InputValue

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • How to load data from matrix report in the base using ODI table data

    Hello

    How to load matrix report data in the base table data using oracle Data Integrator?

    Description of the requirement:

    This is the data from matrix report:
    JOB                       DEPT10                DEPT20  
    ___________________________ _____________
    ANALYST                                           6000
    CLERK                   1300                     1900 
    Need to convert it to the format below:
    JOB                             Dept                        Salary
    _____________________________________________
    ANALYST                  DEPT10      
    ANALYST                  DEPT20                     6000
    CLERK                       DEPT10                    1300
    CLERK                       DEPT20                    1900
        
    Thank you for your help in advance. Let me know if any other explanation is needed.

    Your list seems to be a bit restrictive, you can do much more with the procedures of ODI.

    If you create the new procedure and add a step. In the 'source' tab command you define technology and pattern according to your source database. Use the unpivot operator as described in the link, please, instead of using "SELECT *' use the column names and aliases for example:"

    SELECT workstation,
    deptsal as deptsal,
    saldesc as saledesc
    OF pivoted_data
    UNPIVOT)
    deptsal-<-->
    FOR saldesc-<-->
    IN (d10_sal, d20_sal, d30_sal, d40_sal).<-->
    )

    Then in your tab 'command on target' defined technology and drawing on your target db, then put your INSERT statement for example:

    INSERT INTO job_sales
    (employment,
    deptsal,
    saledesc
    )
    VALUES
    (
    : job,.
    : deptsal,.
    : saledesc
    )

    That's why you use bind variables from source to load data into the target.

    Obviously if the source and target table is in the same database, you can have it all in a single statement to the "command on target' as

    INSERT INTO job_sales
    (employment,
    deptsal,
    saledesc
    )
    SELECT workstation,
    deptsal as deptsal,
    saldesc as saledesc
    OF pivoted_data
    UNPIVOT)
    deptsal-<-->
    FOR saldesc-<-->
    IN (d10_sal, d20_sal, d30_sal, d40_sal).<-->
    )

    also assign the log count "Insert" on the tab corresponding to your INSERT statement, so that you know how many rows you insert into the table.

    I hope this helps.

    BUT remember that this feature is out in Oracle 11 g.

  • SOS! -Error to load data from Oracle 11 g to Essbase using ODI

    Hi all.

    I want to load data from oracle database to essbase using ODI.
    I have set up correctly the Hyperion essbase physical and logical topology manager and got the structure ESSBASE BASIC app DEMO.
    The problem is.
    1. when I try see data right click on the table of essbase.
    va.sql.SQLException: driver must be specified
    at com.sunopsis.sql.SnpsConnection.a (SnpsConnection.java)
    at com.sunopsis.sql.SnpsConnection.testConnection (SnpsConnection.java)
    at com.sunopsis.sql.SnpsConnection.testConnection (SnpsConnection.java)
    at com.sunopsis.graphical.frame.b.jc.bE (jc.java)
    at com.sunopsis.graphical.frame.bo.bA (bo.java)
    at com.sunopsis.graphical.frame.b.ja.dl (ja.java)
    to com.sunopsis.graphical.frame.b.ja. < init > (ja.java)
    to com.sunopsis.graphical.frame.b.jc. < init > (jc.java)
    I got the answer of partisan Oracle it's ok, just omit it. The second problem will appear.
    2 create an interface between the oracle and essbase database, click on the option ' staging of deffirent from the area of the "(ce qui signifie que la mise en scène est créé à la base de données oracle) target, and using IKM SQL for Hyperion Essbase (metadata), run this interface".

    org.apache.bsf.BSFException: exception of Jython:
    Traceback (innermost last):
    "< String >" file, line 61, in there?
    com.hyperion.odi.essbase.ODIEssbaseException: invalid value specified [RULES_FILE] for the load option [null]

    at com.hyperion.odi.essbase.ODIEssbaseMetaWriter.validateLoadOptions (unknown Source)

    at com.hyperion.odi.essbase.AbstractEssbaseWriter.beginLoad (unknown Source)

    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)

    at sun.reflect.NativeMethodAccessorImpl.invoke (unknown Source)

    at sun.reflect.DelegatingMethodAccessorImpl.invoke (unknown Source)

    at java.lang.reflect.Method.invoke (unknown Source)

    at org.python.core.PyReflectedFunction.__call__ (PyReflectedFunction.java)

    at org.python.core.PyMethod.__call__ (PyMethod.java)

    at org.python.core.PyObject.__call__ (PyObject.java)

    at org.python.core.PyInstance.invoke (PyInstance.java)

    to org.python.pycode._pyx1.f$ 0 (< string >: 61)

    to org.python.pycode._pyx1.call_function (< string >)

    at org.python.core.PyTableCode.call (PyTableCode.java)

    at org.python.core.PyCode.call (PyCode.java)

    at org.python.core.Py.runCode (Py.java)

    at org.python.core.Py.exec (Py.java)



    I am very confused by it. Someone give me a solution or other docs.

    Ethan.

    Hello

    You say that you are loading, but you chose the KM for the loading of metadata?
    If you load metadata then you must provide a rules file, if you are loading and then choose the KM - IKM SQL for Hyperion Essbase (DATA)

    Ok?

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Loading Oracle data to essbase using erroneous ODI

    I am trying to load data from oral table to essbase. But I get the following error.

    org.apache.bsf.BSFException: exception of Jython:
    Traceback (innermost last):
    "< String >" file, line 23, in there?
    com.hyperion.odi.essbase.ODIEssbaseException: lack of standard dimension for data loading column
    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData (unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke (unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke (unknown Source)
    at java.lang.reflect.Method.invoke (unknown Source)
    at org.python.core.PyReflectedFunction.__call__ (PyReflectedFunction.java)
    at org.python.core.PyMethod.__call__ (PyMethod.java)
    at org.python.core.PyObject.__call__ (PyObject.java)
    at org.python.core.PyInstance.invoke (PyInstance.java)
    to org.python.pycode._pyx3.f$ 0 (< string >: 23)
    to org.python.pycode._pyx3.call_function (< string >)
    at org.python.core.PyTableCode.call (PyTableCode.java)
    at org.python.core.PyCode.call (PyCode.java)
    at org.python.core.Py.runCode (Py.java)
    at org.python.core.Py.exec (Py.java)
    at org.python.util.PythonInterpreter.exec (PythonInterpreter.java)
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
    at com.sunopsis.dwg.codeinterpretor.k.a (k.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt (SnpSessTaskSqlI.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep (SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession (SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand (DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute (DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i (e.java)
    at com.sunopsis.dwg.cmd.h.y (h.java)
    at com.sunopsis.dwg.cmd.e.run (e.java)
    at java.lang.Thread.run (unknown Source)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: lack of standard dimension for data loading column
    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.validateColumns (unknown Source)
    ... 32 more

    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: lack of standard dimension for data loading column

    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
    at com.sunopsis.dwg.codeinterpretor.k.a (k.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt (SnpSessTaskSqlI.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep (SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession (SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand (DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute (DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i (e.java)
    at com.sunopsis.dwg.cmd.h.y (h.java)
    at com.sunopsis.dwg.cmd.e.run (e.java)
    at java.lang.Thread.run (unknown Source)


    Also, I have the following doubts

    1. what should be the LKM defined for the soruce. What is LKM SQL FOR SQL

    2. I have 6 standard dimensions in the table oracle but Currncy and hsprates in essbase. So, how can I ignore these two dimensions by loading data. Or is there a way where I can have the default options set for these two dimensions.

    3. can I load data from several tables into essbase. If Yes please let me know the procedure.

    Hello

    In the interface and the target data store, you can enter a value in the field of cartography.
    For example, if you have a column called year in your target data store and you still want to load FY09, then you would enter "FY09" in the field of cartography

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Error loading of ODI planning data

    Hello

    I am trying to load data from the planning table. I get the error in the charge at the stage of planning data:

    com.hyperion.odi.planning.ODIPlanningException: com.hyperion.odi.planning.ODIPlanningException: java.lang.RuntimeException: HyperionPlanningBean::beginLoad: could not derive local systemCfg format for loading data.

    Code:

    from com.hyperion.odi.common import ODIConstants
    from com.hyperion.odi.connection import HypAppConnectionFactory
    import java.lang class
    import java.lang Boolean
    import java.sql *.
    from java.util import HashMap
    #
    # Get the select statement in the meeting area:
    #
    SQL = ' "' select ltrim (rtrim (C2_NODE))"Project", ltrim (rtrim (C3_PARENT)) 'Parent', case when ltrim (rtrim (C4_ALIAS_2)) is null then '#missing»»
    of another ltrim (rtrim (C4_ALIAS_2)) end ' Alias: Alias 2 ', case when ltrim (rtrim (C5_ALIAS_1)) is null then '#missing' another
    LTRIM (RTrim (C5_ALIAS_1)) end ' Alias: Default "," #missing ' 'valid for Consolidations ", ltrim (rtrim (C6_DATA_STORAGE))" Data Storage", ltrim (rtrim (C7_TWO_PASS_CALCULATION)) 'Calculation of two', case when ltrim (rtrim (C8_DESCRIPTION)) is null then '#missing' other '.
    LTRIM (RTrim (C8_DESCRIPTION)) end 'Description', "#missing '' formula' ', #missing '" UDA"', #missing ' 'Smart List', C1_DATA_TYPE 'Data Type'," #missing ' "Operation", ltrim (rtrim (C9_AGGREGATION_1)) '(F_AdSale) aggregation' of "C$ _0Project" where (1 = 1) ' "

    srcCx = odiRef.getJDBCConnection ("SRC")

    stmt = srcCx.createStatement)

    srcFetchSize = 30

    stmt.setFetchSize (srcFetchSize)

    RS = stmt.executeQuery (sql)

    #load data
    stats = pWriter.loadData (rs)

    #close the database result set, connection
    RS. Close()
    stmt. Close()

    Any help will be appreciated

    -Kash

    I have not seen that error before but there's a doc in support of Oracle that highlights the error ' could not derive local systemCfg format for loading data [ID 1479461.1].
    It is not directly related to ODI or problem, but the workaround may help in your case.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • SQL Loader loading data into two Tables using a single CSV file

    Dear all,

    I have a requirement where in I need to load the data into 2 tables using a simple csv file.

    So I wrote the following control file. But it loads only the first table and also there nothing in the debug log file.

    Please suggest how to achieve this.

    Examples of data

    Source_system_code,Record_type,Source_System_Vendor_number,$vendor_name,Vendor_site_code,Address_line1,Address_line2,Address_line3

    Victor, New, Ven001, Vinay, Vin001, abc, def, xyz

    Control file script

    ================

    OPTIONS (errors = 0, skip = 1)
    load data
    replace
    in the table1 table:
    fields ended by ',' optionally surrounded "" "
    (
    Char Source_system_code (1) POSITION "ltrim (rtrim (:Source_system_code)),"
    Record_type tank "ltrim (rtrim (:Record_type)),"
    Source_System_Vendor_number tank "ltrim (rtrim (:Source_System_Vendor_number)),"
    $vendor_name tank "ltrim (rtrim (:Vendor_name)),"
    )
    in the Table2 table
    1 = 1
    fields ended by ',' optionally surrounded "" "
    (
    $vendor_name tank "ltrim (rtrim (:Vendor_name)),"
    Vendor_site_code tank "ltrim (rtrim (:Vendor_site_code)),"
    Address_line1 tank "ltrim (rtrim (:Address_line1)),"
    Address_line2 tank "ltrim (rtrim (:Address_line2)),"
    Address_line3 tank "ltrim (rtrim (:Address_line3)).
    )

    the problem here is loading into a table, only the first. (Table 1)

    Please guide me.

    Thank you

    Kumar

    When you do not provide a starting position for the first field in table2, it starts with the following after a last referenced in table1 field, then it starts with vendor_site_code, instead of $vendor_name.  So what you need to do instead, is specify position (1) to the first field in table2 and use the fields to fill.  In addition, he dislikes when 1 = 1, and he didn't need anyway.  See the example including the corrected below control file.

    Scott@orcl12c > test.dat TYPE of HOST

    Source_system_code, Record_type, Source_System_Vendor_number, $vendor_name, Vendor_site_code, Address_line1, Address_line2, Address_line3

    Victor, New, Ven001, Vinay, Vin001, abc, def, xyz

    Scott@orcl12c > test.ctl TYPE of HOST

    OPTIONS (errors = 0, skip = 1)

    load data

    replace

    in the table1 table:

    fields ended by ',' optionally surrounded "" "

    (

    Char Source_system_code (1) POSITION "ltrim (rtrim (:Source_system_code)),"

    Record_type tank "ltrim (rtrim (:Record_type)),"

    Source_System_Vendor_number tank "ltrim (rtrim (:Source_System_Vendor_number)),"

    $vendor_name tank "ltrim (rtrim (:Vendor_name)).

    )

    in the Table2 table

    fields ended by ',' optionally surrounded "" "

    (

    source_system_code FILL (1) POSITION.

    record_type FILLING,

    source_system_vendor_number FILLING,

    $vendor_name tank "ltrim (rtrim (:Vendor_name)),"

    Vendor_site_code tank "ltrim (rtrim (:Vendor_site_code)),"

    Address_line1 tank "ltrim (rtrim (:Address_line1)),"

    Address_line2 tank "ltrim (rtrim (:Address_line2)),"

    Address_line3 tank "ltrim (rtrim (:Address_line3)).

    )

    Scott@orcl12c > CREATE TABLE table1:

    2 (Source_system_code VARCHAR2 (13),)

    3 Record_type VARCHAR2 (11),

    4 Source_System_Vendor_number VARCHAR2 (27),

    5 $vendor_name VARCHAR2 (11))

    6.

    Table created.

    Scott@orcl12c > CREATE TABLE table2

    2 ($vendor_name VARCHAR2 (11),)

    3 Vendor_site_code VARCHAR2 (16).

    4 Address_line1 VARCHAR2 (13),

    5 Address_line2 VARCHAR2 (13),

    Address_line3 6 VARCHAR2 (13))

    7.

    Table created.

    Scott@orcl12c > HOST SQLLDR scott/tiger CONTROL = test.ctl DATA = test.dat LOG = test.log

    SQL * Loader: release 12.1.0.1.0 - Production on Thu Mar 26 01:43:30 2015

    Copyright (c) 1982, 2013, Oracle and/or its affiliates.  All rights reserved.

    Path used: classics

    Commit the point reached - the number of logical records 1

    TABLE1 table:

    1 row loaded successfully.

    Table TABLE2:

    1 row loaded successfully.

    Check the log file:

    test.log

    For more information on the charge.

    Scott@orcl12c > SELECT * FROM table1

    2.

    RECORD_TYPE SOURCE_SYSTEM_VENDOR_NUMBER $VENDOR_NAME SOURCE_SYSTEM

    ------------- ----------- --------------------------- -----------

    Victor Ven001 new Vinay

    1 selected line.

    Scott@orcl12c > SELECT * FROM table2

    2.

    $VENDOR_NAME VENDOR_SITE_CODE ADDRESS_LINE1 ADDRESS_LINE2 ADDRESS_LINE3

    ----------- ---------------- ------------- ------------- -------------

    Vinay Vin001 abc def xyz

    1 selected line.

    Scott@orcl12c >

Maybe you are looking for