Automate the process of loading data using Task Scheduler

Hello

I do the automated processes to load the data into Hyperion Planning application with the help of the Task Scheduler.

I created Data_Load.bat file, but the rest of the process, I cannot complete.

So could you help me, how to automate the loading of the process data using the file Data_load.bat & Task Scheduler.

Thank you

Windows 2008? Open the Task Scheduler, Action > creates a fundamental task, give it a name, select the frequency, select start a program, find the batch script, then you're pretty much done for a basic task.

See you soon

John
http://John-Goodwin.blogspot.com/

Tags: Business Intelligence

Similar Questions

  • How to automate the process of loading data using load file & Task Scheduler

    Hello

    I do the automated processes to load the data into Hyperion Planning application using the file data_Load.bat & Scheduler of tasks.

    I created Data_Load.bat file, but the rest of the process, I cannot complete.

    Could help you me, how to automate the process of loading data using the file Data_load.bat & task Scheduler or what are the rest of the file is require it to achieve.

    Thank you

    In response to your question using the maxl for loading scripts?

    If Yes, I've seen and deliver in the batch (ex: load_data.bat) that is you do not have the path of the maxl script complete with a batch when passing through the event the task scheduler will work, but the log file and / or error will not be created. Which means lots claims it linked task scheduler, although he did not do what you need to.

    If you use maxl use this as the batch

    "essmsh C:\data\DataLoad.mxl" or you can also use the full path for the maxl or work elsewhere. The only reason why I think that the maxl can then not work is if you do not have the updated batch updated to call on all LANE changes maxl or if you need to update your environment variables to correct the command essmsh to work in a command prompt.

  • How to automate the process of loading FDM

    Hello

    I use tax script to extract data from GL to HFM by FDM.

    I know that the automation process if I load a flat file in the system target through FDM.

    But I don't use flat file. I use the script to import to load the GL data directly in HFM.

    Can someone let me know if its possible to automate.

    Thank you
    Hima

    Yes, put an empty file in the directory of open batch named according to the batch of the naming standards. This will trigger the integration import script.

    Please mark this answer as correct if you find it so.

  • Failed to load data using the contour load utility

    I am unable to load data using contour utility charge.
    Assigned data loading dimension such as: account
    Driver-Dimension such as: period
    Member as a pilot: Jan

    Connection file:

    Account, Jan, Point of view, the name of Cube data loading
    Investment, 1234, "India, current, heck, FY13", Plan1


    Utility of contour control
    OutlineLoad A:pract U:admin /I:C:\test1.csv /D:Account /L:C:\lg.log /X:C:\ex.exc


    Exception file

    [Thu Mar 28 18:05:39 GMT + 05:30 2013] Error loading of data record 1: investments, 1234, '' ' India, common, project, FY14' ' ', Plan1
    [Thu Mar 28 18:05:39 GMT + 05:30 2013] com.hyperion.planning.InvalidMemberException: Member India, common, rough, FY14 does not exist or you do not have access to it.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Planning of vector data store processes loaded with exceptions: exceptions have occurred, examine the file exception for more information. 1 data record has been read 1 record of data have been processed, 0 were loaded successfully, 1 was rejected.


    Log file:


    Connected application "Rlap", liberation 11.113, adapter Interface Version 5, supported and not enabled workforce, CapEx no taken care and not active, CSS Version 3
    [Thu Mar 28 18:05:38 GMT + 05:30 2013] Input file located and opened successfully "C:\load.csv".
    [Thu Mar 28 18:05:38 GMT + 05:30 2013] Record header fields: account, Jan, Point of view, the name of Cube data loading
    [Thu Mar 28 18:05:38 GMT + 05:30 2013] Located by using 'Account' dimension and for the loading of data into the application 'Rlap.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Loading dimension 'Account' has been successfully opened.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] A refresh of the cube operation will not be run.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Create filters for safe operation will not be performed.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Look at the files of newspapers of Essbase to status if Essbase data have been loaded.
    [Thu Mar 28 18:05:40 GMT + 05:30 2013] Planning of vector data store processes loaded with exceptions: exceptions have occurred, examine the file exception for more information. 1 data record has been read 1 record of data have been processed, 0 were loaded successfully, 1 was rejected.



    Infact members exist in outline.
    Any help would be appreciated.

    can you double check your csv file, open it in a text editor, because the error is showing additional citations.

    [Thu Mar 28 18:05:39 GMT + 05:30 2013] Error loading of data record 1: investments, 1234, '' ' India, common, project, FY14' ' ', Plan1

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Issue while loading data using the file Rules Essbase

    Hi all

    I am facing problem while loading data using the Rules file. In the rules file, I rejected several members in two areas (two dimensions). Now if I load the data using the rules file I'm getting errors for all members in the dataload.err file. If I reject mutiple members of a single field, the data load without settling errors in the dataload.err file.

    I want to know how rmany members of several fields of ejection for loading data using the file Rules Essbase? Is it possible?

    Okay, okay... I think that you must assign Global Select / reject Boolean in the parameters of loading data as 'Or':

  • How to taskkill all the open running program using Task Scheduler?

    I already know the command "taskkill /f /im notepad.exe" or n *...

    But how to kill any process running in the task (ex > process and automatically process background execution..).

    program to kill only with the windows open...

    If I run taskmgr.exe program alone and add task annex taskkill, add the arguments: /f /im t *.

    It must end only taskmgr.exe, not activate end taskeng.exe, taskhost.exe process...

    Finally, I want to know only to close all windows opened using Task Scheduler or by using another program that can be.

    When you have a number of tasks of the same name, then you must use their PID to kill selectively.

  • Integrate the functionality of loading data Dataworkshop in an Application

    Is it possible to reuse the functionality of loading data from the Dataworkshop inside an application. I write a batch processing application and it would be great to have re-use load capacity that exists already, maybe by default in the target tables and attributes, is there a api is available to enable this?

    Thanks in advance for your help,

    Matt

    I don't think you can use this feature directly from the existing APEX product. However, there are a few people who have posted solutions to load files CSV style for users in the forum...

    Thank you

    Tony Miller
    Software LuvMuffin
    (281) 871-0950

  • How to automate ScanDisk using Task Scheduler tool?

    There is no file c:\windows\scandskw.exe on my XP, does anyone have any suggestions on how to automate using Task Scheduler tool in XP ScanDisk?

    Thanks for all the suggestions

    Saturday, August 14, 2010, 02:18:45 + 0000, SC Tom wrote:

    There is no program Scandisk in Windows XP.

    In order to specify that the person you are you answer to the. There is no program
    called the Scandisk, but there is a program with this feature; There
    Called Chkdsk.

    Ken Blake

  • Name of user and password invalid executes the Plan of loading data

    I get the following error when I try to execute the plan of loading data for loading data from my EBS server.  I don't know what username and password she is claiming is not valid or where to change the value.  Any ideas where I can find it?

    ODI-1519: series step "start load Plan.

    (InternalID:4923500) ' failed, because the child step "Global Variable Refresh.

    (InternalID:4924500) "is a mistake.

    ODI-1529: refreshment of the variable 'BIAPPS.13P_CALENDAR_ID' failed:

    Select CASE WHEN 'The Global Variable Refresh' in (select distinct)

    group_code from C_PARAMETER_VALUE_FORMATTER_V where PARAM_CODE =

    "13P_CALENDAR_ID")

    THEN (select param_value

    of C_PARAMETER_VALUE_FORMATTER_V

    where PARAM_CODE = '13P_CALENDAR_ID. '

    and group_code = 'Global Variable Refresh'

    and datasource_num_id = ' #BIAPPS. WH_DATASOURCE_NUM_ID')

    ON THE OTHER

    (select param_value in the C_GL_PARAM_VALUE_FORMATTER_V where PARAM_CODE =

    '13P_CALENDAR_ID' and datasource_num_id =

    ' #BIAPPS. WH_DATASOURCE_NUM_ID')

    END

    of the double

    0:72000:Java.SQL.SqlException: ORA-01017: name of user and password invalid.

    connection refused

    java.sql.SQLException: ORA-01017: name of user and password invalid. opening of session

    denied

    to

    oracle.odi.jdbc.datasource.LoginTimeoutDatasourceAdapter.doGetConnection(LoginTimeoutDatasourceAdapter.java:133)

    to

    oracle.odi.jdbc.datasource.LoginTimeoutDatasourceAdapter.getConnection(LoginTimeoutDatasourceAdapter.java:62)

    to

    oracle.odi.core.datasource.dwgobject.support.OnConnectOnDisconnectDataSourceAdapter.getConnection(OnConnectOnDisconnectDataSourceAdapter.java:74)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.executeVariableStep(LoadPlanProcessor.java:3050)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.refreshVariables(LoadPlanProcessor.java:4287)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.AddRunnableScenarios(LoadPlanProcessor.java:2284)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.AddRunnableScenarios(LoadPlanProcessor.java:2307)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.SelectNextRunnableScenarios(LoadPlanProcessor.java:2029)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.StartAllScenariosFromStep(LoadPlanProcessor.java:1976)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.startLPExecution(LoadPlanProcessor.java:491)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.initLPInstance(LoadPlanProcessor.java:384)

    to

    oracle.odi.runtime.agent.loadplan.LoadPlanProcessor.startLPInstance(LoadPlanProcessor.java:147)

    to

    oracle.odi.runtime.agent.processor.impl.StartLoadPlanRequestProcessor.doProcessRequest(StartLoadPlanRequestProcessor.java:87)

    to

    oracle.odi.runtime.agent.processor.SimpleAgentRequestProcessor.process(SimpleAgentRequestProcessor.java:49)

    to

    oracle.odi.runtime.agent.support.DefaultRuntimeAgent.execute(DefaultRuntimeAgent.java:68)

    to

    oracle.odi.runtime.agent.servlet.AgentServlet.processRequest(AgentServlet.java:564)

    to

    oracle.odi.runtime.agent.servlet.AgentServlet.doPost(AgentServlet.java:518)

    at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)

    at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)

    to

    weblogic.servlet.internal.StubSecurityHelper$ ServletServiceAction.run (StubSecurityHelper.java:227)

    to

    weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)

    to

    weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:301)

    at weblogic.servlet.internal.TailFilter.doFilter(TailFilter.java:26)

    to

    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)

    to

    oracle.security.jps.ee.http.JpsAbsFilter$ 1.run(JpsAbsFilter.java:119)

    at java.security.AccessController.doPrivileged (Native Method)

    to

    oracle.security.jps.util.JpsSubject.doAsPrivileged(JpsSubject.java:324)

    to

    oracle.security.jps.ee.util.JpsPlatformUtil.runJaasMode(JpsPlatformUtil.java:460)

    to

    oracle.security.jps.ee.http.JpsAbsFilter.runJaasMode(JpsAbsFilter.java:103)

    to

    oracle.security.jps.ee.http.JpsAbsFilter.doFilter(JpsAbsFilter.java:171)

    at oracle.security.jps.ee.http.JpsFilter.doFilter(JpsFilter.java:71)

    to

    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)

    to

    oracle.dms.servlet.DMSServletFilter.doFilter(DMSServletFilter.java:163)

    to

    weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:56)

    to

    weblogic.servlet.internal.WebAppServletContext$ ServletInvocationAction.wrapRun (WebAppServletContext.java:3730)

    to

    weblogic.servlet.internal.WebAppServletContext$ ServletInvocationAction.run (WebAppServletContext.java:3696)

    to

    weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)

    to

    weblogic.security.service.SecurityManager.runAs(SecurityManager.java:120)

    to

    weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2273)

    to

    weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2179)

    to

    weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1490)

    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:256)

    at weblogic.work.ExecuteThread.run(ExecuteThread.java:221)

    Caused by: java.sql.SQLException: ORA-01017: name of user and password invalid.

    connection refused

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:462)

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:397)

    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:389)

    at oracle.jdbc.driver.T4CTTIfun.processError(T4CTTIfun.java:689)

    to

    oracle.jdbc.driver.T4CTTIoauthenticate.processError(T4CTTIoauthenticate.java:455)

    at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:481)

    at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:205)

    to

    oracle.jdbc.driver.T4CTTIoauthenticate.doOAUTH(T4CTTIoauthenticate.java:387)

    to

    oracle.jdbc.driver.T4CTTIoauthenticate.doOAUTH(T4CTTIoauthenticate.java:814)

    at oracle.jdbc.driver.T4CConnection.logon(T4CConnection.java:418)

    to

    oracle.jdbc.driver.PhysicalConnection. < init > (PhysicalConnection.java:678)

    to

    oracle.jdbc.driver.T4CConnection. < init > (T4CConnection.java:234)

    to

    oracle.jdbc.driver.T4CDriverExtension.getConnection(T4CDriverExtension.java:34)

    at oracle.jdbc.driver.OracleDriver.connect(OracleDriver.java:567)

    to

    oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnectionFromDriver(DriverManagerDataSource.java:410)

    to

    oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnectionFromDriver(DriverManagerDataSource.java:386)

    to

    oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnectionFromDriver(DriverManagerDataSource.java:353)

    to

    oracle.odi.jdbc.datasource.DriverManagerDataSource.getConnection(DriverManagerDataSource.java:332)

    to

    oracle.odi.jdbc.datasource.LoginTimeoutDatasourceAdapter$ ConnectionProcessor.run (LoginTimeoutDatasourceAdapter.java:217)

    to

    java.util.concurrent.Executors$ RunnableAdapter.call (Executors.java:441)

    to java.util.concurrent.FutureTask$ Sync.innerRun (FutureTask.java:303)

    at java.util.concurrent.FutureTask.run(FutureTask.java:138)

    to

    java.util.concurrent.ThreadPoolExecutor$ Worker.runTask (ThreadPoolExecutor.java:886)

    to

    java.util.concurrent.ThreadPoolExecutor$ Worker.run (ThreadPoolExecutor.java:908)

    at java.lang.Thread.run(Thread.java:662)

    Found the answer after 4 days of research.  Opening ODI Studio.  Go to the topology, expand-> Oracle Technologies.  Open BIAPPS_BIACOMP.  In my system, he had "NULLBIACM_IO" has changed nothing to my correct prefix and it worked.

    Now, my data load gives me the error that the table W_LANGUAGES_G does not exist.   At least I liked long.

  • Close a program using Task Scheduler

    How can I close a .exe program by using Task Scheduler? Eg - is the case of NetMeter, I want the program to automatically close every day at 02:00. It can b done?

    Waiting for a quick response...

    Kind regards.

    Ah I see.  Do not schedule the cmd.exe to run.  Taskkill itself is already an executable command.  So just plan sound directly in the program/script box:

    Taskkill /f /im "NetMeter.exe.


    Do not put anything in the arguments box.  When the program runs, you will briefly see to open the command prompt, the program will be completed, then the command prompt will disappear on its own.
  • Question to load data using sql loader in staging table, and then in the main tables!

    Hello

    I'm trying to load data into our main database table using SQL LOADER. data will be provided in separate pipes csv files.

    I have develop a shell script to load the data and it works fine except one thing.

    Here are the details of a data to re-create the problem.

    Staging of the structure of the table in which data will be filled using sql loader

    create table stg_cmts_data (cmts_token varchar2 (30), CMTS_IP varchar2 (20));

    create table stg_link_data (dhcp_token varchar2 (30), cmts_to_add varchar2 (200));

    create table stg_dhcp_data (dhcp_token varchar2 (30), DHCP_IP varchar2 (20));

    DATA in the csv file-

    for stg_cmts_data-

    cmts_map_03092015_1.csv

    WNLB-CMTS-01-1. 10.15.0.1

    WNLB-CMTS-02-2 | 10.15.16.1

    WNLB-CMTS-03-3. 10.15.48.1

    WNLB-CMTS-04-4. 10.15.80.1

    WNLB-CMTS-05-5. 10.15.96.1

    for stg_dhcp_data-

    dhcp_map_03092015_1.csv

    DHCP-1-1-1. 10.25.23.10, 25.26.14.01

    DHCP-1-1-2. 56.25.111.25, 100.25.2.01

    DHCP-1-1-3. 25.255.3.01, 89.20.147.258

    DHCP-1-1-4. 10.25.26.36, 200.32.58.69

    DHCP-1-1-5 | 80.25.47.369, 60.258.14.10

    for stg_link_data

    cmts_dhcp_link_map_0309151623_1.csv

    DHCP-1-1-1. WNLB-CMTS-01-1,WNLB-CMTS-02-2

    DHCP-1-1-2. WNLB-CMTS-03-3,WNLB-CMTS-04-4,WNLB-CMTS-05-5

    DHCP-1-1-3. WNLB-CMTS-01-1

    DHCP-1-1-4. WNLB-CMTS-05-8,WNLB-CMTS-05-6,WNLB-CMTS-05-0,WNLB-CMTS-03-3

    DHCP-1-1-5 | WNLB-CMTS-02-2,WNLB-CMTS-04-4,WNLB-CMTS-05-7

    WNLB-DHCP-1-13 | WNLB-CMTS-02-2

    Now, after loading these data in the staging of table I have to fill the main database table

    create table subntwk (subntwk_nm varchar2 (20), subntwk_ip varchar2 (30));

    create table link (link_nm varchar2 (50));

    SQL scripts that I created to load data is like.

    coil load_cmts.log

    Set serveroutput on

    DECLARE

    CURSOR c_stg_cmts IS SELECT *.

    OF stg_cmts_data;

    TYPE t_stg_cmts IS TABLE OF stg_cmts_data % ROWTYPE INDEX BY pls_integer;

    l_stg_cmts t_stg_cmts;

    l_cmts_cnt NUMBER;

    l_cnt NUMBER;

    NUMBER of l_cnt_1;

    BEGIN

    OPEN c_stg_cmts.

    Get the c_stg_cmts COLLECT in BULK IN l_stg_cmts;

    BECAUSE me IN l_stg_cmts. FIRST... l_stg_cmts. LAST

    LOOP

    SELECT COUNT (1)

    IN l_cmts_cnt

    OF subntwk

    WHERE subntwk_nm = l_stg_cmts (i) .cmts_token;

    IF l_cmts_cnt < 1 THEN

    INSERT

    IN SUBNTWK

    (

    subntwk_nm

    )

    VALUES

    (

    l_stg_cmts (i) .cmts_token

    );

    DBMS_OUTPUT. Put_line ("token has been added: ' |") l_stg_cmts (i) .cmts_token);

    ON THE OTHER

    DBMS_OUTPUT. Put_line ("token is already present'");

    END IF;

    WHEN l_stg_cmts EXIT. COUNT = 0;

    END LOOP;

    commit;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    for dhcp


    coil load_dhcp.log

    Set serveroutput on

    DECLARE

    CURSOR c_stg_dhcp IS SELECT *.

    OF stg_dhcp_data;

    TYPE t_stg_dhcp IS TABLE OF stg_dhcp_data % ROWTYPE INDEX BY pls_integer;

    l_stg_dhcp t_stg_dhcp;

    l_dhcp_cnt NUMBER;

    l_cnt NUMBER;

    NUMBER of l_cnt_1;

    BEGIN

    OPEN c_stg_dhcp.

    Get the c_stg_dhcp COLLECT in BULK IN l_stg_dhcp;

    BECAUSE me IN l_stg_dhcp. FIRST... l_stg_dhcp. LAST

    LOOP

    SELECT COUNT (1)

    IN l_dhcp_cnt

    OF subntwk

    WHERE subntwk_nm = l_stg_dhcp (i) .dhcp_token;

    IF l_dhcp_cnt < 1 THEN

    INSERT

    IN SUBNTWK

    (

    subntwk_nm

    )

    VALUES

    (

    l_stg_dhcp (i) .dhcp_token

    );

    DBMS_OUTPUT. Put_line ("token has been added: ' |") l_stg_dhcp (i) .dhcp_token);

    ON THE OTHER

    DBMS_OUTPUT. Put_line ("token is already present'");

    END IF;

    WHEN l_stg_dhcp EXIT. COUNT = 0;

    END LOOP;

    commit;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    for link -.

    coil load_link.log

    Set serveroutput on

    DECLARE

    l_cmts_1 VARCHAR2 (4000 CHAR);

    l_cmts_add VARCHAR2 (200 CHAR);

    l_dhcp_cnt NUMBER;

    l_cmts_cnt NUMBER;

    l_link_cnt NUMBER;

    l_add_link_nm VARCHAR2 (200 CHAR);

    BEGIN

    FOR (IN) r

    SELECT dhcp_token, cmts_to_add | ',' cmts_add

    OF stg_link_data

    )

    LOOP

    l_cmts_1: = r.cmts_add;

    l_cmts_add: = TRIM (SUBSTR (l_cmts_1, 1, INSTR (l_cmts_1, ',') - 1));

    SELECT COUNT (1)

    IN l_dhcp_cnt

    OF subntwk

    WHERE subntwk_nm = r.dhcp_token;

    IF l_dhcp_cnt = 0 THEN

    DBMS_OUTPUT. Put_line ("device not found: ' |") r.dhcp_token);

    ON THE OTHER

    While l_cmts_add IS NOT NULL

    LOOP

    l_add_link_nm: = r.dhcp_token |' _TO_' | l_cmts_add;

    SELECT COUNT (1)

    IN l_cmts_cnt

    OF subntwk

    WHERE subntwk_nm = TRIM (l_cmts_add);

    SELECT COUNT (1)

    IN l_link_cnt

    LINK

    WHERE link_nm = l_add_link_nm;

    IF l_cmts_cnt > 0 AND l_link_cnt = 0 THEN

    INSERT INTO link (link_nm)

    VALUES (l_add_link_nm);

    DBMS_OUTPUT. Put_line (l_add_link_nm |) » '||' Has been added. ") ;

    ELSIF l_link_cnt > 0 THEN

    DBMS_OUTPUT. Put_line (' link is already present: ' | l_add_link_nm);

    ELSIF l_cmts_cnt = 0 then

    DBMS_OUTPUT. Put_line (' no. CMTS FOUND for device to create the link: ' | l_cmts_add);

    END IF;

    l_cmts_1: = TRIM (SUBSTR (l_cmts_1, INSTR (l_cmts_1, ',') + 1));

    l_cmts_add: = TRIM (SUBSTR (l_cmts_1, 1, INSTR (l_cmts_1, ',') - 1));

    END LOOP;

    END IF;

    END LOOP;

    COMMIT;

    EXCEPTION

    WHILE OTHERS THEN

    Dbms_output.put_line ('ERROR' |) SQLERRM);

    END;

    /

    output

    control files -

    DOWNLOAD THE DATA

    INFILE 'cmts_data.csv '.

    ADD

    IN THE STG_CMTS_DATA TABLE

    When (cmts_token! = ") AND (cmts_token! = 'NULL') AND (cmts_token! = 'null')

    and (cmts_ip! = ") AND (cmts_ip! = 'NULL') AND (cmts_ip! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:cmts_token))' cmts_token,

    cmts_ip ' RTRIM (LTRIM(:cmts_ip)) ")". "

    for dhcp.


    DOWNLOAD THE DATA

    INFILE 'dhcp_data.csv '.

    ADD

    IN THE STG_DHCP_DATA TABLE

    When (dhcp_token! = ") AND (dhcp_token! = 'NULL') AND (dhcp_token! = 'null')

    and (dhcp_ip! = ") AND (dhcp_ip! = 'NULL') AND (dhcp_ip! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:dhcp_token))' dhcp_token,

    dhcp_ip ' RTRIM (LTRIM(:dhcp_ip)) ")". "

    for link -.

    DOWNLOAD THE DATA

    INFILE 'link_data.csv '.

    ADD

    IN THE STG_LINK_DATA TABLE

    When (dhcp_token! = ") AND (dhcp_token! = 'NULL') AND (dhcp_token! = 'null')

    and (cmts_to_add! = ") AND (cmts_to_add! = 'NULL') AND (cmts_to_add! = 'null')

    FIELDS TERMINATED BY ' |' SURROUNDED OF POSSIBLY "" "

    TRAILING NULLCOLS

    ('RTRIM (LTRIM (:dhcp_token))' dhcp_token,

    cmts_to_add TANK (4000) RTRIM (LTRIM(:cmts_to_add)) ")" ""

    SHELL SCRIPT-

    If [!-d / log]

    then

    Mkdir log

    FI

    If [!-d / finished]

    then

    mkdir makes

    FI

    If [!-d / bad]

    then

    bad mkdir

    FI

    nohup time sqlldr username/password@SID CONTROL = load_cmts_data.ctl LOG = log/ldr_cmts_data.log = log/ldr_cmts_data.bad DISCARD log/ldr_cmts_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    nohup time username/password@SID @load_cmts.sql

    nohup time sqlldr username/password@SID CONTROL = load_dhcp_data.ctl LOG = log/ldr_dhcp_data.log = log/ldr_dhcp_data.bad DISCARD log/ldr_dhcp_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    time nohup sqlplus username/password@SID @load_dhcp.sql

    nohup time sqlldr username/password@SID CONTROL = load_link_data.ctl LOG = log/ldr_link_data.log = log/ldr_link_data.bad DISCARD log/ldr_link_data.reject ERRORS = BAD = 100000 LIVE = TRUE PARALLEL = TRUE &

    time nohup sqlplus username/password@SID @load_link.sql

    MV *.log. / log

    If the problem I encounter is here for loading data in the link table that I check if DHCP is present in the subntwk table, then continue to another mistake of the newspaper. If CMTS then left create link to another error in the newspaper.

    Now that we can here multiple CMTS are associated with unique DHCP.

    So here in the table links to create the link, but for the last iteration of the loop, where I get separated by commas separate CMTS table stg_link_data it gives me log as not found CMTS.

    for example

    DHCP-1-1-1. WNLB-CMTS-01-1,WNLB-CMTS-02-2

    Here, I guess to link the dhcp-1-1-1 with balancing-CMTS-01-1 and wnlb-CMTS-02-2

    Theses all the data present in the subntwk table, but still it gives me journal wnlb-CMTS-02-2 could not be FOUND, but we have already loaded into the subntwk table.

    same thing is happening with all the CMTS table stg_link_data who are in the last (I think here you got what I'm trying to explain).

    But when I run the SQL scripts in the SQL Developer separately then it inserts all valid links in the table of links.

    Here, she should create 9 lines in the table of links, whereas now he creates only 5 rows.

    I use COMMIT in my script also but it only does not help me.

    Run these scripts in your machine let me know if you also get the same behavior I get.

    and please give me a solution I tried many thing from yesterday, but it's always the same.

    It is the table of link log

    link is already present: dhcp-1-1-1_TO_wnlb-cmts-01-1

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-02-2

    link is already present: dhcp-1-1-2_TO_wnlb-cmts-03-3
    link is already present: dhcp-1-1-2_TO_wnlb-cmts-04-4

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-5

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-01-1

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-8
    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-6
    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-0

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-03-3

    link is already present: dhcp-1-1-5_TO_wnlb-cmts-02-2
    link is already present: dhcp-1-1-5_TO_wnlb-cmts-04-4

    NOT FOUND CMTS for device to create the link: wnlb-CMTS-05-7

    Device not found: wnlb-dhcp-1-13

    IF NEED MORE INFORMATION PLEASE LET ME KNOW

    Thank you

    I felt later in the night that during the loading in the staging table using UNIX machine he created the new line for each line. That is why the last CMTS is not found, for this I use the UNIX 2 BACK conversion and it starts to work perfectly.

    It was the dos2unix error!

    Thank you all for your interest and I may learn new things, as I have almost 10 months of experience in (PLSQL, SQL)

  • Unable to load data using the rules file

    Hi all

    I am facing a strange problem while loading data by using a rules file:

    1. I don't get any error/kick. It gives fair attention No. data values are modified by this data file loading. But nothing will load inside.

    2. I tried with Maxl and directly from EAS but gettting same warning.

    3 has tried to block sending n to entires in the data file and it does not work (so no Dynamics members etc..)

    4. rule file validates correctly.

    5. I tried with different types of different files (Csv, Xls, Txt, etc.) but no luck.

    Any suggestions / ideas of what could be bad?

    Don't know what exactly was the problem, but I recreated the rule files and it solved the problem... May be a case of corrupted file rule?

  • How to automate the process of CopyApplication using the HFMCMD command line shell

    Hello

    Can someone please explain the automation application copy process using HFMCMD command line shell.

    Detailed information on the use of this are here (pg 19 of the referenced PDF):

    https://support.Oracle.com/epmos/faces/DocContentDisplay?ID=968956.1

    Command line shell

    New in 11 versions is a copy request utility that takes the input from the command line. This utility allows scheduling copy operations using third party, planning of utilities such as the built in Windows Scheduler. The HFM copy App Command Line Utility in releases of the + 11.1.2 lies in EPM_ORACLE_HOME\Products\ FinancialManagement \Utilities and for all the previous versions of + 11.1.1, the location is % Hyperion_Home%\FinancialManagement\Server. The utility is named HfmCopyAppCmd.exe or HfmCopyAppCmd_x64.exe.

    The command line utility has the same options that the utility based GUI with the exception being that there is no progress window to watch tasks participated and no option to retry a failed task. In order to run the command line utility, the computer that runs the utility must have installed the Package redistributable Microsoft Visual C++ 2005 SP1. This package is installed by the installer of version 11 EMP, so it doesn't have to be installed manually when executing an HFM application server, but it may be necessary to install the copy utility is run on a server database directly.

    Use: S = HfmCopyAppCmd.exe"" s = d ="" d =

  • Get the error when loading data into the base cube

    Hi all

    I get the error by loading the file of hollow data to the database below.
    Please take suggest me...

    I use ESSBASE 9.3.1.3.


    Error:

    Basic charge database
    The Xchgrate database management
    Load the database Interntl
    Sample application loaded - connection established
    Start the application [example] with the process id [1688]
    Empty invalid character in name (E:\My personal Data\Test\222.txt) based on the ESSAPI EssUpdateFile
    Unexpected error Essbase 1030010

    Empty invalid character in name (E:\My personal Data\Test\222.txt) based on the ESSAPI EssUpdateFile

    Essbase has problems with spaces (white space characters) the path names (my personal data)

    Rename your path or move the data file to a location that has no characters/spaces in the path name.

  • Failed to retrieve the DIAdem DataFinder research data, using data-finder toolkit LabVIEW 2009

    I am facing problem to retrieve the DIAdem Datafinder data.

    At first, I developed this project in LabVIEW 2010. But because of a problem that I met in LabVIEW 2010, I thougth to work on the project in LabVIEW 2009.

    Then, I reused the project that I created using LV 2010 using option economy for the previous version.

    When I tried to retrieve the data from the tiara using data finder toolkit. I got a 305505 warning. Please see the attached message of Warninig.

    But I have not seen this warning while I was working in LabVIEW 2010 and also, I was able to get the data from the tiara.

    Please suggest me a way to solve this problem.

    Hi Nanda,
    There is a bug when using the SDK use in combination with DataFinder Toolkit 2009.
    This bug is already fixed in LabVIEW 2010 and use SDK 2010. But as you said there is another question in LabVIEW 2010, forcing you to LV 2009.
    I suggest you use LabVIEW 2009 with use SDK 2009 and DataFinder Toolkit 2009 and instead to use the "Waveform.vi results" to read the search results, I would recommend to use the screw storage. I spread your example VI and it attached to this message. You will still see the dialog box to search for "lvStorage.dll" but it will automatically disappear and the VI works anyway because in this case LabVIEW will find the dll itself.
    My extended VI uses the '_openWithRoot.vi' to open a file with a different use than CT or PDM. This VI is also described in the following knowledge base article: http://zone.ni.com/devzone/cda/epd/p/id/4181
    With LabVIEW 2010 use support has been enhanced and fully integrated in the standard palette of storage screws. So in LabVIEW 2010, you can replace the "_openWithRoot.vi" with the 'Open Data Storage.vi' standard but with LabVIEW 2010, you can also use "Waveforms.vi results" to read the search results.

    I hope this helps.

    Kind regards
    Eva

Maybe you are looking for

  • How can I remove the Get Firefox for Android box on the homepage?

    Recently, I noticed that I constantly see a box on the home page of Firefox asking if I want to send me a link to Firefox for Android. Is there a way to delete it? I tried the solution with a similar issue in 2012, but the solution doesn't seem to wo

  • Tecra M2 does not start

    I'm new havoing problems with my Tecra M2. can someone point me in the right direction with regard to breakdowns, I had a glance on the forum and the threads associated with the Tecra range but cannot find all threads that match my problem. In a shor

  • S10 - 3 c and Linux

    Is the S10 - 3 c incompatible with all versions of Linux? Lenovo can highlight a distribution custom of Fedora 15/16, 11/12 OpenSuSE or Ubuntu - 11 and following?

  • Router EA2700 cannot be acecssed by the IP after power failure

    I had a power outage, and wireless settings and access the router did not work. How can I reset things if I can't access the router through its IP address. Had a 500-Server error. I am connected by cable to the router via its ports of exit. Thank you

  • How to change default C4680 printing options

    Desktop computer HP, 64-bit processor.   Windows 7 I had many odd order issues with my printer cartridges is not admitted, after reinstalling all the software printer (which does not solve the problem), I finally found a cartridge that "seems" to hav