Plugin Excel2Collection for excel load data

Hi all

using APEX (oracle 10 and 11 g) 4.2

I have a page and using ()Excel2Collection) plugin data from excel file to load. Everything worked very well in Oracle11g, but when I use the same app within the 10 g environment then some fields are not filled or mapping. I run the report from the collection, but some fields are empty (10g), but 11g environment is ok.

for example. C001, c002, c003 and so on. On 11g are all ok, but on 10g, c001 ('%), c002 (ok), c003 missing.

is this a bug?

any help please?

Sincerely, LAUGHED.

Take this version of the demopage

Anton

Tags: Database

Similar Questions

  • Error loading data from Oracle 10 g to Excel

    Hello gurus,

    I get the error when loading data from the Oracle 10 g database table to Excel.
    I use SQL TO ADD SQL IKM.
    Staging area is different from the target and I use the same pattern sense to source and put in scene.
    Please help me...

    Error is

    0: S1000: sun.jdbc.odbc.JdbcOdbcBatchUpdateException: [Microsoft] [ODBC Excel Driver] operation must use an update query.
    sun.jdbc.odbc.JdbcOdbcBatchUpdateException: [Microsoft] [ODBC Excel Driver] operation must use an update query.

    at sun.jdbc.odbc.JdbcOdbcPreparedStatement.emulateExecuteBatch (unknown Source)

    at sun.jdbc.odbc.JdbcOdbcPreparedStatement.executeBatchUpdate (unknown Source)

    at sun.jdbc.odbc.JdbcOdbcStatement.executeBatch (unknown Source)

    at com.sunopsis.sql.SnpsQuery.executeBatch (SnpsQuery.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execCollOrders (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt (SnpSessTaskSqlI.java)

    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask (SnpSessTaskSql.java)

    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep (SnpSessStep.java)

    at com.sunopsis.dwg.dbobj.SnpSession.treatSession (SnpSession.java)

    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand (DwgCommandSession.java)

    at com.sunopsis.dwg.cmd.DwgCommandBase.execute (DwgCommandBase.java)

    at com.sunopsis.dwg.cmd.e.i (e.java)

    at com.sunopsis.dwg.cmd.g.y (g.java)

    at com.sunopsis.dwg.cmd.e.run (e.java)

    at java.lang.Thread.run (unknown Source)


    Kindly help me...

    Thanks in advance.

    Samuel

    Hello

    When you created the DSN for the spreed excel worksheet you dot not uncheck the read-only box.
    Make sure that the read-only checkbox is not checked, otherwise you will get an error message when trying to insert data into the Excel worksheet (error: "[Microsoft] [ODBC Excel Driver] operation must use an update query.").

    Thank you
    Fati

  • FDMEE error data import: No. periods have been identified for the loading of the data in the table "AIF_EBS_GL_BALANCES_STG".

    Hi experts,

    I tried to load the data of EBS in HFM via FDMEE.

    Importing data in the rule of loading, I have encountered an error in loading.

    2014-11-21 06:09:18, 601 INFO [AIF]: beginning of the process FDMEE, process ID: 268

    2014-11-21 06:09:18, 601 [AIF] INFO: recording of the FDMEE level: 4

    2014-11-21 06:09:18, 601 [AIF] INFO: FDMEE log file: D:\fdmee\outbox\logs\TESTING_268.log

    2014-11-21 06:09:18, 601 [AIF] INFO: user: admin

    2014-11-21 06:09:18, 601 INFO [AIF]: place: Testing_loc (Partitionkey:3)

    2014-11-21 06:09:18, 601 [AIF] INFO: name: OCT period (period key: 31/10/14 12:00 AM)

    2014-11-21 06:09:18, 601 INFO [AIF]: name of the category: real (category key: 1).

    2014-11-21 06:09:18, 601 INFO [AIF]: name rule: Testing_dlr (rule ID:8)

    2014-11-21 06:09:19, 877 [AIF] INFO: Jython Version: 2.5.1 (Release_2_5_1:6813, September 26 2009, 13:47:54)

    [JRockit (R) Oracle (Oracle Corporation)]

    2014-11-21 06:09:19, 877 INFO [AIF]: Java platform: java1.6.0_37

    2014-11-21 06:09:19, 877 INFO [AIF]: connect the file encoding: UTF-8

    2014-11-21 06:09:21, 368 [AIF] INFO: - START IMPORT STEP -

    2014-11-21 06:09:24, 544 FATAL [AIF]: error in CommData.insertImportProcessDetailsTraceback (most recent call last): File '< string >", line 2672, in insertImportProcessDetail

    RuntimeError: No periods have been identified for the loading of the data in the table 'AIF_EBS_GL_BALANCES_STG'.

    2014-11-21 06:09:24, 748 FATAL [AIF]: load balances data launch GL error

    2014-11-21 06:09:24, 752 [AIF] INFO: end process FDMEE, process ID: 268

    I found a post related to this error, but did not respond.

    I know I'm missing something, gurus please help me to overcome this error.

    ~ Thank you

    I managed to overcome this problem,

    This was caused due to an error in the map of the time.

    In the mapping of source, the name of period should be defined exactly as displayed in the EBS.

    for example: {EBS--> OCT - 14} FDMEE {mapping source--> OCT - 14}

    The names of the time must be identical.

  • Dynamic action for validation of date with the notification message plugin

    Hi all

    Someone help me please with dynamic action for validation of date with the message notification plugin. I have a form with two elements of the date picker control and message notification plugin.

    The requirement first user selects the exam is finished and then selects the date. So, if the date is greater than the date of the examination is over + 2 years then doesn't trigger the message notification plugin. I tried to create that dynamic action on the date picker date that triggers the scheduled issue notification message but I want to make conditional, I mean displays the message only if date of the selected is greater than the date of the exam is finished more than 2 years.

    In terms simple, notification is displayed only if provided is superior to (date of the exam is completed + 2 years).

    I use oracle apex 4.0 version and oracle 10g r2 database. I tried to reproduce the same requirement in my personal workspace. Here are the details. Please take a look.

    Workspace: raghu_workspace

    username: orton607

    password: orton607

    APP # 72193

    PG # 1

    Any help is appreciated.

    Thanks in advance.

    Orton.

    You can get the value of the date of entry:

    $(ele) .datePicker ('getDate');

    So what to add functions such as:

    function validateNotification (d1, d2) {}

    Date1 var = $(d1) .datepicker ('getDate');

    date2 var = $(d2) .datepicker ('getDate');

    if(date1 && date2) {}

    return ((date2.getTime()-date1.getTime())/(1000*24*60*60))>(365*2);

    } else {}

    Returns false;

    }

    }

    The logic based on setting (I have two years from years of 365 days preceding)

    Then in the D.A. specify a JavaScript expression as:

    validateNotification ('P2_REVIEW_COMPLETED', this.triggeringElement.id)

    Refer to page 2 for example.

  • How we call OLAP to make unnecessary aggredations for loading data?

    Hello

    I am trying to create an OLAP cube relatively simple two-dimensional (you can call it "Square OLAP"). My current environment is 11.2EE with MN for the workspace management.

    A dimension is date,-> year-> month-> day annually, the other is production unit, implemented as a hierarchy with some machine on the lower level. The fact is defined by a pair of low level of these dimensions values; for example, a measure is taken once per day for each machine. I want to store these detailed facts in a cube as well as aggregates, so that they can be easily drilled up to without questioning the original fact table.

    The rules of aggregation are on 'aggregation level = default' (which is the day and the machine respectively) for two of my dimensions, the cube is mapped to the fact with dimension tables table, the data is loaded, and the everything works as expected.

    The problem is with the charge itself, I noticed that it is too slow for my amount of sample data. After some research on the issue, I discovered a query in cube_build_log table, query the data is actually loaded.

    < SQL >

    <! [CDATA]

    SELECT / * + bypass_recursive_check cursor_sharing_exact no_expand no_rewrite * /.

    T4_ID_DAY ALIAS_37,

    T1_ID_POT ALIAS_38,

    MAX (T7_TEMPERATURE) ALIAS_39,

    MAX (T7_TEMPERATURE) ALIAS_40,

    MAX (T7_METAL_HEIGHT) ALIAS_41

    Of

    (

    SELECT / * + no_rewrite * /.

    T1. "" T7_DATE_TRUNC DATE_TRUNC. "

    T1. "" T7_METAL_HEIGHT METAL_HEIGHT. "

    T1. "" T7_TEMPERATURE OF TEMPERATURE. "

    T1. "" POT_GLOBAL_ID "T7_POT_GLOBAL_ID

    Of

    POTS. POT_BATH' T1)

    T7,

    (

    SELECT / * + no_rewrite * /.

    T1. "" T4_ID_DIM ID_DIM. "

    T1. "" ID_DAY "T4_ID_DAY

    Of

    LAUGHED. ("' DIM_DATES ' T1)

    T4,

    (

    SELECT / * + no_rewrite * /.

    T1. "" T1_ID_DIM ID_DIM. "

    T1. "" ID_POT "T1_ID_POT

    Of

    LAUGHED. ("' DIM_POTS ' T1)

    T1

    WHERE

    ((T4_ID_DIM = T7_DATE_TRUNC)

    AND (T1_ID_DIM = T7_POT_GLOBAL_ID)

    AND ((T7_DATE_TRUNC) IN "a long long list of dates for the currently processed cube partition is clipped")))

    GROUP BY

    (T1_ID_POT, T4_ID_DAY)

    ORDER BY

    T1_ID_POT ASC NULLS LAST,

    T4_ID_DAY ASC NULLS LAST]] > >

    < / SQL >


    View T4_ID_DAY, T1_ID_POT in the high level of the column list - here are identifiers of low level of my dimensions, which means that the query is not doing an aggregation here, because there is that one made by each pair of (ID_DAY, ID_POT).

    What I want to do is somehow to load data without doing that (totally useless in my case) the aggregation intermediaries. Basically I want it to be something like

    SELECT / * + bypass_recursive_check cursor_sharing_exact no_expand no_rewrite * /.

    T4_ID_DAY ALIAS_37,

    T1_ID_POT ALIAS_38,

    T7_TEMPERATURE ALIAS_39,

    T7_TEMPERATURE ALIAS_40,

    T7_METAL_HEIGHT ALIAS_41

    Etc...


    without aggregations. In fact, I can live even with this load query, such as the amount of data isn't that great, but I want that things in the right way to work (more or less ).

    Any chance to do?

    Thank you.

    I thought about it. There is a mistake in my correspondence, I've specified a column dim_table.dimension_id in a column field source of my section of the mapping, rather than the fact_table.dimension_id column, and (supposed) building tried to group by keys of the dimension tables. After this the definition of primary key did the trick (just the unique index, however, was not enough).

  • Excellent loading from data / last copy takes all the data from column to single column

    I use oracle apex apex.oraclecorp.com, I set up a new page to load data for one of the table Simple (with 4 column of numeric values and varchar 1) Ex: component (VARCHAR), Q1 and Q2 & Q3 and Q4 are (NUMBER) based in columns

    In the CSV, I gave the data like this when I move the CSV file on the page "under data source to load', click Next

    COMPONENT1ST QUARTER1ST QUARTERQ3Q4
    ABCD100200300400
    DCBA90807060

    I see all the data IS SHOWN IN THE single COLUMN: PFA screenshot

    I also tried with packaged Application: "loading samples" there in the same way

    Please let me know if I m missing something

    Thank you

    CHS

    Hello

    Looks like you need to specify the right column delimiter (comma, tab, etc.). Check which delimiter is used in your csv and make sure that you specify in the field "separator".

    Kind regards

    Vincent

    http://vincentdeelen.blogspot.com

  • Loading data: formatting of data for the timestamp column

    Hi all

    I have a table with a timestamp column, named created_date. I want to download this table data using load data page. but there is a problem when downloading data, I have a csv file in which data column created_date in two different format as follows,

    03/09/2013-03:33

    02/09/2013-15:24

    the above data throws an error ORA-01821: date format not recognized.

    In Data / Table Mapping page, I tried with HH12:MI MM/DD/YYYY: SS AM. What format should I use for am and pm?


    Please help me solve...

    Thanks in advance

    Lacombe

    I solved it by using the format MM/DD/YYYY HH:MIAM.

    Thank you

    Lacombe

  • ODI - SQL for Hyperion Essbase data loading

    Hello

    We have created a 'vision' in SQL Server that contains our data.  The view currently has every year and periods of Jan 2011 to present.  Each period is about 300 000 records.  I want to only load one period at a time.  For example may 2013.  Currently we use ODBC through a rule of data loading, but the customer wants to use ODI to be compatible with the versions of dimension metadata.  Here's the SQL on the view that works very well.   Is there a way I can run this SQL in the ODI Interface so it pulls only what I declare in the Where clause?  If yes where can I do it?

    Select

    CATEGORY, YEAR, LOCATION, SCRIPT, DEPT, PROJECT, EXPCODE, TIME, ACCOUNT, AMOUNT

    Of

    PS_LHI_HYP_PRJ_ACT

    Where

    YEAR > = "2013" AND PERIOD = 'MAY '.

    ORDER BY CATEGORY ASC ASC FISCAL_YEAR, LOCATION ASC, ASC, ASC, ASC, ASC, PERIOD EXPCODE PROJECT DEPT SCENARIO CSA ACCOUNT CSA;

    Hello

    Simply use the following KM to load data - IKM SQL for Hyperion Essbase (DATA) - in an ODI interface that has the view that you created the Source model. You can add filters to the source which are dynamically by ODI variables to create the Where clause based on the month and year. Make sure you only specify a rule of load method to load the data into the KM

  • Load data for a YEAR in a scenario of periodicals

    Dear,

    I think that it is an easy one, but I want to be 100% sure of what I should expect from HFM.

    Need to load balances for a YEAR in a regular scenario (periodic = default view, zeroview = periodic, consolidate the CDA = no)

    I have to specify the "CDA" string in the file, that I download?

    Periodic data are automatically calculated as the difference between consecutive CDA balances?

    Is there some advise/alert should I consider?

    A big thanks to all the world.

    Yes, you must specify the CDA in the files that you upload.  As a general rule, you want to avoid to load data to a recurring scenario for a YEAR, and vice versa.  In your example, a question if you have loaded data from one YEAR to a flow account and then the amount of your VOE in the month following to zero resulting in your not having data file is not a record for this intersection next month.  For example, if you have loaded $1,000 in January to sell your account but then the amount of the CDA in February is zero (rare but it happens) then you would not have a record of this account data from sale in February so that after loading you expect to see a zero number CDA in HFM.  However, with the zeroview the value of periodic, HFM expects a periodic amount and when you're not busy it assumes that the periodical (not a year) is zero and that it derives from the number of CDA.  In this case, it would show a value of February of the periodical of $0 and $1,000 for a YEAR when you expect $0 CDA and - periodical of $1,000.  Bottom line is that you should take periodic data to a periodic script and vice versa.

  • problem loading data from oracle table excel table / oracle

    Hi all

    I'm working on the form of data transformation 2 different sources to a target table only oracle.
    My first source - excel sheet.
    2nd source - oracle table.

    Creation of data servers, dsn(for excel), reversing everything done very well.
    When I try to display the data there is a problem.

    EXCEL - serial number column is displayed as 1.0,2.0,.3.0 etc. need 1,2,3,4,5...
    Oracle tables-> after that reverse it does not display the data types for the selected columns.

    I am not able to create a join and run.
    Help, please.

    Thank you
    Chaitanya.

    Hello

    Just a thought, why not u cast TO_NUMBER to the interface mapping for the serial number column?

    Thank you
    Guru

  • Loading data into excel

    Hello

    I tried to load data via great sheets, but my cube contain account numbers, when these numbers are entered in Excel, Excel automatically considers that these names are numbers.

    How can I get rid of all this help.

    Thank you
    RRB

    Then turn the text into excel or you can put a "in front of the number e.g. 101000"

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Need data files and rules file to create and load data for sample.basi

    I need emergency data files to build dimensions dynamically and also the rule files for loading data to the application Sample.Basic for Essbase 9. Kindly let me know if anyone can provide them. Otherwise any link frm where I can get it?

    Thanks in advance.

    Outline and all the data (calcdat.txt) are included in the zip, so you don't need to rebuild something with the loading rules

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • ODI loading data for Planning

    Hi all

    If you use IKM to load the data into the planning...

    What happens if we load data into 2 different plan types.

    Do need separate configurations through data charge directors in planning for each load.

    So if we have different interfaces, user should configure the planning before the execution of each adapter?

    Thank you

    Yes boths options seem fine if you cannot use the essbase KM.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Problem loading data for Planning Server through ODI

    Hi all

    I am facing problem while loading data for Planning Server through ODI.
    I created the interface with flat as a source of data as file below

    Location data load Cube name POV cost per seat
    Ctbn_PnL of Tidel Equifax, Budget, Working_2010, FY10, Wk1, Local 333
    Ctbn_PnL of Tidel Equifax, Budget, Working_2010, FY10, Wk2, Local 444
    Ctbn_PnL of Tidel Equifax, Budget, Working_2010, FY10, Wk3, 555 Local
    Ctbn_PnL of Tidel Equifax, Budget, Working_2010, FY10, Wk4, Local 666
    Budget, Working_2010, FY10, Wk5, Tidel Equifax, Local 777 Ctbn_PnL
    Ctbn_PnL of Tidel Equifax, Budget, Working_2010, FY10, Wk6, Local 888
    Budget, Working_2010, FY10, Wk1, PADI Ctbn_PnL Equifax, Local 222
    Budget, Working_2010, FY10, Wk2, PADI Ctbn_PnL Equifax, Local 111
    Budget, Working_2010, FY10, Wk3, PADI Ctbn_PnL Equifax, room 345
    Budget, Working_2010, FY10, Wk4, PADI Ctbn_PnL Equifax, Local 346
    Budget, Working_2010, FY10, Wk5, PADI Ctbn_PnL Equifax, Local 347
    Budget, Working_2010, FY10, Wk6, PADI Ctbn_PnL Equifax, Local 349

    And the target as an application of Planinng. Interface has been properly executed, but are not filled with data in the form of Plannning or Essbase.
    I obtained the information in the log file after running the interface:

    2010-05-20 18:45:38, 409 [DwgCmdExecutionThread:null:5] INFO: Oracle Data Integrator adapter for Hyperion Planning - free 9.3.1.1
    2010-05-20 18:45:38, 409 INFO [DwgCmdExecutionThread:null:5]: connection for application [PnL_ODI] [chnveltss03]: [11333] using [admin] username.
    2010-05-20 18:45:39, 800 [DwgCmdExecutionThread:null:5] INFO: successfully connected to the planning application.
    2010-05-20 18:45:39, 800 INFO [DwgCmdExecutionThread:null:5]: loading for the charge of planning options
    Name of the dimension: location like Parent child: false
    Order By entry charge: forgery
    Update the database: true
    2010-05-20 18:45:39, 815 INFO [DwgCmdExecutionThread:null:5]: beginning of the loading process.
    2010-05-20 18:45:39, 815 DEBUG [DwgCmdExecutionThread:null:5]: number of columns in the result set of source does not match the number of columns planning targets.
    2010-05-20 18:45:39, 831 [DwgCmdExecutionThread:null:5] INFO: type of load is a [member of the load dimension].
    2010-05-20 18:45:39, 831 [DwgCmdExecutionThread:null:5] ERROR: file [[Tidel, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 831 [DwgCmdExecutionThread:null:5] ERROR: file [[Tidel, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 847 [DwgCmdExecutionThread:null:5] ERROR: file [[Tidel, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 847 [DwgCmdExecutionThread:null:5] ERROR: file [[Tidel, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 847 [DwgCmdExecutionThread:null:5] ERROR: file [[Tidel, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 847 [DwgCmdExecutionThread:null:5] ERROR: file [[Tidel, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 847 [DwgCmdExecutionThread:null:5] ERROR: file [[Padi, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 862 ERROR [DwgCmdExecutionThread:null:5]: file [[Padi, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 862 ERROR [DwgCmdExecutionThread:null:5]: file [[Padi, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 862 ERROR [DwgCmdExecutionThread:null:5]: file [[Padi, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 862 ERROR [DwgCmdExecutionThread:null:5]: file [[Padi, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 862 ERROR [DwgCmdExecutionThread:null:5]: file [[Padi, null, null, null, null, null, null, null, null, null, null, null, Ctbn_PnL, Budget, null, "Equifax, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 878 [DwgCmdExecutionThread:null:5] ERROR: Record [[null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null]] has been rejected by Planning Server.
    2010-05-20 18:45:39, 878 [DwgCmdExecutionThread:null:5] INFO: planing cube refresh initiated.
    2010-05-20 18:45:44, 284 INFO [DwgCmdExecutionThread:null:5]: planning of the cube refresh operation completed successfully.
    2010-05-20 18:45:44, 284 INFO [DwgCmdExecutionThread:null:5]: load the process is complete.


    Please help me in this grace,
    Srinivas

    You are right. The table that you have described is in the format 'Generation' (the highest to the lowest level of the hierarchy). The KM of planning requires data to load in format "Parent/child". You must create multiple interfaces (one for each pair of generation) or use a view to make the transformation.

    While a view seems to be the easier choice, note that ODI will not be able to perform any analysis of lineage of data on the integration process.

    Hope this helps,
    Matt

  • Failed to load data to Essbase (IKM SQL for Hyperion Essbase (DATA)

    I am trying to load data to Hyperion Essbase. Unfortunately he is not going so well. I followed the instructions but I get this error "BUFFER_SIZE. I have not changed the default BUFFER_SIZE (it is set to < default >: 80) and I have not changed any other settings in the KM.

    Appreciate any thoughts...



    com.hyperion.odi.common.ODIConstants has No attribute * 'BUFFER_SIZE. "

    org.apache.bsf.BSFException: exception of Jython:
    Traceback (innermost last):
    "< String >" file, line 82, inside?
    AttributeError: class 'com.hyperion.odi.common.ODIConstants' has no attribute 'BUFFER_SIZE '.
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
    at com.sunopsis.dwg.codeinterpretor.k.a (k.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt (SnpSessTaskSqlI.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask (SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep (SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession (SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand (DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute (DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i (e.java)
    at com.sunopsis.dwg.cmd.h.y (h.java)
    at com.sunopsis.dwg.cmd.e.run (e.java)
    at java.lang.Thread.run (unknown Source)

    Published by: Chris Rothermel on April 13, 2010 15:44

    Hello

    What ODI and patch version you are running.
    Looks like you are using a KM newer than the java files that are located in the oracledi\drivers directory, I would say that you need to update the files java essbase for a newer version.
    In the last few patch releases memory size buffer has been added to the data load essbase KM for the use of the ASO, even if this meant you had to also update the java files as well as in the version.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

Maybe you are looking for

  • FN + F6/F7 do not work on satellite L755D-11W

    Hello I reinstalled the system, now, I have win 7 Home Basic (64-bit)All Fn functions work fine, but - combinations of keys (Fn + F6/F7) does not work.I wonder, what to do, I want this funtion to work. He is not comfortable when my screen have always

  • Unable to save since the sound card of Qosmio F30

    Hello - I have a Qosmio F30 standard with integrated sound card but can't direct recording of the map (use of Audacity). I have Vista Home premium, but try when it recording using the microphone and picks it up plenty of background noise.Have sorted

  • Unable to locate the UART.lvlib: Configuration.ctl of Visa

    Hi guys,. I run a problem after reinstalling labview 2014. When I open the existing project. He started looking for the control file. The file name is UART.lvlib: Configuration.ctl of Visa Then I click cancle and the errors appear on the front and th

  • Updates Windows 7 has failed: error code 646

    Every time my computer trying to install MS security updates most of the updates fail.  I currently have 19 updates which will not be installed.  I get the error code 646.  I have disbling my firewall while trying to install updates, but that did not

  • HP Simple pass for 8 of windows logon screen...

    Hey,. I just installed HP single pass and everything seems to work fine, but when I want to create a password to logon, it asks me my password and when I get it, it says "windows indicated connection sΘlectionnΘ password do not match. Please try agai