Query Performance problem in the apex

Hi all
I use

Select Address1 address2, address3, city, place, PIN code, siteid, bpcnum_0, contactname, fax, mobile, phone, Web site, rn from (select Address1 address2, address3, city, place, pincode, siteid, bpcnum_0, contactname, fax, mobile, phone, website, dense_rank () (order of contactname, Address1) as rn, row_number() over (partition contactname, Address1 order by contactname, Address1) as rn1 vw_sub_cl_add1 where siteid = v ('P10_SITENO') and bpcnum_0 = ('P10_CLNO') v) emp where rn1 = 1 and rn > = v('P10_) RN')

the query above to extract the details of a view, an area of pl/sql
paging also works very well.
That is to say only 4 records at a time will be displayed.

It comes
It takes 1 minute and 5 seconds to display, then set records.

Please, could any tell me how to reduce the time of rendering the page. ?

Thanks in advance
Good bye
Sonny_starck

If it's really true, then the use of bind variables in the query is fast, you can rewrite your query in your application to use bind variables?
Try to rewrite your query to use: P10_SITENO instead of v ('P10_SITENO') etc. If possible.

You can find more details on the http://www.inside-oracle-apex.com/2006/12/drop-in-replacement-for-v-and-nv.html of Patrick Wolf blog

Tags: Database

Similar Questions

  • query works not in the APEX but works in TOAD

    Hello

    I had major problems trying to build a LOV cascading. I narrowed the problem down to SELECT which it depends.

    I have the following:

    SELECT pasv.vendor_name
    pasv.vendor_site_code
    pasv.vendor_id
    pasv.vendor_site_id
    OF po_asl_suppliers_v pasv
    po_asl_documents_v padv
    po_asl_attributes_v paav
    WHERE pasv.owning_organization_id = 26
    AND pasv.item_id = 131089
    AND pasv.asl_id = padv.asl_id
    AND pasv.asl_id = paav.asl_id
    AND padv.document_type = 'COVERAGE '.
    ORDER BY pasv.vendor_name
    pasv.vendor_site_code

    This query works well as a TOAD, but when I run it in APEX "SQL commands" it comes up with the error "ORA-01722: invalid number ' and when I try to generate a report about this no return value (the values are returned when you run the same query in TOAD and SQL Developer)."

    Any advice would be great.

    Thank you
    Chris

    Work around:

    SELECT RPAD (SUBSTR (pasv.vendor_name, 1, 20), 20,' ') | » '|| RPAD (SUBSTR (padv.document_type, 1, 20), 20,' ') | RPAD (SUBSTR (padv.document_num, 1, 10), 20,' ') PROVIDER
    pasv.vendor_id
    OF po_asl_suppliers_v pasv
    po_asl_documents_v padv
    po_asl_attributes_v paav
    WHERE pasv.owning_organization_id = 26
    AND pasv.item_id = 131089
    AND pasv.asl_id = padv.asl_id
    AND pasv.asl_id = paav.asl_id
    ORDER BY pasv.vendor_name
    pasv.vendor_site_code

    There seems to be a problem with the 'COVER' in the first SQL. Even if it is that a field of VARCHAR2 in the APEX database doesn't. It works somehow the other tool we use.

  • On query performance problem. Need help.

    It is essentially a performance problem. I hope someone can help me with that.

    Basically, I have four old masters (150000 records), (100000 records) Child1, Child2 (50 million records!), child 3 (10000 + records)
    (please forgive the alias).

    Each record in the master has now more than one matching record in each table child (one to many).
    Also there may be any record in any or all of the tables for a particular master record.

    Now, I need to get the maximum of last_updated_date for each master record in each table 3 child and then find the maximum of
    the three obtained last_active_dates from the 3 tables.
    for example: Master ID 100, to interrogate Child1 for all Master ID 100 records and get the max last_updated_date.
    Same for the other 2 tables and get the most out of these three values.
    (I also need to deal with cases where no trace may be found in a child table to a Master ID)

    Write a procedure that uses sliders that the value of each of the performance hits of child table
    evil. And that's, I need to know the last_updated_date for each master file (all 150000 of them). It will probably take days to do this.

    SELECT MAX (C1. LAST_UPDATED_DATE)
    MAX (C2. LAST_UPDATED_DATE)
    MAX (C3. LAST_UPDATED_DATE)
    OF CHILD1 C1
    CHILD2 C2
    CHILD3 C3
    WHERE C1. MASTER_ID = 100
    OR C2. MASTER_ID = 100
    OR C3. MASTER_ID = 100

    I tried the above, but I got an error in tablespace temp. I don't think that the application is good enough at all.
    (The GOLD clause is to take care of any records in a child table. If there is an AND, then the join and then select
    No, not even if there is no record in a child table, but valid values in the other 2 tables).

    Thank you very much.

    Published by: user773489 on December 16, 2008 11:49

    You want alias to this field then.

    SELECT MAX (C.LAST_UPDATED_DATE)
    FROM
    (select child1_master_id MASTER_ID, field2, field3,... field4 from CHILD1 UNION ALL
     select child2_master_id MASTER_ID, field2, field3,... field4 from CHILD2 UNION ALL
     select child3_master_id MASTER_ID, field2, field3,... field4 from CHILD3) C
    WHERE C.MASTER_ID = 100
    

    If do you something like that, and explicitly list the columns you want.

    Edit: for something like a specific query for a MASTER_ID...

    SELECT MAX (C.LAST_UPDATED_DATE)
    FROM
    (select child1_master_id MASTER_ID, LAST_UPDATED_DATE from CHILD1 where child1_master_id = 100 UNION ALL
     select child2_master_id MASTER_ID, LAST_UPDATED_DATE from CHILD2 where child2_master_id = 100 UNION ALL
     select child3_master_id MASTER_ID, LAST_UPDATED_DATE from CHILD3 where child3_master_id = 100) C
    WHERE C.MASTER_ID = 100
    

    That should give you very good performance by raising a record. But a better idea, as indicated, would be to get it all at once with a sql:

    SELECT MASTER_ID, MAX(C.LAST_UPDATED_DATE)
    FROM
    (select child1_master_id MASTER_ID, LAST_UPDATED_DATE from CHILD1 UNION ALL
     select child2_master_id MASTER_ID, LAST_UPDATED_DATE from CHILD2 UNION ALL
     select child3_master_id MASTER_ID, LAST_UPDATED_DATE from CHILD3 ) C
    GROUP BY MASTER_ID
    

    This will give you the max for each MASTER_ID in a sql without a cursor.

    Published by: tk-7381344, December 16, 2008 12:12

  • Performance problems since the update of 2015.6

    Hello together,

    Since I updated lightroom yesterday to version 2015.6 the performance on my machine is incredible bad.

    My system:

    -MacBook Pro (retina, mid-2012)

    -2.3 GHz Intel Core i7

    -16 GB 1600 MHz DDR3

    -NVIDIA GeForce GT 650M 1024Mo

    -OS X El Capitan (10.11.5)

    Lightroom synchronizes actually 1400 photos, but even if I stop this process, LR uses CPU so I can't work in other programs like Photoshop (updated yesterday, too). I detected this performance problem as PS freezes when I used the Healing Brush tool.

    LR.jpg

    I don't know what's happening LR, but is not ideal for my workflow.

    Does anyone have an idea?

    If I close LR I can work in PS, even though the Healing Brush tool is slower than in the version before. But I can't always close and open LR.

    Thank you

    regards Denis

    There are already dozens of threads discussing the same question in this forum as well as in the official comments forum here: https://feedback.photoshop.com/photoshop_family/categories/photoshop_family_photoshop_ligh troom

    You can restore 6.5.1 if you wish. Here's how: How do I roll back to Lightroom 2015.1.1 or Lightroom 6.1.1? Lightroom Queen

    Just replace 6.1.1 with 6.5.1

  • Performance problem with the guest operating system

    Hello gurus - I have one of the most unusual questions, I ran into a few time.   Here it is:

    I have a guest w2008R2 OS - installed in a 5.1 version of VM 7 vSpherer - which has an Emulex LPe 12002-E 8 GB dual HBA port is assigned to the as 2 PCI (1 device per port) via DirectPath devices

    The guest operating system has an application that performs backup tests.   WWPN of the HBA is zoned in the fabric and declared in the interface of Compellent Storage Center and everything is good here. the application creates a replay of a volume, mount it as a drive, read all the blocks, then dismantles and moves then repeat the process with what ever additional volumes should be safeguarded.   It is so cool I have do this with a VM right and not a physical host?  (would be cool if I had work NPIV but that's another story, not with technology HBA vendor support...)

    So, what's the problem?   Well, the bed the mounted disk backup will work extremely slow as slow as 5MB\sec flow, then during any other activity takes place on the system the flow jumps up to 150MB\sec all of a sudden and will continue to run fast for a long time and finish jobs on time with each work taking a little more than an hour.  But when the slow going the same work (all jobs are the same, they are simply by reading all the blocks of a bunch of 200 GB volumes and wrote a set of NULL value, do not forget this is just test) takes up to 10 hours to complete.

    I think that what is the culprit?   POWER MANAGEMENT.

    I have just about each performance counter and the journal of events known to man and I don't see anything abnormal on the table, ESXi virtual machine, nothing.  But what I see in the newspapers is any kind of activity that could happen in a transparent manner to save energy.

    FACTS on this case:

    -Reservation of memory is defined (it must be for directpath otherwise, you can't power on the virtual machine)

    -Booking CPU - has not been defined before an hour ago, I gave her just a small reserve for this rule, so I do not know if this will help alleviate the slow flow, or have an effect on what I think is a question of power (unit linked) management

    -windows power management is set to high performance never-ever - but not until an hour there doesn't I changed advanced settings for USB Disabled------turned off.  PCI peripheral energy saving setting was already off.  Which is in contradiction with my theory.  But maybe, just maybe, something is broken somewhere, or could be a bug... blah blah blah

    -the HBA, Emulex, as it appears in Device Manager in the guest OS doesn't have a power management tab to disable the feature, I don't see that the adapter VMXNET3 had this tab, and I turn off the power save setting to be conservative, do not think that it was the culprit, but as there is no real data running through the NIC.  The data is simply to be read from the disk mounted, not copied on any wire Ethernet.  never the less, in the Emulex properties is a Details tab and a selection of power data, and performance data indicate you what power mode, the device is, as well as feeding supported systems.  D0: full power, but D3, which shows as supported by the device, is an energy saving mode.   I could not yet attend the setting at the time of slow activity, my good intention as soon as possible. He is currently shooting in D0, full power.

    This is so what I have to offer on this, I hope that its just...   If I'm missing something, or if others have experienced this massive loss of unknown flow without reason and have resolutions, please share.

    If there is a way for me to disable power on the Emulex HBA management inform you please, I don't mind if I stop to enter the firmware, if that was even possible.  But if possible I would disable it immediately before any new tests are executed.

    The root cause of this issue has never been identified by VMware support which for me is a question of ugly.  Reservation of CPU for the VM using directpath IO with HBA installed in it corrected the problem.  I'd be leary with direct path IO.

  • Troubleshooting performance problems with the mail server on the virtual computer

    I have a virtual machine with a mail server installed on a windows OS. But all of a sudden, my clients on the server response times decreased significantly.

    This could be the cause of this problem?
    It could be to do with the network, storage, or resources, but I don't know what steps I could take to determine fault

    Would someone mind helping me with the steps I could take to solve this problem?

    Thank you

    The below document page 20 and 21 accurate meters, you should look at the virtual machine level.

    http://www.VMware.com/files/PDF/Exchange_2010_on_VMware_-_Best_Practices_Guide.PDF

    These counters can be used for any virtual computer not only Exchange.  If you are familiar with ESXTOPs I believe that the best way to deep dive in the diagnosis of performance problems.  Also here is a link for these parameters.

    http://KB.VMware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalID=1008205

    This link takes you to http://communities.vmware.com/docs/DOC-9279 which has the description on the stats.

  • tool to diagnose performance problems in the oracle database

    Is there a tool to diagnose performance problems in queries and stored procedures in oracle similar to sql for sql server Profiler

    Thank you

    You can use oem oracle enterprise manager to diagnose and monitor the database.

    Chapter 10: Monitoring and Tuning of the database (see the link oracle obe series, procedures step by step with the presentation of the screenshot)
    This chapter shows you some of the monitoring and tuning of the operations performed by Enterprise Manager.
    http://www.Oracle.com/technology/OBE/2day_dba/monitoring/monitoring.htm

    Reference: monitoring and Tuning of the database

    http://download.Oracle.com/docs/CD/B14117_01/server.101/b10742/montune.htm

    hope this helps you.

    Published by: rajeysh on July 14, 2010 21:28

  • Performance problem on the SQL query that does not use the primary key index

    Hello!

    I have some performance issues on a single SQL query (Oracle 10 g).
    I could solve the problem by using the INDEX indicator, but I would like to know WHY this is happening.

    * Tables *.
    create table jobs)
    ID number (5) not null,
    name varchar2 (100),
    primary key constraint Job_PK (id)
    )
    /
    -Record count: 298

    create table Comp)
    integer ID not null,
    name varchar2 (100),
    primary key constraint Comp_PK (id)
    )
    /
    -Record count: 193

    -Relation m: n
    create table JobComp)
    integer ID not null,
    id_job integer not null,
    id_comp integer not null,
    primary key constraint JobComp_PK (id),
    unique key constraint JobComp_UK (id_job, id_comp),
    Constraint JobComp_FK_Job foreign key (id_job) refers to Job (id),
    Constraint JobComp_FK_Comp foreign key (id_comp) makes reference Comp (id)
    )
    /
    create index JobComp_IX_Comp on JobComp (Cod_Comp)
    /
    create index JobComp_IX_Job on JobComp (Cod_Job)
    /
    -Record count: 6431

    * Ask *.

    When I run this query, the execution plan shows the index using (JobComp_PK and JobComp_IX_Comp).
    No problem.

    Select JobComp.*
    of JobComp
    Join jobs
    on Job.id = JobComp.id_job
    where JobComp.id_comp = 134
    /
    -runs in 0.20 sec

    But when I add the field 'name' of the work table the plan uses full access table to the table of work

    Select JobComp.*, Job.name
    of JobComp
    Join jobs
    on Job.id = JobComp.id_job
    where JobComp.id_comp = 134
    /
    -runs in the 2.70 dry

    With the help of the index

    Select / * + INDEX (Job Job_PK) * /.
    JobComp.*, Job.name
    of JobComp
    Join jobs
    on Job.id = JobComp.id_job
    where JobComp.id_comp = 134
    /
    -runs in 0.20 sec

    * Doubt *.

    This behavior is correct?

    PS. : I tried to recalculate the statistics, but nothing changes:

    analyze the job calculation table statistics.
    /
    change the statistical calculation of index Job_PK reconstruction;
    /
    Start
    dbms_utility.analyze_schema (sys_context ('userenv', 'current_schema'), 'CALCULATE');
    end;
    /

    [of]
    Gustavo Ehrhardt

    Gus.EHR wrote:
    Hello.
    I'm sorry for the plan unformatted.
    The execution time of the querys "without field name' and 'with the field name with suspicion" are equal.
    He has no problem caching, because I get the plans of the sequence different from the querys and repeated the performance. The result is always the same.

    I don't think that there is no problem with oracle crossing LOOP IMBRIQUEE to the HASH JOIN when you include the field name and this should be the expected behavior. But it seems that your WORKING table has a degree of parallelism set against what is causing the query to run in parallel (as JOB table is now available with full table scan, instead of indexed access earlier). It could be that the parallel execution is contributor to extra Runtime.
    (a) do you know why the degree of parallelism on the WORK table has been defined? Do you need it?

    You can see if the following query provides a better response time?

    select /*+ NOPARALLEL(JOB) */ JobComp.*, Job.Name
      from JobComp
      join Job
        on Job.id = JobComp.id_job
     where JobComp.id_comp = 134
    
  • Performance problem - generating the SQL query

    Hi experts,

    We are facing a problem of performance on our dashboards. Used for testing dashboard page contains a prompt with two parameters (month & location) and 3 reports (1 table & 2 graphics). We put in the effort to optimize the speed of the query and the database returns the results after 1 sec. However, after changing values prompt, it takes up to 15 years before the request is sent to the database (based on the log).

    We have done the following tests for an overview. Have tests in a test environment with clustered servers to exclude the possible implications of users numbers.

    (1) raise the reports with the same prompt values
    -> Page is returned immediately and the log shows that the results are retrieved from the cache

    (2) run the page even with fewer reports
    -> Removing a report from the page reduces the time to send the request to the database to 8 s

    (3) check the traffic of a browser with sniffer
    -> No. requests were sent 15-year timespan. Once the request is sent results instantly appear and the page is built in no time.

    Can anyone provide an idea as to what happens between the time the user's request is triggered via the guest of dashboard and the request is sent to the database?

    Hi Karl,
    I'm so glad you have at the bottom of it. You might find that oracle can help you with more information about the ODBC via a SR statement.

  • Performance problem for the ADF calendar

    Hello

    We intend to apply the calendar adf for one of our customers. They will use the calendar as a planner. There will be at least 350-500 activities for a provider and a month.

    We are filtering of calendar data by selecting provider from the drop-down list box. However, execution of the view object takes about 4-5 seconds and activities appear 15 seconds. Total waiting time may be 20 seconds.

    How can I improve performance? Customer will use the calendar as a scheduler of tasks online. Therefore, it must be very fast.

    JDeveloper Version: 11.1.1.6.0

    Please note that all the components are bind to the backup of the attributes of bean whose calendar itself jsff layout.

    Steps for filtering:

    1 combo box action in backing bean:

        public void onResourceChange(ValueChangeEvent valueChangeEvent) {
            valueChangeEvent.getComponent().processUpdates(FacesContext.getCurrentInstance());
            if (valueChangeEvent.getNewValue() != null) {
                if(valueChangeEvent.getNewValue() != valueChangeEvent.getOldValue()) {
                    applyFilter();
                }
            }
            else {
                removeFilter();
            }   
            
            RequestContext context = RequestContext.getCurrentInstance(); 
            context.addPartialTarget(getC1());
        }
    

    2 ApplyFilter and removeFilter in backing bean methods:

        private void applyFilter() {
            DCBindingContainer dcbindings = (DCBindingContainer)ADFUtils.getBindingContainer();
            XxntcCalendarAMImpl am = (XxntcCalendarAMImpl)dcbindings.getDataControl().getApplicationModule(); 
            DCIteratorBinding garageIterator = ADFUtils.findIterator("XxntcGarageInfoVO1Iterator");
            Row currentGarageRow = garageIterator.getCurrentRow();
            Number resourceId = (Number)currentGarageRow.getAttribute("ResourceId");
            am.applyCalendarFilter(resourceId);        
        }
        
        private void removeFilter() {
            DCBindingContainer dcbindings = (DCBindingContainer)ADFUtils.getBindingContainer();
            XxntcCalendarAMImpl am = (XxntcCalendarAMImpl)dcbindings.getDataControl().getApplicationModule();  
            am.removeFilter();
        }
    

    3. the application model Impl applyCalendarFilter removeFilter methods and

        public void applyCalendarFilter(Number resourceId) {
            XxntcServiceCalendarEOVOImpl calendarVo = (XxntcServiceCalendarEOVOImpl)getXxntcServiceCalendarEOVO1();
            System.out.println(new java.util.Date());
            ViewCriteria vc = calendarVo.createViewCriteria();
            vc.setName("fiterCalendar");
            ViewCriteriaRow vcRow = vc.createViewCriteriaRow();
            vcRow.setAttribute("ResourceId", "= " + resourceId);
            vc.insertRow(vcRow);
            calendarVo.appendViewCriteria(vc);
            calendarVo.executeQuery();
            
            /*calendarVo.setWhereClause(null);
            calendarVo.setWhereClause("RESOURCE_ID = :pResourceId");
            calendarVo.defineNamedWhereClauseParam("pResourceId", null, null);
            calendarVo.setNamedWhereClauseParam("pResourceId",resourceId);         
            calendarVo.executeQuery();*/
            System.out.println(new java.util.Date());
        }
        
        public void removeFilter() {
            XxntcServiceCalendarEOVOImpl calendarVo = (XxntcServiceCalendarEOVOImpl)getXxntcServiceCalendarEOVO1();
            calendarVo.removeApplyViewCriteriaName("fiterCalendar");
            //calendarVo.removeNamedWhereClauseParam("pResourceId");
            calendarVo.executeQuery();
        }
    

    Kind regards

    Anil

    Hi all

    Problem is solved.

    There are three important points for the solution.

    1 oracle has proposed to use a separate iterator for suppliers. Here is the explanation of SR Oracle:

    @ The iterator binding of the calendar and the iterator binding of the

    @ suppliers are the same, there is no point at all. Normally, the

    @ the iterator binding calendar is for the events calendar and the provider

    @ the iterator is on another iterator which gives information about the provider, the

    @ provider in the timing of the event is the foreign key to the iterator of provider.

    @You need to create a table for the vendor and then set object entity / view

    @ the object on the table, an iterator will be generated. The provider_id in the

    @ event table should be the foreign key to the table of the provider and the

    @ providerDefinition must point to the iterator of the vo provider.

    @ You cannot use the same table to set the calendar events and providers,

    @ This just doesn't make sense.

    2. I am able to set the iterator range size if I use a separate iterator for suppliers. Otherwise, ADF throws it npe: ' java.lang.NullPointerException: ADFv: calendar activity Id and Vendor Id cannot be null. "

    3. I forgot to use parameters of connection (StartDayTime, EndDayTime and time zones) in my query where clause. Creation of a criteria like this where clause "((HrCalendarEvents.START_TIME > =: StartDayTime) AND (HrCalendarEvents.END_TIME)"<= :enddaytime="" )="" )="" "="" solves="" fetching="" all="" rows="" problem.="" now="" it="" fetches="" just="" current="" time="">

    Kind regards

    Anil

  • Query performance problem

    I have two schemas of two databases.

    When I check the sql plan, both the schema contains diffrently (one is underway for a full table scan and a scan of systematic index range).

    Both the scheme almost similar kind of data, indexes, and charges.

    What is causing the performance of sql

    in the second plan, the optimizer expects the analysis of range on IDX_TSK_ID step 5 to return to only 14 lines and decides it's a good idea to join the second TB_TRANS_MSTR table with a nested loops join (make a loop on 14 TASK_INSTANCE results and do a search on each iteration of the use of the PK_TRANS_ID index).

    In the foreground, the optimizer decides to read TB_TRANS_MSTR (containing 978 lines) and build a table of hash in memory of the results - and then probe against the TASK_INSTANCE second set.

    The next question is: which plan is most suitable and translates into better performance? The chances are high that the best plan is one that includes an estimate more fitting of the cardinalities. These estimates are based on a simple arithmetic (more or less) - and they depend on the table and column statistics. So the dba_tab_column entires Swen W. mentioned would be useful. In addition the text of the query would probably we shed some light on the question.

  • Performance problem of the migration of more than 500 million using OWB records.


    I need to migrate 500 millions of records. The structure of the table is similar to the table of scott.employee with the only exception being that the HIREDATE column is NOT NULL. The table is partitioned by using month-wise column HIREDATE. The target table is the same as the source.

    I used a direct correspondence with the approach based on the game but it run for more than 10 hours. I then tried to create another temporary table to store the date. I've first grouped column date and got the number for number of records by date. I have it stored in a temporary table. Then I took the first 10 dates and it is attached with the source table and inserted the result set into my target table. The performance is slightly better but still not acceptable.

    Can someone please help?

    I have the necessary indexes in place and oracle is using them. I don't know a lot about the ELP language (using OWB 10 g 2 and so I can use it), but if I understand good documentation, it changes the identity simply and does not move the data. In this case, I can't use it because I want the source data too.

    You will appreciate all the assistance. Thanks in advance.

    Concerning

    I ask you to remove data from the source.

    All I'm asking is:
    you said that your target tbal has all the lines.
    So instead of insert source to target why not make your transformation on the source table and rename the source table to target the tbalename and remove the taret table and build your constraints on source table IE

    Staged tbal - CBC - 500 million lines
    Target - TGT - 0 rows table

    create the table src_copy in select * from src - so now you have a copy of your table in the source
    do your cleaning on lines 500 million on the table of data src_copy
    build constraints on table src_copy
    drop table tgt
    Rename src_copy table table tgt.

    Do not see a problem in doing this.

  • a defective battery in a tablet will cause performance problems of speed

    a defective battery in a tablet will cause a matter of output speed

    Hello

    Sometimes it can cause a performance problem, visit the following link to monitor how to take care of the battery:

    Taking care of your laptop battery

    http://Windows.Microsoft.com/en-us/Windows7/taking-care-of-your-laptop-battery

    In order to increase the performance of Windows XP please visit the link below:

    Windows XP performance

    http://TechNet.Microsoft.com/en-us/library/bb457057.aspx

    Reference: PC slow? Optimize your computer for peak performance

    http://www.Microsoft.com/athome/Setup/optimize.aspx

    Note: as long, the Tablet is connected to a power source it shouldn't affect the performance in itself. However, bad battery itself a performance problem.

  • ViewObject range Paging performance problem

    Hi all

    I am facing a performance problem with the implementation of an obligation to programmatically add a number of extra where the parameters of the clause (using bind) variable in combination with range paging.

    My code looks like this

    ...
    
    ApplicationModule am = Configuration.createRootApplicationModule("services.DossierAM", "DossierAMLocal");
    ViewObject vo = am.findViewObject("DossierListView");
    
    // apply programmatic view criteria
    ViewCriteria vc = vo.createViewCriteria();
    ViewCriteriaRow vcr = vc.createViewCriteriaRow();
    vcr.setAttribute("Reference", "15/%");
    vc.addElement(vcr);
    vo.applyViewCriteria(vc, true);
    
    
    // enable range paging
    vo.setAccessMode(RowSet.RANGE_PAGING);
    vo.setIterMode(RowIterator.ITER_MODE_LAST_PAGE_PARTIAL);
    vo.setRangeSize(50);
    vo.scrollToRangePage(5); // Cause a java.sql.SQLException: Parameter IN or OUT missing for index.....debugging learned that the :vc_temp_1 bind variable is not filled
    // vo.scrollToRange(250); // Cause a java.sql.SQLException: Parameter IN or OUT missing for index.....debugging learned that the :vc_temp_1 bind variable is not filled
    
    ... 
      ...
    

    I found 2 solutions, but they both require an application of additional database that is, performance wise, is not acceptable.

    The first solution is to slip into an additional call to exectueQuery() before the call to function scrollToRangePage (int) or scrollToRange (int).

    The second solution is to use the method (int) setRangeStart instead of variants scrollToRange (Page). This method performs also 2 database calls.

    My question to you:

    Is there another way to satisfy the requirement of programming add a certain number of parameters of the additional where clause (using the variable binding) in combination with the pagination of the range without the need to perform queries of database 2?

    The code is tested with JDeveloper, 11.1.2.4.0, and 12.1.3.0.0 and behaves the same on both versions.

    Kind regards

    Steven.

    Have you tried to create truly VC with bind variable (rather than use binding implied var created by frame)?

    Something like: http://www.jobinesh.com/2010/10/creating-view-criteria-having-bind.html

    Dario

  • Deployment of the APEX, for commercial purposes

    Dear Experts,

    I build application APEX for the customer. I have two options for deployment.
    (1) to prepare the necessary infrastructure for a dedicated server in the client's site?
    2) host in one of the APEX hosting companies?

    What are the short-term and long-term benefits and disadvantages of each option from the commercial point of view?

    Best regards
    Mahmoud

    -How to estimate the cost (for example in the proposal to the client) to get the actual benefit?

    Not really the Apex technology related, so I'll keep away from it.
    >
    -Are there a difference between sales application with possibility of access (with administrator access or developer) code on the one hand and sales demand without giving this privilege on the other? I focus on the commercial point of view.
    >
    Involvement are not technological partners.
    >
    -A_ are there some specific problems at the APEX for the comparison between the two approaches (internal supplier or accommodation)?
    >
    A public facing Internet application must be safer than an Intranet, or internal, application. But I see no difference related to the choice of hosting.

    See you soon,.

Maybe you are looking for