Multiple queries in a data model - Performance?

Hi experts,

I created a data model and defined queries 7 (isolated; not connected together). I want to create 7 reports based on queries 7 (for example query1 used in Report1, query2 used into Report2, etc.).

The consolidation of 7 queries in a data model affect performance? How the BEEP collects data - is only the query used processed, or each time the entire data model?

Thanks in advance!

Hello

According to the data that you want to use in your report.

In a report, you can use data from more than queries.

We recommend that you create 1 data model for each report, for a good performance with all applications you need.

7 queries fetch data slower than 1 query in your data model.

Kind regards

Liviu

Tags: Business Intelligence

Similar Questions

  • Need to use values from the first query in other queris to the data model

    Hello
    Here is my requirement-

    I use the data model to run multiple queries. The first query, I get 10 records. Now, I want to use these 10 records in the second query to get my final result. I am not able to use the sub query as the two motions are quite long and complex.

    Select distinct Bishop of emp

    Select empno, emp where Bishop in (: Bishop)

    I can't use: Bishop because it will give only the last value stored at Archbishop. Is it possible to be able to use all the values from the first query in the second query using the data model?

    Hello

    Are you sure that you have your "dataStructure" configured correctly? Try this simple example:

                                                                          
    

    Hope this helps

    Andy

  • Multiple datasets in single data model

    Dear experts,

    We have created a unique data model that has several sets of data sharing common to BI 11 g Publisher parameters.

    We are facing a problem of performance (data recovery is very slow) while to fetch data from database on execution.

    How to improve the performance in this case?

    Your help will be very appreciated.

    Thanks in advance!

    DB BLOB data recovery should not be a problem. But just to make sure that you remove your SQL BLOB type and see to it there improvements?

    Unfortunately I can't find any documents which mention the process of execution of the data sets.

    But you can do a proof for yourself:

    Add sysdate with timestamp on each set of SQL data and you can see the date with timestamp in the output xml (data view tab) which will show you what set of data is performed first and then what that and etc...

    It will be executed in the order of your data set creation. You can see the structure of the data set in the Structure tab.

    Have you considered the approach below?

    I recommend you to do.

    Is there is no adjustment to make in your SQL, then you can use triggers of events (before the data type).

    You can create a package that runs your sql DB and insert the data into a temporary table and then just ask the temporary table on the data set.

  • How to pass multiple values to a data model parameter

    I have a data model, where I put a setting like this.

    Select col1, col2, col3 from table1 where col1 =: param1

    This works very well for a single value that I pass the param1 parameter.

    How can I get this to work when I want to send multiple values for param1. I already checked the box "Multiple values" and tried to change the sql for this code.

    Select col1, col2, col3 from table1 where col1 in: param1

    Looks like I'm missing something here.



    Thanks in advance!
    Kris

    you need to write it as:

    Select col1, col2, col3 from table1 where col1 in (: param1)

    It works for me.

    Klaus

  • How to store the result of a query in a variable in the data model

    In a model of date I want to do something like this

    < name of dataTemplate = than one dataSourceRef "HURDetail" = "BRM_DATA_SOURCE" >
    < Parameters >
    < parameter name = "PARAM_THRESHOLD_VALUE" dataType = "number", defaultValue = "0" / >
    < / Parameter >

    < SQLStatement instance name = "Q1" >
    <! [CDATA]
    SELECT count (*) FROM TABLE_NAME
    []] >
    < / sqlStatement >

    -I want to assign the output of the above query to PARAM_THRESHOLD_VALUE so I can use it in future requests...
    -My query is complex, for reason of performance I want to store the output of the query.

    Please suggest how do...
    Made a package with the PARAM_THRESHOLD_VALUE parameter and the function value entry assigns the done variable to work?

    Thank you
    Shiva

    Hey Shiva,

    If you want to use the value of the counter in the other queries in the data model, you can have an alias for the extraction of County and use it as a bind variable in other queries as


    SELECT count (*) PARAM_THRESHOLD_VALUE
    FROM TABLE_NAME
    ]]>

    and can use it in another query as


    SELECT XYZ
    FROM TABLE_NAME2
    WHERE XXX = *: PARAM_THRESHOLD_VALUE *.
    ]]>

    Hope this is what you want.
    Thank you.

  • Question by showing the data model in reports of 11 g

    Hello

    I'm working on forms and reports on 11 g. I have a problem with the reports.

    After opening an existing report, I am unable to view the queries in the data model. The data model page is open, but immediately, it pops up a window saying generator stopped working. He has no problem working with the rest of reports such as the use of the document layout, etc.

    Could someone help me with this problem?

    (Sorry, i, m not able to add any image with this)


    Thanks in advance.


    You use the 64-bit Oracle Forms and reports 11 GR 2 version? If you are, it's a known bug with 64-bit report designer. As long as you use not Sources of pluggable data (PDS) in your reports, you can implement workaround in the following article:

    http://Pitss.com/us/2012/10/10/reports-Builder-11gr2-crashes-when-viewing-data-model/

    This should solve the problem with the help of the data model in the generator of reports without it crashing.

    Thank you

    Scott

  • Is conditional SQL processing of XDL available data models

    Hello

    Is it possible to have a CONDITIONAL STATEMENT in .xml (data model), based on a SQL statement that was going to be executed? Sense, based on a PARAMETER value, only the TWO queries should be run.

    Ex: Lets take a system used, or if the report is run
    SUMMARY = > summary query must be executed
    DETAIL = > detailed queries to run.

    < IF P_REPORT = SUMMARY >
    < DO >
    -THE QUERY SUMMARY
    < END IF >

    < IF P_REPORT = DETAIL >
    < DO >
    -DETAIL QUERY
    < END IF >

    Could if it you please let me know the feasibility of this in XML BI. Ask that you please point me to the documentation, if any.

    Thank you
    Vijaya

    OK, I withdraw,

    If it is required to be achieved. then that,

    1. you need to add a report parameter values valid as 'SUMMARY' and 'DETAIL '.
    2. you have 2 Add two separate petitions, no need to join them.
    3 just put separate way queries in the data model or concatenated sql same datasource.
    4 each of the query, to equate them with the setting something like
    5 ' SUMMARY' =: PARAM_REPORT_TYPE in the query summary
    6 ' DETAIL' =: PARAM_REPORT_TYPE in the query in detail
    7. in execution, that which, one of the query to execute and extract the data.
    8. Similarly, in the model, you can add the conditional creation of models.

  • Impacts on the performance of the attributes from the features of data model design

    I'm trying to understand the implications of the performance of two possible data model design.

    Here is my structure of the entity:

    Global > person > account > option

    Generally, when running, I instantiated a person, a single accountand five option's .

    There are various amounts determined according to the age of the person who should be assigned to the correct option.

    Here are my two designs:

    Design a

    attributes on the entity of the person :
    age of the person
    its option 1 amount
    its option 2 amount
    its option 3 amount
    its option quantity 4
    its option 5 amount

    attributes on the option endity:
    amount of the option

    support table rules:
    option = amount
    its option 1 amount if the option is number 1
    its option 2 amount if the option number 2
    its option 3 amount if the option number 3
    its 4 option amount if the option is number 4
    its option 5 amount if the option is number 5

    Two design

    attributes on the entity of the person :
    age of the person

    attributes on the entity of the option :
    amount of the option
    of the option option 1 amount
    of the option option 2 amount
    of the option option 3 amount
    of the option quantity 4
    of the option option 5 amount

    support table rules:
    option = amount
    of the option option 1 amount if the option is number 1
    option 2 amount option if the option number 2
    of the option option 3 amount if the option number 3
    the option amount 4 If the option is number 4
    option 5 option amount if the option is number 5

    Given two models, I can see what looks like an advantage for a design that, when running, you have less attributes (6 on retirement member + 1 on each of the 5 options = 11) as two Design (1 on retirement members + 6 on each of the 5 options = 31), but I'm not sure. An advantage to design two might be that the algorithm must do less through the structure of the entity: the table of rules support everything for the amount of the option option.

    Anyway there is a table of rules to determine the amounts:

    Design a
    its option 1 amount =
    2 if age = 10
    5 if age = 11
    7 if age = 12, etc..

    Design two
    of the option option 1 amount =
    2 if age = 10
    5 if age = 11
    7 if age = 12, etc..

    Here, it seems that the one would have to cross over the structure of the entity for the design of two.

    The design will have a better performance with a large amount of rules, or it would make a difference at all?

    Hello!

    In our experience, just think about this kind of stuff if you were dealing with 100's or 1000 instances (usually through ODS). You have a very low number, the differences will be negligible, as you should (in general) go with the solution that is most similar to the material of origin or the understanding of the business user. Also, I guess that's an OWD project? Which may be even better, the inference is performed gradually when new data are added to the modules, rather than in a 'big bang' as ODS.

    It seems that the model 1 is the easiest to understand and explain. I wonder why you have the option at all entity, because it seems to be a relationship to one? If the person cannot have only a single amount of option 1, option 2 amount etc, and there's only ever going to be (up to) 5 options... is this assumption correct? If so, you can keep just like the attributes at the level of the person without the need for bodies. If there are other requirements of an instance of the option then, of course, use them, but given the information here, the option feature doesn't seem to be necessary. It would be the fastest of all :-)

    Whatever it is, that the number of instances is so low, you should have nothing to fear in terms of performance.

    I hope this helps! Write back if you have more info / questions.
    See you soon,.
    Ben

  • Need help to create "no data available" message with multiple data models

    I tried several ways to have a "no documents found" generic message based on certain criteria however it doesn't seem to work. I have two models of data, each only show the information if something is created, deleted, etc. in a certain number of days. If nothing happens for these days, I don't want to report table to display, only a message saying something like "no record".

    I tried the following different types:
    <?if:count(ROW/opened_last_7_days) = 0?>There weren’t any projects created within the past week.<?end if?>
    <?if:count(ROW/opened_last_7_days) > 0?> (report table) <?end if?>
    <?if:count(DATA/opened_last_7_days) = 0?>There weren’t any projects created within the past week.<?end if?>
    <?if:count(DATA/opened_last_7_days) > 0?> (report table) <?end if?>
    The names of the two data models are:
    opened_last_7_days
    closed_last_7_days

    I be recover anything on the page or both records not found and the table appear.

    Sql queries work fine, I get all the answer I need.

    Any help would be appreciated. Thank you.

    Take a look at this: http://winrichman.blogspot.com/2009/05/no-data-found.html

    Thank you!

  • Break with multiple queries

    I built a relationship with the data model and two requests. I set up the break-up on it. The two queries contain the Department and select it. The Department is the key to the outbreak. The breakdown by item in the bursting may be alone. So if I specify, for the split by element the Department in the first query, only the first request is divided by Department, but the second query is not distributed. Breaking for queries longer divide? How to divide multiple queries in the logic of failure?

    Can you help me please?

    Hello Martini,

    Using concatenated sql and select the single lines or datatemplate

  • -You also get an invalid identifier error when executing this query sql for a data model, but not in TOAD/SQL Developer?

    Hello OTN.

    I don't understand why my sql query will pass by in the data model of the BI Publisher. I created a new data model, chose the data source and type of Standard SQL = SQL. I tried several databases and all the same error in BI Publisher, but the application works well in TOAD / SQL Developer. So, I think it might be something with my case so I'm tender hand to you to try and let me know if you get the same result as me.

    The query is:

    SELECT to_char (to_date ('15-' |)) TO_CHAR(:P_MONTH) | » -'|| (To_char(:P_YEAR), "YYYY-DD-MONTH") - 90, "YYYYMM") as yrmth FROM DUAL


    Values of the variable:

    : P_MONTH = APRIL

    : P_YEAR = 2015

    I tried multiple variations and not had much luck. Here are the other options I've tried:

    WITH DATES AS

    (

    Select TO_NUMBER (decode (: P_MONTH, 'JANUARY', '01',))

    'FEBRUARY', '02',.

    'MARCH', '03'.

    'APRIL', '04'

    'MAY', '05'.

    'JUNE', '06'.

    'JULY', '07',.

    'AUGUST', '08'.

    'SEPTEMBER', '09'.

    'OCTOBER', '10',.

    'NOVEMBER', '11'.

    "DECEMBER", "12."

    '01')) as mth_nbr

    of the double

    )

    SELECT to_char (to_date ('15-' |)) MTH_NBR | » -'|| (TO_CHAR(:P_YEAR), 'DD-MM-YYYY') - 90, "YYYYMM")

    OF DATES

    SELECT to_char (to_date ('15-' |: P_MONTH |)) » -'|| ((: P_YEAR, 'MONTH-DD-YYYY')-90, "YYYYMM") as yrmth FROM DUAL

    I'm running out of ideas and I don't know why it does not work. If anyone has any suggestions or ideas, please let me know. I always mark answers correct and useful in my thread and I appreciate all your help.

    Best regards

    -Konrad

    So I thought to it. It seems that there is a bug/lag between the guest screen that appears when you enter SQL in the data model and parameter values, to at model/value data.

    Here's how I solved my problem.

    I have created a new data model and first created all my settings required in the data model (including the default values without quotes, i.e. APRIL instead "Of APRIL") and then saved.

    Then I stuck my sql query in the data model and when I clicked ok, I entered my string values in the message box with single quotes (i.e. "in APRIL' instead of APRIL)

    After entering the values of string with single quotes in the dialog box, I was able to retrieve the columns in the data model and save.

    In the data tab, is no longer, I had to enter the values in single quotes, but entered values normally instead, and the code worked.

    It seems the box prompted to bind the values of the variables when the SQL text in a data model expects strings to be wrapped in single quotes, but no where else. It's a big headache for me, but I'm glad that I solved it, and I hope this can be of help to other institutions.

    See you soon.

  • [SOLVED] Export Oracle SQL Data Modeler is missing a PRIMARY KEY on the DDL script

    I use data 4.1.888 maker to create an ER diagram and generate a DDL her script.

    The diagram contains more than 40 paintings, most of them have a primary key defined.

    For some reason any there is a table that has a primary key defined, but which is ignored when I export the model to a DDL script.

    It is the "wrong" key (even if it is checked that it is not found on the generated DDL script):

    4faS5.png

    This is where the key is set:

    O5mPb.png

    And it is the preview of the DDL (Yes, primary key up there shows):

    SrwMu.png

    This is what happens if I try to generate the DDL for just this (still not generated primary key) table:

    MljUm.png

    Has anyone had the same problem? Any ideas on how to solve it?

    There is no error in the log file, but when I run the generated DDL script there, and then I realized that I was doing something wrong:

    The table MEMBERS had a mandatory foreign key from another table, which in turn had a mandatory key against MEMBERS himself.

    So even if I could generate this primary key on members myself, and then run the the constraint definition that returned an error on the DDL script, I could not perform an insert operation on any of these two tables because of the constraint.

    I revised my design and realized relationships was not mandatory. I unchecked the mandatory box on the definition of the constraint and everything went well.

    I could reproduce the problem and the solution on a diagram with only two tables, so I'm sure that's it.

    Anyway, the Data Modeler is "a failed" silently in this kind of situation. It should be fairly obvious to an experienced designer that I was doing something wrong, but it is not so obvious when you deal with dozens of tables and all their relations and this is your first time using the Modeler.

    Thanks for your reply :-)

  • Data Guard Performance: apply Lag

    In Cloud control there is a measure called "Lag (seconds) apply" under a group called 'Data Guard Performance '.  I created an alert on this metric rule.  However, I want to know what thresholds are defined for this measure, if I look under the supervision of models.  He isn't here.  There are also a number of other Performance measures of custody of data that is, I can't find or add in templates.  Can anyone help?

    From the home page of a database on hold, you can navigate to the (Oracle database > surveillance >) metrics and Collection settings page and select all the metrics view to check or set thresholds for (Data Guard Performance) metrics apply Lag. Note that by default, no limit is set for this measure.

    Kind regards

    -Loc

  • Generate the Sql Data model ER diagram

    Hello

    I want to use the Oracle Sql Developer Data Modeler to generate ER diagram for my schema. There are a lot of tables in this schema, so I would like to exclusively identify the tables that should be selected to generate my ER diagram.

    Basically, I want these table that are related to other tables here. The reason being, if I select all the tables in the schema, and then I went to these tables in the ER diagram which have nothing to do with other tables.

    Please can anyone suggest write queries that produce this data dictionary?

    Thank you.

    select uc1.constraint_name,
           uc1.table_name,
           uc1.constraint_type,
           uc2.constraint_name,
           uc2.table_name,
           uc2.constraint_type
    from   user_constraints uc1,
           user_constraints uc2
    where  uc1.constraint_type='R'
           and uc1.r_constraint_name=uc2.constraint_name
    

    This gives you the relationship between the primary and foreign key

  • The maximum size of the model in SQL Data Modeler?

    Hello

    The number of objects is the maximum value that can be used in a model in SQL data maker. I reverse engineered a scheme (see previous posts - thank you to all that helped) which contains 1000 + tables (a candidate for remanufacturing if ever I saw a!) but the Data Modeler is struggling to display them and performs very slowly. I can pull the same schema in the Oracle Designer and that works well, as ERwin-t - y at - it something I can do to improve the performance of the Data Modeler?

    Or people would recommend cutting the model into smaller pieces, which will be a little difficult because it's a bit of a rat's nest.

    John

    Hi John,.

    You can try to fix the memory usage - Re: problem of memory with large model

    And you may have better performance if divide you large diagram in subviews.

    Philippe

Maybe you are looking for