Need to dimension aggregated using the summary tables

Hello

I have two made tables workdetail and worksummary. Worksummary is grouped in time Sun and workdetail is at the level of day Timedim.

Now, I set up my business model with Timedim and secondary table work. (creates a hirerchy for time (year-month-day) Sun). Now I want to use the tables of worksummary, how can I include this in my business model. I know to create a new Source of logic and mentioning levels.

My important question is what do I have to create another physical table for time-months? or can I use same calendar dim physical to use with the summary fact tables?

the answer to your question is YES, to use aggregated summary tables, we have grouped the dimensions. Other wise data will be redundant and return values incosistent.

In your case if you use the same table of Timedim-day level with summaries, data tables will be multiplied by 30 days due to the time-Sun monthkey will be repeated in several lines.

the simplest solution is to create Time_Dim table view, select separate year, month, monthkey. This view returns only unique year-month. Thus, each month will have only one line.
-> view to import in your physical layer and create a join with the fact summary table.
-> The table of months (which is the point of view) in the logic time_dim as another source, mention levels.
-> and include your table of facts in logical fact table and mention that the levels at months time Sun

It will work. Let me know if I'm not clear. Also, we can expect further comments of experts.

-Madan

Tags: Business Intelligence

Similar Questions

  • Do I need a camera to use the Animation of characters?

    Do I need a camera to use the Animation of characters?

    You need not use a camera. If you want to use triggers on keyboard to move the puppet and a WAV for speech, it can bring very good results.

  • I tried to download Adobe Premier on my Macbook Pro and it says I don't have enough room.  How much room do I need?  If I use the cloud version, how much room do I need?

    I tried to download Adobe Premier on my Macbook Pro and it says I don't have enough room.  How much room do I need?  If I use the cloud version, how much room do I need?

    Mac OS

    • Processor Intel multicore with 64-bit support
    • Mac OS X v10.9 or v10.10
    • 4 GB of RAM (8 GB recommended)
    • 4 GB of disk space available for installation. additional space required during installation (cannot install on a volume that uses a case-sensitive file system or on removable flash storage devices)
    • Additional space required for preview files and other files to work (10 GB recommended)
    • 1280 x 800 display
    • Hard drive 7200 RPM (multiple fast disks configured in RAID 0 recommended)
    • QuickTime 7.6.6 software required for QuickTime features
    • In option: GPU card Certified Adobe for GPU acceleration
    • Recording and Internet connection are necessary for activation of the software required, validation of subscriptions, online access to services *.
  • Using the external table error

    Hi all

    This is the first time I use the external table for loading the data.

    I did these steps

    In the D:\sqlloade folder, I have my text file containing data example.txt

    I ran these queries

    CREATE OR REPLACE DIRECTORY  ext_tab_data AS 'D:\sqlloader';
    
    

    Directory created

    CREATE TABLE fxops.t_ext (
    t number
    )
    ORGANIZATION EXTERNAL (
      TYPE ORACLE_LOADER
      DEFAULT DIRECTORY ext_tab_data
      ACCESS PARAMETERS (
        RECORDS DELIMITED BY NEWLINE
        FIELDS TERMINATED BY ','
        MISSING FIELD VALUES ARE NULL
        (
         t number
        )
      )
      LOCATION ('sample.txt')
    )
    PARALLEL 5
    REJECT LIMIT UNLIMITED;
    
    

    create table

    When I tried to create a view on that table, he throws an error

    CREATE OR REPLACE VIEW fxops.t_view  AS
      SELECT *
      FROM   fxops.t_ext;
    
    SQL Error: ORA-06564: object EXT_TAB_DATA does not exist
    06564. 00000 -  "object %s does not exist"
    *Cause:    The named object could not be found.  Either it does not exist
               or you do not have permission to access it.
    *Action:   Create the object or get permission to access it.
    
    

    Sign in, and then run this:

    Select HOST_NAME from v$ instance;

    is this the same hostname that you are connected?

    Kind regards

    Harry

  • Dimension to build using the SQL table and process to populate the SQL table

    I have a dimension in a cube that is manually* built by one of our power users. Now, I have to get all the information of members of this dimension in a SQL table (example: with columns... Level0, level0property, level1, level1property etc...) to use this table in the STUDIO for the responsibility of the Member.

    Is there any easy process to do this? Currently I am building each line manually in the SQL table and there are 1100 + members in that manually built dimension. Please advice.

    Look at the outline of applied olap Extractor is a free utility

  • insert into the summary table of the table.

    I was wondering if there is a smart way to do it with SQL, but I'm not sure. I would like to consult you yo guys.

    I have a table like this
     CREATE TABLE "TEMPLE_FINANCE"."TEST" 
       (     "COLUMN1" VARCHAR2(10 BYTE), 
         "COLUMN2" VARCHAR2(10 BYTE), 
         "COLUMN3" VARCHAR2(10 BYTE), 
         "COLUMN4" VARCHAR2(10 BYTE)
       ) ;
    
    
    Insert into TEST (COLUMN1,COLUMN2,COLUMN3,COLUMN4) values ('1','2','50.00',null);
    Insert into TEST (COLUMN1,COLUMN2,COLUMN3,COLUMN4) values ('1','2','50.00',null);
    I would like to at the rate of an update and you end up with
    "COLUMN1"     "COLUMN2"     "COLUMN3"     "COLUMN4"
    "1"                  "2"          "100.00"     
    What update statement can run for this?
    I would enter in the summary of the lines based on the column 1 and column2 and get rid of repeative lines.
    I have try this insert statement, but it's bascially just adding an extra record in the table.
    INSERT INTO  TEST (SELECT COLUMN1, COLUMN2, SUM(COLUMN3), COLUMN4 FROM TEST GROUP BY COLUMN1, COLUMN2,COLUMN4);
    any help would be grateful.

    Published by: mlov83 on January 25, 2013 12:45

    Published by: mlov83 on January 25, 2013 13:03

    Hello

    I can't help but wonder if you have the best design of table for what it is, you need to do.

    The best solution would be to create a new table, using the SUM (TO_NUMBER (Column3)) and GROUP BY, and then delete the original table and rename a new one to the old name. While you're at it, change Column3 as a NUMBER and add a primary key.

    In collaboration with just the existing table, INSERT one won't work. INSERT always adds new lines. You want something that can INSERT new lines (or update some existing routes) and DELETE lines at the same time. FUSION can do it all. Here's a way to use the MERGE:

    MERGE INTO     test     dst
    USING   (
              SELECT column1, column2, column4
              ,      SUM (TO_NUMBER (column3))
                                 OVER (PARTITION BY  column1, column2, column4)
                                AS column3_total
              ,      ROWID             AS r_id
              ,      MIN (ROWID) OVER (PARTITION BY  column1, column2, column4)
                                          AS min_r_id
              FROM    test
         )          src
    ON     (src.r_id     = dst.ROWID)
    WHEN MATCHED THEN UPDATE
    SET     dst.column3       = TO_CHAR ( src.column3_total
                                , 'FM9999999.00'
                            )
    DELETE
    WHERE     src.r_id     != src.min_r_id
    ;
    

    Published by: Frank Kulash on January 25, 2013 16:07

  • ListField using the hash table

    Hi guys, new here so please, be gentle.

    I was wondering if there is way to a ListField (or something similar) but using a hash instead of a vector table. I searched through the forums without success. And if it is would it be possible to display the incredible list. (List in the hash table and sort by alphabetical order, via the keys)

    The hash table structure K = String (name of player), V = object reader.

    If this can be done using a hash table, is there a way to do this using a vector. So that I could search the data structure for a players name and return the object.

    Sorry if it's confusing or vague. I'm not not used to describe my problems!

    Thank you very much for the help.

    You can also move forward and persist in the hash table, and then build an index using a few stores of SimpleSortingVector who keys in order.

    There are a lot of options... you just need to decide which is best for you.

  • Use the same table queue send two different messages?

    Hello

    We have a reconciliation and an outgoing queue in our database!

    Now, we need to create another message content to send to IBM MQ series, so it's just an outgoing messages (no incoming message!)

    To the exsisting two queues, we have the gateway messages in place!

    I was wondering if I can use the same outgoing queue table and infrastructure Message gateway to send another message charge (contect) to the IBM MQ.

    Don't you think it's possible?

    More info made me know!

    Thanks a lot for your help!


    Hello Louis,.

    first of all I didn't detect something strange or even incorrect - perfect!

    Your questions:

    > 1. have I defind is correct? "multiple_consumers => TRUE);"

    Yes, that's correct. We must define this order to use MGW.

    > 2. I failed: gv_mq_inbound_log_queue: = 'ORACLE.002 ';     part because I have not all incoming messages to oracle, I just sent messages in my case! > Is this correct?

    Yes, that's also OK. Since you're carrying data from AQ to MQ in an outgoing queue. Incoming queue would be

    the other way around. You have configured your MQ as a remote queue (sys.mgw_property ('MQ_openOptions', 16')).

    You wishes to clarify this yet once with your MQ management group if this is correct. You may not know at this stage.

    The alternative would be a local queue (value = 2066).

    You don't show the part for the queue. This should be something like:

    BEGIN

    v_message_text: = p_message_text. GETCLOBVAL().    -using the XMLTYPE as payload type

    v_text_body: = SYS. MGW_TEXT_VALUE_T (NULL, v_message_text);

    v_payload: = SYS. MGW_BASIC_MSG_T (v_header, v_text_body, NULL);

    DBMS_AQ. ENQUEUE (queue_name-online p_queue_name,

    enqueue_options-online enqueue_options,

    message_properties-online message_properties,

    payload-online v_payload,

    -Online v_msgid msgid);

    > 3. The last of them, do you think I'll reach my goal with all these codes? Is that what you think may be missing or... ?

    Yes, you will reach your goal using this code. Try it and tell me what happened.

    Kind regards

    WoG

  • Issue using the ENGINE table MEMORY SUNOPSIS (high priority)

    Hi gurus,

    It is to launch an operator, like any error by using the SUNOPSIS MEMORY ENGINE to generate a .csv file using the database as a source table.

    ODI-1228: SrcSet0 (load) task fails on the target connection SUNOPSIS ENGINE table SUNOPSIS MEMORY ENGINE.
    Caused by: java.sql.SQLException: unknown token

    (LKM used: LKM Sql to Sql.)
    IKM used: IKM Sql to add files.)

    can you please help me about this ASAP so that it became a show-stopper for me to go further.

    Any help will be greatly appreciated.

    Thank you very much
    Pavan

    Published by: Pavan. on July 11, 2012 10:22

    Hi all

    The issue has been resolved successfully.

    The solution is
    We need to change the E$ _, I have$ _, J$ _,... E_, I_ , J_... ((c'est à dire, enlever le symbole '$')) in the PHYSICAL SCHEMA of ENGINE table SUNOPSIS MEMORY according to the information provided below.

    When interfaces running and using an XML schema or file complex as the transit area, the 'Unknown token' error. This error is caused by the HSQL (2.0) updated version. This new version of HSQL requires that the names of tables that contains a dollar sign ($) are enclosed in quotes. Temporary tables (loading, integration, etc.) that are created by the knowledge Modules do not respect this requirement on the technologies of complex and HSQL records.

    To work around the problem, change the physical schema definitions to remove the dollar symbol ($) of all paintings of prefixes work. Scenarios must be regenerated with these new settings.

    It has worked well for me.

    Thank you
    Pounet

  • problem in the form of master-detail when you use the ADF table for detail

    Hello

    jdev version - 11.1.2.1.0


    I create master shape detail using datacontrol drag as ADF master shape secondary Table.


    Now when I create a new line in the detail table using the key CreateInsert a new empty row created on top of the secondary table.

    and other show line that the previous record data based on the master.

    problem is I want to when I click on the createInsert button all the line of the secondary table must be empty and what line to fill two or three user then validate.



    Thanks in advance

    Hello

    If a secondary table has data, then createInsert adds to them. If you want to hide the existing lines, create a new instance of the View object and set the option "extract database" to "No. Rows. Use an af:switcher to change the specified table when the user clicks the createInsert button. There is some coding needed to have this use case in the ADF, but its essentially declarative. Bottom line, is that there is no option automated other than to create new lines in a separate page or dialog box if you are bothered by existing lines

    Frank

  • Two filters in two dimensions without constraining the fact table

    Hi all

    does anyone know how to avoid the factual constraint when you create a report with two filters on different dimensions?

    I have a fact table big with more than 10 million lines. In the starmodel is the customer of the dimension and the products. I create a filter on the customer atrribute 'Status' and set to 'active '. Now I add the "Product Type" column of the dimension 'Product' in the filter section. When I want to choose a value OBIEE executes a select statement in the fact table. So I have to wait very long to select a value. Is it possible to say OBIEE only to select the dimension table without joining the fact table?

    Thank you much in advance.

    Kind regards
    Stefan

    Use is implied.

    Create a table of facts of the DUMMY and make also implied made for this dimension column. It will solve the problem.
    http://obiee11gqna.blogspot.com/2011/01/implicit-fact-column-in-OBIEE.html

    Published by: MK on January 17, 2012 07:08

    Published by: MK on January 17, 2012 07:08

  • Type of Dimension set using the Rules file

    Hello

    I tried to build a dimension (ASO) of accounts using the build parent-child and wanted to know if we can specify the dimension type "Account" to in the rules file?
    Also all the codes of the property can be given in a single column or placing them in separate columns a good practice? I need to make this hierarchy of dimension as 'dynamic '.

    I use the definition of dimension of the rules file option name.

    Thanks for your contributions!

    Yes you can do it in a rules file. In the setting dimension dialog box, go to the third tab. You can see the list of dimensions and the possibility of adding dimensions. Right-click on the desired size and then select Edit properties. Then, you can set the dimension type and dimension member properties.
    With regard to the properties. I always keep in separate columns of my source.

  • Needing a maximum date using the group value of

    Create table student (dept number(10), dep_name varchar2(10),join_date date,years_attended number(10),end_date date);
     
    insert into student values (1,'I',to_date('3/7/1917','MM/DD/YYYY'),4,to_date('8/26/1987','MM/DD/YYYY'));
    insert into student values (1,'I',to_date('1/1/1900','MM/DD/YYYY'),4,to_date('8/26/1932','MM/DD/YYYY'));
    insert into student values (1,'D',to_date('1/1/1920','MM/DD/YYYY'),5,to_date('8/26/1994','MM/DD/YYYY'));
    insert into student values (1,'C',to_date('1/1/1920','MM/DD/YYYY'),6,to_date('8/26/1945','MM/DD/YYYY'));
    insert into student values (2,'I',to_date('7/1/1900','MM/DD/YYYY'),3,to_date('8/26/1932','MM/DD/YYYY'));
    insert into student values (2,'I',to_date('8/16/1916','MM/DD/YYYY'),9,to_date('8/26/1923','MM/DD/YYYY'));
    insert into student values (2,'D',to_date('8/16/1916','MM/DD/YYYY'),10,to_date('8/26/1987','MM/DD/YYYY'));
    insert into student values (3,'I',to_date('3/7/1917','MM/DD/YYYY'),4,to_date('8/26/1987','MM/DD/YYYY'));
    insert into student values (3,'D',to_date('7/28/1920','MM/DD/YYYY'),6,to_date('8/26/1945','MM/DD/YYYY'));
    insert into student values (3,'I',to_date('7/28/1920','MM/DD/YYYY'),8,to_date('8/26/1965','MM/DD/YYYY'));
    insert into student values (4,'I',to_date('12/31/1924','MM/DD/YYYY'),2,to_date('8/26/1998','MM/DD/YYYY'));
    insert into student values (4,'I',to_date('6/10/1929','MM/DD/YYYY'),1,to_date('8/26/1943','MM/DD/YYYY'));
    insert into student values (4,'C',to_date('1/17/1927','MM/DD/YYYY'),4,to_date('8/26/1955','MM/DD/YYYY'));
    insert into student values (4,'C',to_date('6/10/1929','MM/DD/YYYY'),30,to_date('8/26/1967','MM/DD/YYYY'));
    insert into student values (5,'D',to_date('2/10/1931','MM/DD/YYYY'),2,to_date('8/26/1943','MM/DD/YYYY'));
    insert into student values (5,'I',to_date('2/10/1931','MM/DD/YYYY'),24,to_date('8/26/1962','MM/DD/YYYY'));
    commit;
    I need a join_date of date value maximum for each Department. If max (join_date) has two records for each dept max (end_date) are to be considered. I used a select query
    select * from student where join_date in (select 
    max(join_date) from student group by dept);
    gives me the following result
    1     D     1/1/1920     5     8/26/1994
    1     C     1/1/1920     6     8/26/1945
    2     I     8/16/1916     9     8/26/1923
    2     D     8/16/1916     10     8/26/1987
    3     D     7/28/1920     6     8/26/1945
    3     I     7/28/1920     8     8/26/1965
    4     I     6/10/1929     1     8/26/1943
    4     C     6/10/1929     30     8/26/1967
    5     D     2/10/1931     2     8/26/1943
    5     I     2/10/1931     24     8/26/1962
    But I'm looking for the result that gives me only a maximum value for each column dept. First of all, it should look like to a maximum value of join_date, so two records even join_date max (end_date) are to be considered. The result should be sumthing like this
    1     D     1/1/1920     5     8/26/1994
    2     D     8/16/1916     10     8/26/1987
    3     I     7/28/1920     8     8/26/1965
    4     C     6/10/1929     30     8/26/1967
    5     I     2/10/1931     24     8/26/1962
    Can you please tell me how to rewrite the select query for results above.

    Published by: user11872870 on August 2, 2011 17:29

    Published by: user11872870 on August 2, 2011 17:36

    Hello

    This is called a Query Top - N , and this is a way to do it:

    WITH     got_r_num     AS
    (
         SELECT     student.*
         ,     ROW_NUMBER () OVER ( PARTITION BY  dept
                                   ORDER BY          join_date     DESC
                             ,                end_date     DESC
                           )      AS r_num
         FROM    student
    )
    SELECT       dept, dep_name, join_date, years_attended, end_date
    FROM       got_r_num
    WHERE       r_num     = 1
    ORDER BY  dept
    ;
    

    Another way is similar to what you posted:

    SELECT    *
    FROM       student
    WHERE        (dept, join_date, end_date)
                   IN (
                        SELECT    dept
                   ,       MAX (join_date)
                   ,       MAX (end_date) KEEP (DENSE_RANK LAST ORDER BY join_date)
                   FROM      student
                   GROUP BY     dept
                   );
    

    I suspect that the first way (using ROW_NUMBER) will be faster.
    In addition, the ROW_NUMBER approach is guaranteed to return only 1 line by Department using the approach of GROUP BY, if there is a link on join_date and end_date then it will return all the contenders in this Department. using ROW_NUMBER, it is easy to add expressions to tie-break as much as you want, and, if there is still a tie, it will be arbirarily pick, one lines involved in the tie as #1.

    Thanks for posting the CREATE TABLE and INSERT! It is very useful.

    Published by: Frank Kulash, August 2, 2011 21:00
    Added GROUP BY alternative

  • Error importing CSV files with "hidden" characters using the external Table

    Hi people

    Bit of a strange here.

    Well, we are accustomed to the use of the external Table method to load data from CSV files in the database, but a recent event presented us a problem.

    We have received some CSV files that "look like" regular CSV files, but Oracle will not load them.

    When we looked at the CSV using VIM on a UNIX machine, we saw the following characters 'hidden' between each regular character in the file.
    ^@
    If a string that looks like this when opened in Excel/Wordpad etc.
    "TEST","TEXT"
    Looks like this when exmained with VIM
    ^@"^@T^@E^@S^@T^@"^@,^@"^@T^@E^@X^@T^@"
    Has anyone encountered this before?

    Thank you very much

    Simon Gadd
    Oracle 11g 11.2.0.1.0

    Hi Simon,.

    ^ @ represents the ZERO character (0x00).
    So, most likely, you have a file encoded in Unicode.

    You need to specify the character set in the record specification (and if necessary the byte order mark), for example:

    CREATE TABLE ext_table
    (
      col1 VARCHAR2(10),
      col2 VARCHAR2(10)
    )
    ORGANIZATION EXTERNAL
    (
      TYPE ORACLE_LOADER
      DEFAULT DIRECTORY dump_dir
      ACCESS PARAMETERS
      (
       RECORDS DELIMITED BY '
    ' CHARACTERSET 'UTF16'
      FIELDS TERMINATED BY ','
      )
      LOCATION ('dump.csv')
    )
    REJECT LIMIT UNLIMITED;
    

    http://download.Oracle.com/docs/CD/E11882_01/server.112/e16536/et_params.htm#i1009499

  • How seeds Cache using the Event Table of the poll

    Hello Experts

    I have configured the Event Table of the poll in my PC
    Steps to follow:
    (1) table created using SAEPT. Oracle.SQL scheme in 10g
    (2) any SAEPT. Oracle.SQL by creating a new connection pool in the physical layer
    (3) configured in repository - tools - utility - OBI event Polling tables by selecting SAEPT. Oracle.SQL and gave 15 min in the polling frequency
    (4) inserted data in SAEPT. Oracle.SQL
    (5) NQserver.log checked after the time of the poll which is 15 minutes. NQserver.log was not getting updated or details about the event that occurred.

    You think that I missed a few steps in the configuration of EFA, if so... Please help me I need to implement on the client machine soon.

    Enjoy your time guys.

    Published by: Newby on July 1, 2010 10:34

    As much as I know the vote event tables are used to purge the cache automatically. It is not used to start the cache.
    If you want to start the cache, the ibots use or you do like a program...

    Published by: user8000915 on July 1st, 2010 07:37

Maybe you are looking for

  • key control (ctrl) and scrolling problem

    Key control (ctrl) zooms window can not copy, paste, etc. using the function without the Zoom window key before, this does not happen with other browsers, it started about 3 weeks, possibly after the last update of firefix, windows 8 When I scroll do

  • Ports high speed USB Equium A100

    I installed a webcam Advent (before exchanging it for a better creative). Part of the installation process involves installing a Microsoft's USB controller. (OS is Windows XP Home Edition). I uninstalled the ADVENT software when I got back to the web

  • Satellite A200 - cannot downgrade the BIOS to v2.40

    Please help me. I bought a brand new toshiba Satellite A200 series laptop with vista home premimum inside. For some reason, I'm not a fan of big vista.So I lowered my XP OS and as I was looking for pilots and all this for XP I saw an update of the Bi

  • error code 80070641

    updates not responding error 80070641 code

  • Color LaserJet CM1312nfi MFP that gives the message 'unexpected paper size ".

    In recent months, when I print a set of envelopes, after EACH envelope my printer gives me a message "unexpected paper size".  This forces me to press the 'OK' button to continue to the next task.  After that, the printer goes into a long routine of