Can date in the fact table as a measure?

Dear all,

I need to migrate a form of dimensional model database relational model. IT sort of a human resources database. I don't know WHAT should I keep in the fact table. Ago only dates, such as the date, the employee joined the institution and the date that he will leave. Most of the other fields are not digital. much is also not digital, but we can calculate the duration, the employee worked from these dates.

What do you suggest me?

Hello.

I suspect that you can add as a column to a minimum as calculation:
case that no employee.leave_date is zero then (employee.start_leave - employee.start_date) or null;

which returns the work period.

Alex.

Published by: Alex B 07/29/2009 16:12

Tags: Business Intelligence

Similar Questions

  • Slow down execution SQL - extracting data from the FACT Table

    taHi,

    We have a SQL that runs slowly.

    Select / * + ALL_ROWS PARALLEL (F 8) * /.

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_GRP_ID,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_BIL_UNT_ID,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_BNF_PKG_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_UNDWR_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_CAC_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_RTMS_HLTH_COV_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_ACTUR_RSRV_CATEG_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_BNF_TYP_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_GRP_BGIN_DT,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_GRP_END_DT,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_PLN_SEL_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_SBGRP_NM,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_SBGRP_TYP_CD,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_SBGRP_TYP_DESC,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_COBRA_IND,

    IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_SBGRP_POL_NBR,

    IA_OASIS_GRP_CAPTR_D.GRP_POL_CD,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_SBSCR_ALTN_ID,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_RLNSP_CD,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_GNDR_CD,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_BTH_DT,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_MBR_ID,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_SBSCR_SCTR_CD,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_SBSCR_RGN_CD,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_SBSCR_ZIP_CD,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_BDGT_RPT_CLS_CD,

    IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_WORK_DEPT_CD,

    IA_OASIS_MBR_CAPTR_D.BSC_DUAL_IN_HOUSE_IND,

    IA_OASIS_MBR_CAPTR_D.MEDCR_HICN_NUM,

    IA_OASIS_CLM_SPEC_D.CLM_PRI_DIAG_CD,

    IA_OASIS_CLM_SPEC_D.CLM_PRI_DIAG_BCKWD_MAP_IND,

    IA_OASIS_CLM_SPEC_D.CLM_PRI_ICD10_DIAG_CD,

    IA_OASIS_CLM_SPEC_D.ICD_CD_SET_TYP_CD,

    IA_OASIS_CLM_SPEC_D.CLM_APG_CD,

    IA_OASIS_CLM_SPEC_D.CLM_PTNT_DRG_CD,

    IA_OASIS_CLM_SPEC_D.CLM_PROC_CD,

    IA_OASIS_CLM_SPEC_D.CLM_PROC_MOD_1_CD,

    IA_OASIS_CLM_SPEC_D.CLM_PROC_MOD_2_CD,

    IA_OASIS_CLM_SPEC_D.CLM_POS_CD,

    IA_OASIS_CLM_SPEC_D.CLM_TOS_CD,

    IA_OASIS_CLM_SPEC_D.CLM_PGM_CD,

    IA_OASIS_CLM_SPEC_D.CLM_CLS_CD,

    IA_OASIS_CLM_SPEC_D.CLM_ADMIT_TYP_CD,

    IA_OASIS_CLM_SPEC_D.CLM_BIL_PROV_PREF_STS_CD,

    IA_OASIS_CLM_SPEC_D.CLM_BIL_PROV_PRTCP_STS_CD,

    IA_OASIS_CLM_SPEC_D.CLM_ATTND_PROV_PRTCP_STS_CD,

    IA_OASIS_CLM_SPEC_D.CLM_BIL_PROV_PLN_STS_CD,

    IA_OASIS_CLM_SPEC_D.CLM_NDC,

    IA_OASIS_CLM_SPEC_D.CLM_BRND_GNRC_CD,

    IA_OASIS_CLM_SPEC_D.CLM_ACSS_PLS_OOP_IND,

    IA_OASIS_CLM_SPEC_D.CLM_BNF_COV_CD,

    IA_OASIS_CLM_SPEC_D.CLM_TYP_OF_BIL_CD,

    IA_OASIS_CLM_SPEC_D.CLM_OOA_CD,

    IA_OASIS_CLM_SPEC_D.CLM_SANCT_LAT_CALL_IND,

    IA_OASIS_CLM_SPEC_D.CLM_SANCT_DED_PNLTY_IND,

    IA_OASIS_CLM_SPEC_D.CLM_SANCT_COPAY_IND,

    IA_OASIS_CLM_SPEC_D.CLM_SANCT_PCT_COPAY_IND,

    IA_OASIS_CLM_SPEC_D.CLM_SANCT_FLAT_DLR_COPAY_IND,

    IA_OASIS_CLM_SPEC_D.CLM_SANCT_HMO_ACCUM_COPAY_IND,

    IA_OASIS_CLM_SPEC_D.CLM_BIL_ALLOW_APPL_CD,

    IA_OASIS_CLM_SPEC_D.CLM_TMLY_FIL_APPL_CD,

    IA_OASIS_CLM_SPEC_D.CLM_TPLNT_APPL_CD,

    IA_OASIS_CLM_SPEC_D.CLM_ADJ_CD,

    IA_OASIS_CLM_SPEC_D.GRP_PRVDR_TIER_CD,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_SS_CD_CD,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_HMO_COV_LVL_CD,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_FUND_POOL_CD,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_HMO_PRDCT_CD,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_BNF_CATEG_CD,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_ADJ_TYP_CD,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_IPA_ACSS_PLS_IND,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_FUND_ID,

    IA_OASIS_CAPITN_CLM_D.CAPITN_CLM_FUND_DESC,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_GRP_ID,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_CLS_ID,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_CLS_DESC,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_PLN_ID,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_PLN_DESC,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_PRDCT_CATEG_CD,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_PRDCT_CATEG_DESC,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_PRDCT_ID,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_PRDCT_DESC,

    CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_MTL_LVL_CD,

    CLM_CLS_PLN_CAPTR_D.HIOS_PLN_ID,

    CLM_CLS_PLN_CAPTR_D.HIX_GRP_ID,

    OASIS_CLM_MBR_XREF. LGCY_SBSCR_ID,

    OASIS_CLM_MBR_XREF. LGCY_MBR_SFX_ID,

    OASIS_CLM_MBR_XREF. FACETS_SBSCR_ID,

    OASIS_CLM_MBR_XREF. FACETS_MBR_SFX_ID,

    OASIS_CLM_MBR_XREF. LGCY_CUST_ID,

    OASIS_CLM_MBR_XREF. LGCY_GRP_ID,

    OASIS_CLM_MBR_XREF. LGCY_BIL_UNT_ID,

    OASIS_CLM_MBR_XREF. FACETS_PRNT_GRP_ID,

    OASIS_CLM_MBR_XREF. FACETS_GRP_ID,

    OASIS_CLM_MBR_XREF. FACETS_CLS_ID,

    OASIS_CLM_MBR_XREF. FACETS_CONVER_MBR_EFF_DT,

    IA_OASIS_TOT_PROV_CUR_D.PROV_ID AS BEN_BILLING_PROV,

    IA_OASIS_TOT_PROV_CUR_D.PROV_ID AS BEN_ATTENDING_PROV,

    IA_OASIS_TOT_PROV_CUR_D.PROV_ID AS BEN_PCP_NUMBER,

    IA_OASIS_TOT_PROV_CUR_D.PROV_ID AS BEN_IPA_NUMBER,

    EDW. CLM_CAPITN_F.CLM_CAPITN_FRST_SVC_DT_SK,-CAL_DT Dimension is associated and had CAL_DT_SK column

    EDW. CLM_CAPITN_F.CLM_CAPITN_LST_PRCS_DT_SK,-CAL_DT Dimension is associated and had CAL_DT_SK column

    EDW. CLM_CAPITN_F.CLM_CAPITN_FRST_LOC_DT_SK,-CAL_DT Dimension is associated and had CAL_DT_SK column

    EDW. CLM_CAPITN_F.CLM_CAPITN_CHK_DT_SK,-CAL_DT Dimension is associated and had CAL_DT_SK column

    EDW. CLM_CAPITN_F.CLM_CAPITN_ADMIT_DT_SK,-CAL_DT Dimension is associated and had CAL_DT_SK column

    EDW. CLM_CAPITN_F.CLM_CAPITN_ICN,

    EDW. CLM_CAPITN_F.CLM_CAPITN_LN_NBR,

    EDW. CLM_CAPITN_F.CLM_CAPITN_PD_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_BIL_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_ALLOW_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_ALLOW_DY_NBR,

    EDW. CLM_CAPITN_F.CLM_CAPITN_UNT_OF_SVC_QTY,

    EDW. CLM_CAPITN_F.CLM_CAPITN_COB_SAV_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_COINS_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_RSN_CHRG_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_WTHLD_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_DED_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_SANCT_LAT_CALL_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_SANCT_DED_PNLTY_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_SANCT_COPAY_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_SANCT_PCT_COPAY_AMT,

    EDW. CLM_CAPITN_F.SANCT_FLAT_DLR_COPAY_AMT,

    EDW. CLM_CAPITN_F.SANCT_HMO_ACCUM_COPAY_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_TOT_NEGOT_RT_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_LN_NEGOT_RT_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_SEC_ALLOW_AMT,

    EDW. CLM_CAPITN_F.CLM_CAPITN_GATWY_ID,

    EDW. CLM_CAPITN_F.CLM_CAPITN_PROV_AGR_ID,

    EDW. CLM_CAPITN_F.CLM_CAPITN_F_SNPSHOT_MO_SK

    OF EDW. Partition CLM_CAPITN_F (JAN2012),

    EDW. IA_OASIS_GRP_CAPTR_D IA_OASIS_GRP_CAPTR_D,

    EDW. IA_OASIS_MBR_CAPTR_D IA_OASIS_MBR_CAPTR_D,

    EDW. IA_OASIS_CLM_SPEC_D IA_OASIS_CLM_SPEC_D,

    EDW. IA_OASIS_CAPITN_CLM_D IA_OASIS_CAPITN_CLM_D,

    EDW. CLM_CLS_PLN_CAPTR_D CLM_CLS_PLN_CAPTR_D,

    EDW. OASIS_CLM_MBR_XREF OASIS_CLM_MBR_XREF,

    EDW. IA_OASIS_TOT_PROV_CUR_D IA_OASIS_TOT_PROV_CUR_D

    WHERE

    EDW. CLM_CAPITN_F.CLM_CAPITN_IA_GRP_CAPTR_EDW_SK = IA_OASIS_GRP_CAPTR_D.GRP_CAPTR_EDW_SK (+) and

    EDW. CLM_CAPITN_F.CLM_CAPITN_IA_MBR_CAPTR_EDW_SK = IA_OASIS_MBR_CAPTR_D.MBR_CAPTR_EDW_SK and

    EDW. CLM_CAPITN_F.CLM_CAPITN_IA_CLM_SPEC_EDW_SK = IA_OASIS_CLM_SPEC_D.CLM_SPEC_EDW_SK and

    EDW. CLM_CAPITN_F.CLM_CAPITN_IA_CLM_EDW_SK = IA_OASIS_CAPITN_CLM_D.CLM_EDW_SK and

    EDW. CLM_CAPITN_F.CLM_CLS_PLN_CAPTR_EDW_SK = CLM_CLS_PLN_CAPTR_D.CLM_CLS_PLN_CAPTR_EDW_SK and

    EDW. CLM_CAPITN_F.MBR_XREF_SK = OASIS_CLM_MBR_XREF. MBR_XREF_SK (+) and

    EDW. CLM_CAPITN_F.CLM_CAPITN_IA_BIL_PROV_EDW_SK = IA_OASIS_TOT_PROV_CUR_D.PROV_EDW_SK and

    EDW. CLM_CAPITN_F.CLM_CAPITN_ATTND_PROV_EDW_SK = IA_OASIS_TOT_PROV_CUR_D.PROV_EDW_SK and

    EDW. CLM_CAPITN_F.CLM_CAPITN_IA_PCP_EDW_SK = IA_OASIS_TOT_PROV_CUR_D.PROV_EDW_SK and

    EDW. CLM_CAPITN_F.CLM_CAPITN_IA_IPA_PROV_EDW_SK = IA_OASIS_TOT_PROV_CUR_D.PROV_EDW_SK (+);

    --------------------------------------------------------------------------------

    Oracle Database 11 g Enterprise Edition Release 11.2.0.4.0 - 64 bit Production

    PL/SQL Release 11.2.0.4.0 - Production

    CORE Production 11.2.0.4.0

    AMT for Solaris: 11.2.0.4.0 - Production Version

    NLSRTL Version 11.2.0.4.0 - Production

    VALUE OF TYPE NAME

    ------------------------------------ ----------- ------------------------------

    OPTIMIZER_CAPTURE_SQL_PLAN_BASELINES boolean FALSE

    optimizer_dynamic_sampling integer 2

    optimizer_features_enable string 11.2.0.4

    optimizer_index_caching integer 0

    OPTIMIZER_INDEX_COST_ADJ integer 100

    the string ALL_ROWS optimizer_mode

    optimizer_secure_view_merging Boolean TRUE

    optimizer_use_invisible_indexes boolean FALSE

    optimizer_use_pending_statistics boolean FALSE

    optimizer_use_sql_plan_baselines Boolean TRUE

    Please suggest how this can be optimized.

    VALUE OF TYPE NAME

    ------------------------------------ ----------- ------------------------------

    db_file_multiblock_read_count integer 32

    VALUE OF TYPE NAME

    ------------------------------------ ----------- ------------------------------

    DB_BLOCK_SIZE integer 32768

    CURSOR_SHARING EXACT string

  • Dependence on the data cube on the fact tables?

    Hello

    We have a cube that is built based on a table of facts (with prior calculation of 35%)

    The fact table has approximately 400 000 records.

    Now I don't want to disturb my cubes but wants to move forward and still changing the data in the fact table, which could include the removal of thousands of records.

    SO my question here is how dependent are cube on the fact table data.

    The cube stores all the data? Can I go ahead and even to truncate (not drop) the fact table?

    The contents of the cube is not changed until it is built again. So, if you do the following

    (A) complete the table of facts
    (B) create the cube fact table
    (C) Truncate table facts

    Then your cube should always contain the data, and you can question him. The content will change only when you

    (D) create the cube again. At this point my previous answer comes into play.

  • Updating the fact table

    Hello

    I'm just starting with data warehousing. I need to design a warehouse that will record sales on a daily basis. This seems pretty much standard task. However, in my case an order may change several times before it is completed. I intend to handle similar to evolution slow type dimensions 2, adding the date effective date/expiration (Ref. for the dimension date) to the fact table. When an order is updated would be brand in the current record as having expired and I add a new. I can not just replace or delete the previous record, because it would make incorrect historical information. I also really like the idea of storing Let's say not historical orders and periodic snapshots separately - it seems too complicated.
    I thought about the partitioning of the fact table so that only the records in the most recent partition would be updated. In addition, I would create a view for users "where"Date of Expiration"=" N/a"" to work with the current information.
    I'm sure this is a common situation in the data warehouse, however, I couldn't find any useful information on the face. I'm on the right track? OWB does support this kind of functionality (find existing facts by a criterion-> update-> load records new) well?

    Thank you.

    Still a 'RESERVATION_TRANSACTION' which provides for a new record of fact related to the current three-dimensional image would be better than a fact of type 2. This way you have added facts (by type of operation - new/changed/cancelled/come/infiltrated etc., by time period, etc.) and the timestamps in place for their appearance takes place. But you do NOT want hunting through historic images in the table of facts at least quite inevitable - which I think is the case here.

    Just my opinion anyway.

    Good luck with that!

  • How to remove the fact Table

    Hi all

    If I have to restart my fact table on the same day, more than once a day, and he had already stored in it, I want to remove these lines and reload the fact with the current date. I want to create a procedure and include it in the package, the process must check the current_timestamp and if the lines with the date and if there is then it should delete it. Please let me know how I can do this. I am running SQL Server - 2008.

    Thanks for your time and your help.

    You should have to date in your primary key (ex: in a varchar as YYYYMMDD format).

    Then you have 2 ways to implement:

    create an ODI procedure that will remove all data where this date = today. Perform this procedure before your interface.
    * or change your IKM: Add a step that will erase the data in the target table if date = today.

  • Why do you say that the FACT table is off standard? I just don't understand.

    Why do you say that the FACT table is off standard? I just don't understand.

    According to my understanding, normalization is the process of getting rid of redundant. So, the fact table have redundant. So how is it that they say it is off standard.

    What dimension tables? They are also OF normailzed?

    Can you explain in words simple pls.

    Fact tables are generally standardized.

    Dimensions are usually denormalised because they often contain descriptive and rollup/hierarchical data repeated. Snow flake off a dimension will normalize the data.

    See you soon
    If

  • Understand the fact Tables

    Hello

    I'm new to OBIEE, I created several dimensions but am a bit confused with the measures, and the fact tables. I have a table of facts already populated in the database that contains the dimension by each dimension keys, 3 measures. There are a few additional columns (agreement_code, agreement_type, schedule_number). These additional columns do not bind to the dimensions. These columns must be included in the table of facts within the layer of OBIEE BM or could be deleted. My question is the fact table should contain only measures or can it be other columns not aggregated.

    Thank you

    Hello

    If you take certain columns actually that depends on the requirements of the company. If you think that these columns will be used in the reports, then you should take these columns in the data warehouse.

    Should not that only columns of measure should go in fact. You can take these columns not aggregated in your fact. But as a best practice, always try to move the columns not aggregated to your dimension. If you can't, as a last resort to it in fact.

    Thank you
    Sandeep

  • Historical data in the audit tables

    We are facing the following requirement of our current customer.

    They run a number of workflow manually for a few years, and they have retained the historical data in Excel spreadsheets. Now, they want to start using the BPM Suite.

    They would like to query power and display the track Audit Suite BPM of all workflows, i.e. not only those who will be executed with BPM Suite in the future, but also those who have been stored in spreadsheets Excel over the years.

    I have a few questions about this:

    What is a reasonable demand?
    My guess is that we need to load historical data in the tables of audit BPM (BPM_AUDIT_QUERY, etc.), is - that correct?
    Any suggestions on how to get there?

    I was reading the doc, Googling the web and searching this forum and I have not found yet any relevant material on this subject.

    Concerning

    Juan

    Published by: Juan Algaba Colera on October 11, 2012 04:14

    It would be very difficult to directly load their data in the audit tables. Also check tables are stored in the soa-infra schema which is usually configured to be served after a certain period of time so that it grows indefinitely, because it contains all the runtime information as well. You will need to determine how long they need these historical data and configure properly the purge and the size of the DB correctly for this amount of data.

    If data in excel are truly one, the new BPM process the best way to get the audit data in the process can be use the APIs to create instances of the process and then automate execution of each task as defined in the spreadsheets. This way these past instances have been run through the system.

    In most cases this is not done however. Most of the customers usually somewhere archived these worksheets in case they need to recover, but will not try to make that old data visible from the BPM workspace and to view an audit trail.

    Thank you
    Adam DesJardin

  • Code to get the data in the child table in composite SOA using the IOM APIs

    Hi all

    I am a new bie to IOM. I have a query related to obtaining data in the child table using the API of the IOM in the task of embeded java SOA composite.

    I've created a workflow that has a form of child of providing service of slef. I created a composite SOA custom also approval. In the composite approval I have embedded java code and I want to get the values entered in the child form using the API IOM inot the java code embeded in composite SOA.

    I tried gettting the child form data by using getChildAttributes(), but I'm getting the following exception.

    The local Exception stack:
    Exception [EclipseLink-7242] (Eclipse - 2.1.3.v20110304 persistence Services - r9073): org.eclipse.persistence.exceptions.ValidationException
    Description of the exception: an attempt was made to navigate a relationship using indirection that had a null Session. This often happens when an entity has a relationship of LAZY not instantiated is serialized and this lazy relationship is crossed after serialization. To avoid this problem, instantiate the LAZY relationship before serialization.
    at org.eclipse.persistence.exceptions.ValidationException.instantiatingValueholderWithNullSession(ValidationException.java:994)
    at org.eclipse.persistence.internal.indirection.UnitOfWorkValueHolder.instantiate(UnitOfWorkValueHolder.java:218)
    at org.eclipse.persistence.internal.indirection.DatabaseValueHolder.getValue(DatabaseValueHolder.java:83)
    at oracle.iam.request.vo.RequestBeneficiaryEntityAttribute.getChildAttributes(RequestBeneficiaryEntityAttribute.java:100)
    to com. CASApproval.main (CASApproval.java:137)

    Please suggest me if I'm following the correct procedure of the child form data or if we can use another approach.

    Can we get the data using formInstanceOperationsIntf.getProcessFormChildData ().

    Thanks in advance for the help.

    Thank you
    PT

    When you're in approvals, there is no form of process data. Process form data would come only when approvals are completed. If you want to read data from the child form of the DataSet (i.e. the form object as in OIM9.x) you can use the RequestService on the id of the request and read the data. Or another approach would be to the child the dataset data value in the payload of the request and read the XML payload in the composite.
    Let me know if you need more information about the second approach. As for the first approach, search through the forums here and you should find my previous posting on how to reach child dataset values using the ask service API.

    -Marie

    Found these for you:
    OIM11G: Way to get values from dataset of the application for approval
    Re: How to get the value of the AD details of payload of SOA user group

  • Limit the rows of the fact Table by using a table Dim - 10 g

    Hello

    I'm having some trouble trying to restrict the result to a fact table using a Dim picture (assume that the example below).

    -----DIM A-------------------------------FACT B--------------
    ID-Code - Id_Date - Id_Dim - value
    1---ABCD---01-01-2011---1---10
    2---XYZ---02-01-2011---2---20
    3---RST---03-01-2011---1---30
    ----------------------------------04-01-2011-------3------------40

    I want to show only the rows where Dim.Code = 'ABCD '. I know on the MDB, I choose the LTS of the fact table and on the content tab, on where clause insert: Dim.Id = 1, but I don't want to be limited by Dim.Id, I want to limit by Dim.Code and who cannot do this way.

    What I did on the LTS of the fact table, on the tab general got the fact Table on the mapped tables and I added the table Dim doing a join internal between the fact Table and the Table Dim. This way when I go to the content tab, I can do: Dim.Code = "ABCD" because the tables are attached now.

    Is it bad to do? Is there a better way to solve this problem?

    Before ask you I can't do it directly on the table of the Sun because this chart Dim is used in other Tables of facts. Creates an alias for the Table Dim and limit there the way forward?

    I hope that I was clear, thank you

    The way you do it is correct. If you have only one or two measures, then you could also do it using logical columns with the filter function.

    table of facts, for example:
    value = (unfiltered, do not show in the layer close)
    Value of ABCD = (filter (value using dim code = "ABCD"))

    Then you can expand this without having to create a table of facts for each variation:
    value = (unfiltered, do not show in the layer close)
    Value of ABCD = (filter (value using dim code = "ABCD"))
    Value of XYZ = (filter (value using dim code = "XYZ"))
    First value = (filter (value using dim code = "RST"))

    etc. Contrary to the statements of case, it pushes the return filter logic to the database (you get a where clause clause). Kind regards

    Robert

  • How to insert 10,000 records test data into the emp table

    Hi I am new to oracle can someone please help me write a program so that I can insert the test data into the emp table
    SELECT       LEVEL empno,
                 DBMS_RANDOM.string ('U', 20) emp_name,
                 TRUNC (DBMS_RANDOM.VALUE (10000, 100000), 2) sal,
                 DBMS_RANDOM.string ('U', 10) job,
                 TO_DATE ('1990-01-01', 'yyyy-mm-dd')
                 + TRUNC (DBMS_RANDOM.VALUE (1, 6000), 0)
                   hiredate,
                 CASE
                   WHEN LEVEL > 10 THEN TRUNC (DBMS_RANDOM.VALUE (1, 11), 0)
                   ELSE NULL
                 END
                   mgr,
                 TRUNC (DBMS_RANDOM.VALUE (1, 5), 0) deptno
    FROM         DUAL
    CONNECT BY   LEVEL <= 10000
    
  • How to insert the legacy data into the QP_RLTD_MODIFIERS table?

    How insert Legacy data into the QP_RLTD_MODIFIERS table in the instance of R12.

    I would use the QP_Modifiers_PUB API. Process_Modifiers to push the old data on prices in R12.  QP_RLTD_MODIFIERS is only used for certain types of discounts (in my prod environment, only the promos are given in this table).

  • Two filters in two dimensions without constraining the fact table

    Hi all

    does anyone know how to avoid the factual constraint when you create a report with two filters on different dimensions?

    I have a fact table big with more than 10 million lines. In the starmodel is the customer of the dimension and the products. I create a filter on the customer atrribute 'Status' and set to 'active '. Now I add the "Product Type" column of the dimension 'Product' in the filter section. When I want to choose a value OBIEE executes a select statement in the fact table. So I have to wait very long to select a value. Is it possible to say OBIEE only to select the dimension table without joining the fact table?

    Thank you much in advance.

    Kind regards
    Stefan

    Use is implied.

    Create a table of facts of the DUMMY and make also implied made for this dimension column. It will solve the problem.
    http://obiee11gqna.blogspot.com/2011/01/implicit-fact-column-in-OBIEE.html

    Published by: MK on January 17, 2012 07:08

    Published by: MK on January 17, 2012 07:08

  • How can I do a new insert data in the other table extrancting?

    Hello, I'm looking how to do an insertion of two tables save values and apply to other tables like this:

    The insert I'm looking for is something like this:

    Have two tables, the first table is emp1 and contains the column: data 1 and 2 with line of data1: 1234 and line data 2:5555, therefore,.

    I want to create an insert on a new table called (emp2) that contain the data in the column: data 1, data 2 (from the first table demo1), data 3, data 4, etc.

    How insert again I can put / call data from another table insert in this table (emp2).

    I don't know if my example is correct (my English is very bad).

    This is an example: (sorry for the points).

    table EMP2:

    1 data... 2 data... 3 data... data 4... data 5
    (from emp1)... (from emp1)... 8534... 2347... 4544

    Thanks for the help.

    Published by: user10182692 on 24-May-2009 08:39

    Published by: user10182692 on 24-May-2009 08:40

    Published by: user10182692 on 24-May-2009 08:45

    Welcome! on the forum.

    You can include the code and data in the {code} tags

    If I'm good, and I know what is your condition

    insert into emp2 ( data1, data2)
    select data1, data2
    from emp1
    

    SS

  • How can I publish the data to the hash table using regulator netui?

    Could someone help me on how to display the data of the hash to netui in the controller.

    There are altogether 18 survey questions (each question will have a list of surveyquestionanwer)
    but I see only 14 issues of the investigation during the validation in the controller and also I see all the answers of the survey instead of the one that is selected on the jsp.

    Here is my sample code.
    setQuestionSurveyAnswers(Set<Entry<String, Collection<SurveyQuestionAnswer>>> surveyAnswers)
    
    Map<String, Collection<SurveyAnswer>> surveyQuestionAnswerList = new HashMap<String, Collection<SurveyAnswer>>();
    
    surveyQuestionAnswerList contains <questionId, List<SurveyQuestionAnswer>>
    
    ---------------------------------------------------------------------------------------------------------------
    <netui-data:repeater dataSource="pageFlow.surveyForm.surveyAnswers">
    <netui-data:repeaterItem>
       <netui:span value="${container.item.key}" />
         <netui:radioButtonGroup dataSource="container.item.key" orientation="horizontal"> // is anything wrong here?
              <netui-data:repeater dataSource="container.item.value">
                   <netui-data:repeaterItem>
                        <netui:radioButtonOption value="${container.item.answerText}">
                             <netui:label value="${container.item.answerText}" />
                        </netui:radioButtonOption>
                   </netui-data:repeaterItem>
              </netui-data:repeater>
         </netui:radioButtonGroup>
    </netui-data:repeaterItem>
    </netui-data:repeater>
    ---------------------------------------------------------------------------------------------------------------
    data structure:
    ---------------------------------------------------------------------------------------------------------------
    questionId  answerId    answerText      surveyQuestionLabel
    ---------------------------------------------------------------------------------------------------------------
    1              1           Yes          Do you want to take the survey?  <SurveyQuestionAnswer>
    1              2           No           Do you want to take the survey?     <SurveyQuestionAnswer>
    2              1           Beginner     Your expertese Level?    <SurveyQuestionAnswer>
    2              2           Expert       Your expertese Level?    <SurveyQuestionAnswer>
    Appreciate your help.
    Thank you

    Hello
    It won't work, you bind the data of the hashmaps key, which does not work (you have a vo as the key)
    I believe that your structure should be something like

    [QuestionAnswer] where QuestionAnswer has the label of the Question as well as a map of string, the string (AnswerId v/s AnswerText) (this may be an implicit method that traverses all objects and creates a map of AnswerId, AnswerText) and a single field of selectedAnswer.

    Your outer loop will make a loop on QuestionAnswer and print the label of the Question and the inner loop will be the RadioButtonGroup, the data source must be the field of selectedAnswer (because you want to fill only) and the optionsDataSource is the AnswerId card, AnswerText

    concerning
    Deepak

Maybe you are looking for

  • old hard drive in the laptop double

    I worked for several years on a nc8000 HP with Windows XP, and I love it.Unfortunately, the keyboard and the screen age.I just bought another HP nc8000, second hand; This laptop of twins has the same processor, also Windows XP.Question: can I place m

  • RV082 - cannot VPN over subnet 2

    I enabled Multiple on a RV082 subnet option. VPN works on the subnet but not on subnet 2. I use the Shrew Soft VPN client. Any help would be really appreciated. Thank you

  • property of graphics rendering and panelgrouplayout does not work?

    Hello worldMy version of Jdeveloper is 11.1.2.3.0.I have three graphs G1, G2, and G3 (dvt:comboGraph). I have a menu drop down say 1,2 and 3.If I select 1 then G1 must appear, if I select 2 then G2 must be posted and if I select 3 drop down G3 should

  • Mode Live view and attachment

    Using a Canon 70 d, I can attach reliably.  But is there a switch/change what I should do in LR or camera order to use mode live view?  I only managed to do work once, think it's a stroke of luck.

  • data transfer of Oracle Data miner from one machine to the other?

    I had done my project using Oracle 10 g data miner, now I want this project to another machine. Source machine: Windows Vista, Oracle 10 g, Oracle Data Miner 10 g.Destination machine: Windows 7, Oracle 10 g, Oracle Data Miner 10g. Please suggest me h