Export data - with the colunn attribute

Hello, we have a cube with 12 decreases intensity, we have a Sun cost center that we have assigned a dim (SGA) attribute. The Sun CC has ~ 3,000 and the dim SGA has approximately 80 members with each CC with an LMS are entrusted to him.

I need to export data to power a different cube based on the Sun BMG. I tried to do this with a report Script and it worked very well for a member of the SGA and one account, but when I tried to do it for all members of the SGA that he brought EAS (v9.3.1.4) as well as the report never came back. I tried it twice, the people in charge the servers weren't happy either time.

I searched the web looking for solutions and see a lot of start analytics and JExport. No money to start analytics and I can't find the source of JExport. Does anyone have an suggestions.

I've been watching the DataExport command, but I don't think I can have the attribute member is displayed in the output file.

My thought is to get rid of the protected attribute in Essbase and let the FDM to manage the translation as the protected attribute is only intended to be used to export data.

Any suggestion would be appreciated.

Thanks in advance.
Steve

JExport is here: http://essbaselabs.blogspot.com/2008/10/cdf-jexport-and-dynamic-members.html
^ ^ ^ Those are the guys who wrote it.

I already tried to put a Sun in a DATAEXPORT attribute. Have you tried and failed?

Kind regards

Cameron Lackpour

Tags: Business Intelligence

Similar Questions

  • How to export data from the table with the colouring of cells according to value.

    Hi all

    I use jdeveloper 11.1.1.6

    I want to export data from the table with a lot of formatting. as for color cells based on value and so much. How to do this?

    You can find us apache POI-http://poi.apache.org/

    See this http://www.techartifact.com/blogs/2013/08/generate-excel-file-in-oracle-adf-using-apache-poi.html

  • Foreign key constraint, not recognized during the synchronization of data with the model dictionary

    Hello

    Data Modeler is a foreign key constraints do not recognize when synchronizing data with the model dictionary, although the foreign key is there (in the database that a data dictionary is read). I can't find any criterion when a foreign key is not recognized by the Data Modeler. Are there limits to the length of the attribute, or the number of columns in a foreign key, or other limitations which may lead to this behavior not to recognize a fk by Data Modeler? I have columns more than 32 characters. I compared with the fk is recognized by DM, but I can't find anything that indicates why it is not recognized.

    I wonder if someone also has constraints of foreign keys that are not recognized in the comparison of data bases and model?

    Thank you

    Robert

    Hi Robert,.

    Thanks for the comments, I logged a bug.

    Philippe

  • Failure to EXE to EXPORT DATA to THE Clipboard

    LV 2010, Win Vista, Win7.

    A TABLE indicator has a context menu that can EXPORT DATA to the Clipboard.

    In LV2010 (and LV2010 SP1), this function does not work in an executable file - it works very well in the dev system. LV.

    It was fixed in LV 2011, but we can't move to that at the present time.

    What is the best way to take advantage of this feature in LV2010?

    See this JING

    Actually, there is an easier way.  The table has a method called "EXPORT DATA to THE Clipboard".  Apparently, the method is OK, it is not only called on the menu, when in an EXE file. This is a bug of LABVIEW.

    I added only a button called 'EXPORT', an event to respond to a CHANGE OF VALUE on this button and a NŒUD INVOKE to call this method.  It exports very well selected cells.

    Thanks for your idea, if.

  • all my data is all save on the local disk c, how to share the data with the other drive, local drive d.

    all my data records on the local disk c, how to share the data with the other drive, local drive d.

    Hi Jasonbichard,

    1. what type of drive is D? Is - this another partition on the same disk?

    2 Windows operating system you are using?

    You can change the location of the disk to save the data in the d: instead of C: and check if it helps.

    a. navigate to the location (username) C:\Users\.
    b. right click on the folder that you want to change the location, and then select Properties.
    c. click on the location tab and change the location to D: drive.

    d. click on apply and Ok.

  • Dynamic action for validation of date with the notification message plugin

    Hi all

    Someone help me please with dynamic action for validation of date with the message notification plugin. I have a form with two elements of the date picker control and message notification plugin.

    The requirement first user selects the exam is finished and then selects the date. So, if the date is greater than the date of the examination is over + 2 years then doesn't trigger the message notification plugin. I tried to create that dynamic action on the date picker date that triggers the scheduled issue notification message but I want to make conditional, I mean displays the message only if date of the selected is greater than the date of the exam is finished more than 2 years.

    In terms simple, notification is displayed only if provided is superior to (date of the exam is completed + 2 years).

    I use oracle apex 4.0 version and oracle 10g r2 database. I tried to reproduce the same requirement in my personal workspace. Here are the details. Please take a look.

    Workspace: raghu_workspace

    username: orton607

    password: orton607

    APP # 72193

    PG # 1

    Any help is appreciated.

    Thanks in advance.

    Orton.

    You can get the value of the date of entry:

    $(ele) .datePicker ('getDate');

    So what to add functions such as:

    function validateNotification (d1, d2) {}

    Date1 var = $(d1) .datepicker ('getDate');

    date2 var = $(d2) .datepicker ('getDate');

    if(date1 && date2) {}

    return ((date2.getTime()-date1.getTime())/(1000*24*60*60))>(365*2);

    } else {}

    Returns false;

    }

    }

    The logic based on setting (I have two years from years of 365 days preceding)

    Then in the D.A. specify a JavaScript expression as:

    validateNotification ('P2_REVIEW_COMPLETED', this.triggeringElement.id)

    Refer to page 2 for example.

  • What is always "stay up to date with the latest innovations in the field of photography from Adobe"?

    Hello

    I am considering buying stand-alone version 6 of Lightroom, but I wonder what does that "Always stay up to date with the latest innovations in photography from Adobe" is not available?

    Will be the updated standalone for free with the new camera support?

    Will be free updated standalone version with correction of defects?

    If there are new features (minor version of lightroom), will not be available for the stand-alone version before the next major release?

    Thanks for your help

    My understanding is that LR6 will continue to receive bug fixes and new camera support updates, but not feature/news which receives from the CRD. See question 2 below on this post: CC Lightroom / Camera Raw FAQ

  • Export data from the zip with PC?

    Is it possible to export the data of the folio (creation with digital publishing suite) with a PC?

    Whenever I try to export the data-folio, the zip data disappear after completing the download.

    What do I need create the datas for PC/Mac, Tablet and Smartphone?

    You will need to have a Mac to build your application. As Bob mentions, in InDesign select Create App in the context menu of the Folio Builder Panel.

    You can design your content on a PC and then use the download command in the Folio Builder Panel to put it on our servers. Then, the Mac, you can save in the Panel with the same account and Create App from there.

    Neil

  • Loading data with the size of the attribute.

    I have a currency attribute dimension asociated to dimension of entities and I would download data using the dimension of the associated attribute. I have a file with a column with the name of the entity and also a column with the name of the attribute. Could someone tell me how to set the column attribute of the rule of load?

    Published by: user5170363 on March 22, 2012 12:55

    I guess that, dimensions of attribute are not County while you load data... Data file always have the standard dimension members...

    Thus, you can filter your features according to your attribute or load only on these entities...

    -----
    Vivek jerbi

  • Export data from the database Table in the CSV file with OWB mapping

    Hello

    is it possible to export data from a database table in a CSV with an owb mapping. I think that it should be possible, but I didn't yet. Then someone can give me some tips how to handle this? Someone has a good article on the internet or a book where such a problem is described.

    Thank you

    Greetings Daniel

    Hi Daniel,.

    But how do I set the variable data file names in the mapping?

    Look at this article on blog OWB
    http://blogs.Oracle.com/warehousebuilder/2007/07/dynamically_generating_target.html

    Kind regards
    Oleg

  • Export DataPump with the query option

    Hi all

    My environment is IBM AIX, Oracle 10.2.0.4.0 database.

    I need a few sets of records using a query in export production. Request is attached to several tables. Since we have the BLOB data type, we export using datapump.

    We have weaker environments, but have not the same set of data and tables, and therefore not able to simulate the same query in lower environment. But created a small table and faked the query.

    My order is

    expdp system / < pwd > @orcl tables = dump.dump1 query = dump.dump1:' ' where num < 3 ' ' directory = DATA_PUMP_DIR dumpfile = exp_dp.dmp logfile = exp_dp.log

    Query in the command pulls two records directly. By running the command above, I see the size 80KO dump file,
    In the export log file.

    I see Total estimation using BLOCKS method: 64 KB.
    export Dump.Dump1 = 4,921 KB 2 rows.

    My doubts are,
    (1) is the correct command that I am running.
    (2) estimate said 64 k, considering that it says also exported 4,921 KB. But the dump file created is 80KO. It is exported correctly?
    (3) given that I run with the use of the system, apart from 2 rows, export all data. We must send the dump file to the other Department. We should not export all of the data other than the query output.
    (4) in the order if I am not using "tables = dump.dump1), the export file big mess." Don't know which is the right.

    Your answers will be more useful.

    The short answer is 'YES', he did the right thing.

    The long answer is:

    Query in the command pulls two records directly. By running the command above, I see the size 80KO dump file,
    In the export log file.

    I see Total estimation using BLOCKS method: 64 KB.
    export Dump.Dump1 = 4,921 KB 2 rows.

    My doubts are,
    (1) is the correct command that I am running.

    Yes. As long as you query is correct. DataPump will export on the lines that match this query.

    (2) estimate said 64 k, considering that it says also exported 4,921 KB. But the dump file created is 80KO. It is exported correctly?

    Estimate is made using the full picture. Since you specify, he used the method of estimation of block. Basically, how many blocks have been attributed to this table. In your case, I guess it was 80KB.

    (3) given that I run with the use of the system, apart from 2 rows, export all data. We need to send the dump file to other > Department. We should not export all of the data other than the query output.

    I will export all the data, but going to export metadata. It exports the table definition, all indexes on it, all the statistics on tables or indexes, etc. This is why the dump file could be bigger. There is also a 'main' table that describes the export job who gets exproted. This is used by export and import to find what is in the dumpfile, and where in the dumpfile these things are. It is not user data. This table needs to be exported and will take place in the dumpfile.

    (4) in the order if I am not using "tables = dump.dump1), the export file big mess." Don't know which is the right.

    If you only want this table, then you order export is right. If you want to export more, then you need to change your export command. From what you say, it seems that you order is correct.

    If you do not want any expoirted metadata, you can add:

    content = data_only

    at the command line. This will only export the data and when the dumpfile is imported, it must have the table already created.

    Dean

  • Explorer Windows in Windows 7 change suddenly last updated Date of the files attribute

    I'm moving from XP Pro to Win 7 of last month during this process, I came across a disturbing behavior in Windows Explorer in Windows 7.

    In Windows 7 (and I hear Vista as well) Windows Explorer will sometimes change the file Date changed when slide it / file to a new location.  I spent hours researching this issue.  In some discussions, some participants said it is a reasonable thing, since copying a folder with files creates a new folder and new files, so a new modified Date is guaranteed.  I can understand that as a basis to change the creation Date.  This was the behavior under XP and I think that NT as well.  However, changing the Date of change is a fundamental departure from how the system of files are featured in the past, and it is also a serious departure from good sense.  At the very least, it will make it impossible to find or organize files by date of update, since the original modification Date (date the actual content of the file may have been modified) is lost.  It also seriously decreases the possibility of finding files duplicated by this attribute when simply copy a file it modifies.

    Change the folder view options can prevent this behavior (Options: view folders: check always show icons, never thumbnails).  In addition, edit a file to read-only will also stop this change of Date of modification.  However, none of these options provide a viable solution to this problem.  All first, change the option to display file then prevents look a thumbnails of the image files.  Change once the copy of file causes the copied files to change Date change date created.  The same is true for the readonly attribute (removing it causes the modification Date change), not mention that it is often impractical to effectively apply this strategy.

    I also note that this behavior is not consistent in all areas.  Files copied from a shared disc on a PC network do not seem to this behavior.  But, files and folders transferred by using Dropbox (www.dropbox.com) show the unwanted behavior.  Worse still, the behaviour was not consistent within the folder.  I transferred a file with the 2007 via Dropbox eml files.  In most tests, once the folder has been moved and dropped on the Win 7 machine (whether copied or moved) the Date of editing files has been changed to take account of the date and time that they were dropped.  However, in one case, two of the files has not changed the Date of change.  By clicking on the file or even open it to read has not made a difference.  The modification Date remained unchanged.  I don't really know why this was the case in this instance, given that these same files has been modified Date change in previous experiences.

    Frankly, I'm puzzled and dismayed that this did not become a more important issue.  Do not have a lot of imagination to see how this problem might affect users old files of e-mail to archiving as LME and anyone copy older files, who later need to search or sort them by the Date of change.  There are solutions that I know, as the compression of the files before any copy or use Robocopy with the appropriate options, but they are much less practical as drag - drop.  Furthermore, drag and drop is now so embedded in users what their likely natural choice for file copy or move.  Once the modification Date is lost, it cannot be found without a restore of backup files - once again, an unrealistic option and a lot of time.

    I think it is a serious problem that needs to be processed quickly by Microsoft.  Good design keeps logical and mitigates the factors that can lead to human error.  The above behavior fails both these needs.  Anyone found a solution for this (maybe change registry) or to know if and when Microsoft plans to approach this?

    I've just dealt with the same problem when copying all the contents of my hard drive and found an impressive (and without :-)) Microsoft Utility called RichCopy which solves this problem and makes it much easier to copy a large number of files/folders :-))

    Here is a link to the article about this in Microsoft TechNet Magazine described the tool, which has a link to download it as well: http://technet.microsoft.com/en-us/magazine/2009.04.utilityspotlight.aspx

  • Pump diagram remapping of data with the default value of sequence (12 c feature)

    Hello

    in the 12 c Oracle database there is a new feature that allows you to set default values directly from a sequence:

    http://docs.Oracle.com/database/121/NEWFT/chapter12101.htm#NEWFT155

    This helps us to save the use of triggers to get the next value of an ID column for example the DDL of such a table might look something like this:

    CREATE TABLE "FOO"."MY_TABLE" ("ID" NUMBER(10,0) DEFAULT "FOO"."MY_TABLE_ID"."NEXTVAL", ...   );

    During the pattern FOO with Data Pump export and import the schema with the remap_schema option in the BAR diagram for example, mapping works fine for all the tables, triggers, etc. except those defaults (as Oracle always writes the schema information in these default values).

    The error in the output log that says 'FOO. MY_TABLE_ID"sequence is unknown, because the sequence is now called 'BAR. MY_TABLE_ID' - mapping so does not work for default values.

    Someone knows how to fix this?

    Thanks in advance

    Concerning

    Hello

    Read this note, I think that it is an expected behavior:

    http://docs.Oracle.com/CD/E11882_01/server.112/e22490/dp_import.htm#SUTIL927

    REMAP_SCHEMA

    .

    Restrictions

    * The mapping cannot be 100 percent complete because there are some references to diagrams that importing is not able to find. For example, import will not find of references to patterns incorporated into the body of the definitions of triggers, types, views, procedures and packages.  -I think in your case it is the default value.

    * If any table in the schema are remapped contains object types defined by the user and this table changes between the time wherever it is exported and the time you are trying to import, and then import this table will fail. However, the import operation will continue.

    HTH,

    Pradeep

  • Extraction of data from the xmlclob attribute

    Hi all

    I am trying to extract the value of the attribute below clob.

    kindly advice how to retrieve the value by using the xml functions.

    I use version 11.2.0 and toad 9 oracle database.

    I've highlighted the comment tag. the highlighted, I need to recover ' time changed according to the dtd VSG mail 25/01/2013 14:27.» SVK, & #xD; & #xA; Recipient added by mail from Richard dtd 20/02/2013 09:45. SVK" of the value attribute.

    Please suggest how to achieve using SQL. Data stored in xml format in the clob column.



    <? XML version = "1.0"? >

    " < ValuationRequest xmlns: xsi =" http://www.w3.org/2001/XMLSchema-instance "container =" " http://www.w3.org/2001/XMLSchema "ValueID ="169821"Version ="2.6" "

    " xmlns =" http://XBank.FxVals/ProcessValuationRequest.xsd "Status ="AC"> "

    < ValueID '5' = ClientRelationShipType > < / ClientRelationShipType >

    < ValueID owner = '14539"Name ="JEREMY LEVINE"> < / owner >

    < ValueID customer = "6892067" Name = "XXXXXXXXXX, NEW YORK" > < / customer >

    < ValueID = "1" CMVRegion > < / CMVRegion >

    < ValueID = "1" ValuationTypeID > < / ValuationTypeID >

    < CSRContacts >

    < CSRContact Value = "917" Name = "EVALUATIONS NY[valuations.ny@gmail... com]' > < / CSRContact >

    < / CSRContacts >

    < title >

    < ValueID security = PrimaryID "900158" = "NY-518362" ItemLevel = '2' ItemType = "38" >

    "< rank EffectiveDate =" "2013-01 - 24 T 00: 00:00 ' TerminationDate = ' 0001-01 - 01 T 00: 00:00" comments = "" SecurityDetailID = "429190" > < / row >

    < / security >

    < ValueID security = PrimaryID "900158" = "NY-518362" ItemLevel = '1' ItemType = "38" >

    "< rank EffectiveDate =" "2013-01 - 24 T 00: 00:00 ' TerminationDate = ' 0001-01 - 01 T 00: 00:00" comments = "" SecurityDetailID = "429191" > < / row >

    < / security >

    < / securities >

    < created Value = "2013-01 - 24 T 12: 20:48" > < / creation >

    < ValueID '37126' = CreatedBy > < / CreatedBy >

    < LastModified Value = "2013-02 - 21 T 01: 56:00" > < / LastModified >

    < ValueID = '18732' LastModifiedBy > < / LastModifiedBy >

    < ReportDefinition ValueID = "1" > < / ReportDefinition >

    < StartDate = value "2013-01 - 24 T 12: 00:00" > < / StartDate >

    <Comment Value = "time changed according to the dtd VSG mail 25/01/2013 14:27." SVK, & #xD; & #xA; Recipient added by mail from Richard dtd 20/02/2013 09:45. SVK">< / comment >

    < ValueID = "100" PricePercentage > < / PricePercentage >

    < ISDSL Value = "F" > < / ISDSL >

    < ValueID effects = '0' > < / Effects >

    < WorkFlowAction >

    ActiveChangesApproved < action > < / Action >

    < / WorkFlowAction >

    < ValueID = "4" RecertificationFrequency > < / RecertificationFrequency >

    < RecertificationDate Value = "0001-01 - 01 T 00: 00:00" > < / RecertificationDate > "".

    < RecertificationNextDate Value = 'January 24, 2014' > < / RecertificationNextDate >

    < StatementFrequency >

    < SFrequency Value = '1' Type = 'Daily' date limit indicator = "1005" = "F" PDeadline = '990' > < / SFrequency > "

    < / StatementFrequency >

    < value AutoGenerate = "T" > < / AutoGenerate >

    < monetary value = "Us $" > < / currency >

    < SLanguage ValueID = "1" > < / SLanguage >

    < GP2AccountID Value = "480352" > < / GP2AccountID >

    < SalesApproval Value = "F" > < / SalesApproval >

    < RegenerateAfterPartial Value = "F" > < / RegenerateAfterPartial >

    < SalesCoverageTeam >

    < = 'Sales' sales type Value = "14539" NotificationFlag = "N" name ="JEREMY LEVINE [[email protected]]" > < / sales > "

    < / SalesCoverageTeam >

    < OffSystem Value = "F" > < / OffSystem >

    < HolidayCalendar >

    < country value = 'US' > < / country >

    < / HolidayCalendar >

    < GP2AccountStatus Value = "P" > < / GP2AccountStatus >

    < CurrencySecondary / >

    < GP2ParentID Value = '16649' > < / GP2ParentID >

    < StatementPerspective Value = "D" > < / StatementPerspective >

    < DBEntity ValueID = "2" > < / DBEntity >

    < ValueID DisplayClientName = "6892067" Name = "XXXXXXX, NEW YORK" > < / DisplayClientName >

    < FXRates >

    < FXRate >

    < of Value = "EUR" > < / From >

    < value = "Us $" > < /pour >

    < / FXRate >

    < / FXRates >

    < ShowFXRates Value = "T" > < / ShowFXRates >

    < OfficialMTMs Value = "F" > < / OfficialMTMs >

    < ExcludeStructuredTrades Value = "F" > < / ExcludeStructuredTrades >

    < ValueID = "1" DisclaimerID > < / DisclaimerID >

    < TradeRisk Value = "F" > < / TradeRisk >

    < MTMsInTradeCurrency Value = "F" > < / MTMsInTradeCurrency >

    < / ValuationRequest >

    Thanks in advance.

    with t as)

    Select xmltype (')

    http://www.w3.org/2001/XMLSchema-instance"container ="http://www.w3.org/2001/XMLSchema"ValueID ="169821"Version ="2.6"xmlns ="http://XBank.FxVals/ProcessValuationRequest.xsd"Status ="AC">

    ActiveChangesApproved

    [email protected]]">

    ') double x

    )

    Select comment_text

    t,.

    XMLTable)

    XmlNamespaces (default 'http://XBank.FxVals/ProcessValuationRequest.xsd'),

    «ValuationRequest / / comment»

    go x

    columns

    comment_text varchar2 (200) path "@Value"

    )

    /

    COMMENT_TEXT
    --------------------------------------------------------------------------------------------------------------------------------------
    Time changed according to the dtd VSG mail 25/01/2013 14:27. Beneficiary SVK added according to mail from Richard dtd 20/02/2013 09:45. SVK

    SQL >

    SY.

  • Reg: Export data of the physical database system standby.

    Hi all

    We have a standard edition one 11 GR 1 material oracle environment, I need to export the data from the physical monitoring system.

    If anyone can suggest me, how to do it safely (up state).

    Kind regards

    Konda.

    Oracle Data Guard is available only as a feature of Oracle Database Enterprise Edition. It is not available with Oracle database Standard edition.

    Then you must export data only from primary or you use EXP instead of EXPDP on the standby database. Because EXPDP create a temporary table export process duration.

    Concerning

    Mr. Mahir Quluzade

Maybe you are looking for

  • How to save all e-mail sent and received in their folder structure

    I want to just save all emails on my system, sent and received, keeping the structure of folders. How can I do this?

  • USB2 on Satellite P30

    Hello How do I install USB2 on my computer? I read the manual and the satellite P30 and USB2 is normally available. But when I look in system/hardware I can not find any USB2. I also try to uninstall all the USB driver, but still the same situation:+

  • write data 2D table spreadsheet with good alignment

    With 'write in the spreadsheet file' function to write the 2D table, I still have the problem that headers may not align well wiith data. See attached snapshot please (there are 7 column headers). WT LV2013sp1

  • nc4400 driver problems

    Hi, just did a clean install of XPpro on a nc4400, downloaded the drivers from the HP site and installed. Ethernet will not install and there are 3 unknown devices in Device Manager. Any help would be appreciated more Thank you urban

  • Question about the features of the browser inner in the storm.

    BB Storm JDE 4.7 Can someone tell me if the internal browser is entirely html 5.0 compatible?, or comment on my next statement... I managed to run 'certain' iphone web applications that use the tag, if clearly who is supported to a certain extent, as