Building ColdFusion output grouping by Date report

I build a report in Report Builder CF where he documents will appear are grouped and sorted by date in the CF query

I need to be able to view the date under the heading grouped on each page, if there is a change of date in the results on the same page, I am eager to the subtitle for the new date grouped to appear on the page as a subtitle.

the output I need is the same as the output that would be generated from the following code in a cfm output

< cfoutput query = "queryname" group = "TransactionDate" >

#TransationDate # < br >

< cfoutput >

#transationTime #, #transactiontype #, #transactionamount #.

< / cfoutput >

< / cfoutput >

When I try to generate output similar in CF Builder above I'm not able to have the TransationDate appear as the subheading once the date changes

the report simply shows the lines containing

#transationTime #, #transactiontype #, #transactionamount #.

However in the change of date there is no subtitle appearing to separate the transaction from one date to another via a titre3 void

Gezahegn on how to solve this problem would be appreciated and or reference to examples of a case fatality rate of the file which produces its results grouped within a page as described above

the answer to that is that you add the field names in the field p [aramaters so they are referenced in the report as query.fieldname not simply fieldname, then group with the titles of sub as value changes grouped reports]

Tags: ColdFusion

Similar Questions

  • between operator to group by date in the Apex of the interactive reports

    Hello

    In the filter of interactive reports, I could not find the "between" for the date field (got a "group by date" in my (source) sql query.) I wonder, is - this in view of the group by clause date?. Is there a way to show the operator "between" in the interactive reports filter.

    Thank you

    I just opened an existing style IR report, went to the actions, filter, selected a date column and found at the bottom of the list of values... Are you sure of the date that you want to filter on is a real date column?

    Thank you

    Tony Miller
    Webster, TX

    What happens if you were really stalking a paranoid schizophrenic... They would know?

    If you answer this question, please mark the thread as closed and give points where won...

  • View threads AND grouped by date/conversation

    Hello

    I try to have messages see the thunderbird grouped by date (with a thread on today, yesterday, last week, etc.) and within these threads, see email conversations. A conversation that began three weeks ago, but has received a new email today is expected to manifest itself in the thread today, but from the first email of the conversation.
    I did not understand if it is possible, as grouped by (Date) and Threads are mutually exclusive. Any ideas?

    Thank you

    As you say, it is there only one group so at the time. So I guess that your out of luck.

    You could fill a better bug report and see if one of the developers interested. https://Bugzilla.Mozilla.org/

  • GROUP BY date range to identify duplicates revisited!

    Good afternoon

    It is a continuation of the previous discussion, I previously created GROUP BY date range to identify duplicates

    Your help with the following would be appreciated (examples of data below)

    I've highlighted what I mark as returned to double as below:

    example4.jpg

    Definition of duplicate (However, this is slightly different to the previous post)

    the same account_num

    maximum 20 days apart tran_effective_date

    tran_processed_date maximum 10 days apart

    the same tran_amount

    However, I do not want to return a duplicate if they have both a tran_priced_date filled.

    So, in light of the foregoing, I don't expect the following account_numbers to be marked as duplicate:

    N100283 - one of the records has populated trab_priced_date

    N101640 - none of the records have the tran_priced_date filled

    N102395 - same as N101640

    N102827 - same as N101640

    N108876 - although the two documents have the populated tran_priced_date, the tran_effective_dates are more than 20 days apart.

    BUT for the rest of the accounts, N100284 and N102396 I want to execute the following logic

    Compare the 3rd rank in 4th place and ask the following questions:

    Is tran_effective_date to a maximum of 20 days out?

    Is tran_processed_date maximum 10 days apart?

    If yes then report it as dupe

    Compare line 4 to 5, then ask the same question until you get to the line 4 or 5. When everything is done, I want to examine only the transactions that have the status of normal and if the above question is true for both and then return to my game of result as dupes.

    I hope that makes sense!

    BEGIN
      EXECUTE IMMEDIATE 'DROP TABLE samp_data';
    EXCEPTION
      WHEN OTHERS THEN
        IF SQLCODE = -942 THEN
          DBMS_OUTPUT.put_line('');
        ELSE
          RAISE;
        END IF;
    END;
    /
    
    
    CREATE TABLE samp_data (
      ACCOUNT_NUM             VARCHAR2(17),
      TRAN_ID                 NUMBER(10),
      TRAN_TYPE               VARCHAR2(50),
      TRAN_EFFECTIVE_DATE     TIMESTAMP(6),
      TRAN_PROCESSED_DATE     TIMESTAMP(6),
      TRAN_STATUS             VARCHAR2(17),
      TRAN_PRICED_DATE        TIMESTAMP(6),
      TRAN_AMOUNT             NUMBER(13,2)
      );
    /
    
    
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N100283',140119178,'Regular With',to_timestamp('01-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('22-MAY-15 07.00.34.235000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', to_timestamp('21-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),200);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE,TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N100283',140158525,'Regular With',to_timestamp('13-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('26-MAY-15 08.39.14.090000000 AM','DD-MON-RR HH.MI.SS.FF AM'),'Normal', null,200);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N100284',140118826,'Regular With',to_timestamp('03-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('22-MAY-15 07.00.19.072000000 AM','DD-MON-RR HH.MI.SS.FF AM'),'Normal', to_timestamp('20-MAY-15 03.25.05.438000000 AM','DD-MON-RR HH.MI.SS.FF AM'),450);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N100284',140158120,'Regular With',to_timestamp('06-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('23-MAY-15 08.38.42.064000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Reversed', to_timestamp('21-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),450);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N100284',140158120,'Regular With',to_timestamp('06-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('02-JUN-15 08.38.42.064000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', to_timestamp('31-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),450);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N101640',140118957,'Regular With',to_timestamp('18-MAY-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('22-MAY-15 07.00.25.015000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', null,120);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N101640',140158278,'Regular With',to_timestamp('22-MAY-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('26-MAY-15 08.38.56.228000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', null,130);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102395',140118842,'Regular With',to_timestamp('03-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('22-MAY-15 07.00.19.665000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', null,250);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102395',140158235,'Regular With',to_timestamp('03-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('26-MAY-15 08.38.53.093000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', null,250);
    
    
    
    
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102396',140118823,'Regular With',to_timestamp('09-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('18-MAY-15 07.00.18.931000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', to_timestamp('19-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),750);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102396',140158099,'Regular With',to_timestamp('16-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('24-MAY-15 08.38.39.443000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Reversed', to_timestamp('21-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),750);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102396',140158099,'Regular With',to_timestamp('16-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('29-MAY-15 08.38.39.443000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', to_timestamp('30-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),750);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102396',140158099,'Regular With',to_timestamp('12-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('22-MAY-15 08.38.39.443000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Reversed', to_timestamp('30-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),750);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102396',140158099,'Regular With',to_timestamp('14-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('23-MAY-15 08.38.39.443000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Reversed', to_timestamp('30-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),750);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102827',140118850,'Regular With',to_timestamp('03-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('22-MAY-15 07.00.20.045000000 AM','DD-MON-RR HH.MI.SS.FF AM') , 'Normal',null,157.84);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N102827',140158118,'Regular With',to_timestamp('03-JUN-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('26-MAY-15 08.38.41.861000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', null,157.84);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N108876',139840720,'Regular With',to_timestamp('01-MAY-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('11-MAY-15 08.35.34.646000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', to_timestamp('20-MAY-15 03.25.05.438000000 AM','DD-MON-RR HH.MI.SS.FF AM'),1000);
    Insert into samp_data (ACCOUNT_NUM,TRAN_ID,TRAN_TYPE,TRAN_EFFECTIVE_DATE,TRAN_PROCESSED_DATE, TRAN_STATUS, TRAN_PRICED_DATE,TRAN_AMOUNT) 
    values ('N108876',139889880,'Regular With',to_timestamp('22-MAY-15 12.00.00.000000000 AM','DD-MON-RR HH.MI.SS.FF AM'),to_timestamp('12-MAY-15 08.49.29.080000000 AM','DD-MON-RR HH.MI.SS.FF AM'), 'Normal', to_timestamp('21-MAY-15 03.26.18.954000000 AM','DD-MON-RR HH.MI.SS.FF AM'),1000);
    /
    
    
    select * from samp_data
    ORDER BY account_num, tran_effective_date, tran_processed_date;
    

    PL continue the discussion in your original post

  • VMware Capacity Planner data report - shows don't not all systems in the dashboard Version 2.7

    I worked on an evaluation with over 1300 machines for a few weeks now. I just ran a VMware capacity Planner data report today and it shows that 37 systems. Anyone else having this problem? I started a Monday SR something else and still have no answer, so I thought I would try the forum first this time.

    Dave Convery

    VMware vExpert 2009

    http://www.dailyhypervisor.com

    http://communities.vmware.com/servlet/JiveServlet/download/38-20623/vExpert_logo_100x57.jpg

    Prudent. We do not want to make of this.

    Bill Watterson, "Calvin and Hobbes".

    Try to create a group of man LAN and add the systems and re - run the report.

    Which fixed the problem with a report, I ran...

  • Group by Date and place of DateTime

    Hello

    I'm looking for some counts that are grouped by date / time to date.  For example, I run the query below.

    ALTER session SET NLS_DATE_FORMAT = "mm/dd/yyyy";

    SELECT count (distinct contact_id) as charges, created in the created_date

    of s_con_chrctr

    where created > = TO_CHAR('07/01/2015') and created < = TO_CHAR('07/10/2015')

    and created_by = 1-J9455J'

    Group by created;

    The results are as follows.

    COUNTIES CREATED_DATE
    102/07/2015
    102/07/2015
    102/07/2015
    102/07/2015

    I think that what is happening because this column is a datetime column and his back one record for each time specific in this day.  The results, I'm looking for are...

    COUNTIES CREATED_DATE
    1502/07/2015
    1703/07/2015
    3504/07/2015
    1105/07/2015

    How to change my query to group by date and not by date time?

    So created is a date, then do not compare with TO_CHAR literals.

    If you want to skip the part time in your account, use TRUNC:

    SELECT count (distinct contact_id) as charges, TRUNC (created) as created_date

    of s_con_chrctr

    where created > = TO_DATE (July 1, 2015 "," MM/DD/YYYY ') and created< to_date('07/11/2015',="">

    and created_by = 1-J9455J'

    TRUNC Group (created);

  • Call stored procedures and using its output parameters in a report

    Hello

    I have a procedure defined in a package on the database. This procedure returns a number of output parameters. How to call this procedure and use the output parameters in my report?

    Thank you, Mark

    Depends on the level at which the procedure should be called. If its at the State level, call the form in the trigger-REPORT-BEFORE (or AFTER-PARAMFORM) and store the off-values space reserved columns. These you can use anywhere in your report.

  • Event for groups of data listeners

    Hello

    It's really a silly question, but nobody knows how to define an event listener for a group of data? Is there a 'method called' property for control of presentation "Custom Layout". I don't know how to define a method in a business object so that it shows in the dropdown of listener event.

    The good links on the use of data groups are also appreciated.

    Thank you very much in advance.

    Yes - when you click on the table/group widget in the form editor, there is a section "Event listener" with a property called "invoked method.

    I think you've thought this out on your own, but if you have trouble seeing a method from the drop-down list, create a method that has a type argument Fuego.Util.GroupEvent. If you look at this class, you can see the different types of events that can be received (ADD, INSERT_DOWN, INSERT_UP, REMOVE, and REMOVE_LAST).

    Dan

  • Report of the Group of dates for end of month

    Hi guys

    I need to get 6 months of data grouped by months end dates...

    I need to automate this query so that it looks always to date for the last 6 months

    How do I get there?

    The reason for writing the automated query, it's that this sql script is assigned to the business objects reporting.

    I use the pl/sql

    Here's the DOF

    create the table #Something
    (
    base_date datetime, int connections
    )

    Insert #Something
    Select ' 01 jul 2013', 21 Union all the
    Select ' 02 jul 2013', all 22 union
    Select ' 03 jul 2013', 210 Union all the

    ....

    ...

    Select July 31, 2013,"Union 498
    ' Select ' 01 August 2013 ', 44 Union all the
    ' Select ' 05 August 2013 ', 66 Union all the
    ...

    ....

    Select ' 03 dec 2013", 456 union all
    .

    .

    'Select December 31, 2013,"788

    ..............................................................................

    Desired output

    Connection of base Date

    31 jul 500 (not exact, just randomly)

    August 31 600

    30 sep 356

    31 Oct 676

    30 Nov 544

    31 Dec 456

    ..............................................................................

    Hope this helps

    Hello

    In Oracle, you can do this:

    SELECT LAST_DAY (TRUNC (base_date)) AS last_day_of_month

    SUM (connections) AS total_connections

    OF something

    WHERE base_date > = TRUNC (ADD_MONTHS (SYSDATE,-6), 'MONTH')

    AND base_date<  trunc="" (           ="" sysdate,     ="">

    GROUP OF LAST_DAY (TRUNC (base_date))

    ORDER BY last_day_of_month

    ;

    If you would care to post some CREATE TABLE and INSERT statements so you want the results of work from these data of the sample, then I could test this.

    Using Oracle?  #Something (with a sign # at the beginning) is not valid a table name, and datetime is not a valid data type in Oracle.

    Always say what version of Oracle you are using (for example, 11.2.0.2.0).

    See the FAQ forum: https://forums.oracle.com/message/9362002

    The query above works if base_date is a DATE or TIMESTAMP.

    Since it is now January 2014, what are "the last 6 months?  The above query assumes that they are the last 6 months, that is, July to December 2013.

    If you mean the current (incomplete) and 5 months before it (i.e. August 2013 to January 2014), then add 1 month to the two deadlines in the WHERE clause:

    WHERE base_date > = TRUNC (ADD_MONTHS (SYSDATE,-5), 'MONTH')

    AND base_date<  trunc="" (add_months="" (sysdate, ="" 1) ="">

  • How to associate cluster esx with the data report store

    Hi all

    I'm running a repore that displays data stores less than 50 free concerts. The storage team would like a coloum for esx clusters associated with the LUN. I can't see to find it and what is possible.

    Thanks in advance,

    Scott.

    I would create a WCF function, script type, with a timeRange parameter and a new data store VMWDatastore context parameter, required, not a list.  Set the output in common: string.

    Add the following script:

    QS = server. QueryService

    Query = "!" "VMWCluster where esxServers.datastores.uniqueId = ' ${datastore.uniqueId}"

    cluster = server. QueryService.queryTopologyObjects (query)

    clusterNames = clusters? collect {cluster-> cluster?. name}?. Join(', ')

    return clusterNames

    You can then use this in your report, feeding the entrance of context for the current row data store.

  • The matrix - unescaped data report

    Following my recent question on the matrix report proposed by fac586

    I was wondering how you could modify the XML not escape content coming from 'nominative' (query below)

    for example: I added an html tag similar to the empno field, to contrive to script at the client level.

    When analyzed by magical xmltransform etc.

    with thedata as (
      select
          deptno, job
         ,htf.bold('deptno:')||deptno||';'||count(empno) over (partition by deptno order by null) empno
      from scott.emp o
    )
    ,order_pivot as (
        select
            deptno
          , job_xml
        from  thedata
        pivot xml
            (max(empno) n_orders for job in (select distinct job from thedata)))
    select
        deptno
      , xmlserialize(
          content
          xmltransform(
              job_xml
            , xmltype(q'{
                <xsl:stylesheet version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform">  
                  <xsl:output method="html" />
                    <xsl:template match="/PivotSet">
                      <xsl:for-each select="item">
                        <td>
                          <xsl:attribute name="headers">
                            <xsl:value-of select="column[@name='JOB']" />
                          </xsl:attribute>
                          <xsl:value-of select="column[@name='N_ORDERS']" />
                        </td>
                      </xsl:for-each>
                    </xsl:template>
                </xsl:stylesheet>}'))) matrix
    from     order_pivot
    order by deptno;
    
    
    

    the data is escaped.

    <td headers="ANALYST"></td><td headers="CLERK">&#60;B&gt;deptno:&#60;/B&gt;10;3</td><td headers="MANAGER">&#60;B&gt;deptno:&#60;/B&gt;10;3</td><td headers="PRESIDENT">&#60;B&gt;deptno:&#60;/B&gt;10;3</td><td headers="SALESMAN"></td>
    
    
    


    I want to keep the markup for use in the final report, that is to say:

    <td headers="ANALYST"></td><td headers="CLERK"><b>deptno:</b>10;3</td>
    
    
    

    I would also like to know if there is a better way to do this kind of pivot, re: MAX (expr)

    See you soon

    Scott

    Scott.Wesley wrote:

    Following my recent question on the matrix report proposed by fac586

    I was wondering how you could modify the XML not escape content coming from 'nominative' (query below)

    for example: I added an html tag similar to the empno field, to contrive to script at the client level.

    When analyzed by magical xmltransform etc.

    1. with the nominative as)
    2. Select
    3. DEPTNO, job
    4. htf.bold('deptno:') | DEPTNO | « ; » || Count (empno) on empno (partition sort null deptno)
    5. from scott.emp o
    6. )
    7. order_pivot as)
    8. Select
    9. DEPTNO
    10. job_xml
    11. of the crossroads
    12. Pivot xml
    13. (max (empno) n_orders for employment (select distinct nominative use)))
    14. Select
    15. DEPTNO
    16. xmlserialize)
    17. content
    18. xmlTransform)
    19. job_xml
    20. , xmltype (q' {})
    21. http://www.w3.org/1999/XSL/transform">
    22. (((}'))) matrix
    23. of order_pivot
    24. order by deptno;

    the data is escaped.

    1. <><><>

    I want to keep the markup for use in the final report, that is to say:

    1. DEPTNO:10; 3

    Apply the utl_i18n.unescape_reference method on the xmlserialized output:

    with thedata as (
      select
          deptno, job
        ,htf.bold('deptno:')||deptno||';'||count(empno) over (partition by deptno order by null) empno
      from scott.emp o
    )
    ,order_pivot as (
        select
            deptno
          , job_xml
        from  thedata
        pivot xml
            (max(empno) n_orders for job in (select distinct job from thedata)))
    select
        deptno
      , utl_i18n.unescape_reference(
            xmlserialize(
              content
              xmltransform(
                  job_xml
                , xmltype(q'{
                    
                      
                        
                          
                            
                              
                                
                              
                              
                            
                          
                        
                    }')))) matrix
    from    order_pivot
    order by deptno;
    
  • Questions about the data reporting PowerCLI store

    Hello

    I have a problem with a script that I use and I'm looking for an idea on the problem.

    I have a script that gathers pretty much all the information on the virtual machine in my cluster. (Disk size, free disc, data store).

    Most of the time, it works very well. But there is something that bothers me little. It doesn't report the data store that each VM is linked to. It works for most, but on a select few, it's not.

    I wonder if there is something in my script which is the first cause, or if it is a question of VSphere.

    My script and the output is attached.

    Thanks in advance for the help.

    -Aaron

    I was able to get the warehouses of missing data as well.

    That's when the VMDK are stored on more than 1 data store for a specific virtual machine.

    In this case, the line

        $myDS = Get-Datastore | Where { $_.Id -eq $VM.Datastore}
    

    nothing will store in $myDS because $VM. Data store is an array of MoRefs.

    In fact you try to map the partitions in the guest OS with data warehouses that host the VMDK files.

    There is no easy solution for that I fear (there are number of threads in this community who are trying to do the same thing).

    I fear that there is no real solution for this problem

  • Need help with query outputing group names

    I'm trying to find a way for groups of output headers, then all the records in each group etc header. It would be easy, except there is a key with what I want to do.

    Normally, if I have this data set (that I've ' borrowed' a site that showed the closest to what I was looking for):

    Example table:

    TABLE [number]

    (Name, NUMBER)

    Dave Bower 843-444-4444

    Dave Bower 843-555-5555

    Matthew Small 843-111-1111

    Matthew Small _843-222-2222

    Matthew Small 843-333-3333

    I could use the following code:

    < cfoutput query = "somequery" group = "name" >

    #name # < br >

    < cfoutput >

    #phonenumber # < br >

    < / cfoutput >

    < hr >

    < / cfoutput >

    And get this:

    Dave Bower

    843-444-4444.

    843-555-5555.

    -------------------

    Matthew Small

    843-111-1111.

    843-222-2222.

    843-333-3333.

    -------------------

    BUT my actual tables are not set up like that. Rather than recording of every name of every record, I would have an ID that is the foreign key to another table.

    Current table set up is as follows:

    TABLE [people]

    (ID, NAME)

    1 Dave Bower

    2 small Matthew

    TABLE [Phones]

    (PEOPLE_ID NUMBER)

    1 843-444-4444

    1 843-555-5555

    2 843-111-1111

    2 843-222-2222

    2 843-333-3333

    If this output actually would this give me with my current setup and request above code:

    1

    843-444-4444.

    843-555-5555.

    -------------------

    2

    843-111-1111.

    843-222-2222.

    843-333-3333.

    -------------------

    How can I keep my current setup but create a query that produces the same result from the top? (The table names of people like the group headers, but data from the phones table under that output)

    You must gather the two tables, and then group the output.

    Something along the lines of the (may vary slightly depending on your DB and the exact table structure)

    SELECT ppl.name, ph.number

    PEOPLE ppl

    INNER JOIN phones ph ON ppl.id = ph.people_id

    ORDER BY ppl.name

    See you soon

    Kai

  • Enter in the input area, output of the query / report of the AN_id, to query/display page.

    Hello
    I tried to create a text input area to enter a type of alphanumeric data, however it does not work for me again.
    Can someone help me here is the new code 2, I added to my query results.
    Basically, I have 2 pages
    One is the entry (index)
    And the other is where is the logic. / query

    I was working on a fifth Query but now since I added the code. It doesn't yet?
    Can someone help me?
    Thank you.
    And thank you

    In the box, I have this in the index.cfm

    < label > enter AN ID
    < input type = "text" name = "AN_ID" / >
    < / label >

    In where clause is what I have the requests page or

    AN_ID = ' #form. "AN_ID #

    Here is the code:

    index page

    Sort by:
    < select name = "Orderby" size = "1" >
    < option value = "AN_ID" > YEAR ID < / option >
    < option value = "AN_DATE" > a DATE < / option >
    < / select >
    < br / >
    < br / >
    Output format:
    < input type = "radio" name = "outputFormat" value = "HTML" checked = "checked" / >
    HTML
    < input type = "radio" name = "outputFormat" value = "PDF" / >
    PDF
    < input type = "radio" name = "outputFormat" value = "CSV" / >
    Download/open Excel Spreadsheet < BR / >
    < BR / >
    < /p >
    < / make >
    < p > < / p >
    < / div >

    < form id = "form1" name = "form1" method = "post" action = "" >
    < label > enter AN ID
    < input type = "text" name = "AN_ID" / >
    < / label >
    ----------------------------------------------------------------------------------------- ------------------------------------------------------------

    the query page and view the page


    < cfquery name = "getHIPPOcases" datasource = "oracle10" >
    Select
    Thecasertab.AN_ID as AN ID
    Thecasertab.AN_DATE as a date
    Thecasertab_requi.case_requi_id
    Thecasertab_requi.requi_type_cd
    Thecasertab_requi. App_user_id


    Of
    thold_case
    tloc
    tla_prop
    Thecasertab
    tref_plan_area
    Thecasertab_requi

    where
    Thecasertab_requi.caser_id = Thecasertab.caser_id
    and
    thold_case.caser_id = Thecasertab.caser_id
    and
    AN_ID = ' #form. "AN_ID #



    order by

    < cfswitch expression = "#Form.orderBy #" >
    < cfcase value = "AN_ID" >
    CASE. AN_ID
    < / cfcase >
    < cfcase value = "AN_DATE" >
    CASE. AN_date
    < / cfcase >
    < / cfswitch >


    < / cfquery >


    <! - report html - >
    < cfswitch expression = "#Form.outputFormat #" >
    < cfcase value = "HTML, PDF" >

    < cfsavecontent variable = "htmlData" >
    <! DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional / / IN" " http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd" >
    "< html xmlns =" http://www.w3.org/1999/xhtml "> "
    < head >
    < meta http-equiv = "Content-Type" content = text/html"; charset = iso-8859-1 "/ >"
    getHIPPOcases < title > < /title >
    < / head >
    < style type = "text/css" >

    TD {}
    do-family: Arial, Helvetica, without serif.
    do-size: 12px;
    }
    Th {}
    do-family: Arial, Helvetica, without serif.
    do-size: 12px;
    make-weight: bold;
    background-color: #FFFF99;
    }
    H2 {}
    do-family: Arial, Helvetica, without serif.
    do-size: 14px;
    }
    H3 {}
    do-family: Arial, Helvetica, without serif.
    font-size: 15px;
    }
    < / style >

    < body >
    < cfoutput >
    < table border = "0" cellpadding = "3" cellspacing = "0" >
    < b >

    < td align = "center" >
    getHIPPOcases < h3 > < / h3 > < br > < br > < table >
    < /tr >
    < b >
    < td align = "center" >
    < / h2 > your query returned #getHIPPOcases.RecordCount # entries < / h2 > < table >
    < /tr >
    < b >
    < td >
    < table border = "2" cellpadding = "2" cellspacing = "0" >
    < b >
    Number < /th > < th > Record
    < width th = "120" > YEAR ID < /th >
    date of YEAR < /TH > < TH >

    < /tr >

    < cfloop query = "getHIPPOcases" >


    < tr bgcolor = "< cfif currentrow mod 2 > F8F8FF < cfelse > WHITE < / cfif >" >
    < td > #CurrentRow # < table >
    < td > #AN_ID # < table >
    < td > #dateformat(AN_date,"mm/dd/yyyy") # < table >


    < /tr >
    < / cfloop >

    < /table >
    < table >
    < /tr >
    < /table >

    < / BODY >
    < / HTML >
    < / cfoutput >
    < / cfsavecontent >

    < cfswitch expression = "#Form.outputFormat #" >

    < cfcase value = "HTML" >
    < cfoutput >
    #htmldata #.
    < / cfoutput >
    < / cfcase >
    < cfcase value = "PDF" >
    PDF
    < cfdocument format = "pdf" >
    < cfoutput >
    #htmldata #.
    < / cfoutput >
    < / cfdocument >
    < / cfcase >

    < / cfswitch >





    < / cfcase >

    < cfcase value = "CSV" >


    < NAME CFHEADER = "Content-Disposition" VALUE = "attachment; filename = Citywide.csv">
    < cfcontent type = "application/msexcel" > "YEAR ID ', 'a date '.
    < cfoutput query = "getHIPPOcases" > #ltrim (AN_ID) #, "#dateformat(AN_date,"mm/dd/yyyy")" # ""
    < / cfoutput >


    < / cfcase >
    < / cfswitch >



    Quote:
    Posted by: briankind
    on the entry
    I get home an_id value let lov-1234-oop-ui
    and
    PRESST the button
    It's notreturn results (nothing happens)
    Thank you

    In this case, hard-code the value "lov-1234-oop-ui" in your query instead of the form variable. You get a result?

  • How to build a table of measured data permanently

    Have a good day for everyone,

    I have a problem about the construction of a table permanently. I tried to find the solution for a long time, even if it's a task really simle.

    I want to build a table permanently, virtually:

    I like to read data from a measuring devices and I would like to sotre it one by one, at all times. I made a simple VI to present one where my problem is. The random generator represents the data given the measuring device.

    At currently I am able to store data only 1. When I want to store then that it is crushed.

    Thanks for anyone who can help me.

    Hi LIG.

    no need for a structure business when you can use autoindexing:

    Also, the code snippet shows how to convert a number in your reading string. You probably prefer tanks instead of integers...

Maybe you are looking for