We are a global group with divisions using a single instance of Eloqua

Any advice on that would be great.

We are a group of Gloabl, with a number of different divisions, offering different solutions and IT services. Each has its own roadmap marketing whereby a single database will be used to target and segment via marketing activity.

Just after a few tips and best practices that I guess, I hope that there is someone out there who has tackled a similar problem...

(i) what is the best way to target these divisions within a single database, without causing too many headaches.

(II) we have one instance of ELQ, in terms of differentiation of the divisions, is the easy approach just set up different email for camapigns groups? That is to say, different headers, footers and preference centers etc...?

Have not begun to consider the side ELQ of things yet, but I was wondering if someone has approached this recently, before I deep dive and if they give me advice?

Thanks in advance.

Thanks for the concise answers. Excellent advice that will for sure and is straightforward. Quickly discover that Topliners is the place to come for any "staggering" Eloqua emits...

Thanks again.

Tags: Marketers

Similar Questions

  • Can we use 2 proceedings of Essbase with Shared Services of single instance

    Hello

    We have Essbase installed on 2 different servers and shared on the third server services. Can I use the same SSP for the 2 instances of Essbase?

    Thanks in advance
    PM

    Yes, multiple instances of essbase can be entered against one of shared services.

    See you soon

    John
    http://John-Goodwin.blogspot.com/

  • Management of location mutli using vCAC single instance

    Hello

    I'm surprised to see there is no much information about how to deploy vCAC instead of managing Multi environments and deploy location-based as well.

    I read somewhere in the vCAC documentation referred to on installation of DEM to help rental multi-site management, but nothing more is mentioned in the documents in detail.

    I would really appreciate if can get this information in details

    Thanks in advance

    ARO

    MG

    From therefore have two server vCenter, I the vCAC unique environment.

    1. Deploy the proxy agent of vSphere on a virtual computer in each site that has the server vCenter, when done installing specify endpoint vCenter to point to the proxy of local agents vCenter server.
    2. Edit the c:\Program Files (x86)\VMware\vCAC\Server\Website\XMLData\DataCenterLocations.xml - replace the names of your data center, London and Boston
    3. Change the cluster in your business groups in respect of computing resources and assign them to the location of the actual location, as they are.
    4. Change your plan for the seizure of location in the query
    5. You can probably use the same name of action plan on the two sites, and allowing you to have a service plan of action for each site. Problem is perhaps that you have different prefixes of host names for each site, and you will have to address this problem and then maybe to keep two different blue print. In this case, you can use reservation policies to force a plan to go to a special reserve, which also give the appropriate site.
  • Using a single instance of a Clip to load graphics library

    Hi all

    Another question... How would go on the loading of a variable in a movie clip and draw a different graphic in my library at different places in the main timeline?

    Thank you!

    Patrick

    attachMovie is AS2 code, it does not exist in AS3.  Here is a link to a sample that I have made for you with the ealrier code, dig and see what you missed.  I can't open your file, which means that you probably use CS4.

    http://www.nedwebs.com/Flash/AS2_attach.fla

  • Regarding the installation of a single instance with ASM in AIX

    Hello
    I try to install a single instance with the help of ASM in AIX. I made the requirements for drives prior to installation.
    I'm running Universal Installer and I choose the option of database to create.
    During the stage of creation of the database, there is a page to choose the types of storage.
    I choose the DSO and then click Next.
    Caution is open and he tell ASM must CSS and it gives the following to run.
    reset of the/Home/App/tuba2/prod / 10 GR 2/bin/localconfig.

    I run but my CSS is not started and it gives the following.
    # localconfig reset
    Successfully accumulated necessary OCR keys.
    Creating OCR keys for the user 'root', privgrp 'system '...
    Successful operation.
    Configuration for local CSS has been initialized
    Stale demon CSS is running... kill now
    Adding inittab
    Startup will be queued to init within 30 seconds.
    Checking the status of the new Oracle init process...
    Expected CRS demons to be in place within 600 seconds.
    Abandon: Oracle CSS stack seems to NOT be running.
    Oracle CSS service would not start as installed
    Automatic storage management (ASM) cannot be used until the Oracle CSS service is started

    How can I solve these problems. I'm new in ASM.
    Is there anyone to help?
    Is also the way I was using the single instance of the correct installation with ASM? or there is another way that is more efficient for the installation of the database with ASM?

    Technically, there is nothing wrong in what you do. Since your log message, it seems, has already CSS configured on your system and is not in a working state.
    I would recommend, reconfiguration of the CSS.

    Complete the following as root user.

    $ORACLE_HOME/bin/localconfig delete
    $ORACLE_HOME/bin/localconfig add
    CD /etc/init.d
    . /init. CSSD run &
    crsctl check crs
    And you can check the logs for more from diagnosis to the suite.

    $ORACLE_HOME/log /<$HOST>/cssd/cssdOUT.log
    $ORACLE_HOME/log /<$HOST>/cssd/ocssd.log
    / var/log/messages / *.

    HTH,
    Arul.

  • What are the advantages and disadvantages of using iCloud drive (or other Cloud services) compared with iTunes game against external hard drive, Solid State etc. to store my songs iTunes, playlists, ratings etc?

    First question: what are the advantages and disadvantages of using iCloud drive (or other Cloud services) compared with iTunes Match against Solid State external hard drive etc to store my songs iTunes, playlists, ratings etc?

    Please let me know what are the alternatives for storage of my songs, playlists, song etc. other than iTunes game ratings ($ 24.99 / year)?  Is there a way to store all my songs with my song ratings, playlists, etc. with a CD song quality? Most of my (currently 130 GB) songs on my computer are CD quality that I prefer to keep as well as my notes of each song (i.e. 5 * vs. 4 * etc.), playlists, groups (music, classical, popular, country etc.).  What are the advantages and disadvantages to store my iTunes songs while CD quality with my personal notes, reading lists, groupings in the cloud?  ICloud of Apple Drive ($ 2.99 per month or $36 / year for 200 GB) would be the best way to automatically update and maintain my rankings iTunes personal including the future editions of the song etc or using Amazon, IBM, Microsoft etc cloud or an external hard drive SSD would work as well and maybe cheaper?  I currently have my iTunes on a PC of 2007 with vista (I use a 1 8T Western Digital external upward by car for my personal info all over HP, songs, photos etc. which is currently close to 170 GB) but also have an Apple i-phone 6 & I Air cushion.  Although I did not need to have all of my music with all their sides, etc. available on all of my devices at all times as iTunes game, I still would have easy access to information on both my PC and my iPhone or iPad?  Thank you.

    In my opinion?

    I back up my data to an external drive for several reasons.  The first is my data - if I put on the cloud, someone else has control over it, they can see what I have and change the preferences to what they see of marketing.

    Almost as important as that - if I had to rely on the cloud whenever I wanted to update my playlists on my phone, I need a better internet connection I have now (working at sea).  To get the best out of the cloud, you need a solid link to home, work or mobile.

    As a spin-off - I hate the fact that Apple take away a feature in their software updates and then allow an app to be sold in their store to replace this feature - or worse still, some features that were available before the updates are now "cloud only."  And Yes - you have to pay for the cloud.  I like most of the people do not appreciate the 'accomplished fact. "

    From the back of the external hard drive ASI downside, is that to be safe, you need to store outside in a separate computer place.  If keep you them together, if one is lost, probably, they are both.  That is to say for me a full backup before you travel, then a new upward as soon as I go home.  Downside is how much data may be lost between the two back ups if I needed to restore.  For me, its worth the risk.  If I was away and say had bought a ton of new music and couldn't go home anytime soon, I can always head to the store and buy a new external drive, they are massively expensive.  Even if my computer has to go for repair and my remote is to halfway around the world - I want to buy a new one.

    Now, some will say that once you buy a new external, you spend more than you would on a cloud subscription.  It's true, but in the 6 years following the purchase of Mac that I've only had to buy a new external once - so with a little luck and planning - it works.

    I left took well - I am not a subscriber of cloud, nor do I want to be.

  • Can I have different AP with SSID groups using a single controller?

    I am designing a wireless network that will provide different SSIDS to various parts of a campus.  Since it is open to the public, I can't use an assignment raius Server vlan dynamic.  Anyone know if I can do with one controller or will it take a controller (or redundant pair) for each zone?  Thank you!

    From liberation WLC 5.2 and later versions, you can have a lot (I'll have a doubt whether it is 255 or 512) SSID configured on the WLC. Given that you can have up to 16 on an AP, you can then assign some SSID for some APs.

    The key is to configure the AP group.

    An AP group defines a list of access points, which SSID these APs are maintenance and what interface they use for each SSID (the default interface mentioned in the config SSID or a substitution).

    So you can have one SSID on your entire campus, but different groups AP it will link with different interfaces. I think that's what you are trying to reach.

    Nicolas

  • Some build-in panels are grouped with my plugin Panel in InDesign CC

    My settings for panelist look like this:

    resources LocaleIndex (kSDKDefPanelResourceID)

    {kViewRsrcType,

    {kWildFS, k_Wild, kSDKDefPanelResourceID + index_enUS}

    };

    / * Definition of panelist.

    */

    resources involved (kSDKDefPanelResourceID)

    {

    {

    1 group in the list

    kSDKDefPanelResourceID, / / resource ID for this Panel (use SDK by default ID rsrc)

    kMyPluginID, / / ID of the plugin which holds this Panel

    kIsResizable,

    kMTPanelWidgetActionID, / / Action ID to show/hide Panel

    "MyPlugin:Test", / / appears in the list window.

    "", / / Substitute the form menu path ' hand: Foo "If you want your palette in the second menu item to place

    0.0, / / menu replacing the position of the alternative Menu to determine the order of the menu

    kMyImageRsrcID, kMyPluginID, / / Rsrc ID, ID Plugin for a PNG icon resource to use for this palette (when it is not active)

    c_Panel

    }

    };

    resources MTPanelWidget (kSDKDefPanelResourceID + index_enUS)

    {

    __FILE__, __LINE__, / / macro location

    kMTPanelWidgetID, / / WidgetID

    kPMRsrcID_None, / / RsrcID

    kBindAll, / / Binding (0 = none)

    0, 0, 320, 386, / / framework: left, top, right, bottom.

    kTrue, kTrue, / / Visible, Enabled

    kTrue, / / clear the background

    kInterfacePaletteFill, / / Color Erase

    kMTPanelTitleKey, / / name of Panel

    {

    }

    "MyPlugin" / / name of contextual menu (internal)

    };

    The problem is that when I click on the window-> MyPlugin-> Test liquid integrated panels layout of the article and references are also open and grouped with my Panel. I don't understand what I'm doing wrong. The situation is the same with all the examples in the sdk that have signs. Help, please. 10 x in advance.

    My previous post was that this panelMgr-> GetPanelFromWidgetID (kMyPanelWidgetID) does not find the Panel when it is called in the initializer.

    I'm now deep - search the trees from the GetRootPaletteNode paletteRef, using PaletteRefUtils methods. For kTabPanelContainerType, I used panelMgr-> GetPanelFromPaletteContainer() to check the ID of widget, fortunately callers already works. For other types, I just go down the children.

    Three steps more to move the paletteRef found in a new floating window:

    columnRef = PaletteRefUtils::NewFloatingTabGroupContainerPalette(); kTabPaneType

    groupRef = PaletteRefUtils::NewTabGroupPalette (column); kTabGroupType

    PaletteRefUtils::ReparentPalette(myPanelRef,groupRef,PaletteRef());

    If you have several panels, Iterate step 2 + 3 to create groups (lines) per Panel, or just step 3 if they all end up stacked.

    Edit: I can find why GetPanelFromWidgetID did not work: apparently initializer priority is considered by plugin, so one of my signs in a plugin different just was not yet registered. Forget all that deep-search tips.

  • How to stop a Web site generating spams are sent to me - it is all associated with the use of viagra and I had 10 different emails today

    How to stop a Web site generating spams are sent to me - it is all associated with the use of viagra and I had 10 different emails today

    Thank you

    You need to ask the experts of Hotmail one.
     
     
    Help for Windows Live Hotmail can be found in this forum.
    http://answers.Microsoft.com/en-us/windowslive/Forum/Hotmail?tab=all 

    Help for Outlook.com is located in this forum
    http://answers.Microsoft.com/en-us/windowslive/Forum/mail?tab=all
     
  • Deployed Adobe CC 2014 32 bits to our neighborhood. Students launch Adobe Flash Professional and are able to start a project using ActionScript 2.0. But as soon as they click on save and try to save their project... boom... an error occurs "met with an in

    Deployed Adobe CC 2014 32 bits to our neighborhood. Students launch Adobe Flash Professional and are able to start a project using ActionScript 2.0. But as soon as they click on save and try to save their project... boom... an error occurs "met with an invalid argument' it happens to any student and any computer. Help, please!

    They can only use the old Adobe CS5 flash professional because it does not crash.

    Thank you

    Raul

    [email protected]

    CC for flash doesn't support AS2, so perhaps that is the reason.  CS6 (or earlier version) supports AS2.

  • Photo constantly order of photos in a slideshow changes no matter how many times the movements of the user the photo back to the good look at an order. Example: Bathroom Plans eventually grouped with pictures of kitchen! I have found no way to stop this o

    Apple Photo 1.3 serious problems - how can I SOLVE all these problems?

    (1) breaks down without rhyme or reason no matter where I am in the workflow.

    (2) pictures will not be Shut Down Every Time, even after several days of waiting.

    (3) aPhoto frequently badly chooses picture in the EDIT picture option, I get a picture different than the one I clicked on which is on a 100 pictures in a row.

    (4) picture constantly order of photos in a slideshow changes no matter how many times the movements of the end user the photo back to the good look at an order. Example: Bathroom Plans eventually grouped with pictures of kitchen! I have found no way to stop this weird behavior! Is there a way to stop this? If I drag the photo again some 7 additional photos in the slide show, after a minute or less, he appears again to where it was it not. !@#$%$#

    (5) If you make any CHANGES to a photo, it often changes the appearance of your complete slideshow of this picture with impatience. So you lose all this not work fix your configuration of the slide show. Even changing the order of photos once more that I had put back where they should be. !@#$$#@

    (6) photo identifies often shades of lamps and long door handles as the faces of the people.

    (7) photo made bad decisions when it comes to brightness, contrast and colors effortlessly around other than to use other software, where as with iPhoto there was a lot of workarounds. I could continue, but will save one who might be reading of this.

    I am up to date on all updates for my Mac. If anyone have REAL answers so please spilling the beans, but according to me, it's the only truth is that Apple has rolled out a product inferrer to replace an exceptional product, called iPhoto, which does not work on my new iMac computer 5K of 27 ".   If I knew what I would have chosen another computer that I use iPhoto to prepare more of fifty to sixty thousand photos in a given year and I use iPhoto to make hundreds of slideshows from it.  Are there plugins for Photo 1.3? I ask because I see where there could be Add-ons, but I can't find.

    Apple has taken a serious decision by turning his back to iPhoto and tens of millions of loyal users.

    Thanks in advance to anyone brave enough to tackle this job.

    James

    First, back up your library of Photos and hold down the command and option keys while launching Photos - repair your database - you have a corrupted database

    LN

  • A Cisco ME-3600 X-24FS-M can support pppoe bba-global group?

    Hello

    I'm new on routing and switching, nicely naked with me. I have a Cisco 1941 router that I'm trying to reproduce on a Cisco 3600 ME X-24FS-M but whenever I try to configure pppoe bba-world group I can't to, configuration see error below:

    xxx_xxx (config) #bba - pppoe global group
    ^
    Invalid entry % detected at ' ^' marker.

    xxx_xxx (config) #.

    Y at - it any other way for me to set up a group of bba or a way to activate on a Cisco ME?

    Thanks again.

    Hello

    1941 is a router and 3600 X is a switch. That's why many devices you need to 1941 will be unavailable on 3600 X, and this is one of them. You can use Cisco Feature Navigator to find the features supported on a specific image of the platform or the software.

    -

    pajaja

  • [8i] grouping with delicate conditions (follow-up)

    I am posting this as a follow-up question to:
    [8i] grouping with tricky conditions

    This is a repeat of my version information:
    Still stuck on an old database a little longer, and I'm trying out some information...

    BANNER

    --------------------------------------------------------------------------------
    Oracle8i Enterprise Edition Release 8.1.7.2.0 - Production
    PL/SQL Release 8.1.7.2.0 - Production
    CORE 8.1.7.0.0-Production
    AMT for HP - UX: 8.1.7.2.0 - Production Version
    NLSRTL Version 3.4.1.0.0 - Production

    Now for the sample data. I took an order of my real data set and cut a few columns to illustrate how the previous solution didn't find work. My real DataSet still has thousands of orders, similar to this one.
    CREATE TABLE     test_data
    (     item_id     CHAR(25)
    ,     ord_id     CHAR(10)
    ,     step_id     CHAR(4)
    ,     station     CHAR(5)
    ,     act_hrs     NUMBER(11,8)
    ,     q_comp     NUMBER(13,4)
    ,     q_scrap     NUMBER(13,4)
    );
    
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0005','S509',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0010','S006',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0020','A501',0.85,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0026','S011',0.58,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0030','S970',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0040','S970',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0050','S003',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0055','S600',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0060','Z108',6.94,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0070','Z108',7,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0080','Z310',4.02,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0085','Z409',2.17,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0090','S500',0.85,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0095','S502',1.63,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0110','S006',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0112','S011',0.15,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0117','S903',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0118','S900',0,9,1);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0119','S950',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0120','S906',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0140','S903',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0145','S950',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0150','S906',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0160','S903',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0170','S900',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0220','S902',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0230','S906',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0240','S903',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0250','S003',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0260','S006',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0270','S012',0.95,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0280','Z417',0.68,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0285','Z417',0.68,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0290','Z426',1.78,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0300','Z426',2.07,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0305','Z426',1.23,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0310','Z402',3.97,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0315','Z308',8.09,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0410','Z409',4.83,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0430','S500',3.6,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0435','S502',0.43,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0450','S002',0.35,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0460','S001',1,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0470','Z000',2.6,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0476','S011',1,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0478','S510',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0480','S903',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0490','S003',1.2,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0500','S500',1.37,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0530','B000',0.28,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0536','S011',0.65,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0538','S510',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0540','S923',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0560','S003',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0565','S001',0.85,0,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0570','S012',2.15,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0575','S509',0,0,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0580','B000',3.78,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0590','S011',0.27,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0001715683','0600','S510',0,9,0);
    Instead of grouping all sequential steps with 'OUTPR' station, I am gathering all the sequential steps with "S9%" station, then here is the solution changed to this:
    SELECT       item_id
    ,        ord_id
    ,       MIN (step_id)          AS step_id
    ,       station
    ,       SUM (act_hrs)          AS act_hrs
    ,       MIN (q_comp)          AS q_comp
    ,       SUM (q_scrap)          AS q_scrap
    FROM       (          -- Begin in-line view to compute grp_id
               SELECT  test_data.*
               ,           ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                                ORDER BY        step_id
                              )
                  - ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                          ,                CASE
                                            WHEN  station LIKE 'S9%'
                                            THEN  NULL
                                            ELSE  step_id
                                           END
                                ORDER BY      step_id
                              )     AS grp_id
               ,           ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                                ORDER BY        step_id
                              )                    AS r_num1
               ,       ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                          ,                CASE
                                            WHEN  station LIKE 'S9%'
                                            THEN  NULL
                                            ELSE  step_id
                                           END
                                ORDER BY      step_id
                              )                    AS r_num2
               FROM    test_data
           )          -- End in-line view to compute grp_id
    GROUP BY  item_id
    ,            ord_id
    ,       station
    ,       grp_id
    ORDER BY  item_id
    ,            step_id
    ;
    If just run the subquery to calculate grp_id, you can see that it sometimes affects the same number of group in two stages that are not side by side. For example, the two step 285 and 480 are they assigned group 32...

    I don't know if it's because my orders have many more steps that the orders of the sample I provided, or what...

    I tried this version too (by replacing all the names of the stations "S9%" by "OUTPR"):
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0005','S509',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0010','S006',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0020','A501',0.85,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0026','S011',0.58,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0030','OUTPR',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0040','OUTPR',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0050','S003',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0055','S600',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0060','Z108',6.94,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0070','Z108',7,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0080','Z310',4.02,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0085','Z409',2.17,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0090','S500',0.85,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0095','S502',1.63,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0110','S006',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0112','S011',0.15,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0117','OUTPR',0,10,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0118','OUTPR',0,9,1);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0119','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0120','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0140','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0145','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0150','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0160','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0170','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0220','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0230','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0240','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0250','S003',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0260','S006',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0270','S012',0.95,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0280','Z417',0.68,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0285','Z417',0.68,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0290','Z426',1.78,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0300','Z426',2.07,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0305','Z426',1.23,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0310','Z402',3.97,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0315','Z308',8.09,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0410','Z409',4.83,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0430','S500',3.6,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0435','S502',0.43,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0450','S002',0.35,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0460','S001',1,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0470','Z000',2.6,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0476','S011',1,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0478','S510',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0480','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0490','S003',1.2,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0500','S500',1.37,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0530','B000',0.28,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0536','S011',0.65,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0538','S510',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0540','OUTPR',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0560','S003',0,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0565','S001',0.85,0,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0570','S012',2.15,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0575','S509',0,0,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0580','B000',3.78,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0590','S011',0.27,9,0);
    INSERT INTO test_data
    VALUES ('abc-123','0009999999','0600','S510',0,9,0);
    
    
    SELECT       item_id
    ,        ord_id
    ,       MIN (step_id)          AS step_id
    ,       station
    ,       SUM (act_hrs)          AS act_hrs
    ,       MIN (q_comp)          AS q_comp
    ,       SUM (q_scrap)          AS q_scrap
    FROM       (          -- Begin in-line view to compute grp_id
               SELECT  test_data.*
               ,           ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                                ORDER BY        step_id
                              )
                  - ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                          ,                CASE
                                            WHEN  station = 'OUTPR'
                                            THEN  NULL
                                            ELSE  step_id
                                           END
                                ORDER BY      step_id
                              )     AS grp_id
               ,           ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                                ORDER BY        step_id
                              )                    AS r_num1
               ,       ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                          ,                CASE
                                            WHEN  station = 'OUTPR'
                                            THEN  NULL
                                            ELSE  step_id
                                           END
                                ORDER BY      step_id
                              )                    AS r_num2
               FROM    test_data
           )          -- End in-line view to compute grp_id
    GROUP BY  item_id
    ,            ord_id
    ,       station
    ,       grp_id
    ORDER BY  item_id
    ,            step_id
    ;
    and it shows the same problem.

    Help?

    Hello

    I'm glad that you understood the problem.

    Here's a little explanation of the approach of the fixed difference. I can refer to this page later, so I will explain some things you obviously already understand, but I jump you will find helpful.
    Your problem has additional feature that, according to the station, some lines can never combine in large groups. For now, we will greatly simplify the problem. In view of the CREATE TABLE statement, you have posted and these data:

    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0010', 'Z417');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0011', 'S906');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0012', 'S906');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0140', 'S906');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0170', 'Z417');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0175', 'Z417');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0200', 'S906');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0205', 'S906');
    

    Let's say that we want this output:

    `                  FIRST LAST
                       _STEP _STEP
    ITEM_ID ORD_ID     _ID   _ID   STATION  CNT
    ------- ---------- ----- ----- ------- ----
    abc-123 0001715683 0010  0010  Z417       1
    abc-123 0001715683 0011  0140  S906       3
    abc-123 0001715683 0170  0175  Z417       2
    abc-123 0001715683 0200  0205  S906       2
    

    Where each line of output represents a contiguous set of rows with the same item_id, ord_id and station. "Contguous" is determined by step_id: lines with "0200" = step_id = step_id "0205' are contiguous in this example of data because there is no step_ids between '0200' and '0205". "
    The expected results include the step_id highest and lowest in the group, and the total number of original lines of the group.

    GROUP BY (usually) collapses the results of a query within lines. A production line can be 1, 2, 3, or any number of lines in the original. This is obviously a problem of GROUP BY: we sometimes want several lines in the original can be combined in a line of output.

    GROUP BY guess, just by looking at a row, you can tell which group it belongs. Looking at all the 2 lines, you can always know whether or not they belong to the same group. This isn't quite the case in this issue. For example, these lines

    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0140', 'S906');
    INSERT INTO test_data (item_id, ord_id, step_id, station)  VALUES ('abc-123', '0001715683', '0200', 'S906');
    

    These 2 rows belong to the same group or not? We cannot tell. Looking at just 2 lines, what we can say is that they pourraient belonging to the same group, since they have the same item_id, ord_id and station. It is true that members of same groups will always be the same item_id, the ord_id and train station; If one of these columns differ from one line to the other, we can be sure that they belong to different groups, but if they are identical, we cannot be certain that they are in the same group, because item_id, ord_id and station only tell part of the story. A group is not just a bunch or rows that have the same item_id, ord_id and station: a group is defined as a sequence of adjacent to lines that have these columns in common. Before we can make the GROUP BY, we need to use the analytical functions to see two lines are in the same contiguous streak. Once we know that, we can store this data in a new column (which I called grp_id), and then GROUP BY all 4 columns: item_id, ord_id, station and grp_id.

    First of all, let's recognize a basic difference in 3 columns in the table that will be included in the GROUP BY clause: item_id, ord_id and station.
    Item_id and ord_id always identify separate worlds. There is never any point comparing lines with separate item_ids or ord_ids to the other. Different item_ids never interact; different ord_ids have nothing to do with each other. We'll call item_id and ord_id column 'separate world '. Separate planet do not touch each other.
    The station is different. Sometimes, it makes sense to compare lines with different stations. For example, this problem is based on questions such as "these adjacent lines have the same station or not? We will call a "separate country" column of the station. There is certainly a difference between separate countries, but countries affect each other.

    The most intuitive way to identify groups of contiguous lines with the same station is to use a LAG or LEAD to look at adjacent lines. You can certainly do the job, but it happens to be a better way, using ROW_NUMBER.
    Help the ROW_NUMBER, we can take the irregular you are ordering step_id and turn it into a dial of nice, regular, as shown in the column of r_num1 below:

    `                                 R_             R_ GRP
    ITEM_ID ORD_ID     STEP STATION NUM1 S906 Z417 NUM2 _ID
    ------- ---------- ---- ------- ---- ---- ---- ---- ---
    abc-123 0001715683 0010 Z417       1         1    1   0
    abc-123 0001715683 0011 S906       2    1         1   1
    abc-123 0001715683 0012 S906       3    2         2   1
    abc-123 0001715683 0140 S906       4    3         3   1
    abc-123 0001715683 0170 Z417       5         2    2   3
    abc-123 0001715683 0175 Z417       6         3    3   3
    abc-123 0001715683 0200 S906       7    4         4   3
    abc-123 0001715683 0205 S906       8    5         5   3
    

    We could also assign consecutive integers to the lines in each station, as shown in the two columns, I called S906 and Z417.
    Notice how the r_num1 increases by 1 for each line to another.
    When there is a trail of several rows of S906 consectuive (for example, step_ids ' 0011 'by '0140'), the number of s906 increases by 1 each line to another. Therefore, during the duration of a streak, the difference between r_num1 and s906 will be constant. For 3 lines of the first series, this difference is being 1. Another series of S906s contiguous started step_id = '0200 '. the difference between r_num1 and s906 for this whole series is set to 3. This difference is what I called grp_id.
    There is little meaning for real numbers, and, as you have noticed, streaks for different stations can have as by chance the same grp_id. (it does not happen to be examples of that in this game of small sample data.) However, two rows have the same grp_id and station if and only if they belong to the same streak.

    Here is the query that produced the result immediately before:

    SELECT    item_id
    ,        ord_id
    ,        step_id
    ,       station
    ,       r_num1
    ,       CASE WHEN station = 'S906' THEN r_num2 END     AS s906
    ,       CASE WHEN station = 'Z417' THEN r_num2 END     AS Z417
    ,       r_num2
    ,       grp_id
    FROM       (          -- Begin in-line view to compute grp_id
               SELECT  test_data.*
               ,           ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                                ORDER BY        step_id
                              )
                  - ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id, station
                                ORDER BY      step_id
                              )     AS grp_id
               ,           ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                                ORDER BY        step_id
                              )                    AS r_num1
               ,       ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id, station
                                ORDER BY      step_id
                              )                    AS r_num2
               FROM    test_data
           )          -- End in-line view to compute grp_id
    ORDER BY  item_id
    ,            ord_id
    ,       step_id
    ;
    

    Here are a few things to note:
    All analytical ORDER BY clauses are the same. In most of the problems, there will be only an ording regime that matters.
    Analytical PARTITION BY clauses include the columns of 'distinct from the planet', item_id and ord_id.
    The analytical PARTITION BY clauses also among the column 'split the country', station.

    To get the results we want in the end, we add a GROUP BY clause from the main query. Yet once, this includes the columns of the 'separate world', column 'split the country', and the column 'fixed the difference', grp_id.
    Eliminating columns that have been includied just to make the output easier to understand, we get:

    SELECT    item_id
    ,        ord_id
    ,        MIN (step_id)          AS first_step_id
    ,       MAX (step_id)          AS last_step_id
    ,       station
    ,       COUNT (*)          AS cnt
    FROM       (          -- Begin in-line view to compute grp_id
               SELECT  test_data.*
               ,           ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id
                                                ORDER BY        step_id
                              )
                  - ROW_NUMBER () OVER ( PARTITION BY  item_id, ord_id, station
                                ORDER BY      step_id
                              )     AS grp_id
               FROM    test_data
           )          -- End in-line view to compute grp_id
    GROUP BY  item_id
    ,            ord_id
    ,       station
    ,       grp_id
    ORDER BY  item_id
    ,            ord_id
    ,       first_step_id
    ;
    

    This prioduces the output displayed much earlier in this message.

    This example shows the fixed difference indicated. Specific problem you is complicated a little what you should use an expression BOX based on station rather than the station iteself.

  • Get members of the group with the Java API in LiveCycle

    I work with the Java API in LiveCycle (LiveCycle 8.2) and I am looking


    an example of how read the members of a group.

    I am able to get a group (Type GroupImpl) using the PrincipalSearchFilter.


    But how do I get the members of this group. I checked the
    DirectoryManagerServiceClient and classes of different search filters
    without success?

    As a solution, I can retrieve all users of my application area and
    check their membership to see if they are in the group concerned. But this
    is a dirty hack, I really don't want to use.

    Any ideas?

    Hello!

    I know you mentioned that you were watching in this without success, but it should work really.

    First make a GroupMemebershipSearchFilter, defining the groupoid be the ID of the group itself.  Using the DirectoryManagerServiceClient you can then call findGroupMembers using the searchFilter that will return a list of, as follows:

    DirectoryManager DirectoryManagerServiceClient = new DirectoryManagerServiceClient (serviceConnection.getFactory ());

    List userList = new ArrayList ();

    Try
    {
    userList = directoryManager.findGroupMembers (searchFilter);
    }
    catch (System.Exception e)
    {

    Animals mistake
    }

    return userList;

  • Add the group with the ES2 Java API

    Hi all

    There is an example that shows how to create a new group using the ES2 API here:

    http://help.Adobe.com/en_US/LiveCycle/9.0/programLC/help/index.htm

    However, if you run the code to create a group, the code works fine and a new OID is returned. But if you go in the /AdminUI ES2 and the list of all groups, you don't see the 'AdobeGroup' group here! Also if you are looking to the /AdminUI for the group 'AdobeGroup' it is not found.

    If the code does not work or the Admin UI does not show the newly created group.

    What's not here?

    Thank you.
    Paul.

    After the code snippet should work for you

    public String createSampleGroup() throws UMException{        String groupOid = checkGroupExist(groupName);
    
            if(groupOid != null){            return groupOid;        }        String groupCanonicalName = groupName;
    
            GroupImpl group = new GroupImpl();        group.setCanonicalName(groupCanonicalName);        group.setDomainName(domainName);        group.setGroupType(Group.GROUPTYPE_PRINCIPALS);        group.setLocal(true);        group.setPrincipalType(Principal.PRINCIPALTYPE_GROUP);
    
            groupOid = directoryManager.createLocalGroup(group);        log("Sample group created with name %s",groupName);
    
            return groupOid;    }
    
        /**     * Search the groupwith the given name in the sample domain      */    private String checkGroupExist(String groupName) throws UMException{        PrincipalSearchFilter psf = new PrincipalSearchFilter();        psf.setCommonName(groupName);        psf.setSpecificDomainName(domainName);
    
            //By default the filter causes like search unless you are using the absolute version        //Setting this ensures that search is exact        psf.setMatchExactCriteria(true);
    
            //By default search returns obsolete users also. Set this to ensure that        //only active users are returned        psf.setRetrieveOnlyActive(); 
    
            //PrincipalReference are lightweight user objects and searching for them would be more performant        //compared to the User search. If you do not require any other user attribute then prefer this        //mode of search        List result = directoryManager.findPrincipalReferences(psf);        if(result.isEmpty()){            log("Sample group with name [%s] does not exist",groupName);            return null;        }else{            String oid = result.get(0).getOid();             log("Sample group with name [%s] already exist",groupName);            return oid;        }    }
    

Maybe you are looking for