Best practices for integrating flex application in html

Hello guys

I don't know what the best practice is to embed the flex application. We couldn't use the html code in flex builder. You guys edit html of other tools files to integrate the swf file when the application is completed or when there are other ways to do it. Thanks for the reply.

Hi Jerry,

Why you couldn't use the html generated code in flex builder. In the html wrapper file generated in FlexBuilder you have SWF embed code directly, as well as javascript Flash Player detection code... If your users do not have the FlashPlayer installed or if they not have the latest version of Flash Player installed, then the application invite users to install the latest Flash player plugin...

What is the problem that you have obtained the help of the html wrapper file generated in FlexBuilder.

Thank you

Jean Claude Chari

Tags: Flex

Similar Questions

  • HTML and CSS best practices for Eloqua?

    Hello Topliners community,

    My name is Ben and I am a Web Designer. Currently looking for guidance on best practices in HTML and CSS when you work with Eloqua. I am interested in best practices for email and landing pages.

    Thank you

    Ben

    For landing pages, you can use a bit of HTML, CSS, and Javascript like on any other page. For example, we use the Bootstrap on a couple of our Eloqua landing pages.

    Support for HTML and CSS is much more limited. It is one of the best resources I've seen for CSS support:

    http://www.campaignmonitor.com/CSS/

  • Best practices for the integration of the Master Data Management (MDM)

    I work on the integration of MDM with Eloqua and are looking for the best approach to sync data lead/Contact changes of Eloqua in our internal MDM Hub (output only). Ideally, we would like that integration practically in real time but my findings to date suggest that there is no option. Any integration will result in a kind of calendar.

    Here are the options that we had:

    1. "Exotic" CRM integration: using internal events to capture and queue in the queue changes internal (QIP) and allows access to the queue from outside Eloqua SOAP/REST API
    2. Data export: set up a Data Export that is "expected" to run on request and exteernally annex survey via the API SOAP/REST/in bulk
    3. API in bulk: changes in voting that has happened since the previous survey through the API in bulk from Eloqua outside (not sure how this is different from the previous option)

    Two other options which may not work at all and who are potentially antimodel:

    • Cloud connector: create a campaign questioning changes to schedule and configure a connector of cloud (if possible at all) to notify MDM endpoint to query contact/lead "record" of Eloqua.
    • "Native" integration CRM (crazy): fake of a native CRM endpoint (for example, Salesforce) and use internal events and external calls to Eloqua push data into our MDM

    Issues related to the:

    1. What is the best practice for this integration?
    2. Give us an option that would give us the close integration in real-time (technically asynchronous but always / event-based reminder)? (something like the outgoing in Salesforce e-mail)
    3. What limits should consider these options? (for example API daily call, size response SOAP/REST)

    If you can, I would try to talk to Informatica...

    To imitate the integrations of native type, you use the QIP and control what activities it validated by internal events as you would with a native integration.

    You will also use the cloud api connector to allow you to set up an integration CRM (or MDM) program.

    You have fields of identification is added objects contact and account in Eloqua for their respective IDs in the MDM system and keep track of the last update of MDM with a date field.

    A task scheduled outside of Eloqua would go to a certain interval and extract the QAP changes send to MDM and pull the contacts waiting to be sent in place of the cloud connector.

    It isn't really much of anything as outgoing unfortunately use Messaging.  You can send form data shall immediately submit data to Server (it would be a bit like from collections of rule of integration running of the steps in processing of forms).

    See you soon,.

    Ben

  • Best practices for the application of sharpness in your workflow

    Recently I tried to get a better understanding of some of the best practices for sharpening in a workflow. I guess that I didn't know, but there are several places to sharpen. Who are the best? They are additive?

    My typical workflow involves capture an image with a professional DSLR in RAW or JPEG or import into Lightroom and export to a JPEG file for the screen or printing of two lab and local. 

    There are three places in this workflow to add sharpening. In the DSLR manually into Lightroom and when exporting a JPEG file or print directly from Lightroom

    It is my understanding that sharpening is added to RAW images even if you add sharpening in your DSLR. However sharpening will be added to the JPEG from the camera. 

    Back to my question, it is preferable to manually the sharpness in the SLR in Lightroom or wait until you export or output of your printer or final JPEG file. And additive effects? If I add sharpening to the three places I am probably more sharpening?

    You have to treat them differently both formats. Data BULLIES never have any sharpening applied by the camera, only JPEG files. Sharpening is often considered to be a workflow where there are three steps (see here for a seminal paper on this idea).

    I. a step sharpening of capture which compensates for the loss of sharp in detail due to the Bayer matrix and the filter anti-repliement and sometimes the lens or diffraction.

    II. A creative sharpening step where some details in the image are 'highlighted' sharpness (do you eyelashes on a model's face), and

    III. output sharpening, where you fix the loss of sharpness due to the scale/resampling for the properties of the average output (as blur because of an impression how process works or blurry because of the way that a screen LCD sets out its pixels).

    These three are implemented in Lightroom. I. and II. are essential, and basically must always be performed. II. until your creative minds. I. is the sharpening that see you in the Panel to develop. You need zoom at 1:1 and optimize the settings. The default settings are OK but quite conservative. Usually you can increase the value of the mask a little so that you're not sharpen noise and play with the other three sliders. Jeff Schewe gives an overview of a strategy simple to find the optimal settings here. It is for the cab, but the principle remains the same. Most of the photos will benefit from optimization a bit. Don't go overboard, but just OK for smoothness to 1:1.

    Stage II as I have said, is not essential, but it can be done using the local adjustment brush, or you can go to Photoshop for this. Stage III, however, is very essential. This is done in the export, the print panel Panel or the web. You can't really get a glimpse of these things (especially the sharpening printing-oriented) and it will take a little experimentation to see what you like.

    For jpeg, sharpening is done already in the camera. You could add a small extra capture sharpening in some cases, or simply lowering the sharpening in camera and then have more control in the post, but generally, it is best to leave it alone. Stage II and III, however, are still necessary.

  • Best practices for tags

    Hello

    In the bundled applications Tags are used in most applications. For example. in App Customer Tracker, we can add tags to a customer where these tags are stored in a varchr2 column in the Customers Table.
    In my case, I've pre-defined tags real estate (Real Estate) in a table of research called TAGS. For example, Full floor, furnished, equipped, duplexes, attached... What is the best practice for tag properties:
    1 - to store these tags in a varchar column in the table of PROPERTIES using the Shuttle box.
    OR
    2. to store in a third table Eg, PROPERTIES_TAGS (ID PK, FK property-ID, TAG_ID FK), then use the LISTAGG function to show tags in one line in the report properties.
    OR
    You have a better option?

    Kind regards
    Fateh

    Fateh says:
    Hello

    In the bundled applications Tags are used in most applications. For example. in App Customer Tracker, we can add tags to a customer where these tags are stored in a varchr2 column in the Customers Table.
    In my case, I've pre-defined tags real estate (Real Estate) in a table of research called TAGS. For example, Full floor, furnished, equipped, duplexes, attached...

    These seem to be two different use cases. In the bundled applications tags allow end users to join free-form metadata to the data for their own needs (they are sometimes called "folk taxonomies"). Users can use tags for different purposes or different tags for the same purpose. For example, I could add 'Wednesday', 'Thursday' or 'Friday' tags customers because these are the days that they receive their deliveries. For the same purpose, you could mark the same customers '1', '8' and '15' by the numbers of road trucks making deliveries. You can use 'Monday' to indicate that the client is closed on Mondays...

    In your application you assign to the known properties of predefined attributes. It is a model of standard attribute 1:M. their view using the metaphor of the label is not equivalent to the user of free-form tags.

    What is the best practice for tag properties:
    1 - to store these tags in a varchar column in the table of PROPERTIES using the Shuttle box.

    If you do this, how can you:

  • Search for furnished duplex properties effectively?
  • Change in the world "mounted" to "integrated"?
  • Ratio of the number of properties, broken down by full floor, double-sided, equipped...

    OR
    2. to store in a third table Eg, PROPERTIES_TAGS (ID PK, FK property-ID, TAG_ID FK), then use the LISTAGG function to show tags in one line in the report properties.

    As Why use Look up Table, shows the correct way to proceed. It allows the data to be indexed for efficient extraction, and issues such as those above should be dealt with simply by using joins and grouping.

    You might want to examine the possibility of eliminating the PK ID and use an index table organized for this.

    OR
    You have a better option?

    I'd also look carefully your data model. Make sure that you're not flirting with the anti-pattern VAE. Some/all of these values are not simply the attributes on the property?

  • [ADF, JDev12.1.3] Best practices for maintaining a form validation

    Hallo,

    in my application, I need to create a registration form which contains fields that must be validated (for example they should follow a format like e-mail, phone number, tax code,...).

    If the data inserted by the user are ok, a new record in my custom db table Users will be created.

    I would like to know which are the best practices for maintaining the validation, which means the place where the controls must be made and a message to the user who fills out the form when something goes wrong.

    The best vo or EO or managed bean? Or some controls should be put in the OS, others in the VO and other in the managed bean?

    I would be happy if you could give me some examples.

    Thank you

    Federico

    Assuming you want the validation on the value of the field to any screen data can be entered in (and possibly web services that rely on the same BC ADF) then put the validation on the definition of the attribute in the EO.

    If you want to add a little more friendliness and eliminate some of the network traffic to the server, you can also implement the validation client in your page - for example by using the regular expression validator.

    https://blogs.Oracle.com/Shay/entry/regular_expression_validation

  • Best practices for color use in Adobe CC?

    Hi all

    Is there an article that describes the best practices for use of color in Adobe CC?

    I produce a mixture of viewing online (PDF, for the most part) and real world print projects - often with the obligation for both. I recently updated my PANTONE + bridge books for the first time in ages and I am suddenly confused by the use of Lab colors in the Adobe Suite (Illustrator and InDesign).

    Everything I found online, looks like Lab color mode preferred to use because it is device independent. And perceptual (on screen), it looks much closer to the color, it is trying to represent. But when I mark a Spot color Illustrator rectangle using laboratory coordinates, to the sides of a rectangle using PANTONE + bridge CP and then export it to PDF, the version of CP to mix CMYK color corresponds exactly to my Pantone book - while the version of laboratory (after converted to CMYK using the ink Manager) is far away.

    I have this fantasy to manage only a single Illustrator or InDesign file for both worlds (PDF) printed and online. Is not possible in practice?

    Any info describing the basic definitions of the color modes - or even a book tracing more than use them in the real world - would be much appreciated!

    Thank you

    Bob

    Here are a few best practices you can already do.

    1 make sure that your color settings are synchronized on all applications.

    2. use a CMYK profile appropriate for your print output. Lab spot colors convert to CMYK values based on the CMYK icc profile.

    3. include icc profiles when save or export pdf files

    In theory, your imagination is possible today. It requires color management and the use of icc profiles. You can place RGB images in InDesign and use Pantone colors in your objects. The problem lies in the printers. If a printer uses a RIP with built in Pantone library, the colors will match when printing. Unfortunately, this kind of CUT is more expensive and not enough printers use them. Most of them is always manually approximate CMYK values composition given Pantone colors.

  • Best practices for managing exceptions and success messages.

    Hey people,

    These days I've been shooting packages to clean my application. And question came to my mind, ' should I treat my exceptions the right way?


    So I want to ask you met guys, what is the best practice for this? (I want to learn it until it's too late )

    Currently I have a function that returns "OK" if all goes well.


    return('OK');  

    Can I manage my exceptions like this

      EXCEPTION
        WHEN OTHERS THEN
          ROLLBACK;
          RETURN (SQLERRM);
    
    

    At THE SUMMIT, I have a process that calls the function and then checks if the function returned "OK".

         IF cRet not LIKE 'OK%' THEN
          RAISE_APPLICATION_ERROR(-20000,cRet);
         END IF;
    
    

    And in 'process Error Message' I put "#SQLERRM_TEXT #" so that I can see what error occurred.

    Question aside, how do you manage your messages of success?

    Currently in 'process success Message' put something along the lines "Action completed successfully". What to do about all the processes.

    Do you want to do differently?

    I really want to hear what you have to say since I'm tired of feeling like this is a terrible way to manage these things.

    Hi Para,

    Para wrote:

    I don't know of situations where my service throw exceptions like no_data_found.

    and I need to know that the process is not so I can get to see my # #SQLERRM_TEXT.

    I got this by increasing the error in the application (which I think is a bad way to go) if the return is anything other than 'OK '. I get my application error in the process of the apex, and not in my service.
    And I want to show the inline error in the notification. (Which I am currently with my approach).

    You can use APEX_ERROR. ADD_ERROR in your PL/SQL process to throw exceptions in Oracle APEX.

    Reference: Re: Re: error in the processing of the page management

    Kind regards

    Kiran

  • What are the best practices for a new employee to learn inside the instance of their business of Eloqua as efficiently as possible?

    We have companies everything changed at some point in our lives. And we all go through the process in the first weeks, where you feel new and are just trying to figure out how not to get lost on your way in the mornings.

    On top of that, trying to familiarize yourself with your new company Eloqua instance can be a daunting task, especially if it's a large organization.

    What are the best practices for new employees to learn as efficiently and effectively as possible?

    I am in this situation right now. Moved to a much larger organization. It is a huge task trying to understand all the ins and outs not only society, but also of the eloqua instance, especially when she is complex with many points of integration. I find that most of the learning happens when I really go do the work. I spent a ton of time going through the programs, documentation, integrations, etc., but after awhile, it's all just words on a page and not absorbed.

    The biggest thing that I recommend is to learn how and why things are made the way they are currently, ask lots of questions, don't assume not that things work the same as they did with your previous employer.

    Download some base in place level benchmarks to demonstrate additional improvement.

    Make a list of tasks in the long term. As a new pair of eyes, make a list of things you'd like to improve.

  • Best practices for building an infrastructure of APEX for 12 c

    Hi all

    Have we not the docs on best practices for building an infrastructure of APEX?

    Which means, for the production, it is acceptable to use Embedded PL as the listener, or we stick with the Listenr tested on Weblogic?

    Thank you

    Hi JCGO,.

    JCGO wrote:

    Hi all

    Have we not the docs on best practices for building an infrastructure of APEX?

    Which means, for the production, it is acceptable to use Embedded PL as the listener, or we stick with the Listenr tested on Weblogic?

    Thank you

    I agree with Scott's response '' it depends. '' It starts with the appropriate choice of a web listening port.

    You should discourage use EPG facility based in Production environments in accordance with the recommendation of the Oracle.

    Reference: See security considerations when you use the Embedded PL/SQL Gateway section.

    ADR (APEX Listener) + Oracle Weblogic Server sounds good, if you already have tried and have appropriate expertise to manage it.

    Also, you might consider what other facilities ADR based ADR + Apache Tomcat with Apache HTTP Server reverse proxy as described here:

    Dimitri Gielis Blog (Oracle Application Express - APEX): Prepare the architecture for the APEX 5.0 upgrade

    But it depends on Apache skills, you have within your organization.

    I hope this helps!

    Kind regards

    Kiran

  • What is the best way or best practices for access to the session defined in servelt pipeline component?

    Hi Experts,

    What is the best way or best practices for access to the session defined in servelt pipeline component?

    Please, share your ideas.

    Thank you

    ankV

    As performance is concerned, a lot would depend on as to how your design & logic is implemented and you presented search operations to support the specific performance problems. In fact the operation itself can be not expensive than in-house it is somewhat like a search for key/value of the objects in session/application context name. But having said that, Yes, a potential performance problem could be because it's a synchronized operation. So to avoid searches to be performed during each request, caching the results of a search of a component. And in the majority of cases to solve a 'A' component within your component, you must configure a property reference 'A' in the file properties for your component.

  • best practices for the storage of the vm and vhd

    no doubt this question has been answered not once... Sorry

    I would like to know the best practice for the storage of the vm and its virtual hard disk to a SAN.

    Any show advantage does make sense to keep them on separate LUNS?

    Thank you.

    It will really depend on the application of the virtual machine - but for most of the applications no problem by storing everything on the same data store

  • BEST PRACTICES FOR PL/SQL

    Hi all
    I'm looking for advice on best practices to create c# applications that have Oracle as backend.
    I noticed this insertion, update, and delete procedures are not so dynamic because they are static:

    Insertion procedure:
     Create or replace procedure insertion (param1 in varchar2 default '', 
    param2 in number default 0, param3 in float default 0,
     table_to_insert in varchar2, Some_error out varchar2) is
    begin
    if (table_to_insert='table1_spec') then
    insert into table1 (val1,val2,val3) values (param1, param2,param3);
    elsif (ect) then
    .
    .
    .
    end;
    /
    Updated and delete procedures also have the same behavior, I wonder if you use a cursor to do more dynamic values return and the use of native dynamic SQL could not help the developers and I have to have something more dynamic.
    Kind regards

    Published by: user650358 on June 9, 2011 08:37

    user650358 wrote:

    I'm looking for advice on best practices to create c# applications that have Oracle as backend.

    What John said. (even if I would have been more blunt and used the expression "+ approach silly +" several times) {noformat} ;-) {noformat}

    The most flexible approach is to abstract the complexities of SQL, the relational database and the physical implementation of it, of the code c# developer.

    No need to know the database design, the joints, the structures table, SQL and others. Instead the abstract layer takes care of that - where this layer is a suite of packages of PL/SQL procedures and functions.

    As the c# developer would use the Win32 API for creating threads, or the use of Sockets of Win, it now uses the layer of abstraction of PL/SQL in a similar way.

    You want to add an invoice? No need to know the name of the table, the structure of the table and SQL - c# code calls simply the CreateInvoice()PL/SQL procedure.

    This code performs the validation. Business rules and logic. The SQL.

    More, the underlying database model can change, the new rules of business introduced and so on - and the call from c# to CreateInvoice() will remain the same and the c# code not affected by these changes in dorsal.

    Oh yes - it does not using PL/SQL code to extract data from line in the PL/SQL variables and then push these data in turn to the c# client code.

    It means that a procedure or a function like GetCustomertInvoices() returns a ref cursor to the appellant to consume. (the SQL to create the slider is abstract - not the actual slider itself because it is correct the only way to access the data in row SQL)

  • Measurement on the side time server? Best practices for the turn-based game

    Hello

    What would be the best practice for measuring time in a turn based game?

    I was looking at the timeout of the room, but to use that it would mean that, for each round, I put users in a new room?

    Is there a way where I can measure time serverside and keep users in the same room?

    If so, I could use php, otherwise, we would need java that allows to measure a time race.

    See you soon,.

    G

    Hello

    You can definitely use PHP or Java - we provide integration of server

    libraries for either. I don't know exactly what is the use case, so I can't

    comment on what makes the most sense, but if it is not information which must be

    totally secure, grading on the client can be a viable approach also.

    Nigel

  • Best practices for deferred loading collection once but ensuring there?

    I'm confused on best practices for managing the "setup" of the form, where I need a remote call to occur once only once for the form, but I also need to make use of this collection for a combobox that will change when different lines in the DataGrid or clicked. Easier if I just explain...

    1. You click on a row in a datagrid control to modify an object (for this example we will say it is an "employee")
    2. The form you go must have a collection of objects 'Department' charged by a remote call. This collection of departments should only occur once, since it is not common for them to change. The departments collection is used to fill a combobox in form.
    3. You need to understand what Department of the comboBox control is the property selectedIndex by iterating over the departments and find one that fits the employee.department.id

    Individually, I know how I can do all of the above, but because of the asynchronous nature of Flex, I'm having hard time setting up things. Here are a few questions...

    My first thought was just put the loading of the departments in an init() method on the employeeForm who would load as an event on the form creationComplete(). On the component page grid when the event handler by clicking a line of fire, I then calls the setup() method on my employeeForm which stands at which selectedIndex to set to the combobox control looking at the departments.

    The problem is the resultHandler for the load of the departments could not returned (so departments could not be there when "setUp" is called), but I can't put my business logic to determine the correct combobox in the departmentResultHandler because it would mean that I would always have whenever I don't want the fire of the call to the remote server object.

    I have to miss a single best practice? Suggestions welcome.

    Hi there rickcr

    It is pretty rough and you need to make a few storage upward, but take a look below.


    http://www.Adobe.com/2006/mxml"layout ="absolute">
       
            Import mx.controls.Alert;
    Import mx.collections.ArrayCollection;
               
    private var comboData:ArrayCollection;
               
    private void Setup (): void {}
    If {(comboData)
    Alert.Show ("data are present")
    populateForm()
    } else {}
    Alert.Show ("data not '")
    getData();
    }
    }
               
    private void getData (): void {}
    comboData = new ArrayCollection();
    The result of this call, the installer again
    }
               
    private function populateForm (): void {}
    fill out your form
    }
    ]]>
       

       
       
           
           

           
           

       

    I think this example type of watch you want.  When you first click on tab 2 there is no data.  When you click on tab 2 once again, there is. The data for your combo will be stored in comboData.  When the component gets created first the comboData is not instansiated, just romance.  This allows you to say

    If (comboData)

    This means that if the variable contains your data, you can fill out the form.  Initially it is not so now the else condition, you can call your data and return the result of your data then you can say

    comboData = new ArrayCollection(), put the data in it and remember the installation again.  This time comboData is populayed and is so it will run the method populate the form and you can decide which selected item to affect.

    If it is on a large scale you want to look into the creation of a suitable handler class to handle this, but this simple demo shows you can test to see if the data is different.

    Hope it helps and gives you some ideas.

    Andrew

Maybe you are looking for

  • DeskJet All in One 2540 series. How can I find the name and address of the fax server?

    I'm working on my Windows 7 64 bit computer laptop and I try to send a fax with Microsoft Fax and Scan and HP Deskjet 2542 all-in - one network. When it comes to the name and the address of the fax server, I can't find it. Can someone tell me please

  • Script on web pages and email error

    I looked at your solutions and they do not seem to work for me so I'll tell my story. I have a model HP Photosmart Premium 309 a print-scan-copy-fax works with Windows 7 and IE9. My cartridge black Hp stopped working and I replaced all the with the n

  • Access denied to administrator records

    It is a new machine, Windows 7 Pro.  After installing TN3270 Plus SDI Software, which created folders in C:\Program Files (x 86), I copied my config from my old XP machine files in the program folder.  This succeeded.  One of the config files should

  • Catalyst Control Centre: host application has stopped working - Windows 7 Pro

    My father has a Toshiba laptop with Windows 7 Professional and a 2nd monitor.  He has downloaded an update from Microsoft and now gets the error of the Catalyst Control Center and his mouse no longer works.  This isn't the first time.  What causes th

  • AutoConfig is destructive?

    Hi allI am responsible for carrying out autoconfig pending PROD due to an error of access on the other modules.I ran this script several times on TEST, CRP, UAT environment but I'm afraid to run it on PROD.According to your experience, here was a cha