Best practices for managing strategies of path

Hello

I get conflicting advice on best practices for managed paths.

We are on version 4.0 of ESXi connection to a HP EVA8000. Best practices guide HP recommends setting the strategy of railways handle on Round Robin.

This seems to give two active paths to the optimized controller. See: http://h20195.www2.hp.com/v2/GetPDF.aspx/4AA1-2185ENW.pdf

We used certain consultants and they say that the best practices of Vmware for this solution is to use the MRU policy which translates a single path to the optimized controller.

So, any idea what good practice is best practice? Does make a difference?

TIA

Rob.

Always go with the recommendation of the storage provider.  VMware recommendation is based on the characteristics of the generic array (controller, capable ALUA failover methods, etc.).  The storage provider's recommendation is based on their performance and compatibility testing.  You may want to review their recommendations carefully, however, to ensure that each point is what you want.

With the 8000, I ran with Round-Robin.  This is the option of creating more robust paths available to you from a failover and performance point of view and can provide performance more even through the ports on the storage controller.

While I did of the specific tests/validation, the last time that I looked at the docs, the configuration of HP recommends that you configure each IO to the ports in the switch configuration.  This adds the charge to the ESX host, the switch to other ports, but HP claims that their tests showed that it is the optimal configuration.  It was the only parameter I wondered in their recommendation.

If you haven't done so already, be sure to download the HP doc on configuring ESX and EVA bays.  There are several parameters that you must configure the policy path, as well as a few scripts to help make the changes.

Virtualization of happy!

JP

Please consider awarding points to useful or appropriate responses.

Tags: VMware

Similar Questions

  • Best practices for managing exceptions and success messages.

    Hey people,

    These days I've been shooting packages to clean my application. And question came to my mind, ' should I treat my exceptions the right way?


    So I want to ask you met guys, what is the best practice for this? (I want to learn it until it's too late )

    Currently I have a function that returns "OK" if all goes well.


    return('OK');  

    Can I manage my exceptions like this

      EXCEPTION
        WHEN OTHERS THEN
          ROLLBACK;
          RETURN (SQLERRM);
    
    

    At THE SUMMIT, I have a process that calls the function and then checks if the function returned "OK".

         IF cRet not LIKE 'OK%' THEN
          RAISE_APPLICATION_ERROR(-20000,cRet);
         END IF;
    
    

    And in 'process Error Message' I put "#SQLERRM_TEXT #" so that I can see what error occurred.

    Question aside, how do you manage your messages of success?

    Currently in 'process success Message' put something along the lines "Action completed successfully". What to do about all the processes.

    Do you want to do differently?

    I really want to hear what you have to say since I'm tired of feeling like this is a terrible way to manage these things.

    Hi Para,

    Para wrote:

    I don't know of situations where my service throw exceptions like no_data_found.

    and I need to know that the process is not so I can get to see my # #SQLERRM_TEXT.

    I got this by increasing the error in the application (which I think is a bad way to go) if the return is anything other than 'OK '. I get my application error in the process of the apex, and not in my service.
    And I want to show the inline error in the notification. (Which I am currently with my approach).

    You can use APEX_ERROR. ADD_ERROR in your PL/SQL process to throw exceptions in Oracle APEX.

    Reference: Re: Re: error in the processing of the page management

    Kind regards

    Kiran

  • Request for advice: generally speaking, what is the best practice for managing a paid and a free application?

    Hi all

    I recently finished my first app of cascades, and now I want to inspire of having a more feature rich application that I can then sell for a reasonable price. However, my question is how to manage the code base for both applications. Any have any "best practices", I would like to know your opinion.

    You use a revision control system? This should be a prerequisite...

    How the different versions of the application will be?

    Generally if you have two versions that differ only in terms of having a handful of features disabled in the free version, you must use exactly the same code base. You could even just it for packaging (build command) was the only difference, for example by adding an environment variable in one of them that would be checked at startup to turn paid options.

  • Advice on best practices for managing analysis listener logs and logs of the list

    Hi friends,

    I just started a job as a dba CARS for some big 24 * 7 systems administrator, I've never worked with RAC and clusterware.

    2 space issues

    (1) very large listener_scan2.log in the folder /u01/11.2.0/grid/log/diag/tnslsnr/ < servername >/listener_scan2/trace
    (2) bunch of files log_nnn.xml /u01/11.2.0/grid/log/diag/tnslsnr/ < servername >/listener_scan2/alert folder (4 GB out of stock)

    Welcome advice on how best to manage these short-term (i.e. remove manually) and recommended practice and the safest way (adri maybe not sure how it works with listeners to scan)

    Welcome advice and commands that could be used in these cleaning and a robust mechanism in place for the management of the log file in RAC and CLusterware systems safely.


    Finally should I check the log files in /u01/11.2.0/grid/log/diag/tnslsnr/ < servername >/listener_scan2/alert regulalrly?
    My experience with newspapers in the listener, is that they are only considered when there are major problems of connectivity and overall are ignored.

    Thanks for your help,
    Cheers, Rob

    Have you had problems in need for the purpose of investigation? If this is not the case, just remove them. Are the logs required for some sort of verification process? If so, gzip to a place where you can use your operating system from tape backup strategies to preserve their n-days. Once you delete a file active, he must recreate the file and continue without interruption.

  • What is the best practice for managing for results in doubles because of several sessions

    I work with PowerCLI 5.0.  I quite often find that I have a few sessions open at the same time connected to a server that hosts a particular virtual machine and vCenter.  I ran into a problem when I tried the cmdlet Add-USBController of VMware vSphere PowerCLI Reference book.

    The following lines are (roughly) in the extract cmdlet of the book:

    $vmView get - vm 'test ' | Get-View

    $vmView.ReconfigVM_Task ($spec)

    I got the following error "method call failed because [System.Object []] contains a method named"ReconfigVM_Task"" etc.

    It turns out that I was getting an array of vmViews for 'test' (instead of one), because I had a connection to the host and vCenter.  Once I deleted the host sessions, the script worked well.

    So now my question - what is the best way to mitigate this in my scripts? At least, I guess I need to make sure that I received only an object in the view of the GET. should I remove all duplicates? kill all other sessions when I run the script?

    Thank you

    Maureen

    The best method is to use the server parameter on cmdlets where it is available, in this way, you can limit the objects returned to this server for vSphere.

    If you have multiple sessions on the same server from vSphere open, you could direct the result of a Select-Object cmdlet.

    Like this

    Get-VM-name MyVM-Server $vcenter | Select - 1 first

  • Best practices for deferred loading collection once but ensuring there?

    I'm confused on best practices for managing the "setup" of the form, where I need a remote call to occur once only once for the form, but I also need to make use of this collection for a combobox that will change when different lines in the DataGrid or clicked. Easier if I just explain...

    1. You click on a row in a datagrid control to modify an object (for this example we will say it is an "employee")
    2. The form you go must have a collection of objects 'Department' charged by a remote call. This collection of departments should only occur once, since it is not common for them to change. The departments collection is used to fill a combobox in form.
    3. You need to understand what Department of the comboBox control is the property selectedIndex by iterating over the departments and find one that fits the employee.department.id

    Individually, I know how I can do all of the above, but because of the asynchronous nature of Flex, I'm having hard time setting up things. Here are a few questions...

    My first thought was just put the loading of the departments in an init() method on the employeeForm who would load as an event on the form creationComplete(). On the component page grid when the event handler by clicking a line of fire, I then calls the setup() method on my employeeForm which stands at which selectedIndex to set to the combobox control looking at the departments.

    The problem is the resultHandler for the load of the departments could not returned (so departments could not be there when "setUp" is called), but I can't put my business logic to determine the correct combobox in the departmentResultHandler because it would mean that I would always have whenever I don't want the fire of the call to the remote server object.

    I have to miss a single best practice? Suggestions welcome.

    Hi there rickcr

    It is pretty rough and you need to make a few storage upward, but take a look below.


    http://www.Adobe.com/2006/mxml"layout ="absolute">
       
            Import mx.controls.Alert;
    Import mx.collections.ArrayCollection;
               
    private var comboData:ArrayCollection;
               
    private void Setup (): void {}
    If {(comboData)
    Alert.Show ("data are present")
    populateForm()
    } else {}
    Alert.Show ("data not '")
    getData();
    }
    }
               
    private void getData (): void {}
    comboData = new ArrayCollection();
    The result of this call, the installer again
    }
               
    private function populateForm (): void {}
    fill out your form
    }
    ]]>
       

       
       
           
           

           
           

       

    I think this example type of watch you want.  When you first click on tab 2 there is no data.  When you click on tab 2 once again, there is. The data for your combo will be stored in comboData.  When the component gets created first the comboData is not instansiated, just romance.  This allows you to say

    If (comboData)

    This means that if the variable contains your data, you can fill out the form.  Initially it is not so now the else condition, you can call your data and return the result of your data then you can say

    comboData = new ArrayCollection(), put the data in it and remember the installation again.  This time comboData is populayed and is so it will run the method populate the form and you can decide which selected item to affect.

    If it is on a large scale you want to look into the creation of a suitable handler class to handle this, but this simple demo shows you can test to see if the data is different.

    Hope it helps and gives you some ideas.

    Andrew

  • Best practices for the integration of the Master Data Management (MDM)

    I work on the integration of MDM with Eloqua and are looking for the best approach to sync data lead/Contact changes of Eloqua in our internal MDM Hub (output only). Ideally, we would like that integration practically in real time but my findings to date suggest that there is no option. Any integration will result in a kind of calendar.

    Here are the options that we had:

    1. "Exotic" CRM integration: using internal events to capture and queue in the queue changes internal (QIP) and allows access to the queue from outside Eloqua SOAP/REST API
    2. Data export: set up a Data Export that is "expected" to run on request and exteernally annex survey via the API SOAP/REST/in bulk
    3. API in bulk: changes in voting that has happened since the previous survey through the API in bulk from Eloqua outside (not sure how this is different from the previous option)

    Two other options which may not work at all and who are potentially antimodel:

    • Cloud connector: create a campaign questioning changes to schedule and configure a connector of cloud (if possible at all) to notify MDM endpoint to query contact/lead "record" of Eloqua.
    • "Native" integration CRM (crazy): fake of a native CRM endpoint (for example, Salesforce) and use internal events and external calls to Eloqua push data into our MDM

    Issues related to the:

    1. What is the best practice for this integration?
    2. Give us an option that would give us the close integration in real-time (technically asynchronous but always / event-based reminder)? (something like the outgoing in Salesforce e-mail)
    3. What limits should consider these options? (for example API daily call, size response SOAP/REST)

    If you can, I would try to talk to Informatica...

    To imitate the integrations of native type, you use the QIP and control what activities it validated by internal events as you would with a native integration.

    You will also use the cloud api connector to allow you to set up an integration CRM (or MDM) program.

    You have fields of identification is added objects contact and account in Eloqua for their respective IDs in the MDM system and keep track of the last update of MDM with a date field.

    A task scheduled outside of Eloqua would go to a certain interval and extract the QAP changes send to MDM and pull the contacts waiting to be sent in place of the cloud connector.

    It isn't really much of anything as outgoing unfortunately use Messaging.  You can send form data shall immediately submit data to Server (it would be a bit like from collections of rule of integration running of the steps in processing of forms).

    See you soon,.

    Ben

  • best practices for backup of bi

    Hello

    Who can suggest me best practices for backup and restore of dashboard reports integer bi, permissions and etc.?

    Ed,

    Hello

    If you want to move entire dashboards, reports, permissions
    Zip the *Web / catalog folder and move new environment. In new relax unwind this catalogue and Instanceconfig.xml mention the path for this new catalogue.

    If you want to move a few reports or dashboads, do so by Catalog Manager.

    Thank you.

  • Just improved m tips on best practices for sharing files on a Server 2008 std.

    The field contains about 15 machines with two domain controllers, one's data is the app files / print etc...  I just upgraded from 2003 to 2008 and want to get advice on best practices for the establishment of a group of file sharing. Basically I want each user to have their their own records, but also a staff; folder. Since I am usually accustomed to using windows Explorer, I would like to know if these actions can be done in the best conditions. Also I noticed on 2008 there is a feature of contacts. How can it be used? I would like to message or send an email to users their file locations. Also, I want to implement an admin at a lower level to handle the actions without making them far in on the server, not sure.

    I read a certain bbut I don't like test direct more because it can cause problems. So basically a way short and neat to manage shares using the MMC, as well as the way that I approach their mail from the server of their actions. Maybe what kind of access cintrol or permissions are suitable also for documents. Also how can I have them use office templates without changing the format of the model.

    THX

    g

    Hello 996vtwin,

    Thank you for visiting the Microsoft Answers site. The question you have posted is related to Windows Server and would be better suited to the Windows Server TechNet community. Please visit the link below to find a community that will support what ask you:

    http://social.technet.Microsoft.com/forums/en-us/category/WindowsServer

    Hope this helps J

    Adam
    Microsoft Answers Support Engineer
    Visit our Microsoft answers feedback Forum and let us know what you think

  • [ADF, JDev12.1.3] Best practices for maintaining a form validation

    Hallo,

    in my application, I need to create a registration form which contains fields that must be validated (for example they should follow a format like e-mail, phone number, tax code,...).

    If the data inserted by the user are ok, a new record in my custom db table Users will be created.

    I would like to know which are the best practices for maintaining the validation, which means the place where the controls must be made and a message to the user who fills out the form when something goes wrong.

    The best vo or EO or managed bean? Or some controls should be put in the OS, others in the VO and other in the managed bean?

    I would be happy if you could give me some examples.

    Thank you

    Federico

    Assuming you want the validation on the value of the field to any screen data can be entered in (and possibly web services that rely on the same BC ADF) then put the validation on the definition of the attribute in the EO.

    If you want to add a little more friendliness and eliminate some of the network traffic to the server, you can also implement the validation client in your page - for example by using the regular expression validator.

    https://blogs.Oracle.com/Shay/entry/regular_expression_validation

  • Best practices for color use in Adobe CC?

    Hi all

    Is there an article that describes the best practices for use of color in Adobe CC?

    I produce a mixture of viewing online (PDF, for the most part) and real world print projects - often with the obligation for both. I recently updated my PANTONE + bridge books for the first time in ages and I am suddenly confused by the use of Lab colors in the Adobe Suite (Illustrator and InDesign).

    Everything I found online, looks like Lab color mode preferred to use because it is device independent. And perceptual (on screen), it looks much closer to the color, it is trying to represent. But when I mark a Spot color Illustrator rectangle using laboratory coordinates, to the sides of a rectangle using PANTONE + bridge CP and then export it to PDF, the version of CP to mix CMYK color corresponds exactly to my Pantone book - while the version of laboratory (after converted to CMYK using the ink Manager) is far away.

    I have this fantasy to manage only a single Illustrator or InDesign file for both worlds (PDF) printed and online. Is not possible in practice?

    Any info describing the basic definitions of the color modes - or even a book tracing more than use them in the real world - would be much appreciated!

    Thank you

    Bob

    Here are a few best practices you can already do.

    1 make sure that your color settings are synchronized on all applications.

    2. use a CMYK profile appropriate for your print output. Lab spot colors convert to CMYK values based on the CMYK icc profile.

    3. include icc profiles when save or export pdf files

    In theory, your imagination is possible today. It requires color management and the use of icc profiles. You can place RGB images in InDesign and use Pantone colors in your objects. The problem lies in the printers. If a printer uses a RIP with built in Pantone library, the colors will match when printing. Unfortunately, this kind of CUT is more expensive and not enough printers use them. Most of them is always manually approximate CMYK values composition given Pantone colors.

  • Best practices for building an infrastructure of APEX for 12 c

    Hi all

    Have we not the docs on best practices for building an infrastructure of APEX?

    Which means, for the production, it is acceptable to use Embedded PL as the listener, or we stick with the Listenr tested on Weblogic?

    Thank you

    Hi JCGO,.

    JCGO wrote:

    Hi all

    Have we not the docs on best practices for building an infrastructure of APEX?

    Which means, for the production, it is acceptable to use Embedded PL as the listener, or we stick with the Listenr tested on Weblogic?

    Thank you

    I agree with Scott's response '' it depends. '' It starts with the appropriate choice of a web listening port.

    You should discourage use EPG facility based in Production environments in accordance with the recommendation of the Oracle.

    Reference: See security considerations when you use the Embedded PL/SQL Gateway section.

    ADR (APEX Listener) + Oracle Weblogic Server sounds good, if you already have tried and have appropriate expertise to manage it.

    Also, you might consider what other facilities ADR based ADR + Apache Tomcat with Apache HTTP Server reverse proxy as described here:

    Dimitri Gielis Blog (Oracle Application Express - APEX): Prepare the architecture for the APEX 5.0 upgrade

    But it depends on Apache skills, you have within your organization.

    I hope this helps!

    Kind regards

    Kiran

  • vSpere 5 Networking of best practices for the use of 4 to 1 GB NIC?

    Hello

    I'm looking for a networking of best practices for the use of 4-1 GB NIC with vSphere 5. I know there are a lot of good practice using 10 GB, but our current config does support only 1 GB. I need to include the management, vMotion, Virtual Machine (VM) and iSCSi. If there are others you would recommend, please let me know.

    I found a diagram that resembles what I need, but it's for 10 GB. I think it works...

    vSphere 5 - 10GbE SegmentedNetworks Ent Design v0_4.jpg(I had this pattern HERE - rights go to Paul Kelly)

    My next question is how much of a traffic load is each object take through the network, percentage wise?

    For example, 'Management' is very small and the only time where it is in use is during the installation of the agent. Then it uses 70%.

    I need the percentage of bandwidth, if possible.

    If anyone out there can help me, that would be so awesome.

    Thank you!

    -Erich

    Without knowing your environment, it would be impossible to give you an idea of the uses of bandwidth.

    That said if you had about 10-15 virtual machines per host with this configuration, you should be fine.

    Sent from my iPhone

  • Best practices for vSphere 5 Networking

    Hi all

    Given the following environment:

    (1) 4 physical servers, each server has 16 (Gigabit) network interface cards will install vSphere 5 std.

    (2) 2 switches with SAN storage battery function

    (3) 2 Equallogic PS4000 SAN (two controllers)

    (4) 2 switches for the traffic of the virtual machine

    As for networking, I intend create some vSwitches on each physical server as follows

    1 vSwitch0 - used for iSCSI storage

    6 network adapters are associated with IP-hash, multitracks with iSCSI Storage consolidation policy. and storage load balancing is round Rodin (vmware)

    (vmware suggests to use 2 cards for 1 traget of IP storage, I'm not sure)

    2 vSwitch1 - used for the virtual machine

    6 Teamed network adapters for the traffic of the virtual machine, with the hash intellectual property policy

    3 vSwitch2 - management

    2 network cards are associated

    4 vSwitch3 - vMotion

    2 network cards are associated

    You would like to give me some suggestions?

    Alex, the standard set by the storage and VMware is used by dell for their servers and tested sound on their equipment and the publication of the document, it is recommended...

    This sound is the best practice for dell server with this model mentioned in the document to use.

    Hope that clarifies...

  • AF best practices for video

    I'm using an EOS 7 d for 1080 p HD video shooting. New to SLR camera and digital video. I have watched several tutorials but do not yet have a good understanding of what the best practices for the development.

    If I put the camera in AI Servo it will auto focus if the subject is moving away or toward me with depress me back AF-ON or the shutter button? Or should I press one of these buttons to make the camera continues to focus?

    You know tutorials or video lessons that cross using update with the video?

    Thanks in advance!

    I don't know any tutorials, but more serious videographers say it is best to use only the manual focus.

    Trial and error is a good teacher, when it comes to video DSLR. You'll quickly learn that video Auto Focus is really bad with the 7 d and most of the people do not use it all together.

    The method that works for me is to use AF before the filming of the video and then adjust manually if you want to you to shoot the video. With the help of f/8 or f/11 to keep everything in short without having to keep the setting.

    More recent Canon DSLR's like the T4i, T5i, 70 d and SL1 have a "hybrid AF" that does a much better job of AF during video, but it is not yet as good as a camcorder

Maybe you are looking for