Storage vMotion of 4.1 to 5.5?

I am in the process of upgrading our environment vSphere. I need critical servers vmotion ESX 4.1 at the glomerulus ESXi 5.5 cluster. The two groups are now managed by a 5.5 vcenter server.

See attachment. The PROD4 cluster hosts are all ESX 4.1 and the Production cluster hosts are the ESXi 5.5

Can I install 5.5 on another host and add it to the cluster PROD4 so that I can vmotion VMs critical for her, then vmotion them to the 5.5 cluster?

Bottom line, it is how to make virtual machines on hosts 4.1 5.5 guests without interruption?

I can think in this case is to add one of the hosts of the old cluster/storage, vMotion the VMs on that host, ESXi 5.5 and then use the Web Client vSphere to migrate hosts and data store to the new cluster. With vSphere 5.5 and the Web Client, you can do this even without shared storage.

André

Tags: VMware

Similar Questions

  • Storage vMotion with Syncrep on Equallogic San

    Hello

    We recently bought 2 San Equallogic (PS6510ES) and turned on synchronous replication between them.

    We use ESXi 5.5 and followed the best practices guide to disable TCP ack and 60 logintimeout value.

    Everything seems to work fine until we tried Storage vMotion/migration and cloning VMS on the San.

    When we did the migration of 2 volumes on the EQL SAN storage (the target volume has active syncrep), the target volume will be out-of-sync and then the storage vMotion will hang sometimes.

    The worst is when such a problem has occurred, all other volumes on the EQL constantly to respond also. All volumes will not respond until we rebooted the corresponding ESXi host (NOTE: simple shut down/power outside the host won't help.) The San EQL will only return to normal after a reboot of the ESXi host).

    If we pause / off the syncrep before Storage vMotion, the problem will not happen.

    Is this normal or not?

    In addition, whenever the problem occurred, the latency of writing the corresponding volume will be dry and we will receive alerts similar to the following:

    connection iSCSI target ' xx.xx.xx.xx:3260, iqn.2001 - 05.com.equallogic:xxxxx - test-volume ' initiator 'xx.xx.xx.xx:xxxxx, eqlinitiatorsyncrep' failed for the following reason:

    Initiator, disconnected from the target when connecting.

    Thanks in advance.

    Hi Joerg,

    What is a #SR? I opened a record of support to 24 June and still waiting for Dell to contact me.

    Ryan

  • question of vMotion and storage vMotion

    I have everything set, UCS 5108, ESXi host, VMs, I don't have a SAN connection yet... My question is if I can enable vMotion? Let's say I have a cluster in vCenter, within this group, I have attached two ESXi host, win2008 VM in an ESXi and running, I can migrate (using vMotion) as win2008 VM to an another ESXi without connection SAN?

    How storage vMotion to an ESXi host hardrive to an another ESXi host hardrive? Is this possible?

    Note: I have not connection SAN in this configuration

    One of the requirements for vMotion is shared storage that would allow visibility between source and destination ESXi hosts. (SAN, iSCSI or NFS).

    http://www.VMware.com/PDF/vSphere4/R41/vsp_41_dc_admin_guide.PDF

    pages 210 to 215

    You can use OpenFiler as an iSCSI for lab tests SAN

    http://www.Openfiler.com/community/download/

    Thank you

    Zouheir

  • VM VCS-E drop off air after storage vMotion

    We had a VCS-E, which dealt with a few calls.  VMware has performed a task of vMotion of storage on the server and the VCS-E completely dropped out of the air.

    Cisco has said that vMotion is taken in charge even if you may lose packets (although he said nothing about storage vMotion).  However, this time fallen completely out of the air box until I restarted the VCS-E interface vSphere, after what he came back fine.

    Versions of the software:

    ESXi 6, VCS X8.6

    Any ideas?

    8.6 do not officially support deployed on ESXi 6.0.

    Official support is not added for ESXi 6.0 up to 8.7 - looks too that there is a known issue with ESXi 6.0 which is exceeded in the Notes on the network connectivity that can align with this situation.

    8.7 release notes for Page 4-

    http://www.Cisco.com/c/dam/en/us/TD/docs/Telepresence/infrastructure/VCs...

    VM KB-

    https://KB.VMware.com/selfservice/microsites/search.do?language=en_US&cm...

    Hans

  • Question of Storage vMotion

    Hi all

    Newbie VMware admin here, I hope it's an easy question. I have several boxes of Microsoft SQL I need to move the drive on which the operating system is installed on, they have all RDM attached when I try a storage vmotion, it looks like her wanting to move all the drives instead of one, i have never attempted to move a system with RDM attached to it, but I need to get away from this former storage we have urged. All tips are appreciated thank you!

    Thank you

    You can migrate both at the same time, even if you choose a different destination.

    You're welcome and don't forget brand useful and/or correct answers.

  • Storage vmotion very slow

    Storage vmotions run very slowly. The virtual machine that has a 60 GB HARD drive. I tried different storage and still the same result. I have even upgraded to a network of 10 GB with the same results. I opened an evidence of support and the engineer told me vmotion will not use the link in full speed but it seems quite slow for the numbers I get. I hope I have provided enough information below for assistance. I hope that it is something I am doing wrong...

    Hosts:

    3 x ESXi host. Two are updated to version 6.0 1 and a version 5.5.

    Dell PowerEdge R610

    Xeon E5540 - 8 x 2.53GHz

    48 GB OF RAM

    10 GB for storage and vmotion traffic

    4 x 1 Gb NIC for vm traffic interconnection

    All storage is on the shared storage. The only local thing is the operating system

    VMware EVC mode for the production of Intel's Nehalem.

    Storage

    Supermicro NAS 2 to 4 x 4 RAID 10 storage pools

    Integrated 10 GB NIC

    QNAP TS - 469U - PR to 4 x 4 RAID 5 NAS

    2 x 1 GB NIC aggregated link

    Network

    All 3 hosts and NAS are connected with a Netgear dedicated 10 GB switch (low-end model).

    Using a standard vswitch

    Problem

    1 ESXi host has a computer virtual stored on storage on NAS Supermicro 1 pool. When a storage vmotion is running on a virtual machine that has a 60 GB VMDK, it takes a long time to complete. He started at 09:36, and is now 61% fine at 10:58.

    2. the network of 10 GB is where the vmotion traffic is allowed.

    3. If I use a link to 10 GB or a 1 GB link, I get the same speed.

    4. I have disabled all VM at a time given for the trial and who has not helped either.

    5. I tried a QNAP so storage vmotion, and I get the same speed.

    I use connection CAT6A cables. I have fine tuned for storage, and it seems to work better now. Seems to be linked to "writing - sync" as ESXi.

  • Storage vMotion fails with the error below

    I'm trying to run a virtual machine's storage vMotion. It fails with the error below.

    This method is disabled by 'moref = 17705'

    #1 method - manually remove the entry from the VCDB.

    • Stop the Service of VC.
    • Connect to the SQL Instance.
    • Run the SQL for VCDB.
    • Select * from VPX_VM WHERE name LIKE '% VM_Name % '.
    • Note the ID of VM in column 1.
    • Validate the stale entry.

    Select * from VPX_DISABLED_METHODS WHERE ENTITY_MO_ID_VAL = "vm - ID".

    • Remove obsolete entries.

    delete from VPX_DISABLED_METHODS WHERE ENTITY_MO_ID_VAL = "vm - ID"

    • Start the VC service.
    • Try Storage vMotion.

    Method #2-si this does not work, save the *.vmx file once again which will fix the problem.

    Note: To do this, you will be required to server unavailability.

    • Power off the virtual computer.
    • VM unregistered inventory.
    • Navigate to the path of the data store and VMX re-register.
    • Start the virtual computer.
    • Try Storage vMotion.
  • Sequence of migration for storage vMotion?

    Hello

    I'm moving a couple of virtual machine data to another store.

    I noticed that different virtual machines get different status, such as:

    Copy of the files of virtual machines

    The active state of VM migration

    Resources currently used by other operations... Pending

    I know that it should migrate the active state of VM and then copy the files from the virtual machine.  However, for a machine virtual, it is already 67% current but status is always "the active state of the Virtual Machine migration."

    Is there that an article mentions the details of Storage vMotion?

    See you soon

    Resources currently used by other operations... Pending

    This is because for each host, we have 2 simultaneous storage vMotion migration limits, so if you raise more, system will be queued.

    about the documents, please see suite PDF, this technical white paper concerns the performance and best practices, but it has rather nice, explanations on memory storage Migration and VM migration.

    https://www.VMware.com/files/PDF/Techpaper/VMware-vSphere51-VMotion-perf.PDF

  • Storage vMotion a VM except disk that is in another data store

    Hello

    I'm looking to convert a large number of virtual machines clones related to thick.  I decided the best way to do this is probably a simple a storage vMotion, but all the virtual machines have a secondary drive that is in another data store.  I have to make sure that disk #2 stays at its current location, while the machine virtual (Config and disk #1 files) are transferred to a new data store data store.

    I figured out how to move a virtual machine complete or a single disc but not a virtual machine under a disc.  I am able to make it through the advanced options of puissancela vSphere client so I guess it is techinically possible.

    Pointers or other ideas would be greatly appreciated.

    Kind regards

    Ty-

    You will need to use the API method, see for example Re: using several data stores, cannot migrate the vmx file in a data store

  • Storage vmotion esxi 5.5 to 6.0 does not not for some vm

    In a test cluster with 2 nodes with no storage storage shared (local only), host vmotion 5.5 to 6.0 host does not work.

    before the upgrade, I moved all the vm to a host.

    Then I upgraded to 6.0 host and tried to move all the VMS to the host 6.0 updated (moving host + storage).

    20 VM, 5 have been moved to the new host.

    Everyone always get error below.

    The VM Client are all the vm version 10 (but some of them improved 8 or 9).

    ClientOS Windows 2008, 2012 r2 and linux. Some with vmx3, some with e1000.

    The machines can be moved when turned off.

    After the off-vmotion, storage vmotion from 6 to 5.5 and return of 5.5 to 6 works again.

    < code >

    Failed to copy one or more of the disks in the virtual machine. See the journal of the virtual machine for more details.

    Impossible to implement disks on the destination host computer.

    vMotion-Migration [-1408040447:1427943699552189] konnte nicht den Stream-Keepalive read: Connection closed by remote host, probably due to timeout

    Beim Warten auf Daten Fehlgeschlagen. Fehler 195887167. Connection closed by remote host, probably due to delay.

    < code >

    VM - log:

    < code >

    2015-04 - T 02, 03: 01:34.299Z | VMX | I120: VMXVmdb_SetMigrationHostLogState: transits State hostlog of immigration to migrate "at source" mid 1427943699552189

    2015-04 - T 02, 03: 01:34.307Z | VMX | I120: MigratePlatformInitMigration: file DiskOp set to /vmfs/volumes/526d0605-8317b046-f37e-0025906cd27e/test1/test1-diskOp.tmp

    2015-04 - T 02, 03: 01:34.311Z | VMX | A115: ConfigDB: parameter migration.vmxDisabled = 'TRUE '.

    2015-04 - T 02, 03: 01:34.326Z | VMX | I120: MigrateWaitForData: waiting for data.

    2015-04 - T 02, 03: 01:34.326Z | VMX | I120: MigrateSetState: transition of State 8 to 9.

    2015-04 - T 02, 03: 01:34.467Z | VMX | I120: MigrateRPC_RetrieveMessages: was informed of a new message to the user, but cannot process messages from State 4.  Leaving the message in queue.

    2015-04 - T 02, 03: 01:34.467Z | VMX | I120: MigrateSetState: transition of State 9 to 10.

    2015-04 - T 02, 03: 01:34.468Z | VMX | I120: MigrateShouldPrepareDestination: remote host does not support an explicit step to prepare migration destination.

    2015-04 - T 02, 03: 01:34.469Z | VMX | I120: MigrateBusMemPrealloc: BusMem complete pre-allocating.

    2015-04 - T 02, 03: 01:34.469Z | Worker #0 | I120: SVMotion_RemoteInitRPC: completed.

    2015-04 - T 02, 03: 01:34.471Z | Worker #0 | W110: SVMotion_DiskSetupRPC: related disc expected but not found for the file: /vmfs/volumes/526d0605-8317b046-f37e-0025906cd27e/test1/test1-ctk.vmdk

    2015-04 - T 02, 03: 01:34.471Z | Worker #0 | W110: MigrateRPCHandleRPCWork: callback RPC DiskSetup returned UN-successful State 4. RPC fault.

    2015-04 - T 02, 03: 01:34.493Z | VMX | I120: VMXVmdb_SetMigrationHostLogState: hostlog transits of State to fail to migrate 'from' 1427943699552189 environment

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: MigrateSetStateFinished: type = 2 new State = 12

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: MigrateSetState: transition of State 10 to 12.

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: Migrate: list cached migration error message:

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: [msg.migrate.waitdata.platform] failed waiting for data.  Error bad003f. Connection closed by remote host, probably due to delay.

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: migration of vMotion [vob.vmotion.stream.keepalive.read.fail] [ac130201:1427943699552189] failed to read streams keepalive: Connection closed by remote host, probably due to timeout

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: Migrate: cleaning of the migration status.

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: SVMotion_Cleanup: cleaning of XvMotion State.

    2015-04 - T 02, 03: 01:34.502Z | VMX | I120: Closing all disks in the virtual machine.

    2015-04 - T 02, 03: 01:34.504Z | VMX | I120: Migrate: Final status, who reported the force.

    2015-04 - T 02, 03: 01:34.504Z | VMX | I120: MigrateSetState: transition of the State 12-0.

    2015-04 - T 02, 03: 01:34.505Z | VMX | I120: Migrate: Final status, which reported VMDB.

    2015-04 - T 02, 03: 01:34.505Z | VMX | I120: Migrate Module switch has failed.

    2015-04 - T 02, 03: 01:34.505Z | VMX | I120: VMX_PowerOn: ModuleTable_PowerOn = 0

    2015-04 - T 02, 03: 01:34.505Z | VMX | I120: SVMotion_PowerOff: does not not Storage vMotion. Nothing to do

    2015-04 - T 02, 03: 01:34.507Z | VMX | A115: ConfigDB: parameter replay.filename = «»

    2015-04 - T 02, 03: 01:34.507Z | VMX | I120: Vix: [291569 mainDispatch.c:1188]: VMAutomationPowerOff: power off the power.

    2015-04 - T 02, 03: 01:34.507Z | VMX | W110: /vmfs/volumes/526d0605-8317b046-f37e-0025906cd27e/test1/test1.vmx: cannot remove a link symbolic/var/run/vmware/root_0/1427943694053879_291569/configFile: no such file or directory

    2015-04 - T 02, 03: 01:34.507Z | VMX | I120: WORKER: asyncOps = 4 maxActiveOps = 1 maxPending = 0 maxCompleted = 0

    2015-04 - T 02, 03: 01:34.529Z | VMX | I120: Vix: [291569 mainDispatch.c:4292]: VMAutomation_ReportPowerOpFinished: VarEtat = 1, newAppState = 1873, success = 1 additionalError = 0

    2015-04 - T 02, 03: 01:34.529Z | VMX | I120: Vix: [291569 mainDispatch.c:4292]: VMAutomation_ReportPowerOpFinished: VarEtat = 0, newAppState = 1870, success = 1 additionalError = 0

    2015-04 - T 02, 03: 01:34.529Z | VMX | I120: Past vmx/execState/val of poweredOff

    2015-04 - T 02, 03: 01:34.529Z | VMX | I120: Vix: [291569 mainDispatch.c:4292]: VMAutomation_ReportPowerOpFinished: VarEtat = 0, newAppState = 1870, success = 0 additionalError = 0

    2015-04 - T 02, 03: 01:34.529Z | VMX | I120: Vix: [291569 mainDispatch.c:4331]: error VIX_E_FAIL in VMAutomation_ReportPowerOpFinished(): unknown error

    2015-04 - T 02, 03: 01:34.529Z | VMX | I120: Vix: [291569 mainDispatch.c:4292]: VMAutomation_ReportPowerOpFinished: VarEtat = 0, newAppState = 1870, success = 1 additionalError = 0

    < code >

    Any idea to solve this problem? Maybe it's a bug with vsphere 6.0

    Given that some machines were functioning and vmotion later in both directions from 6.0 to 5.5 works, I think that is no configuration error.

    Hello

    It seems that there could be a bug or that guests have difficulties to negotiate the transfer of drive on a network. VMotions storage are carried out via the management network (limitation of hard due to the codification)-are you sure that management networks are not overloaded or that they operate at the desired speed? This also happens with machines that have actually disabled followed block? It seems that this could be a problem.

    In addition, you use the 'old' client or Web client? It is possible that some features in vSphere 6.0 may have been omitted in the old thick client.

    Below are the errors:

    2015-04 - T 02, 03: 01:34.468Z | VMX | I120: MigrateShouldPrepareDestination: remote host does not support an explicit step to prepare migration destination.

    2015-04 - T 02, 03: 01:34.471Z | Worker #0 | W110: SVMotion_DiskSetupRPC: related disc expected but not found for the file: /vmfs/volumes/526d0605-8317b046-f37e-0025906cd27e/test1/test1-ctk.vmdk

    Thanks for the answers in advance and good luck!

  • Storage test 55 certification vmotion, storage vmotion fails with "Could not lock file" on the virtual machine during storage vmotion

    In my view, that it is a known problem in ESXi 5.5 Update 2 release notes, someone knows a solution yet? This is an excerpt from the release notes:

    • Attempts to perform live storage vMotion of virtual machines with RDM disks can fail
      Storage vMotion of virtual machines with RDM disks can fail and virtual machines can be considered in turned off of the State. Attempts to turn on the virtual machine fails with the following error:

      Cannot lock the file

      Workaround: No.

    VMware support for the record said that the patch will update 3, it is known as an intermittent problem. Finally, it happened to me on the 9th attempt.

  • storage vmotion 4 1 TB vmdk to a new data store 4TB on 5.5

    All,

    Benefit in performance if the 4 storage vmotion VMDK located on 4 separate data warehouses in a newly created 4TB data store. with 5.5 capacity 62, I was hoping that someone could tell me.

    I appreciate it

    Well well, the response - as before - is "it depends...". ».

    If you want to use a large drive or multiple small disks depends on the structure of data. If you can not divide the data into departments for example, projects, home/profile,... you can be better with a single large disk. Don't forget to use a GPT partition table, and you need to check if for example your backup software can handle that. Some image/VM based backup applications can backup and restore virtual drives very well, but do not have the ability to file level restore! Also don't forget that - disaster - restoration such a large virtual disk will take a significant amount of time during which data is not available for users. If the data structure allows to distribute to several virtual disks (you can always use DFS to present the data through a single share), I prefer this over the large virtual disk. With the smallest disks virtual, you will certainly have to watch each of them for free disk space, but you can increase the size of disk individually and restoration of a single virtual disk can be done much more quickly (with other discs still online / available to users).

    André

  • Storage vMotion

    virtual machines from one storage to another does that 2 simultaneous connections. If I understand well ESXi5,.0 should support up to 8 with 10 GB. Will be setup here to twist?

    Thank you

    The maximum Storage vMotion simultaneous operations by host is 2... You may think that the maximum number of vMotion (not Storage vMotion) which is really 8.

    http://www.VMware.com/PDF/vsphere5/R50/vSphere-50-configuration-maximums.PDF

  • Please explain SVmotion (Storage Vmotion) and DRS for storage in simple words

    Friends,

    I little confusing... could you please contact Storage vmotion and DRS for storage in simple words and the important points to remember in this regard...

    Kind regards

    Sirot

    9884106697

    Storage DRS is a feature that you allow warehouses of data aggregated on a single object (cluster) and recommend the implementation of disks in virtual machine based on the performance of the store of data and the available space on the data store.

    The DRS storage uses Storage vMotion to migrate virtual (online) computers from a data store to another to balance data store performance and space.

    You can use Storage vMotion without storage DRS too, moving virtual machine between clusters of DRS storage and warehouses of data even from the local level to the data store shared.

  • double control system - storage vmotion with or without storage in common?

    If I have a cluster of existing (vsphere 5.1), where 5 guests are all connected to the FC storage, and then I create a new cluster (managed by the same vcenter) with its own storage space, and leaders of each group DO NOT access other cluster storage:

    I can storage vmotion a computer virtual from one cluster to another?

    Y at - he of the opposition? Takes if press the company or company longer? The servers are in the same cluster, or is simply to be managed by the same sufficient vcenter?

    Hello

    It is possible to check this: VMware vSphere 5.1

Maybe you are looking for