Storage of da Migrazione VM FC ISCSI storage
Hello to all,
Abbiamo due separate infrastrutture:
UNA composed da due uno condividono storage server esx 4.1 company (S1 and S2) che FiberChannell 4 GB;
the Altra composed da due 4.1 esx Server enterprise (S3 S4 e) che condividono uno storage ISCSI 1 GB;
Infrastrutture sono currently separate e non hanno some punto di outdo, pure resiedendo nello stesso CED due.
Un'esigenza di dismettere storage FiberChannell, migrare e lo abbiamo tutti i vmdk (has all the VM) dallo storage FiberChannell a quello ISCSI.
Soluzione più semplice ed economica che ipotizzo di offers:
(1) collegare, per mezzo di hba iscsi items, ed S1 S2 agli SAN ISCSI a cui sono treatment S3, S4 switch e lo storage ISCSI;
(2) add I have datastore dello storage ISCSI ad ed S1 S2;
(3) do the dei dai dello CF IA dello storage ISCSI datastore datastore vmdk migrazione.
Che doesn't think?
Ciao,.
Confermo gli sono quelli corretti steps.
Verification prima di avere nic available su e S1 S2 e ovviamente door release sugli switch.
Oltre an IOC, to Magdalene vorrai he multipathing known iSCSI, ti consiglio di leggere questo post.
Ciao
Tags: VMware
Similar Questions
-
Hi all
Can I ask a question regarding the connection of iSCSI storage for use with UCS. Look at us with Nimble iSCSI based storage & want to understand best practice recommendations on how to connect it to UCS to get the best level of performance & reliability / resilience etc.
Another issue is more closely how VMware deals with loss of connectivity on a path (where double connections are the installation program from the warehouse to the tissues), would he re - route traffic to the path running?
Any suggestion would be appreciated.
Kassim
Hello Kassim,
Currently the agile iSCSI storage is certified with UCS 2.0.3 firmware version.
http://www.Cisco.com/en/us/docs/unified_computing/UCS/interoperability/matrix/r_hcl_B_rel2.03.PDF
The following guide can serve as a reference.
Virtualization solution Cisco with the agile storage reference Architecture
http://www.Cisco.com/en/us/solutions/collateral/ns340/ns517/ns224/ns836/ns978/guide_c07-719522.PDF
In above installation, ESXi software iSCSi multipath with PSP Round Robin algorithm is implemented to take care of the IO and failover with load balancing two paths.
HTH
Padma
-
We are poised to implement a UCS B Series. I have a question for FI 6248UP. I have read and used the emulator for UCS Manager and noticed that you can configure ports FI as ways for storage material. Is it limited to a certain protocol storage or the seller? We use Dell EQ and plan to be there to connect the EQ 6510 X. I was wondering if that is supported and if iSCSI upstream traffic would be able to access the storage?
The 6248UPs will be passed to a pair of Catalyst 4506-E race vs. We have an IBM chassis and servers should be able to access the EQ6510X too. I guess I just need master iSCSI VLANS for the 6248UP and the servers would have access?
Thank you!
Hi Cowetac,
Yes, not all storage arrays are supported by UCS, but your DELL Equalogic is supported. If you have any questions about the other compatibility of storage arrays, you can take a look at the UCS (see link below) storage interoperability matrix.
UCS storage interoperability Matrix (matrix of the storage of UCS - B of table 10-2)
http://www.Cisco.com/en/us/docs/switches/Datacenter/MDS9000/interoperability/matrix/Matrix8.html
I guess I just need master iSCSI VLANS for the 6248UP and the servers would have access?
Depends, if you use one port of the device the UCS for you connect directly to storage, you can use a single vlan.
If you are connected via the Ethernet uplinks via a switch, you need to set your switches as a trunk port.
-
Cisco UCS Direct Attached ISCSI Storage access upstream to storage array
I'm doing my first installation of UCS with a chassis 5108) and (4) B200 M3 blade and (2) 6248 fabric interconnects. The plan is to connect our EMC VNX directly on the fabric of interconnections as the chassis will be the only system for access to the SAN, after migration from our current system. The fabric interconnects will be transmitted to (2) 4500-X switches configured in a vs.
In my research of best practices and general deployment guidelines I noticed the need to create (2) separate iscsi VLANS for fabric and a fabric in B. I'm guessing that this requirement is the lack of STP as YEW is configured in EHM, but I was not able to find an exact answer. Any idea would be appreciated.
My main question and my concern is that if I create a second VLAN ISCSI on our X 4500 and then create both of these VIRTUAL on FI networks A & B I'll be able to connect the SAN to the UCS as well as our former VMware cluster (we are migrating our system HP Compute and storage UCS 5.5 environment) in order to facilitate the vMotions of storage?
I noticed the need for iscsi separated (2) VLAN a fabric and a fabric B
Are you referring to
https://www.EMC.com/collateral/hardware/technical-documentation/h8229-VN...
p 36
If I understand your question: can a host attached to 4500 (connected to the UCS FI), access LUNS on the VNX, which is directly connected to the UCS fabric interconnect through lanes (EHM FI) material? YES! I would accept this for a migration, but not forever, because it means that traffic on the fabric of interconnection.
-
12 c grid rhel6 iscsi asm Storage Infrastructure
I'm working on the construction of my 12 c Grid Infrastructure database by using the following:
- RHEL 6.6
- node rac1
- node rac2
- I use Synology NAS storage Oracle Clusterware RAC shared storage. The following objectives were discovered by node rac1 and rac2. (/ dev/sda/dev/sdb/dev/sdc)
- LUN/iSCSI shared CRS 10 g
- LUN/iSCSI shared DATA 400 G
- LUN/iSCSI shared FRA 400 G
My question is?
How do these iSCSI disks for ASM 12 c?
Should I format each drive on each node and then oracleasm create discs?
Please notify.
BECAUSE two-node Oracle 12 c Grid Infrastructure installed on RHEL6
Shared Storage for laboratory: Synology NAS which I created three LUNS/iSCSI devices:
1 for the CRS
1 for DATA
1 for FRA
Each node (initiator) discovered and connect on the objectives, the DATA and FRA
I then used fdisk/dev/sda and/dev/sdb, / dev/sdc, which created
/ dev/sda1
/ dev/sdb1
/ dev/sdc1
Initialized oracleasm on both RAC nodes and created the CBC records using the oracleasm createdisk CRS/dev/sda1
Then oracleasm scandisk
Then oracleasm listdisks
And CBC, DATA and FRA discs seem to be working now.
- RHEL 6.6
-
iSCSI storage presented to the hosts managed by different vCenter Server - questions?
I currently have three hosts (esxi 5.0) that are managed by vcenter server 5.5 U2. The hosts are attached to iSCSI LUNS (data warehouses) 10.
I'm migrating to esxi 6.0 U2. For this, there are three new hosts have esxi 6.0 installed and managed by vcenter server 6.0 U2 U2.
The plan is to detach/remove 5.0 5.5 U2 vcenter esxi hosts. Then import it into vcenter 6.0 U2 (a new cluster will be created). Once imported, uninstall the 5.5 vcenter u2. Then turn off the resident VMs imported esxi 5 hosts and move them to 6.0 esxi hosts.
My query is regarding regarding storage.
At present, three new guests see all storage the old 5.0 esxi hosts to see (guests are not in the cluster. "I'm still trying to put things up). That's because the new hosts were deposited to the same group on the side iSCSI initiator storage. Now things, data warehouses are visible by the hosts with esxi 5.0 (managed by vcenter 5.5 u2) and also the esxi hosts 6.0 (managed by vcenter 6.0 u2). The only VMs residing in esxi environment 6 is vcenter 6.0 u2 and update manager 6.0 u2. These are in a data store that is not any other virtual machines.
That's no problem during the migration? I have not created a cluster for the esxi 6.0 hosts yet and plan to do after obtaining your entries.
Thank you!
No problem whatsoever, regardless whether if you do or that you add no vSphere 6 hosts in a HA cluster.
If you temporarily enable EVC on the vSphere hosts 6, once all hosts are connected to the same vCenter you can VMotion all VMs to new hosts even without any stop. Clear CVS once the migration is complete.
-
Slow iSCSI-IP connection between ESXi and DataCore Virtual storage via 10 Gbit cards
Hi all
at the moment, I have test the following configuration:
Site #1: DELL PowerEdge R730xd with 10 Gbit NIC and DataCore-V DataCore Virtual Storage under ESXi (vSphere 6.0.0 build Standard, 3380124)
Site #2: Apple MacPro6, 1 with 10 Gbit NIC and ESXi (6.0.0 vSphere Standard, build of 3380124)
DataCore # 1 server has a disk via iSCSI to ESXi on #2. The connection is running. Up here, so good, but for example when I start a Storage vMotion on #2 of local SSD on the iSCSI disk, speed is exactly 1 Gbps. interesting is that the speed increase when I start a second Storage vMotion and still when I start a third Storage vMotion etc. This behavior is clearly visible in my attached screenshot.
To me that sounds like to it no matter what iSCSI-IP connection is limited to 1 Gbps.
All the components used are certified VMware.
Any ideas that I can check for this problem?
Thank you very much
Migo
The reason for this behavior is that the MacPro6, 1 has two interfaces of iSCSI software that points to a DataCore frontend. It is not allowed. After you have disabled the second interface of software iSCSI, growth up to full 10 Gbps speed.
-
subnet VMkernel iSCSI for direct-attached storage
Hello
I have two hosts esxi 6.0 that I connect to a shared via iSCSI storage. I followed this document technical http://www.vmware.com/files/pdf/techpaper/vmware-multipathing-configuration-software-iSCSI-port-binding.pdf#sthash.7bztDoAS.dpf to configure interfaces iSCSI on the esxi with binding ports host. No problem.
I use HP MSA 2040 with a dual controller and I'll connect two hosts directly on the shared storage. As for the hardware, according to the documents on best practices, it is recommended to follow a vertical subneting for ports of controllers. However, I have not found no requirement for iSCSI ports in vmware, all of the examples I found configure each vmkernel port in a different subnet, as I did:
vSwitch0
iSCSI1 - vmk2 - 192.168.1.11 (unused vmk3)
iSCSI2 - vmk3 - 192.168.2.11 (unused vmk2)
So, my question is vmkernel ports are in different subnets? or could they be in the same subnet with no difference performance wise?
Any help would be appreciated.
Thank you!
in your case, because you are trying to achieve Multipathing to iSCSI Software initiator, the vmkernel ports should be in the same subnet.
-
The upgrade to ESXi 4.1 with storage iSCSI software to ESXi 5.5
I intend to upgrade ESXi 4.1 with which I have attached to ESXi 5.5 iSCSI software storage
I've updated all my guests who were on the SAN storage with no problems.
I would like to know that if it there's nothing I should take care before I update the host connected iSCSI to ESXi 5.5, as for the hosts with SAN attached storage I made to remove the cables of SAN before upgrading the host.
Also, if all known problems, you will be informed during the upgrade to ESXi 4.1 to ESXi 5.5 with iSCSI Software storage, please let me know?
If there is no different process to upgrade coz in ESXi 4.1 I do not see this link ports of iSCSI vmkernel ports software iSCSI but I know in ESXi5.5 we have to or I just launch a procedure of standard upgrade via the Update Manager and everything will be taken care?
Thanks in advance
With ESXi prior to version 5, linking ports had to be done via command line. However, if it has been configured correctly, you should be able to upgrade to ESXi 5.5 (assuming that the host has not been previously updated since version 3). BTW. the last time that I disconnected the host of storage for an upgrade has been with ESX 3.0.
André
-
Cannot add the iSCSI storage unless it is formatted
Hello
I have therefore 2 boxes of ESX 4.0 and I migrate all the machines to a new box of 5.5 ESXi. Box 5.5 cannot see one of iSCSI targets that can the other 2. The only way I can add by using the application client vSphere to add storage, where I can see it, but I'm not able to add without formatting. Of course, this is not desirable. Why this storage can be used without formatting? All iSCSI targets are the same machine, they all were created the same way I can say, so why is it a unable to add without a format?
Well here is the solution to my problem:
http://www.experts-exchange.com/Software/VMware/Q_27032673.html
I had to mount the storage using the CLI instead of the vSphere client.
-
Creation of the NSX for IP (iSCSI) storage
Hi all
I considered just the implications of VMware NSX on storage over IP (iSCSI) Design.
If we create three groups: Edge, Mgmt and Compute, each in its own grid:
- Each cluster/rack is associated with (2) sheet of 10 GB switches
- There are 2 two 10 GB backbone switches for the installation
- Each sheet is connected to each switch of the spine (sheet ports 47 to the spinal #1 switch) and leaves 48 to the #2 of the spine switch
- 10Gb iSCSI without are configured in the cluster of MGMT/rack and connected to the 10 GB sheet past in the cluster/rack of Mgmt.
Then, how facilitate us the IP storage for ESXi hosts who are not in the same basket as the SAN?
Suppose we define VLAN 99 for IP storage (by Figure 4 - NSX v2.1 Design Guide) and follow the recommendations 'trunking VLAN on the link between the sheet and the spine is not allowed' (p. 65 - NSX v2.1 Design Guide), then the inescapable conclusion is that the SAN are not available since the Compute cluster/rack or edge.
How to set an exception that will allow us to the trunk VLAN 99 (no gateway, without calculation, not able rout) of the spine, so IP storage is accessible from all ESXi hosts?
THX in ADV!
Here, you will need to use the IP routing for communication inter rack configured at the ESXi host level.
For example on the grid 1 VMKernel Interface for storage is 10.77.1.10 in Vlan 77.
switch sheet of TOR, we terminate that VLAN as IVR with IP address of 10.77.1.1
To Inter Communication IP Rack storage, let's add command to host ESXi on racks 1 hosts next-level:
esxcli ip network route ipv4 add - n 10.77.0.0/16 g 10.77.1.1
(Note VMkernel Interface for different media guests will be in different subnets. So in rack 2 VMKernel Interface for storage will be 10.77.2.10)
(Source Guide design v2.1, page 79-80)
-
All share storage iscsi connect same target and volume
Hi Alls,
I have openfiler and do 1 iscsi volume share in two ESX hosts with even described as iscsi target.
So a (first) ESX have dynamic config discover iscsi target and the format of finishing, so ESX block size have new data store come from iscsi.
Another ESX must sign in with the same ip address of the iscsi target and even target iqn. then discover iscsi and rescan datastore have see the name of name identical to that found on ESX first data store.
(Note for second ESX is not block format size but rescan HBAS/VMSF found eponymous store data like first ESX)
Issue.
1. for iscsi target that have one objective and correspond to a volume on openfiler can operate with 2 ESXs at the same time with the same target?
2 if not, how to configure shared storage with openfiler that can used data volume on openfiler?
Kind regards
Bancha
Like I said before, vMotion can saturate an uplink, so you will need to keep an eye on it. If you want to perform vMotion on the same vSwitch anyway, I suggest you create a group of separate ports with private IP (e.g. 192.168.x.y) addresses for her and let the two active vmnic for this group of ports allow failover in the event of an uplink failure.
André
-
I have not yet connect to an iSCSI target device.
I can ping my host
when I open http:// "hostip" in a web browser, I get a 503 service not available.
restart the host gets me nowhere.
SSH opens somehow, but can not connect
Console seems OK
vSphere Client can not connect
If I reset to the default values of the console it is ok, but when I reconfigure the host, this error will be returned.
I tried to reinstall from DVD
I'm completely corrected to date via SSH esxcli
This happens on both my hosts, although they are almost identical Lenovo thinkserver TS140s with broadcom 10Gig NIC and intel NETWORK card integrated
It almost always seems to happen the next time I reboot after enabling iscsi support
The only weird thing I have is that my Integrated NIC is a processor intel 217's and I have to use a special VIB so that it can be used in ESXi
Customer's winning 8.1
Here is my installation notes
Install on USB/SSD stick with custimized ISO i217 NIC driver, reset the configuration and reboot
Management NIC set to NIC0:1Gig
IP management: hostIP/24 GW: my gateway
DNS:DNS on windows vm1, vm2 Windows dns
HostName:ESXi1.Sub.myregistereddomainname custom DNS Suffixes: sub.myregistereddomainname
Reset
Patch to date (https://www.youtube.com/watch?v=_O0Pac0a6g8)
Download the VIB and .zip in a data store using the vSphere Client
To get them (https://www.vmware.com/patchmgr/findPatch.portal)
Start the SSH ESXi service and establish a Putty SSH connection to the ESXi server.
Put the ESXi server in maintenance mode,
example of order: software esxcli vib install /vmfs/volumes/ESXi2-2/patch/ESXi550-201404020.zip d
Re install the Intel 217 NIC driver if removed by patch
Change acceptance ESXi host sustained community level,
command: esxcli software - acceptance-level = CommunitySupported
Install the VIB
command:esxcli vib software install - v /vmfs/volumes/datastore1/net-e1000e-2.3.2.x86_64.vib
command: restart
Connect via VSphere client
-Storage
Check/fix/create local storage. VMFS5
-Networking
vSwitch0
Check vmnic0 (1)
Network port group rename VM to the 'essential '.
Rename management group of network ports for management of basic-VMkernel traffic.
-Configuration time
Enable NTP Client to start and stop the host. ntp.org set 0-3 time servers
DNS and routing
Start the virtual machine and stop
-enable - continue immediately if tools start - stop - prompted action Shutdown - the two delay to 10 seconds
Security profile
Services
SSH - startup - enable the start and stop with host
Cache host configuration
-Properties to start SSD - allocate 40GB for the cache host.
Flashing warnings SSH:
Advanced settings, UserVars, UserVars.SuppressShellWarning, change from 0 to 1.
Storage adapters
-Add - add-in adapter software iSCSI
I think I saw that I was wrong. In fact, I applied two patches when only suited. I started with 5.5u1rollup2 and then applied the ESXi550-201404001 and ESXi550-201404020. Strangely I did not t o had problems until I worked with iSCSI.
-
Hello
We will be consolidating the 2 sites in one soon, and in the context of that we'll have the blades moving and storage iSCSI from one site to the other and re - IP them in the process.
That's what I had in mind, please let me know if Miss me something:
-Stop everything and moving equipment
-Fuel storage iSCSI and re - IP
-Power of blades and change their IPs
(at this point all data warehouses and virtual machines on show them also inaccessible)
-Add the IPs blade storage unit to allow access
-In storage adapters / iSCSI / dynamic discovery, remove the old storage IPs and add new IPs, rescan storage
Here, we should have our data storages and all VMs back on their show normal, that we should be good to start turning on VMs.
Anything I missed? I did similar work a few times, but not in the last year or so, just wanted to do a validation test. We lack ESXi 4.6 build 1050704.
Thank you
If it worked very well, just as described above. I have to award points to myself :-)
Thank you
-
Need urgent help: loss of storage after rename of iSCSI target
Hello guys,.
I need help, I had the problem with one of my iSCSI targets with all virtual machines on it, this iSCSI target on the storage iomega NAS device, is suddenly the storage is no longer available, I tried to rename the 'VmwareStorage' to 'VmwareStorage1' iSCSI target, but his is more appropriate (his was visible on a vsphere servers) but now visible sound like a DOS partition, please help to recover to vsphere without losing any data and the virtual machines inside. Note that I use Vsphere 5.5. see attached photo:
the selected one is 16-bit DOS > = 32 M, its must be VMFS, as all other stores, I don't want to not loose my vms, the company stops and I'm fired
I fixed it... I followed this vmware kb
Thank you vmware, linux thanks<>
Maybe you are looking for
-
Pavilion 15-n027sl: type of motherboard and RAM upgrade
Hello worldI would like to know the maximum amount of ram (and type) that can be installed on the pc of my father, a 15-n027sl Pavilion. At the moment it has 1x4Go slots DDR3L 1600. Furthermore, the pc has two slots, but I noticed that the a slot doe
-
model not found on the lenovo Web site
I recently purchased ideapad lenovoZ580, but I do not find this laptop model in the Web site. I have need for support and software update for this model. model no: 59-383215 3rd generation of Intel ci3 3110 m Windows 8 64-bit oem 4 GB of ram hard dri
-
Infobox last updated in background
I have several years of experience of LabVIEW. I just started using LabWindows 8.5 and have a question about the update of data in an infobox. I will carry out my program in LabWindows as opposed to creating an executable, so this behavior may be dif
-
Two applications of CVI each calling another version of TestStand
We have a Test Manager software written in LabWindows CVI. A version of the Test Manager wrote in a version more old of the CVI and calls TestStand 3.x. A newer version was written in CVI 7.x and also calls TestStand 3.x. The latest version of the Te
-
Windows Media Center ruins my games
Windows Media Center is to give me a real kick in the face. I tried to play Guild Wars and Combat arms, both are online games only. But when I run games, it takes a few seconds and all of a sudden, housing starts Windows Media Center and disable the