Quantcast
Channel: VMware Communities : Discussion List - ESXi
Viewing all 8132 articles
Browse latest View live

esxcli network firewall refresh: More than one instance of tag found, expected only one.

$
0
0

We decided to change the SSH port for ESXi following the instructions in this article:

 

Change Port 22 SSH on ESXi 6.5

 

After doing so, we realized that actually the command

esxcli network firewall refresh

returns with an error

"More than one instance of tag found, expected only one."

 

I checked the XML, and actually there is something wrong with the example configuration given in the article above:

        <protocol>tcp</protocol>

        <protocol>dst</protocol>

 

The second "protocol" entry must actually be changed to "porttype" to conform to the service.xml specification.

 

Now the crazy thing is, with the malformed TCPALLOUT entry and refresh throwing an error, I can access via SSH on the newly defined port.

Without the TCPALLOUT entry, I have no SSH access.

 

And if I  fix "protocol" to "porttype" in the entry and refresh, I instantly lose SSH access completely.

 

How does that fit together?


Migrate vm from host with dvswitch to a host with stanrd virtual switch

$
0
0

I need to migrate a VM from a host with a distributed switch to a host with a standard virtual switch. I get an error: Network card adapter 1 ahs DVPort backing, which is not supported. This could be because the host does not support VDS or because the host has not joined a VDS. I was hoping I could do a live migration.

 

Any ideas?

 

Thank you

Threadripper 3970x GPU and USB Passthrough, esxi 6.7 U3, nvidia RTX 2080, TRX40

$
0
0

Hi all-- wanted to describe progress on an update to my former threadripper system.

 

Starting point: 4 Vm's  on a threadripper 1950, each with GPU passthrough (1 x 2080, 3 x 2070).  64 GB RAM, esxi 6.7 U3. System was quite stable (see prior thread).

 

Target: Threadripper 3970x (double the cores), 128 GB RAM, on an Asrock Creator TRX40 motherboard.

 

I started by validating the new hardware under a temporary (non-virtualized) windows build.  Stuff worked.

 

BIOS settings used: defaults, except:

Changed some fan settings to make them quieter

Turned on XMP

Turned on SR-IOV

left PBO OFF (default, but I changed it to disabled. PBO sucks up huge amounts of power for little performance benefit, to say nothing of validation!)

Used current BIOS version, not the beta for 3990x.

 

esxi installation: Used previous installation.  This had passthru.map entries for AMD and NVIDIA as detailed in my last post. It also had the epyc-recommended configuration change previously recommended (which I removed) and preinstalled aquantia NIC driver.

 

Moved 2 x m.2 SSd's  from old into new system. System booted nicely into esxi. All hardware passthrough vanished as expected. Of note, *neither* of the NIC's on this board has a native driver. I used the aquantia driver and live off the 10Gb aquantia.  I have no idea if there is a realtek Dragon 2.5g driver out there.

 

Redid the hardware passthrough.  All GPU's passed back through to their VM's, yay.  Only two of them would boot, boo. Eventually, after much gnashing of teeth, remade three VM's from scratch: they would all keep crashing immediately upon booting windows.

 

This was interesting.  Esxi would report that the GPU's had violated memory access and advised adding a passthrough.map entry, which didn't fix the problem.  Changing BIOS on the host to remove CSM support and enable 4G support, and enabling 64 bit MMIO in the vm didn't fix it either.  A new vm with fresh windows install worked.

 

There were several other interesting changes from previous system:

 

disabling msi on the gpu made them keep crashing , unlike previously when it fixed stuttering

No cpu pinning or NUMA settings were used or needed

The mystical cpuid.hypervisor setting remains required to avoid error 43

 

With these caveats, I got 4 bootable VM's each using the Nvidia card's own USB-c connector for keyboard/mouse. 8 cpus/vm.  Which led to the next problem, which I haven't been able to solve:

 

The mice/keyboards would all intermittently freeze for moments to minutes, and sometimes not come back. Lots of testing inside of windows showed no cause. Interestingly, the problem was 1)worse with a highend G502 mouse, and 2) much worse inside of windows UI -- and never happened for example in demanding real time full screen apps.  I was sure it was going to be some bizarre windows problem. Rarely (every few hours) systems would crash completely (while idle!) with the same memory access violation.  Also rebooting one of the vm's would make other vm's momentarily stutter. None of this ever happened on the 1950x system where these controllers were reliable.

 

I eventually worked around the problem with the motherboard's USB controllers.  There are 5: 2 x Matisse, 2 x Starship, and 1 x asmedia.  The Matisse ones are lumped in the same IOMMU group and won't pass through (They are perpetually "reboot needed.") The Asmedia chip worked with no problems (usb-c port on the back of the motherboard).  The Starship USB 3.0 controllers both worked IF you had a passthru.map entry moving them to d3d0 reset method.  Otherwise, booting a VM with one of these controllers failed AND crashed a different VM with a GPU memory access violation and the controller then permanently disappeared until the system was then powered down (not just a reboot).  Wow, talk about bad crashes.

 

Using these 3 motherboard controllers an 3 vms appears rock stable (I haven't tested the fourth yet.) One of them has 64but mmo enabled which probably isn't needed.

 

Things I haven't gotten around to testing yet:

 

1. Does isolating the vm to one ccx fix anything?

2. If only one vm is running, does the usb-c nvidia controller become reliable?

3. Does turning off XMP or using the latest beta BIOS change anything?

 

Other advice -- I'm obviously waaay off of HCL here-- but don't even try DRAMless SSD's.  Datastore *vanishes* under high load. Bad. Same thing happens with my OEM samsung until I updated the firmware, but that's another story well documented elsewhere.

 

I'm really puzzled by the nvidia usb-c thing. Would also be nice if the Matisse controllers worked.  Otherwise mostly pleased-- many of the kludges needed on older esxi versions and the 1950x with its wacky NUMA configuration are no longer needed and the new system is *much* faster.

 

 

Hope this helps someone else.  If anyone can tell me what's going on (or at least that it's not just me) would be much appreciated. I speculate it's a BIOS bug.

 

Thanks LT

datastore as shared storage

$
0
0

Greetings,

I'm fairly new to VMWare. I have set up a server with VMWare ESXI and have created two logical devices that are controlled by a RAID controller.

1 x 6TB RAID 5 for storage (hopefully) - currently a datastore [activedata]

1 x 2TB RAID 1 for virtual machines - currently a datastore [datastore1]

 

I have installed Windows server on the 2TB datastore and I need to add the other data store a shared storage. I have been beating my head against the wall over this. I have no way to create a drive on the second datastore. Am I going about this wrong? Also, when I try and add an existing hdd to the virtual machine, there is none listed in the selection pool.

VMWare.pngvmware settings.pngvmware datastore browser.png

How to Configure Physical Network Adapter

$
0
0

See image for more clarification:

 

I have a 2 port 10GB network card in a physical server. 1 of the ports is used for management and is connected to our physical network. The other port I want to set to use iSCSI to a synology storage server and access from one of the Virtual Machines in VMWare.

I expect that in order to be able to communicate with the Synology box, I would need to be able to set the IP address on that physical network adapter port. Am I going about this the wrong way?

Where are these settings as I do not find them in the Network settings in the web management interface.

vmware networking.png

FNIC Abort errors on different hosts in a cluster

$
0
0

Hi,

 

 

While I examine the VMkwasrning log for a host, its showing this error.

 

2020-01-25T14:29:07.390Z cpu48:33390)WARNING: LinScsi: SCSILinuxAbortCommands:1890: Failed, Driver fnic, for vmhba2

2020-01-25T14:29:09.391Z cpu41:137434)WARNING: NMP: nmp_DeviceRequestFastDeviceProbe:237: NMP device "naa.514f0c539f60000f" state in doubt; requested fast path state update...

2020-01-25T14:53:44.761Z cpu40:35708)WARNING: LinuxThread: 340: sfcb-HTTP-Daemo: Error cloning thread: -28 (bad0081)

2020-01-25T14:53:44.762Z cpu40:35708)WARNING: LinuxThread: 340: sfcb-HTTP-Daemo: Error cloning thread: -28 (bad0081)

2020-01-25T14:53:44.763Z cpu40:35708)WARNING: LinuxThread: 340: sfcb-HTTP-Daemo: Error cloning thread: -28 (bad0081)

2020-01-25T16:04:55.996Z cpu4:37993)WARNING: VSCSI: 3711: handle 8237(vscsi0:0):WaitForCIF: Issuing reset;  number of CIF:2

2020-01-25T16:04:55.996Z cpu4:37993)WARNING: VSCSI: 2632: handle 8237(vscsi0:0):Ignoring double reset

2020-01-25T16:05:50.685Z cpu36:33390)WARNING: LinScsi: SCSILinuxAbortCommands:1890: Failed, Driver fnic, for vmhba2

2020-01-25T16:05:52.686Z cpu14:129249)WARNING: NMP: nmp_DeviceRequestFastDeviceProbe:237: NMP device "naa.514f0c539f60000f" state in doubt; requested fast path state update...

2020-01-25T16:22:06.044Z cpu45:35708)WARNING: LinuxThread: 340: sfcb-HTTP-Daemo: Error cloning thread: -28 (bad0081)

2020-01-26T19:12:36.741Z cpu21:33188)WARNING: LinScsi: SCSILinuxAbortCommands:1890: Failed, Driver fnic, for vmhba2

2020-01-26T19:12:36.741Z cpu21:33188)WARNING: Cow: 1707: Failed to reset level 2, fid 2688425

2020-01-26T19:12:36.741Z cpu29:55666)WARNING: NMP: nmp_DeviceRequestFastDeviceProbe:237: NMP device "naa.514f0c5279200013" state in doubt; requested fast path state update...

2020-01-27T18:44:04.647Z cpu54:32956)WARNING: Uplink: 5427: requesting cap 0x0!

2020-01-27T18:44:04.648Z cpu54:32956)WARNING: Uplink: 5427: requesting cap 0x0!

0:00:00:15.042 cpu0:32768)WARNING: VMKAcpi: 2448: Bus 10 (0b) is already defined

0:00:00:17.046 cpu0:32768)WARNING: PCI: 1275: No resources for device: 0000:ff:12.0, BAR[4]: 0x0, size: 64, type: 0x2, flags: 0x0

0:00:00:17.046 cpu0:32768)WARNING: PCI: 1275: No resources for device: 0000:ff:12.4, BAR[2]: 0x0, size: 64, type: 0x2, flags: 0x0

0:00:00:19.050 cpu0:32768)WARNING: PCI: 1275: No resources for device: 0000:7f:12.0, BAR[4]: 0x0, size: 64, type: 0x2, flags: 0x0

0:00:00:19.051 cpu0:32768)WARNING: PCI: 1275: No resources for device: 0000:7f:12.4, BAR[2]: 0x0, size: 64, type: 0x2, flags: 0x0

0:00:00:19.053 cpu0:32768)WARNING: PCI: 160: device 0000:7f:1e.3 pciBar 0 (t=1 f=0x3 0x14-0x18) registration failed (Bad address range)

2020-01-27T18:48:05.577Z cpu46:33541)WARNING: LinuxSignal: 541: vmkdevmgr: ignored unexpected signal flags 0x2 (sig 17)

2020-01-27T18:48:15.722Z cpu2:33477)WARNING: NetDVS: 660: portAlias is NULL

2020-01-27T18:48:15.727Z cpu2:33477)WARNING: NetDVS: 660: portAlias is NULL

2020-01-27T18:48:15.731Z cpu2:33477)WARNING: NetDVS: 660: portAlias is NULL

 

 

Can anyone please explain why its happening ?

 

Is this something we need to take it serious ?

 

Thanks,

 

Vivek

Virtual Networks on ESX 6.7

$
0
0

Hi everyone,

 

I recently installed ESX 6.7 in a lab on an Intel NUC with one NIC built-in and added a separate NIC by installing a USB-C NIC adapter. Both NICs are connect to a single switch and (successfully) receive their internal IP addresses via DHCP. The USB NIC is used for management, while the "real" NIC will then be used for the production network traffic (DMZ, once I have set everything up).

 

Now, I have created two virtual NICs (one for each real NIC), two virtual switches (again, one for each) and two port groups (all with vlan 0) and connected them so that I ideally would have two separate TCP/IP stacks - one for production use and one for management only. However, I cannot add network adapters to virtual machines (dropdown is blank) and on the dashboard page, the section "Networks" is empty...can anyone tell me how to configure this correctly?

 

Many thanks,

 

Andreas

Unknown storage device 0B

$
0
0

I have 4 brand new HPE Gen10 servers, and onbly on one of them I get this device :

How to get rid of it ?


Install only ESXi security patches

$
0
0

Hi,

 

is it possible to install only the current security patches but don't go to a higher version(U3)?

 

For Example: im running "ESXi U2 13006603" and i want to install the current security patches.

 

Do I have to upgrade to the current version U3 to get the current security update/patches or is there a option to install only security patches?

 

For example: if i want to have the Security patchtes from "ESXi 6.7 Patch 1 ESXi670-201912001 2019-12-05 15160138"  - so i have to go to U3?

 

 

Thanks

VM stops pinging after vmotion

$
0
0

I know there are many, many previous discussions on this topic and have not yet found out the issue.

 

I have two hosts in a cluster, which both were running ESXi, 6.5.0, 7388607. These hosts are HP BL460c Gen9 in an HP c7000 enclosure with flex fabric virtual connects. I wanted to upgrade these hosts but b/c these hosts were using the old net-bnx2x driver and not the qfle3 driver the normal VUM upgrade would fail when the host would come back up to a black screen. So instead of doing alot of manual driver updates, etc. my team opted to do a "fresh" install of 6.5.0, 11925212 (latest) since we've had multiple versions upgraded over the years. This fresh install would reuse the same management IP and vMotion IP's configured. I did a fresh install on the first host to 6.5.0, 11925212. Everything went great as any previous host reinstall that I've done. When I was ready to start on the second host, when the VM's vMotioned to the first host (now upgraded to latest), half of the VM's lost network connection when they landed on this host. Some VM's in the same VLAN were not affected after vMotion. A quick vmotion back didn't seem to fix the issue (my recollection on this part is fuzzy, I was just trying to do everything I could to get VM's back on network). I had another cluster of hosts using 6.5.0, 11925212 that see's the same storage for these VM's and when I vmotioned the VM's to that cluster the VM's randomly still would not have a network connection (but some would automatically ping). Once the VM Windows Guest was restarted or powered off and on, the network connection would come back. Since then I have upgraded the second host in the previous two host cluster to 6.5.0, 11925212, so now the hosts match versions. I have a few VM's left in this cluster still that if I vmotion to the other host (either host back and forth) that it randomly loses network connection. Sometimes it doesnt. For ex - Scenario 1 - VM1 on Host1 is pinging. I migrate VM1 to Host2, it will stop pinging. Migrate VM1 back to Host1, still not pinging. Power off VM or restart and it will resume pinging. Scenario 2 - VM1 on Host1 is pinging. I migrate VM1 to Host2, it will still ping.

 

Before 6.5.0, 11925212 we did not have this issue on these hosts with the same set of VM's in this cluster. Only after upgrading to 6.5.0, 11925212. For additional troubleshooting, I can edit the VM settings and sometimes choose a different port on the network adapter that is open and the VM resumes its network connection. Sometimes not. I am at a loss as to what the issue is here, but it seems related to 6.5.0, 11925212??? The same vDS is used across ALL clusters in the environment so its not that. I'm less inclined to think its out upstream physical switch b/c we did not have these issues before 6.5.0, 11925212. For now I turned off DRS so VM's do not move around. It does seem like both hosts have this issue. I do have a VMware SR open and am waiting to hear back, but I am very curious what the communities thoughts are.

 

Other observations in VM1's current network state:

--------------------------------

From my workstation:

-Cannot ping VM1

 

From Host2 VM1 is on:

-Cannot ping VM1 from its host (host mgmt is in different vlan)

 

From VM1 with connectivity issues:

-Cannot ping its host

-CAN ping another VM on the same vlan and same host (interesting)

-Cannot ping its gateway

-Cannot ping another VM on the same vlan on a different host in the cluster

 

From another (lets say VM2) without network connectivity issues on the same host:

-CAN ping the problematic VM successfully

-----------------------------------

An ESXTOP -n on the host indicates VM1 (problematic) and VM2 (no problems) are both using vmnic4

vmnic4 shows 100 under %DRPTX which from the surface seems alarming

vmnic5 shows 0.00 under %DRPTX

ESXi 6.5 Deploy ova from Datastore

$
0
0

Hello to all,

 

can you please guys help me on how to deploy an ova file from my datastore on the ESXi that I have? [6.5.0 Update 1 (Build 6765664)]

I cant find a way how to do that!

 

Thank you

FS TX2550M4 / CIM Provider

$
0
0

Hi,

i trying found some solution but no luck.

 

Have FS TX2550 with EP420i raid card.

Have installed ESXi 6.7u3

Win2019 STD.

 

But in FS RAID manager im unable to add ESXi host for monitor RAID. Error - no CIM provider found.

 

I have downloaded  ServerView ESXi CIM Provider 9.10.04 (27/11/2019) from FS website.

Upload to TMP via WinSCP

Host in maintanenc mode

esxcli software vib install -d “/TMP/Datastore/DirectoryName/PatchName.zip“

but after this command i get only  "  > "

Same end of command when i unpack vib file from zip.

 

Can somebody help me?

 

Thank you Petr

 

 

Persistent firewall rule

$
0
0

hey guys,

 

   I took over some ESXi servers and I noticed they were not logging correctly.  So I added the firewall rule to open port 10514 through esxcli using an article on git hub. However, when I reboot the rule disappears.  I tried added the firewall rule in a separate xml file and adding it directly to the service.xml file.  yet each time  I reboot the ESXi host the rule and the xml file is not there when it comes back up.   Is this normal behavior for the /etc/vmware/firewall/ directory or am I missing something here. Also if it is normal how do I a persistant firewall rule to allow port 10514 because all I can find in the docs is to edit the service.xml file.

L1 Terminal Fault mitigation setting does not take

$
0
0

I have a Dell R740 with the latest BIOS. I am swapping them with R730s. Both are running ESXi v6.5U2h.

 

I noticed my R730s had the warning about the the L1 fault and so I set this to True.

VMkernel.Boot.hyperthreadingMitigation

 

My R740s never gave the CVE warning. I set the mitigation setting anyway and rebooted. However, it still had 72 Logical CPUs. This is a 36core box.

 

My R730s went from 72Logical CPUs to 36 after the fix.

 

Is the BIOS on my R740 not susceptical and thus the setting wont take?

 

 

 

 

 

VMWare ESXI 6.5 - How to create local datastore (where ESXI 6.5 is installed) ? Error "No partition..."

$
0
0

Hello,

 

I have ESXi 6.5 installed on a server.

I use 2 x 1.8 TB hard drive. (I do NOT use raid).

 

How to create a partition on the same (primary) disk as ESXI OS ?

 

If i select the same "physical" disk as the ESXI installation, i get an error: "No partition has been selected. Please select a partition on the left hand side diagram."

See picture below: datastore-hdd_01.png

 

My question is very simple (and stupid ?! sorry), but after some research, i can't find a solution (for dummies ).

I have been using ESXI 4, 5 and 5.5 for years and have never had this problem, i could use the same physical disk for esxi and some VMFS data.

 

If i create a partition on the second physical disk, everything goes well and i get "the alias link (datastore2 -> for mine)" in the console.

(The command "df -h" in console with ssh give the result below: datastore-hdd_02.png).


The summary:

 

I have two physical drives.

 

Disk 1 where i installed VMWare ESXI 6.5 (and where i'd like to use the remaining 1.81TB):

 

Partition diagram

1. EFI System (3.97 MB)

5. Basic Data (249.98 MB)

6. Basic Data (249.98 MB)

7. VMware Diagnostic (109.98 MB)

8. Basic Data (285.98 MB)

9. VMware Diagnostic (2.5 GB)

2. Basic Data (4 GB)

3. VMFS (1.81 TB)

 

and Disk 2 (for future):

 

Partition diagram

1. VMFS (1.82 TB)

Free space (1.09 MB)

 

Would an expert (or less lame than me :p) have a simple solution (to use the first disk (where esxi is installed)) ?

 

I thank you in advance for your time

 

NB: Sorry for my English, is not my natal language


---

 

 

Physical DISK 1 in HTML5:
datastore-hdd_01.png

300119-datastore-hdd_01.png


Physical DISK 2 (and other) in console (command "df -h"):
datastore-hdd_02.png
300119-datastore-hdd_02.png


Installed ESXI 6.7 in 32GB micro SD card , forgot the root password , when I tried reinstalling , it is not asking for password and coming up with same configuration.

$
0
0

Installed ESXI 6.7 in 32GB micro SD card and created the local VMFS store  , forgot the root password , when I tried reinstalling , it is not asking for password and coming up with same configuration. Need to know reinstall freshly and delete the local VMFS datastore.

CPU Issues

$
0
0

Hi All,

 

I have a small eSXI host with 16 logical processors. I am running 3 VMs on it each assigned 5 CPUs.

Something doesn't seem right. For what is running on this, the CPU should not be struggling like it is. I was wondering if anyone had any ideas on what I can look at to see if there may be an underlying issue that would make CPU performance sub-par? Running just one or two apps pegs it at 99%....I cannot find anything glaring on the OS side and this evening I upgraded to the latest Dell image. Could this be a driver issue or the way I have the CPUs assigned? Thanks.

VMware ESXi decrease vmdk size

$
0
0

With vmware esxi product, below version details.

 

1.33.3

 

12923304

 

6.7.0

 

13006603

 

 

I have increased size of a virtual machine, with machine --> edit (edit settings) -->

WHen I increase the size, it is increased.

ANd it came into unpartitioned space of the windows 10 machine, I increased that using disk management to 400 GB.

Later, I shrinked the partition and kept the 200 GB as unparititoned space.

 

Now, when I try to decrease the space in the esxi settings, I am getting below error.

 

SO, please suggest me how to decrease the disk space.

 

Even, I tried by connecting to ssh to the esxi server and ran the below command to decrease disk size.

 

vmkfstools -X <new size> <virtual disk>.vmdk

 

But getting error as not a valid vmdk file.

 

Please suggest.

 

Same error with any machine.

SSD flash HGST-Ultrastar-SN260-NVMe problem

$
0
0

I have a SSD Flash HGST-Ultrastar-SN260-NVMe in an SuperMicro server.

I install VMWare 6.7. And i did not se any device for this adapter.

I see that the adapter is on HCL and have the right firmware and drivers.

 

I see in the Configure the Model: Ulstrastar SN200 Series NVMe SSD on the vmhba4.

But i not see anything in the Storage Devices. Only my HDD SAS drive that i have also.

 

In the Windows server installd ont this server i see alls hdd on this Adapter.

1.6 Tb size.

 

Did i forgett someting for this adatper to show up in the Configure?

Or is there a driver issue or tomething?

 

Thanks

Transport (VMDB) error -45: Failed to connect to peer process.

$
0
0

Good Morning,

 

I am receiving an error on my Esxi server when starting any vm. It started when I tried to deploy a macos machine. I was using the patch updater for the macos. It worked once, then when I updated a src setting, all of my machines give the following error.

 

haTask-2-vim.VirtualMachine.powerOn-233
Power On this virtual machine
Failed - Transport (VMDB) error -45: Failed to connect to peer process.

I asked our vmware guy at the it place I work at and he mentioned that I may not have the correct version of esxi on my machine. It is a Dell OptiPlex machine running an i7 quad core with 64gb of ram. Right now I just have the general esxi vmware u3 6.7. If I can add anymore information, let me know. Just to clarify as well, I have completely dumped, nuked and paved the machine as of last night with the same error as soon as I start the mac one up after it is created.

Viewing all 8132 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>