Created
November 21, 2014 01:44
-
-
Save michaelrice/ace75264c3874ffb2bb2 to your computer and use it in GitHub Desktop.
VCR recording of issue 190
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| interactions: | |
| - request: | |
| body: null | |
| headers: | |
| Accept: ['*/*'] | |
| Accept-Encoding: ['gzip, deflate'] | |
| Connection: [keep-alive] | |
| User-Agent: [python-requests/2.4.3 CPython/2.7.8 Darwin/13.4.0] | |
| method: GET | |
| uri: https://10.12.254.137:443//sdk/vimServiceVersions.xml | |
| response: | |
| body: {string: !!python/unicode "<?xml version=\"1.0\" encoding=\"UTF-8\" ?>\n\ | |
| <!--\n Copyright 2008-2012 VMware, Inc. All rights reserved.\n-->\n<namespaces\ | |
| \ version=\"1.0\">\n <namespace>\n <name>urn:vim25</name>\n <version>5.5</version>\n\ | |
| \ <priorVersions>\n <version>5.1</version>\n <version>5.0</version>\n\ | |
| \ <version>4.1</version>\n <version>4.0</version>\n <version>2.5u2</version>\n\ | |
| \ <version>2.5</version>\n </priorVersions>\n </namespace>\n <namespace>\n\ | |
| \ <name>urn:vim2</name>\n <version>2.0</version>\n </namespace>\n</namespaces>\n"} | |
| headers: | |
| connection: [Keep-Alive] | |
| content-length: ['530'] | |
| content-type: [text/xml] | |
| date: ['Thu, 20 Nov 2014 23:40:45 GMT'] | |
| status: {code: 200, message: OK} | |
| - request: | |
| body: '<?xml version="1.0" encoding="UTF-8"?> | |
| <soapenv:Envelope xmlns:soapenc="http://schemas.xmlsoap.org/soap/encoding/" | |
| xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xsd="http://www.w3.org/2001/XMLSchema" | |
| xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"> | |
| <soapenv:Body><RetrieveServiceContent xmlns="urn:vim25"><_this type="ServiceInstance">ServiceInstance</_this></RetrieveServiceContent></soapenv:Body> | |
| </soapenv:Envelope>' | |
| headers: | |
| Accept-Encoding: ['gzip, deflate'] | |
| Content-Type: [text/xml; charset=UTF-8] | |
| Cookie: [''] | |
| SOAPAction: ['"urn:vim25/5.5"'] | |
| method: POST | |
| uri: https://10.12.254.137:443/sdk | |
| response: | |
| body: {string: !!python/unicode "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n\ | |
| <soapenv:Envelope xmlns:soapenc=\"http://schemas.xmlsoap.org/soap/encoding/\"\ | |
| \n xmlns:soapenv=\"http://schemas.xmlsoap.org/soap/envelope/\"\n xmlns:xsd=\"\ | |
| http://www.w3.org/2001/XMLSchema\"\n xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\"\ | |
| >\n<soapenv:Body>\n<RetrieveServiceContentResponse xmlns=\"urn:vim25\"><returnval><rootFolder\ | |
| \ type=\"Folder\">group-d1</rootFolder><propertyCollector type=\"PropertyCollector\"\ | |
| >propertyCollector</propertyCollector><viewManager type=\"ViewManager\">ViewManager</viewManager><about><name>VMware\ | |
| \ vCenter Server</name><fullName>VMware vCenter Server 5.5.0 build-2001466</fullName><vendor>VMware,\ | |
| \ Inc.</vendor><version>5.5.0</version><build>2001466</build><localeVersion>INTL</localeVersion><localeBuild>000</localeBuild><osType>linux-x64</osType><productLineId>vpx</productLineId><apiType>VirtualCenter</apiType><apiVersion>5.5</apiVersion><instanceUuid>B4DA3BE6-54CC-4712-9602-A7F038620624</instanceUuid><licenseProductName>VMware\ | |
| \ VirtualCenter Server</licenseProductName><licenseProductVersion>5.0</licenseProductVersion></about><setting\ | |
| \ type=\"OptionManager\">VpxSettings</setting><userDirectory type=\"UserDirectory\"\ | |
| >UserDirectory</userDirectory><sessionManager type=\"SessionManager\">SessionManager</sessionManager><authorizationManager\ | |
| \ type=\"AuthorizationManager\">AuthorizationManager</authorizationManager><serviceManager\ | |
| \ type=\"ServiceManager\">ServiceMgr</serviceManager><perfManager type=\"\ | |
| PerformanceManager\">PerfMgr</perfManager><scheduledTaskManager type=\"ScheduledTaskManager\"\ | |
| >ScheduledTaskManager</scheduledTaskManager><alarmManager type=\"AlarmManager\"\ | |
| >AlarmManager</alarmManager><eventManager type=\"EventManager\">EventManager</eventManager><taskManager\ | |
| \ type=\"TaskManager\">TaskManager</taskManager><extensionManager type=\"\ | |
| ExtensionManager\">ExtensionManager</extensionManager><customizationSpecManager\ | |
| \ type=\"CustomizationSpecManager\">CustomizationSpecManager</customizationSpecManager><customFieldsManager\ | |
| \ type=\"CustomFieldsManager\">CustomFieldsManager</customFieldsManager><diagnosticManager\ | |
| \ type=\"DiagnosticManager\">DiagMgr</diagnosticManager><licenseManager type=\"\ | |
| LicenseManager\">LicenseManager</licenseManager><searchIndex type=\"SearchIndex\"\ | |
| >SearchIndex</searchIndex><fileManager type=\"FileManager\">FileManager</fileManager><datastoreNamespaceManager\ | |
| \ type=\"DatastoreNamespaceManager\">DatastoreNamespaceManager</datastoreNamespaceManager><virtualDiskManager\ | |
| \ type=\"VirtualDiskManager\">virtualDiskManager</virtualDiskManager><snmpSystem\ | |
| \ type=\"HostSnmpSystem\">SnmpSystem</snmpSystem><vmProvisioningChecker type=\"\ | |
| VirtualMachineProvisioningChecker\">ProvChecker</vmProvisioningChecker><vmCompatibilityChecker\ | |
| \ type=\"VirtualMachineCompatibilityChecker\">CompatChecker</vmCompatibilityChecker><ovfManager\ | |
| \ type=\"OvfManager\">OvfManager</ovfManager><ipPoolManager type=\"IpPoolManager\"\ | |
| >IpPoolManager</ipPoolManager><dvSwitchManager type=\"DistributedVirtualSwitchManager\"\ | |
| >DVSManager</dvSwitchManager><hostProfileManager type=\"HostProfileManager\"\ | |
| >HostProfileManager</hostProfileManager><clusterProfileManager type=\"ClusterProfileManager\"\ | |
| >ClusterProfileManager</clusterProfileManager><complianceManager type=\"ProfileComplianceManager\"\ | |
| >MoComplianceManager</complianceManager><localizationManager type=\"LocalizationManager\"\ | |
| >LocalizationManager</localizationManager><storageResourceManager type=\"\ | |
| StorageResourceManager\">StorageResourceManager</storageResourceManager><guestOperationsManager\ | |
| \ type=\"GuestOperationsManager\">guestOperationsManager</guestOperationsManager></returnval></RetrieveServiceContentResponse>\n\ | |
| </soapenv:Body>\n</soapenv:Envelope>"} | |
| headers: | |
| cache-control: [no-cache] | |
| connection: [Keep-Alive] | |
| content-length: ['3599'] | |
| content-type: [text/xml; charset=utf-8] | |
| date: ['Thu, 20 Nov 2014 23:40:45 GMT'] | |
| set-cookie: [vmware_soap_session="522df684-0c61-5c3b-a3a7-ddedd4013b79"; Path=/; | |
| HttpOnly; Secure;] | |
| status: {code: 200, message: OK} | |
| - request: | |
| body: '<?xml version="1.0" encoding="UTF-8"?> | |
| <soapenv:Envelope xmlns:soapenc="http://schemas.xmlsoap.org/soap/encoding/" | |
| xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xsd="http://www.w3.org/2001/XMLSchema" | |
| xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"> | |
| <soapenv:Body><Login xmlns="urn:vim25"><_this type="SessionManager">SessionManager</_this><userName>[email protected]</userName><password>password</password></Login></soapenv:Body> | |
| </soapenv:Envelope>' | |
| headers: | |
| Accept-Encoding: ['gzip, deflate'] | |
| Content-Type: [text/xml; charset=UTF-8] | |
| Cookie: [vmware_soap_session="522df684-0c61-5c3b-a3a7-ddedd4013b79"; Path=/; | |
| HttpOnly; Secure;] | |
| SOAPAction: ['"urn:vim25/5.5"'] | |
| method: POST | |
| uri: https://10.12.254.137:443/sdk | |
| response: | |
| body: {string: !!python/unicode "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n\ | |
| <soapenv:Envelope xmlns:soapenc=\"http://schemas.xmlsoap.org/soap/encoding/\"\ | |
| \n xmlns:soapenv=\"http://schemas.xmlsoap.org/soap/envelope/\"\n xmlns:xsd=\"\ | |
| http://www.w3.org/2001/XMLSchema\"\n xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\"\ | |
| >\n<soapenv:Body>\n<LoginResponse xmlns=\"urn:vim25\"><returnval><key>52d3a5a7-29a1-414a-9256-abeacb0a3c04</key><userName>VSPHERE.LOCAL\\\ | |
| Administrator</userName><fullName>Administrator vsphere.local</fullName><loginTime>2014-11-20T23:40:46.00032Z</loginTime><lastActiveTime>2014-11-20T23:40:46.00032Z</lastActiveTime><locale>en</locale><messageLocale>en</messageLocale><extensionSession>false</extensionSession><ipAddress>10.12.254.102</ipAddress><userAgent></userAgent><callCount>0</callCount></returnval></LoginResponse>\n\ | |
| </soapenv:Body>\n</soapenv:Envelope>"} | |
| headers: | |
| cache-control: [no-cache] | |
| connection: [Keep-Alive] | |
| content-length: ['827'] | |
| content-type: [text/xml; charset=utf-8] | |
| date: ['Thu, 20 Nov 2014 23:40:46 GMT'] | |
| status: {code: 200, message: OK} | |
| - request: | |
| body: '<?xml version="1.0" encoding="UTF-8"?> | |
| <soapenv:Envelope xmlns:soapenc="http://schemas.xmlsoap.org/soap/encoding/" | |
| xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xsd="http://www.w3.org/2001/XMLSchema" | |
| xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"> | |
| <soapenv:Body><RetrieveServiceContent xmlns="urn:vim25"><_this type="ServiceInstance">ServiceInstance</_this></RetrieveServiceContent></soapenv:Body> | |
| </soapenv:Envelope>' | |
| headers: | |
| Accept-Encoding: ['gzip, deflate'] | |
| Content-Type: [text/xml; charset=UTF-8] | |
| Cookie: [vmware_soap_session="522df684-0c61-5c3b-a3a7-ddedd4013b79"; Path=/; | |
| HttpOnly; Secure;] | |
| SOAPAction: ['"urn:vim25/5.5"'] | |
| method: POST | |
| uri: https://10.12.254.137:443/sdk | |
| response: | |
| body: {string: !!python/unicode "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n\ | |
| <soapenv:Envelope xmlns:soapenc=\"http://schemas.xmlsoap.org/soap/encoding/\"\ | |
| \n xmlns:soapenv=\"http://schemas.xmlsoap.org/soap/envelope/\"\n xmlns:xsd=\"\ | |
| http://www.w3.org/2001/XMLSchema\"\n xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\"\ | |
| >\n<soapenv:Body>\n<RetrieveServiceContentResponse xmlns=\"urn:vim25\"><returnval><rootFolder\ | |
| \ type=\"Folder\">group-d1</rootFolder><propertyCollector type=\"PropertyCollector\"\ | |
| >propertyCollector</propertyCollector><viewManager type=\"ViewManager\">ViewManager</viewManager><about><name>VMware\ | |
| \ vCenter Server</name><fullName>VMware vCenter Server 5.5.0 build-2001466</fullName><vendor>VMware,\ | |
| \ Inc.</vendor><version>5.5.0</version><build>2001466</build><localeVersion>INTL</localeVersion><localeBuild>000</localeBuild><osType>linux-x64</osType><productLineId>vpx</productLineId><apiType>VirtualCenter</apiType><apiVersion>5.5</apiVersion><instanceUuid>B4DA3BE6-54CC-4712-9602-A7F038620624</instanceUuid><licenseProductName>VMware\ | |
| \ VirtualCenter Server</licenseProductName><licenseProductVersion>5.0</licenseProductVersion></about><setting\ | |
| \ type=\"OptionManager\">VpxSettings</setting><userDirectory type=\"UserDirectory\"\ | |
| >UserDirectory</userDirectory><sessionManager type=\"SessionManager\">SessionManager</sessionManager><authorizationManager\ | |
| \ type=\"AuthorizationManager\">AuthorizationManager</authorizationManager><serviceManager\ | |
| \ type=\"ServiceManager\">ServiceMgr</serviceManager><perfManager type=\"\ | |
| PerformanceManager\">PerfMgr</perfManager><scheduledTaskManager type=\"ScheduledTaskManager\"\ | |
| >ScheduledTaskManager</scheduledTaskManager><alarmManager type=\"AlarmManager\"\ | |
| >AlarmManager</alarmManager><eventManager type=\"EventManager\">EventManager</eventManager><taskManager\ | |
| \ type=\"TaskManager\">TaskManager</taskManager><extensionManager type=\"\ | |
| ExtensionManager\">ExtensionManager</extensionManager><customizationSpecManager\ | |
| \ type=\"CustomizationSpecManager\">CustomizationSpecManager</customizationSpecManager><customFieldsManager\ | |
| \ type=\"CustomFieldsManager\">CustomFieldsManager</customFieldsManager><diagnosticManager\ | |
| \ type=\"DiagnosticManager\">DiagMgr</diagnosticManager><licenseManager type=\"\ | |
| LicenseManager\">LicenseManager</licenseManager><searchIndex type=\"SearchIndex\"\ | |
| >SearchIndex</searchIndex><fileManager type=\"FileManager\">FileManager</fileManager><datastoreNamespaceManager\ | |
| \ type=\"DatastoreNamespaceManager\">DatastoreNamespaceManager</datastoreNamespaceManager><virtualDiskManager\ | |
| \ type=\"VirtualDiskManager\">virtualDiskManager</virtualDiskManager><snmpSystem\ | |
| \ type=\"HostSnmpSystem\">SnmpSystem</snmpSystem><vmProvisioningChecker type=\"\ | |
| VirtualMachineProvisioningChecker\">ProvChecker</vmProvisioningChecker><vmCompatibilityChecker\ | |
| \ type=\"VirtualMachineCompatibilityChecker\">CompatChecker</vmCompatibilityChecker><ovfManager\ | |
| \ type=\"OvfManager\">OvfManager</ovfManager><ipPoolManager type=\"IpPoolManager\"\ | |
| >IpPoolManager</ipPoolManager><dvSwitchManager type=\"DistributedVirtualSwitchManager\"\ | |
| >DVSManager</dvSwitchManager><hostProfileManager type=\"HostProfileManager\"\ | |
| >HostProfileManager</hostProfileManager><clusterProfileManager type=\"ClusterProfileManager\"\ | |
| >ClusterProfileManager</clusterProfileManager><complianceManager type=\"ProfileComplianceManager\"\ | |
| >MoComplianceManager</complianceManager><localizationManager type=\"LocalizationManager\"\ | |
| >LocalizationManager</localizationManager><storageResourceManager type=\"\ | |
| StorageResourceManager\">StorageResourceManager</storageResourceManager><guestOperationsManager\ | |
| \ type=\"GuestOperationsManager\">guestOperationsManager</guestOperationsManager></returnval></RetrieveServiceContentResponse>\n\ | |
| </soapenv:Body>\n</soapenv:Envelope>"} | |
| headers: | |
| cache-control: [no-cache] | |
| connection: [Keep-Alive] | |
| content-length: ['3599'] | |
| content-type: [text/xml; charset=utf-8] | |
| date: ['Thu, 20 Nov 2014 23:40:46 GMT'] | |
| status: {code: 200, message: OK} | |
| - request: | |
| body: '<?xml version="1.0" encoding="UTF-8"?> | |
| <soapenv:Envelope xmlns:soapenc="http://schemas.xmlsoap.org/soap/encoding/" | |
| xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xsd="http://www.w3.org/2001/XMLSchema" | |
| xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"> | |
| <soapenv:Body><RetrievePropertiesEx xmlns="urn:vim25"><_this type="PropertyCollector">propertyCollector</_this><specSet><propSet><type>ServiceInstance</type><all>false</all><pathSet>content</pathSet></propSet><objectSet><obj | |
| type="ServiceInstance">ServiceInstance</obj><skip>false</skip></objectSet></specSet><options><maxObjects>1</maxObjects></options></RetrievePropertiesEx></soapenv:Body> | |
| </soapenv:Envelope>' | |
| headers: | |
| Accept-Encoding: ['gzip, deflate'] | |
| Content-Type: [text/xml; charset=UTF-8] | |
| Cookie: [vmware_soap_session="522df684-0c61-5c3b-a3a7-ddedd4013b79"; Path=/; | |
| HttpOnly; Secure;] | |
| SOAPAction: ['"urn:vim25/5.5"'] | |
| method: POST | |
| uri: https://10.12.254.137:443/sdk | |
| response: | |
| body: {string: !!python/unicode "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n\ | |
| <soapenv:Envelope xmlns:soapenc=\"http://schemas.xmlsoap.org/soap/encoding/\"\ | |
| \n xmlns:soapenv=\"http://schemas.xmlsoap.org/soap/envelope/\"\n xmlns:xsd=\"\ | |
| http://www.w3.org/2001/XMLSchema\"\n xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\"\ | |
| >\n<soapenv:Body>\n<RetrievePropertiesExResponse xmlns=\"urn:vim25\"><returnval><objects><obj\ | |
| \ type=\"ServiceInstance\">ServiceInstance</obj><propSet><name>content</name><val\ | |
| \ xsi:type=\"ServiceContent\"><rootFolder type=\"Folder\">group-d1</rootFolder><propertyCollector\ | |
| \ type=\"PropertyCollector\">propertyCollector</propertyCollector><viewManager\ | |
| \ type=\"ViewManager\">ViewManager</viewManager><about><name>VMware vCenter\ | |
| \ Server</name><fullName>VMware vCenter Server 5.5.0 build-2001466</fullName><vendor>VMware,\ | |
| \ Inc.</vendor><version>5.5.0</version><build>2001466</build><localeVersion>INTL</localeVersion><localeBuild>000</localeBuild><osType>linux-x64</osType><productLineId>vpx</productLineId><apiType>VirtualCenter</apiType><apiVersion>5.5</apiVersion><instanceUuid>B4DA3BE6-54CC-4712-9602-A7F038620624</instanceUuid><licenseProductName>VMware\ | |
| \ VirtualCenter Server</licenseProductName><licenseProductVersion>5.0</licenseProductVersion></about><setting\ | |
| \ type=\"OptionManager\">VpxSettings</setting><userDirectory type=\"UserDirectory\"\ | |
| >UserDirectory</userDirectory><sessionManager type=\"SessionManager\">SessionManager</sessionManager><authorizationManager\ | |
| \ type=\"AuthorizationManager\">AuthorizationManager</authorizationManager><serviceManager\ | |
| \ type=\"ServiceManager\">ServiceMgr</serviceManager><perfManager type=\"\ | |
| PerformanceManager\">PerfMgr</perfManager><scheduledTaskManager type=\"ScheduledTaskManager\"\ | |
| >ScheduledTaskManager</scheduledTaskManager><alarmManager type=\"AlarmManager\"\ | |
| >AlarmManager</alarmManager><eventManager type=\"EventManager\">EventManager</eventManager><taskManager\ | |
| \ type=\"TaskManager\">TaskManager</taskManager><extensionManager type=\"\ | |
| ExtensionManager\">ExtensionManager</extensionManager><customizationSpecManager\ | |
| \ type=\"CustomizationSpecManager\">CustomizationSpecManager</customizationSpecManager><customFieldsManager\ | |
| \ type=\"CustomFieldsManager\">CustomFieldsManager</customFieldsManager><diagnosticManager\ | |
| \ type=\"DiagnosticManager\">DiagMgr</diagnosticManager><licenseManager type=\"\ | |
| LicenseManager\">LicenseManager</licenseManager><searchIndex type=\"SearchIndex\"\ | |
| >SearchIndex</searchIndex><fileManager type=\"FileManager\">FileManager</fileManager><datastoreNamespaceManager\ | |
| \ type=\"DatastoreNamespaceManager\">DatastoreNamespaceManager</datastoreNamespaceManager><virtualDiskManager\ | |
| \ type=\"VirtualDiskManager\">virtualDiskManager</virtualDiskManager><snmpSystem\ | |
| \ type=\"HostSnmpSystem\">SnmpSystem</snmpSystem><vmProvisioningChecker type=\"\ | |
| VirtualMachineProvisioningChecker\">ProvChecker</vmProvisioningChecker><vmCompatibilityChecker\ | |
| \ type=\"VirtualMachineCompatibilityChecker\">CompatChecker</vmCompatibilityChecker><ovfManager\ | |
| \ type=\"OvfManager\">OvfManager</ovfManager><ipPoolManager type=\"IpPoolManager\"\ | |
| >IpPoolManager</ipPoolManager><dvSwitchManager type=\"DistributedVirtualSwitchManager\"\ | |
| >DVSManager</dvSwitchManager><hostProfileManager type=\"HostProfileManager\"\ | |
| >HostProfileManager</hostProfileManager><clusterProfileManager type=\"ClusterProfileManager\"\ | |
| >ClusterProfileManager</clusterProfileManager><complianceManager type=\"ProfileComplianceManager\"\ | |
| >MoComplianceManager</complianceManager><localizationManager type=\"LocalizationManager\"\ | |
| >LocalizationManager</localizationManager><storageResourceManager type=\"\ | |
| StorageResourceManager\">StorageResourceManager</storageResourceManager><guestOperationsManager\ | |
| \ type=\"GuestOperationsManager\">guestOperationsManager</guestOperationsManager></val></propSet></objects></returnval></RetrievePropertiesExResponse>\n\ | |
| </soapenv:Body>\n</soapenv:Envelope>"} | |
| headers: | |
| cache-control: [no-cache] | |
| connection: [Keep-Alive] | |
| content-length: ['3739'] | |
| content-type: [text/xml; charset=utf-8] | |
| date: ['Thu, 20 Nov 2014 23:40:46 GMT'] | |
| status: {code: 200, message: OK} | |
| - request: | |
| body: '<?xml version="1.0" encoding="UTF-8"?> | |
| <soapenv:Envelope xmlns:soapenc="http://schemas.xmlsoap.org/soap/encoding/" | |
| xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xsd="http://www.w3.org/2001/XMLSchema" | |
| xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"> | |
| <soapenv:Body><RetrievePropertiesEx xmlns="urn:vim25"><_this type="PropertyCollector">propertyCollector</_this><specSet><propSet><type>EventManager</type><all>false</all><pathSet>description</pathSet></propSet><objectSet><obj | |
| type="EventManager">EventManager</obj><skip>false</skip></objectSet></specSet><options><maxObjects>1</maxObjects></options></RetrievePropertiesEx></soapenv:Body> | |
| </soapenv:Envelope>' | |
| headers: | |
| Accept-Encoding: ['gzip, deflate'] | |
| Content-Type: [text/xml; charset=UTF-8] | |
| Cookie: [vmware_soap_session="522df684-0c61-5c3b-a3a7-ddedd4013b79"; Path=/; | |
| HttpOnly; Secure;] | |
| SOAPAction: ['"urn:vim25/5.5"'] | |
| method: POST | |
| uri: https://10.12.254.137:443/sdk | |
| response: | |
| body: {string: !!python/unicode "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n\ | |
| <soapenv:Envelope xmlns:soapenc=\"http://schemas.xmlsoap.org/soap/encoding/\"\ | |
| \n xmlns:soapenv=\"http://schemas.xmlsoap.org/soap/envelope/\"\n xmlns:xsd=\"\ | |
| http://www.w3.org/2001/XMLSchema\"\n xmlns:xsi=\"http://www.w3.org/2001/XMLSchema-instance\"\ | |
| >\n<soapenv:Body>\n<RetrievePropertiesExResponse xmlns=\"urn:vim25\"><returnval><objects><obj\ | |
| \ type=\"EventManager\">EventManager</obj><propSet><name>description</name><val\ | |
| \ xsi:type=\"EventDescription\"><category><label>info</label><summary>Information</summary><key>info</key></category><category><label>warning</label><summary>Warning</summary><key>warning</key></category><category><label>error</label><summary>Error</summary><key>error</key></category><category><label>user</label><summary>User</summary><key>user</key></category><eventInfo><key>ExtendedEvent</key><description>Import\ | |
| \ certificate success</description><category>info</category><formatOnDatacenter>Import\ | |
| \ certificate succeeded.</formatOnDatacenter><formatOnComputeResource>Import\ | |
| \ certificate succeeded.</formatOnComputeResource><formatOnHost>Import certificate\ | |
| \ succeeded.</formatOnHost><formatOnVm>Import certificate succeeded.</formatOnVm><fullFormat>ad.event.ImportCertEvent|Import\ | |
| \ certificate succeeded.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="ad.event.ImportCertEvent"> <description> \ | |
| \ Import certificate succeeded </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Import\ | |
| \ certificate failure</description><category>error</category><formatOnDatacenter>Import\ | |
| \ certificate failed.</formatOnDatacenter><formatOnComputeResource>Import\ | |
| \ certificate failed.</formatOnComputeResource><formatOnHost>Import certificate\ | |
| \ failed.</formatOnHost><formatOnVm>Import certificate failed.</formatOnVm><fullFormat>ad.event.ImportCertFailedEvent|Import\ | |
| \ certificate failed.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="ad.event.ImportCertFailedEvent"> <description>\ | |
| \ Import certificate failed </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Join\ | |
| \ domain success</description><category>info</category><formatOnDatacenter>Join\ | |
| \ domain succeeded.</formatOnDatacenter><formatOnComputeResource>Join domain\ | |
| \ succeeded.</formatOnComputeResource><formatOnHost>Join domain succeeded.</formatOnHost><formatOnVm>Join\ | |
| \ domain succeeded.</formatOnVm><fullFormat>ad.event.JoinDomainEvent|Join\ | |
| \ domain succeeded.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="ad.event.JoinDomainEvent"> <description> \ | |
| \ Join domain succeeded </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Join\ | |
| \ domain failure</description><category>error</category><formatOnDatacenter>Join\ | |
| \ domain failed.</formatOnDatacenter><formatOnComputeResource>Join domain\ | |
| \ failed.</formatOnComputeResource><formatOnHost>Join domain failed.</formatOnHost><formatOnVm>Join\ | |
| \ domain failed.</formatOnVm><fullFormat>ad.event.JoinDomainFailedEvent|Join\ | |
| \ domain failed.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="ad.event.JoinDomainFailedEvent"> <description>\ | |
| \ Join domain failed </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Leave\ | |
| \ domain success</description><category>info</category><formatOnDatacenter>Leave\ | |
| \ domain succeeded.</formatOnDatacenter><formatOnComputeResource>Leave domain\ | |
| \ succeeded.</formatOnComputeResource><formatOnHost>Leave domain succeeded.</formatOnHost><formatOnVm>Leave\ | |
| \ domain succeeded.</formatOnVm><fullFormat>ad.event.LeaveDomainEvent|Leave\ | |
| \ domain succeeded.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="ad.event.LeaveDomainEvent"> <description> \ | |
| \ Leave domain succeeded </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Leave\ | |
| \ domain failure</description><category>error</category><formatOnDatacenter>Leave\ | |
| \ domain failed.</formatOnDatacenter><formatOnComputeResource>Leave domain\ | |
| \ failed.</formatOnComputeResource><formatOnHost>Leave domain failed.</formatOnHost><formatOnVm>Leave\ | |
| \ domain failed.</formatOnVm><fullFormat>ad.event.LeaveDomainFailedEvent|Leave\ | |
| \ domain failed.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="ad.event.LeaveDomainFailedEvent"> <description>\ | |
| \ Leave domain failed </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>Added\ | |
| \ License</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.AddLicenseEvent|License\ | |
| \ {licenseKey} added to VirtualCenter</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Assigned\ | |
| \ License</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.AssignLicenseEvent|License\ | |
| \ {licenseKey} assigned to asset {entityName} with id {entityId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Download\ | |
| \ License Information</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.DLFDownloadFailedEvent|Failed\ | |
| \ to download license information from the host {hostname} due to {errorReason.@enum.com.vmware.license.DLFDownloadFailedEvent.DLFDownloadFailedReason}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>License\ | |
| \ assignment faults</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.LicenseAssignFailedEvent|License\ | |
| \ assignment on the host fails. Reasons: {[email protected]}.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.license.LicenseAssignFailedEvent">\ | |
| \ <description> The host license assignment succeeds\ | |
| \ on vCenter Server but can not be successfully pushed down to the\ | |
| \ host. Any license assignment to a host proceeds in two stages. In the \ | |
| \ first stage vCenter Server does preliminary checks on the license\ | |
| \ key, the license state of the host and determines if the requested\ | |
| \ assignment is valid. If so, it stores this assignment locally\ | |
| \ in its database. In the second stage, vCenter Server pushes the newly \ | |
| \ assigned license to the host. During the second stage the host might\ | |
| \ reject the assignment under certain circumstances. These circumstances\ | |
| \ usually result from a mismatch of the information available to\ | |
| \ vCenter Server and the host concerned. Any such discrepancies are notified\ | |
| \ to the user via this event. This event lists the reason because\ | |
| \ of which it was logged and also shows up as a configuration issue\ | |
| \ on the vSphere Client. </description> <cause> \ | |
| \ <description>License expiry information mismatch between\ | |
| \ vCenter Server and host</description> <action>If\ | |
| \ the system time on the machine running vCenter Server and host \ | |
| \ are not in sync then put them in sync</action> </cause>\ | |
| \ <cause> <description>The license key is a\ | |
| \ per Virtual Machine key and the number of powered on Virtual Machines \ | |
| \ is larger than the maximum limit of the key</description>\ | |
| \ <action>Use a different key with a larger capacity</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>License\ | |
| \ Capacity Exceeded</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.LicenseCapacityExceededEvent|The\ | |
| \ current license usage ({currentUsage} {costUnitText}) for {edition} exceeds\ | |
| \ the license capacity ({capacity} {costUnitText})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>License\ | |
| \ Expiry</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Your\ | |
| \ host license expires in {remainingDays} days. The host will disconnect from\ | |
| \ vCenter Server when its license expires.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.LicenseExpiryEvent|Your\ | |
| \ host license expires in {remainingDays} days. The host will disconnect from\ | |
| \ vCenter Server when its license expires.</fullFormat><longDescription> \ | |
| \ <EventLongDescription id="com.vmware.license.LicenseExpiryEvent">\ | |
| \ <description> If a host is assigned a temporary license\ | |
| \ (a license key with an expiry), this event is logged in order\ | |
| \ to provide users an advanced warning on the imminent expiry of the \ | |
| \ license key. The event logging starts 15 days prior to the expiry of\ | |
| \ the license key. This event also shows up on the host summary\ | |
| \ page as a configuration issue on the vSphere Client. </description>\ | |
| \ <cause> <description>License key is about to\ | |
| \ expire or has expired</description> <action>Assign\ | |
| \ a different license key</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>License\ | |
| \ User Threshold Exceeded</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.LicenseUserThresholdExceededEvent|The\ | |
| \ current license usage ({currentUsage} {costUnitText}) for {edition} exceeds\ | |
| \ the user-defined threshold ({threshold} {costUnitText})</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.license.LicenseUserThresholdExceededEvent">\ | |
| \ <description> Users can define thresholds to monitor\ | |
| \ overuse of the product license. This event is logged when the license\ | |
| \ usage threshold defined by the user for a product edition is exceeded. \ | |
| \ </description> <cause> <description>\ | |
| \ License usage of a product edition has exceeded the user-defined\ | |
| \ threshold </description> <action> \ | |
| \ Review license assignments and usage </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>Removed\ | |
| \ License</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.RemoveLicenseEvent|License\ | |
| \ {licenseKey} removed from VirtualCenter</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Unassigned\ | |
| \ License</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.license.UnassignLicenseEvent|License\ | |
| \ unassigned from asset {entityName} with id {entityId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA agent can reach all cluster management addresses</description><category>info</category><formatOnDatacenter>The\ | |
| \ vSphere HA agent on the host {host.name} in cluster {computeResource.name}\ | |
| \ can reach all the cluster management addresses</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ vSphere HA agent on the host {host.name} can reach all the cluster management\ | |
| \ addresses</formatOnComputeResource><formatOnHost>The vSphere HA agent on\ | |
| \ this host can reach all the cluster management addresses</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.AllHostAddrsPingable|The\ | |
| \ vSphere HA agent on the host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} can reach all the cluster management addresses</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.AllHostAddrsPingable">\ | |
| \ <description> The host is able to ping all\ | |
| \ of the vSphere HA management addresses of every other cluster host. \ | |
| \ </description> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA agent can reach all isolation addresses</description><category>info</category><formatOnDatacenter>All\ | |
| \ vSphere HA isolation addresses are reachable by host {host.name} in cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>All\ | |
| \ vSphere HA isolation addresses are reachable by this host</formatOnComputeResource><formatOnHost>All\ | |
| \ vSphere HA isolation addresses are reachable by host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.AllIsoAddrsPingable|All\ | |
| \ vSphere HA isolation addresses are reachable by host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.AllIsoAddrsPingable">\ | |
| \ <description> The host is able to ping all\ | |
| \ of the vSphere HA isolation addresses. </description> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA answered a lock-lost question on a virtual machine</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA answered the lock-lost question on virtual machine {vm.name} on host\ | |
| \ {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA answered the lock-lost question on virtual machine {vm.name} on host\ | |
| \ {host.name}</formatOnComputeResource><formatOnHost>vSphere HA answered the\ | |
| \ lock-lost question on virtual machine {vm.name}</formatOnHost><formatOnVm>vSphere\ | |
| \ HA answered the lock-lost question on this virtual machine</formatOnVm><fullFormat>com.vmware.vc.HA.AnsweredVmLockLostQuestionEvent|vSphere\ | |
| \ HA answered the lock-lost question on virtual machine {vm.name} on host\ | |
| \ {host.name} in cluster {computeResource.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.AnsweredVmLockLostQuestionEvent">\ | |
| \ <description> The virtual machine running on this\ | |
| \ host lost the exclusive lock of its files on disk. This will occur\ | |
| \ if another instance of this virtual machine is running on a different host.\ | |
| \ This situation can happen if a host loses access to both its storage\ | |
| \ and management networks but is not configured to shutdown its\ | |
| \ virtual machines on isolation. The virtual machines on this host\ | |
| \ will continue to run without access to their disks, while vSphere HA will\ | |
| \ start a new instance of the virtual machines on another host in\ | |
| \ the cluster. When the isolated host regains access to the storage\ | |
| \ network, it will try to reacquire the disk locks. This will fail since \ | |
| \ the disk locks are held by another host. The host will then issue\ | |
| \ a question on the virtual machine indicating that disk locks have\ | |
| \ been lost. vSphere HA will automatically answer this question\ | |
| \ to allow the virtual machine instance without the disk locks to power off.\ | |
| \ <description> </EventLongDescription> \ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA answered a question from the host about terminating a virtual machine</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA answered a question from host {host.name} in cluster {computeResource.name}\ | |
| \ about terminating virtual machine {vm.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA answered a question from host {host.name} about terminating virtual machine\ | |
| \ {vm.name}</formatOnComputeResource><formatOnHost>vSphere HA answered a question\ | |
| \ from the host about terminating virtual machine {vm.name}</formatOnHost><formatOnVm>vSphere\ | |
| \ HA answered a question from the host about terminating this virtual machine</formatOnVm><fullFormat>com.vmware.vc.HA.AnsweredVmTerminatePDLEvent|vSphere\ | |
| \ HA answered a question from host {host.name} in cluster {computeResource.name}\ | |
| \ about terminating virtual machine {vm.name}</fullFormat><longDescription>\t\ | |
| \t<EventLongDescription id="com.vmware.vc.HA.AnsweredVmTerminatePDLEvent">\t\ | |
| \t <description>\t\t The virtual machine running on this host had\ | |
| \ a virtual disk which experienced permenant device loss. \t\t The host will\ | |
| \ issue a question if it is configured to terminate the VM automatically under\ | |
| \ \t\t such condition. This event indicates that vSphere HA answered the\ | |
| \ question. After the VM is terminated, \t\t vSphere HA will make a best\ | |
| \ effort to restart it.\t\t <description>\t\t</EventLongDescription>\t\ | |
| \t</longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA disabled the automatic VM Startup/Shutdown feature</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA disabled the automatic Virtual Machine Startup/Shutdown feature on host\ | |
| \ {host.name} in cluster {computeResource.name}. Automatic VM restarts will\ | |
| \ interfere with HA when reacting to a host failure.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA disabled the automatic Virtual Machine Startup/Shutdown feature on the\ | |
| \ host {host.name}. Automatic VM restarts will interfere with HA when reacting\ | |
| \ to a host failure.</formatOnComputeResource><formatOnHost>vSphere HA disabled\ | |
| \ the automatic Virtual Machine Startup/Shutdown feature. Automatic VM restarts\ | |
| \ will interfere with HA when reacting to a host failure.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.AutoStartDisabled|vSphere\ | |
| \ HA disabled the automatic Virtual Machine Startup/Shutdown feature on host\ | |
| \ {host.name} in cluster {computeResource.name} in {datacenter.name}. Automatic\ | |
| \ VM restarts will interfere with HA when reacting to a host failure.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.AutoStartDisabled">\ | |
| \ <description> Virtual Machine Startup/Shutdown\ | |
| \ has been disabled by HA. A host which is contained in an vSphere\ | |
| \ HA cluster is not permitted to have automatic virtual machine startup and\ | |
| \ shutdown since it may conflict with HA's attempts to relocate\ | |
| \ the virtual machines if a host fails. </description> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA did not reset a VM which had files on inaccessible datastore(s)</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA did not reset VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ because the VM had files on inaccessible datastore(s)</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA did not reset VM {vm.name} on host {host.name} because the VM had files\ | |
| \ on inaccessible datastore(s)</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA did not reset VM {vm.name} on this host because the VM had files on inaccessible\ | |
| \ datastore(s)</formatOnHost><formatOnVm>vSphere HA did not reset this VM\ | |
| \ because the VM had file(s) on inaccessible datastore(s)</formatOnVm><fullFormat>com.vmware.vc.HA.CannotResetVmWithInaccessibleDatastore|vSphere\ | |
| \ HA did not reset VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because the VM had files on inaccessible datastore(s)</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.HA.CannotResetVmWithInaccessibleDatastore</longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA cluster contains incompatible hosts.</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA Cluster {computeResource.name} contains ESX/ESXi 3.5 hosts and more recent\ | |
| \ host versions, which isn't fully supported.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA Cluster contains ESX/ESXi 3.5 hosts and more recent host versions, which\ | |
| \ isn't fully supported.</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.ClusterContainsIncompatibleHosts|vSphere\ | |
| \ HA Cluster {computeResource.name} in {datacenter.name} contains ESX/ESXi\ | |
| \ 3.5 hosts and more recent host versions, which isn't fully supported.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA completed a failover action</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA completed a virtual machine failover action in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA completed a virtual machine failover action</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.ClusterFailoverActionCompletedEvent|vSphere\ | |
| \ HA completed a virtual machine failover action in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA initiated a failover action</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA initiated a virtual machine failover action in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA initiated a virtual machine failover action</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.ClusterFailoverActionInitiatedEvent|vSphere\ | |
| \ HA initiated a virtual machine failover action in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA agent is running</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on host {host.name} in cluster {computeResource.name} is running</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on host {host.name} is running</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent is running</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasAgentRunningEvent|vSphere\ | |
| \ HA agent on host {host.name} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name} is running</fullFormat><longDescription> <EventLongDescription\ | |
| \ id=" com.vmware.vc.HA.DasAgentRunningEvent"> \ | |
| \ <description> This event is logged when the\ | |
| \ vSphere HA agent is running on a host. </description>\ | |
| \ <cause> <description>\ | |
| \ This event is reported after vSphere HA is configured\ | |
| \ on a host or after the vSphere HA agent on a host\ | |
| \ starts, such as after a host reboot. </description>\ | |
| \ </cause> </EventLongDescription> \ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA detected a failed failover host</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA detected a possible failure of failover host {host.name} in cluster {computeResource.name}\ | |
| \ failed</formatOnDatacenter><formatOnComputeResource>vSphere HA detected\ | |
| \ a possible failure of failover host {host.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected a possible failure of this failover host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasFailoverHostFailedEvent|vSphere\ | |
| \ HA detected a possible failover of failover host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name} has failed</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.HA.DasFailoverHostFailedEvent">\ | |
| \ <description> This event is\ | |
| \ logged when vSphere HA has detected the failure of a designated failover\ | |
| \ host. </description> <cause> \ | |
| \ <description> If the\ | |
| \ admission control policy specifies one or more failover hosts, this event\ | |
| \ will be generated if vSphere HA detects the failure\ | |
| \ of a failover host. A host is considered to have failed by a vSphere HA\ | |
| \ master agent if it looses contact with the vSphere\ | |
| \ HA agent on the host, the host does not respond to pings on any of the \ | |
| \ management interfaces, and the master does not observe\ | |
| \ any datastore heartbeats. </description> \ | |
| \ <action> Determine the\ | |
| \ cause of the failover host failure, and correct. vSphere HA will make a\ | |
| \ best effort to place VMs on remaining hosts in the\ | |
| \ cluster if the failover host is not running and a host failure occurs. See\ | |
| \ the product documentation for troubleshooting tips.\ | |
| \ </action> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA detected a network-isolated failover host</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA detected that failover host {host.name} is network isolated from cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA detected that failover host {host.name} is network isolated from the\ | |
| \ cluster</formatOnComputeResource><formatOnHost>vSphere HA detected that\ | |
| \ this failover host is network isolated from the cluster</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasFailoverHostIsolatedEvent|Host\ | |
| \ {host.name} has been isolated from cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.HA.DasFailoverHostIsolatedEvent">\ | |
| \ <description> This event is\ | |
| \ logged when vSphere HA has detected the network isolation of a designated\ | |
| \ failover host. </description> <cause>\ | |
| \ <description> If the\ | |
| \ admission control policy specifies one or more failover hosts, this event\ | |
| \ will be generated if vSphere HA detects the network\ | |
| \ isolation of a failover host. vSphere HA reports a host as isolated if there\ | |
| \ are no heartbeats received from the HA agent on\ | |
| \ that host, the host is not pingable on any of the management interfaces,\ | |
| \ yet the host is still alive as determined by the\ | |
| \ the host's datastore heartbeats. </description>\ | |
| \ <action> Determine the\ | |
| \ cause of the failover host isolation, and correct. vSphere HA will make\ | |
| \ a best effort to place VMs on remaining hosts in\ | |
| \ the cluster if the failover host is isolated and a host failure occurs.\ | |
| \ See the product documentation for troubleshooting\ | |
| \ tips. </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA detected a network-partitioned failover host</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA detected that failover host {host.name} in {computeResource.name} is\ | |
| \ in a different network partition than the master to which vCenter Server\ | |
| \ is connected</formatOnDatacenter><formatOnComputeResource>vSphere HA detected\ | |
| \ that failover host {host.name} is in a different network partition than\ | |
| \ the master to which vCenter Server is connected</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected that this failover host is in a different network partition\ | |
| \ than the master</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasFailoverHostPartitionedEvent|Failover\ | |
| \ Host {host.name} in {computeResource.name} in {datacenter.name} is in a\ | |
| \ different network partition than the master</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.HA.DasFailoverHostPartitionedEvent">\ | |
| \ <description> This event is\ | |
| \ logged when vSphere HA has detected a designated failover host is network\ | |
| \ partitioned. </description> <cause>\ | |
| \ <description> If the\ | |
| \ admission control policy specifies one or more failover hosts, this event\ | |
| \ will be generated if a vSphere HA master agent detects a \ | |
| \ failover host is network partitioned. vSphere HA reports a host\ | |
| \ as partitioned if it cannot communicate with a subset \ | |
| \ of hosts in the cluster, yet can determine that the host is alive\ | |
| \ via its datastore heartbeats. </description>\ | |
| \ <action> Determine the\ | |
| \ cause of the partitioned failover host, and correct. vSphere HA will make\ | |
| \ a best effort to place VMs on remaining hosts in\ | |
| \ the cluster if a failover host is partitioned and a host failure occurs.\ | |
| \ See the prodcut documentation for troubleshooting\ | |
| \ tips. </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA agent on a failover host is unreachable</description><category>error</category><formatOnDatacenter>The\ | |
| \ vSphere HA agent on the failover host {host.name} in {computeResource.name}\ | |
| \ is not reachable but host responds to ICMP pings</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ vSphere HA agent on the failover host {host.name} is not reachable but host\ | |
| \ responds to ICMP pings</formatOnComputeResource><formatOnHost>The vSphere\ | |
| \ HA agent on this failover host is not reachable but host responds to ICMP\ | |
| \ pings</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasFailoverHostUnreachableEvent|The\ | |
| \ vSphere HA agent on the failover host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} is not reachable but host responds to ICMP pings</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Host\ | |
| \ complete datastore failure</description><category>error</category><formatOnDatacenter>All\ | |
| \ shared datastores failed on the host {hostName} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>All\ | |
| \ shared datastores failed on the host {hostName}</formatOnComputeResource><formatOnHost>All\ | |
| \ shared datastores failed on the host {hostName}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasHostCompleteDatastoreFailureEvent|All\ | |
| \ shared datastores failed on the host {hostName} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.DasHostCompleteDatastoreFailureEvent">\ | |
| \ <description> A host in a Component Protection-enabled\ | |
| \ cluster has lost connectivity to all shared datastores </description>\ | |
| \ <cause> <description>Connectivity to all shared\ | |
| \ datastores has been lost</description> <action>Reconnect\ | |
| \ at least one shared datastore</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>Host\ | |
| \ complete network failure</description><category>error</category><formatOnDatacenter>All\ | |
| \ VM networks failed on the host {hostName} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>All\ | |
| \ VM networks failed on the host {hostName}</formatOnComputeResource><formatOnHost>All\ | |
| \ VM networks failed on the host {hostName}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasHostCompleteNetworkFailureEvent|All\ | |
| \ VM networks failed on the host {hostName} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.DasHostCompleteNetworkFailureEvent"> \ | |
| \ <description> A host in a Component Protection enabled\ | |
| \ cluster has lost connectivity to all virtual machine networks \ | |
| \ </description> <cause> <description>Connectivity\ | |
| \ to all virtual machine networks has been lost</description> \ | |
| \ <action>Reconnect at least one virtual machine network</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA detected a host failure</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA detected a possible host failure of host {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA detected a possible host failure of host {host.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected a possible host failure of this host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasHostFailedEvent|vSphere\ | |
| \ HA detected a possible host failure of host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id=" com.vmware.vc.HA.DasHostFailedEvent"> <description>\ | |
| \ This event is logged when vSphere HA detects a possible\ | |
| \ host failure. </description> <cause>\ | |
| \ <description> \ | |
| \ A host is considered to have failed by a vSphere HA master agent if it\ | |
| \ looses contact with the vSphere HA agent\ | |
| \ on the host, the host does not respond to pings on any of the management\ | |
| \ interfaces, and the master does not observe\ | |
| \ any datastore heartbeats. </description> \ | |
| \ <action> Determine\ | |
| \ the cause of the host failure, and correct. See the product documentation\ | |
| \ for troubleshooting tips. </action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA detected a network isolated host</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA detected that host {host.name} is network isolated from cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA detected that host {host.name} is network isolated from the cluster</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected that this host is network isolated from the cluster</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasHostIsolatedEvent|vSphere\ | |
| \ HA detected that host {host.name} is isolated from cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id=" com.vmware.vc.HA.DasHostIsolatedEvent"> \ | |
| \ <description> This event is logged when vSphere\ | |
| \ HA has detected the network isolation of a host. </description>\ | |
| \ <cause> <description>\ | |
| \ This event will be generated if there are no heartbeats\ | |
| \ received from the vSphere HA agent on that host, the host is \ | |
| \ not pingable on any of the management interfaces, yet the host\ | |
| \ is still alive as determined by the the host's \ | |
| \ datastore heartbeats. </description> \ | |
| \ <action> Determine the\ | |
| \ cause of the host isolation, and correct. See the product documentation\ | |
| \ for troubleshooting tips. </action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA host monitoring is disabled</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA host monitoring is disabled. No virtual machine failover will occur until\ | |
| \ Host Monitoring is re-enabled for cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA host monitoring is disabled. No virtual machine failover will occur until\ | |
| \ Host Monitoring is re-enabled</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.DasHostMonitoringDisabledEvent|vSphere\ | |
| \ HA host monitoring is disabled. No virtual machine failover will occur until\ | |
| \ Host Monitoring is re-enabled for cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.HA.DasHostMonitoringDisabledEvent">\ | |
| \ <description> This event is\ | |
| \ logged when host monitoring has been disabled in a vSphere HA cluster. \ | |
| \ </description> <cause> \ | |
| \ <description> Host monitoring\ | |
| \ is disabled, so vSphere HA will not perform any failover actions. This\ | |
| \ event is generated to inform the user that their\ | |
| \ cluster is temporarily not being protected against host or VM failures.\ | |
| \ If host or VM failures occur while host monitoring\ | |
| \ is disabled, HA will not attempt to restart the \ | |
| \ the VMs that were running on the failed hosts. Other vSphere HA features\ | |
| \ are not impacted by whether host monitoring is disabled.\ | |
| \ </description> <action>\ | |
| \ Enable host monitoring to resume hosts monitoring.\ | |
| \ </action> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA failed to restart a network isolated virtual machine</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA was unable to restart virtual machine {vm.name} in cluster {computeResource.name}\ | |
| \ after it was powered off in response to a network isolation event</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA was unable to restart virtual machine {vm.name} after it was powered\ | |
| \ off in response to a network isolation event</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA was unable to restart virtual machine {vm.name} after it was powered\ | |
| \ off in response to a network isolation event</formatOnHost><formatOnVm>vSphere\ | |
| \ HA was unable to restart this virtual machine after it was powered off in\ | |
| \ response to a network isolation event</formatOnVm><fullFormat>com.vmware.vc.HA.FailedRestartAfterIsolationEvent|vSphere\ | |
| \ HA was unable to restart virtual machine {vm.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} after it was powered off in response to\ | |
| \ a network isolation event. The virtual machine should be manually powered\ | |
| \ back on.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA changed a host's heartbeat datastores</description><category>info</category><formatOnDatacenter>Datastore\ | |
| \ {dsName} is {[email protected]}\ | |
| \ for storage heartbeating monitored by the vSphere HA agent on host {host.name}\ | |
| \ in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Datastore\ | |
| \ {dsName} is {[email protected]}\ | |
| \ for storage heartbeating monitored by the vSphere HA agent on host {host.name}</formatOnComputeResource><formatOnHost>Datastore\ | |
| \ {dsName} is {[email protected]}\ | |
| \ for storage heartbeating monitored by the vSphere HA agent on this host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HeartbeatDatastoreChanged|Datastore\ | |
| \ {dsName} is {[email protected]}\ | |
| \ for storage heartbeating monitored by the vSphere HA agent on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HeartbeatDatastoreSelected">\ | |
| \ <description> A datastore is selected or deselected\ | |
| \ for storage heartbeating monitored by the vSphere agent on this\ | |
| \ host. vSphere HA employs stroage heartbeating to detect host\ | |
| \ failures when there is network partition. </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA heartbeat datastore number for a host is insufficient</description><category>warning</category><formatOnDatacenter>The\ | |
| \ number of vSphere HA heartbeat datastores for host {host.name} in cluster\ | |
| \ {computeResource.name} is {selectedNum}, which is less than required: {requiredNum}</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ number of vSphere HA heartbeat datastores for host {host.name} is {selectedNum},\ | |
| \ which is less than required: {requiredNum}</formatOnComputeResource><formatOnHost>The\ | |
| \ number of vSphere HA heartbeat datastores for this host is {selectedNum},\ | |
| \ which is less than required: {requiredNum}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HeartbeatDatastoreNotSufficient|The\ | |
| \ number of vSphere HA heartbeat datastores for host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name} is {selectedNum}, which is less\ | |
| \ than required: {requiredNum}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.HeartbeatDatastoreNotSufficient"> \ | |
| \ <description> The number of heartbeat datastores used\ | |
| \ for this host is less than required. Multiple heartbeat datastores\ | |
| \ are needed to tolerate storage failures. The host summary page\ | |
| \ will report a configuration issue in this case. To ignore the\ | |
| \ configuration issue, use the vSphere HA cluster advanced option, das.ignoreInsufficientHbDatastore.\ | |
| \ </description> <cause> <description>\ | |
| \ The host does not have sufficient number of accessible datastores\ | |
| \ that are shared among other hosts in the cluster. \ | |
| \ </description> <action> \ | |
| \ Add more shared datastores to the host or check if any of its \ | |
| \ datastore is currently inaccessible. </action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA agent on a host has an error</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA agent for host {host.name} has an error in {computeResource.name}: {[email protected]}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent for host {host.name} has an error: {[email protected]}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent for this host has an error: {[email protected]}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostAgentErrorEvent|vSphere\ | |
| \ HA agent for host {host.name} has an error in {computeResource.name} in\ | |
| \ {datacenter.name}: {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.AgentErrorEvent">\ | |
| \ <description> This event is logged when the vSphere HA\ | |
| \ agent for the host has an error. </description> <action>\ | |
| \ See product documentation for troubleshooting tips. </action>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA agent is healthy</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on host {host.name} in cluster {computeResource.name} is healthy</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on host {host.name} is healthy</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent is healthy</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostDasAgentHealthyEvent|vSphere\ | |
| \ HA agent on host {host.name} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name} is healthy</fullFormat><longDescription> <EventLongDescription\ | |
| \ id=" com.vmware.vc.HA.HostDasAgentHealthyEvent"> \ | |
| \ <description> This event is logged when\ | |
| \ the vSphere HA agent on a host transitions to a healthy state. \ | |
| \ </description> <cause> \ | |
| \ <description> vSphere HA reports this\ | |
| \ event when the vSphere HA agent on the host is either a master or a slave\ | |
| \ that is connected to the master over the management\ | |
| \ network. </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA agent error</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on host {host.name} has an error: {reason.@enum.com.vmware.vc.HA.HostDasErrorEvent.HostDasErrorReason}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on host {host.name} has an error. {reason.@enum.com.vmware.vc.HA.HostDasErrorEvent.HostDasErrorReason}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent has an error: {[email protected]}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostDasErrorEvent|vSphere\ | |
| \ HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ has an error: {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostDasErrorEvent">\ | |
| \ <description> The vSphere HA agent on this host has\ | |
| \ an error. The event may provide details with extra information\ | |
| \ indicating the cause of the error. </description> <cause>\ | |
| \ <description>There was an error configuring the vSphere\ | |
| \ HA agent on the host</description> <action> \ | |
| \ Look at the task details for the configure vSphere HA task that\ | |
| \ failed. That will provide more details about why the failure\ | |
| \ occurred. Address the problem and reconfigure vSphere HA\ | |
| \ on the host. </action> </cause> <cause>\ | |
| \ <description> There was a timeout while communicating\ | |
| \ with the vSphere HA agent. This can occur if there is a high\ | |
| \ rate of operations being performed on virtual machines in\ | |
| \ the cluster resulting in the vSphere HA agents not being able to process\ | |
| \ the changes fast enough. </description> \ | |
| \ <action> Verify that this is a transient problem\ | |
| \ by stopping operations on virtual machines in the cluster\ | |
| \ for a few minutes to give time to the vSphere HA agents \ | |
| \ to process all their pending messages. If this resolves the problem, \ | |
| \ consider reducing the rate of operations performed on the cluster.\ | |
| \ </action> </cause> <cause> \ | |
| \ <description>There is vSphere HA agent is in a shutdown or\ | |
| \ failed state</description> <action>Reconfigure vSphere\ | |
| \ HA on the host. If this fails, reconfigure vSphere HA on the cluster</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Unsupported\ | |
| \ vSphere HA and vCloud Distributed Storage configuration</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA cannot be configured on host {host.name} in cluster {computeResource.name}\ | |
| \ because vCloud Distributed Storage is enabled but the host does not support\ | |
| \ that feature</formatOnDatacenter><formatOnComputeResource>vSphere HA cannot\ | |
| \ be configured on host {host.name} because vCloud Distributed Storage is\ | |
| \ enabled but the host does not support that feature</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA cannot be configured because vCloud Distributed Storage is enabled but\ | |
| \ the host does not support that feature</formatOnHost><formatOnVm>vSphere\ | |
| \ HA cannot be configured on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because vCloud Distributed Storage is enabled but the\ | |
| \ host does not support that feature</formatOnVm><fullFormat>com.vmware.vc.HA.HostDoesNotSupportVsan|vSphere\ | |
| \ HA cannot be configured on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because vCloud Distributed Storage is enabled but the\ | |
| \ host does not support that feature</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ has no vSphere HA isolation addresses</description><category>error</category><formatOnDatacenter>Host\ | |
| \ {host.name} in cluster {computeResource.name} has no isolation addresses\ | |
| \ defined as required by vSphere HA</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} has no isolation addresses defined as required by vSphere HA</formatOnComputeResource><formatOnHost>This\ | |
| \ host has no isolation addresses defined as required by vSphere HA</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostHasNoIsolationAddrsDefined|Host\ | |
| \ {host.name} in cluster {computeResource.name} in {datacenter.name} has no\ | |
| \ isolation addresses defined as required by vSphere HA.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostHasNoIsolationAddrsDefined">\ | |
| \ <description> The host has an vSphere HA configuration\ | |
| \ issue because there were no IP addresses that vSphere HA could\ | |
| \ use for detecting network isolation. Without at least one, the\ | |
| \ host will not take any isolation response. HA, by default, will\ | |
| \ use the host's default gateway (defined in the host's\ | |
| \ networking configuration), or use the addresses that were specified\ | |
| \ in the cluster's advanced settings. </description>\ | |
| \ <action> Define a default gateway in the host's\ | |
| \ networking configuration. </action> <action>\ | |
| \ If the cluster advanced setting das.usedefaultisolationaddress\ | |
| \ is false, you must define at least one isolation address using the advanced\ | |
| \ options. </action> <action> Define\ | |
| \ one or more cluster advanced options, each containing an IP address\ | |
| \ to be pinged by vSphere HA to detect if it is network-isolated when \ | |
| \ it no longer receives communication with other hosts in the cluster.\ | |
| \ The advanced option is das.isolationAddress[n], where 'n'\ | |
| \ is a number from 1 to 9. You may specify multiple addresses.\ | |
| \ </action> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA cannot be configured on this host because there are no mounted datastores.</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA cannot be configured on {host.name} in cluster {computeResource.name}\ | |
| \ because there are no mounted datastores.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA cannot be configured on {host.name} because there are no mounted datastores.</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA cannot be configured on this host because there are no mounted datastores.</formatOnHost><formatOnVm>vSphere\ | |
| \ HA cannot be configured on {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} because there are no mounted datastores.</formatOnVm><fullFormat>com.vmware.vc.HA.HostHasNoMountedDatastores|vSphere\ | |
| \ HA cannot be configured on {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} because there are no mounted datastores.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA requires a SSL Thumbprint for host</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA cannot be configured on host {host.name} in cluster {computeResource.name}\ | |
| \ because its SSL thumbprint has not been verified. Check that vCenter Server\ | |
| \ is configured to verify SSL thumbprints and that the thumbprint for {host.name}\ | |
| \ has been verified.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA cannot be configured on {host.name} because its SSL thumbprint has not\ | |
| \ been verified. Check that vCenter Server is configured to verify SSL thumbprints\ | |
| \ and that the thumbprint for {host.name} has been verified.</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA cannot be configured on this host because its SSL thumbprint has not\ | |
| \ been verified. Check that vCenter Server is configured to verify SSL thumbprints\ | |
| \ and that the thumbprint for this host has been verified.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostHasNoSslThumbprint|vSphere\ | |
| \ HA cannot be configured on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} because its SSL thumbprint has not been\ | |
| \ verified. Check that vCenter Server is configured to verify SSL thumbprints\ | |
| \ and that the thumbprint for {host.name} has been verified.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostHasNoSslThumbprint">\ | |
| \ <description> The host has an vSphere HA configuration\ | |
| \ issue because it does not have a verified ssl thumbprint. Hosts\ | |
| \ need verified SSL thumbprints for secure vSphere HA communications.\ | |
| \ </description> <action> If the\ | |
| \ host is using self-signed certificates, check that vCenter Server\ | |
| \ is configured to verify SSL certificates, and verify the thumbprints\ | |
| \ for the hosts in the vSphere HA cluster. </action> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ is incompatible with vSphere HA</description><category>error</category><formatOnDatacenter>The\ | |
| \ product version of host {host.name} in cluster {computeResource.name} is\ | |
| \ incompatible with vSphere HA.</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ product version of host {host.name} is incompatible with vSphere HA.</formatOnComputeResource><formatOnHost>The\ | |
| \ product version of this host is incompatible with vSphere HA.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostIncompatibleWithHA|The\ | |
| \ product version of host {host.name} in cluster {computeResource.name} in\ | |
| \ {datacenter.name} is incompatible with vSphere HA.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostIncompatibleWithHA">\ | |
| \ <description> The host is in a vSphere HA cluster\ | |
| \ but its product version is incompatible with HA. </description>\ | |
| \ <action> To fix the situation \ | |
| \ the host should either be moved out of the vSphere HA cluster or upgraded\ | |
| \ to a version supporting HA. </action> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA detected a network-partitioned host</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA detected that host {host.name} is in a different network partition than\ | |
| \ the master to which vCenter Server is connected in {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA detected that host {host.name} is in a different network partition than\ | |
| \ the master to which vCenter Server is connected</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected that this host is in a different network partition than the\ | |
| \ master to which vCenter Server is connected</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostPartitionedFromMasterEvent|vSphere\ | |
| \ HA detected that host {host.name} is in a different network partition than\ | |
| \ the master {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostPartitionedFromMasterEvent">\ | |
| \ <description> This event is logged when the host is in\ | |
| \ a different partition than the master. </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>The\ | |
| \ vSphere HA host availability state changed</description><category>info</category><formatOnDatacenter>The\ | |
| \ vSphere HA availability state of the host {host.name} in cluster {computeResource.name}\ | |
| \ has changed to {[email protected]}</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ vSphere HA availability state of the host {host.name} has changed to {[email protected]}</formatOnComputeResource><formatOnHost>The\ | |
| \ vSphere HA availability state of this host has changed to {[email protected]}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostStateChangedEvent|The\ | |
| \ vSphere HA availability state of the host {host.name} in cluster in {computeResource.name}\ | |
| \ in {datacenter.name} has changed to {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostStateChangedEvent">\ | |
| \ <description> This event is logged when the availability\ | |
| \ state of a host has changed. </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA agent unconfigure failed on host</description><category>warning</category><formatOnDatacenter>There\ | |
| \ was an error unconfiguring the vSphere HA agent on host {host.name} in cluster\ | |
| \ {computeResource.name}. To solve this problem, connect the host to a vCenter\ | |
| \ Server of version 5.0 or later.</formatOnDatacenter><formatOnComputeResource>There\ | |
| \ was an error unconfiguring the vSphere HA agent on host {host.name}. To\ | |
| \ solve this problem, connect the host to a vCenter Server of version 5.0\ | |
| \ or later.</formatOnComputeResource><formatOnHost>There was an error unconfiguring\ | |
| \ the vSphere HA agent on this host. To solve this problem, connect the host\ | |
| \ to a vCenter Server of version 5.0 or later.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostUnconfigureError|There\ | |
| \ was an error unconfiguring the vSphere HA agent on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}. To solve this problem, connect\ | |
| \ the host to a vCenter Server of version 5.0 or later.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostUnconfigureError">\ | |
| \ <description> There was an error unconfiguring\ | |
| \ the vSphere HA agent on this host. </description> \ | |
| \ <cause> <description> The\ | |
| \ vSphere HA unconfiguration task failed to send the updated hostList to vSphere\ | |
| \ HA agent on the host. This condition may interfere with\ | |
| \ the vSphere HA cluster to which the host used to belong and should be corrected.\ | |
| \ </description> <action> \ | |
| \ Add the host back to a vCenter Server of version 5.0 or later.\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ disconnected host has vSphere HA protected VMs</description><category>error</category><formatOnDatacenter>Host\ | |
| \ {host.name} in cluster {computeResource.name} is disconnected from vCenter\ | |
| \ Server, but contains {protectedVmCount} protected virtual machine(s)</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} is disconnected from vCenter Server, but contains {protectedVmCount}\ | |
| \ protected virtual machine(s)</formatOnComputeResource><formatOnHost>This\ | |
| \ host is disconnected from vCenter Server, but contains {protectedVmCount}\ | |
| \ vSphere HA protected virtual machine(s)</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.HostUnconfiguredWithProtectedVms|Host\ | |
| \ {host.name} in cluster {computeResource.name} in {datacenter.name} is disconnected\ | |
| \ from vCenter Server, but contains {protectedVmCount} protected virtual machine(s)</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.HostUnconfiguredWithProtectedVms">\ | |
| \ <description> This host is disconnected and contains\ | |
| \ one or more virtual machine(s) that are still protected by vSphere\ | |
| \ HA. Consequently, these virtual machines could be failed over\ | |
| \ to another host if this host should fail. </description>\ | |
| \ <cause> <description> \ | |
| \ If a vSphere HA-enabled host is disconnected and is unable\ | |
| \ to unprotect the virtual machines currently running on it \ | |
| \ (perhaps due to datastores being unavailable, or not being able to \ | |
| \ communicate with the vSphere HA master host) \ | |
| \ then these virtual machines would still be protected, but reside on \ | |
| \ the disconnected host. Also, if a virtual machine is migrated\ | |
| \ using vMotion to a vSphere HA-enabled host that is currently\ | |
| \ in the process of disconnecting, this can lead to the same\ | |
| \ result. </description> <action>\ | |
| \ To correct this situation, ensure that the host has access\ | |
| \ to the datastores used by these virtual machines, and then\ | |
| \ reconnect the host to an vSphere HA-enabled cluster. The\ | |
| \ virtual machines should become unprotected shortly after\ | |
| \ vSphere HA is configured on the host. </action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA detected an invalid master agent</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on host {remoteHostname} is an invalid master. The host should\ | |
| \ be examined to determine if it has been compromised.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on host {remoteHostname} is an invalid master. The host should\ | |
| \ be examined to determine if it has been compromised.</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.InvalidMaster|vSphere\ | |
| \ HA agent on host {remoteHostname} is an invalid master. The host should\ | |
| \ be examined to determine if it has been compromised.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.InvalidMaster">\ | |
| \ <description> A host in a vSphere HA cluster\ | |
| \ that is claiming to be a master has been determined to be invalid\ | |
| \ be another master host. This occurs when an existing master \ | |
| \ gets a message from another master in the same cluster. The existing\ | |
| \ master verifies that the other master is actually a valid master\ | |
| \ before it considers abdicating to the other master. An invalid\ | |
| \ master is an indication that there may be a compromised host\ | |
| \ on the network that is attempting to disrupt the HA cluster.\ | |
| \ The offending host should be examined to determine if it has been \ | |
| \ compromised. Its also possible a compromised host is impersonating\ | |
| \ a valid host so the reported host may not be the actual host\ | |
| \ that is compromised. </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA agent cannot reach some cluster management addresses</description><category>info</category><formatOnDatacenter>The\ | |
| \ vSphere HA agent on the host {host.name} in cluster {computeResource.name}\ | |
| \ cannot reach some of the management network addresses of other hosts, and\ | |
| \ thus HA may not be able to restart VMs if a host failure occurs: {unpingableAddrs}</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ vSphere HA agent on the host {host.name} cannot reach some of the management\ | |
| \ network addresses of other hosts, and thus HA may not be able to restart\ | |
| \ VMs if a host failure occurs: {unpingableAddrs}</formatOnComputeResource><formatOnHost>The\ | |
| \ vSphere HA agent on this host cannot reach some of the management network\ | |
| \ addresses of other hosts, and HA may not be able to restart VMs if a host\ | |
| \ failure occurs: {unpingableAddrs}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.NotAllHostAddrsPingable|The\ | |
| \ vSphere HA agent on the host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} cannot reach some of the management network addresses\ | |
| \ of other hosts, and thus HA may not be able to restart VMs if a host failure\ | |
| \ occurs: {unpingableAddrs}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.NotAllIsoAddrsPingable"> <description>\ | |
| \ The vSphere HA agent on host cannot reach some of the management\ | |
| \ network addresses of other hosts, and vSphere HA may not be able\ | |
| \ to restart VMs if a host failure occurs. </description>\ | |
| \ <cause> <description> \ | |
| \ There is a network issue preventing this host from communicating with\ | |
| \ some or all of the hosts in the cluster over their vSphere\ | |
| \ HA management networks. vSphere HA reliability ic currently\ | |
| \ compromised in the cluster and failover may not reliably\ | |
| \ occur if a host or hosts should fail during this condition.\ | |
| \ </description> <action> \ | |
| \ Determine and correct the source of the communication problem.\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA failed to start a Fault Tolerance secondary VM.</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA agent failed to start Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name} in cluster {computeResource.name}.\ | |
| \ Reason : {fault.msg}. vSphere HA agent will retry until it times out.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent failed to start Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name}. Reason : {fault.msg}.\ | |
| \ vSphere HA agent will retry until it times out.</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent failed to start Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name}. Reason : {fault.msg}.\ | |
| \ vSphere HA agent will retry until it times out.</formatOnHost><formatOnVm>vSphere\ | |
| \ HA agent failed to start Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost}. Reason : {fault.msg}. vSphere HA agent will retry\ | |
| \ until it times out.</formatOnVm><fullFormat>com.vmware.vc.HA.StartFTSecondaryFailedEvent|vSphere\ | |
| \ HA agent failed to start Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}. Reason : {fault.msg}. vSphere HA agent will retry\ | |
| \ until it times out.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.StartFTSecondaryFailedEvent"> \ | |
| \ <description> vSphere HA agent failed to start a Fault\ | |
| \ Tolerance secondary VM. vSphere HA will retry until either the\ | |
| \ operation succeeds or until the maximum number of restart attempts is reached.\ | |
| \ </description> </EventLongDescription> \ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA successfully started a Fault Tolerance secondary VM.</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA agent successfully started Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name} in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent successfully started Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name}.</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent successfully started Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name}.</formatOnHost><formatOnVm>vSphere\ | |
| \ HA agent successfully started Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost}.</formatOnVm><fullFormat>com.vmware.vc.HA.StartFTSecondarySucceededEvent|vSphere\ | |
| \ HA agent successfully started Fault Tolerance secondary VM {secondaryCfgPath}\ | |
| \ on host {secondaryHost} for primary VM {vm.name} in cluster {computeResource.name}.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.StartFTSecondarySucceededEvent">\ | |
| \ <description> vSphere HA agent successfully started\ | |
| \ a Fault Tolerance secondary virtual machine. </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA removed a datastore from preferred heartbeat datastores</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA removed datastore {dsName} from the set of preferred heartbeat datastores\ | |
| \ selected for cluster {computeResource.name} because the datastore is removed\ | |
| \ from inventory</formatOnDatacenter><formatOnComputeResource>vSphere HA removed\ | |
| \ datastore {dsName} from the set of preferred heartbeat datastores selected\ | |
| \ for cluster because the datastore is removed from inventory</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.UserHeartbeatDatastoreRemoved|vSphere\ | |
| \ HA removed datastore {dsName} from the set of preferred heartbeat datastores\ | |
| \ selected for cluster {computeResource.name} in {datacenter.name} because\ | |
| \ the datastore is removed from inventory</fullFormat><longDescription> \ | |
| \ <EventLongDescription id="com.vmware.vc.HA.UserHeartbeatDatastoreRemoved">\ | |
| \ <description> The datastore is removed from the\ | |
| \ set of preferred heartbeat datastores selected for this cluster. \ | |
| \ </description> <cause> <description>\ | |
| \ The datastore does not exist in the inventory. \ | |
| \ This happens when the datastore is removed from a host in the cluster\ | |
| \ manually or via a rescan. </description> \ | |
| \ <action> Choose a different datastore by reconfiguring\ | |
| \ the vSphere HA cluster. </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA did not perform an isolation response for vm because its VM restart priority\ | |
| \ is Disabled</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA did not perform an isolation response for {vm.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because its VM restart priority is Disabled</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA did not perform an isolation response for {vm.name} because its VM restart\ | |
| \ priority is Disabled</formatOnComputeResource><formatOnHost>vSphere HA did\ | |
| \ not perform an isolation response for {vm.name} because its VM restart priority\ | |
| \ is Disabled"</formatOnHost><formatOnVm>vSphere HA did not perform an\ | |
| \ isolation response because its VM restart priority is Disabled"</formatOnVm><fullFormat>com.vmware.vc.HA.VMIsHADisabledIsolationEvent|vSphere\ | |
| \ HA did not perform an isolation response for {vm.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because its VM restart priority is Disabled</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.HA.VMIsHADisabledIsolationEvent">\ | |
| \ <description> This event is\ | |
| \ logged when a host in a vSphere HA cluster was isolated and no isolation\ | |
| \ response was taken. </description> <cause>\ | |
| \ <description> The VM\ | |
| \ restart priority setting is set to disabled, so vSphere HA did not perform\ | |
| \ any action on this VM when the host became isolated. If the restart priority\ | |
| \ is disabled, HA will not attempt to restart the VM on another host, so HA\ | |
| \ will take no action for this VM on the isolated host. This event is informational\ | |
| \ only. </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA did not attempt to restart vm because its VM restart priority is Disabled</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA did not attempt to restart {vm.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because its VM restart priority is Disabled</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA did not attempt to restart {vm.name} because its VM restart priority\ | |
| \ is Disabled</formatOnComputeResource><formatOnHost>vSphere HA did not attempt\ | |
| \ to restart {vm.name} because its VM restart priority is Disabled"</formatOnHost><formatOnVm>vSphere\ | |
| \ HA did not attempt to restart vm because its VM restart priority is Disabled"</formatOnVm><fullFormat>com.vmware.vc.HA.VMIsHADisabledRestartEvent|vSphere\ | |
| \ HA did not attempt to restart {vm.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because its VM restart priority is Disabled</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.HA.VMIsHADisabledRestartEvent">\ | |
| \ <description> This event is\ | |
| \ logged when a failed VM in a vSphere HA cluster will not be restarted because\ | |
| \ its VM restart priority setting is set to disabled. </description>\ | |
| \ <cause> <description>\ | |
| \ The restart priority for the cluster or VM is disabled,\ | |
| \ so vSphere HA did not perform any action on this VM failed. This event\ | |
| \ is informational only. </description> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vCenter\ | |
| \ Server is unable to find a master vSphere HA agent</description><category>warning</category><formatOnDatacenter>vCenter\ | |
| \ Server is unable to find a master vSphere HA agent in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ Server is unable to find a master vSphere HA agent</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.VcCannotFindMasterEvent|vCenter\ | |
| \ Server is unable to find a master vSphere HA agent in cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.VcCannotFindMasterEvent"> <description>\ | |
| \ This event is logged when vCenter Server is unable to find a master\ | |
| \ vSphere HA agent. </description> <cause> \ | |
| \ <description> </description> <action>\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vCenter\ | |
| \ Server connected to a vSphere HA master agent</description><category>info</category><formatOnDatacenter>vCenter\ | |
| \ Server is connected to a master HA agent running on host {hostname} in {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ Server is connected to a master HA agent running on host {hostname}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.VcConnectedToMasterEvent|vCenter\ | |
| \ Server is connected to a master HA agent running on host {hostname} in {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.VcConnectedToMasterEvent"> <description>\ | |
| \ This event is logged when vCenter Server is connected with a master\ | |
| \ vSphere HA agent. </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vCenter\ | |
| \ Server disconnected from a master vSphere HA agent</description><category>warning</category><formatOnDatacenter>vCenter\ | |
| \ Server is disconnected from a master HA agent running on host {hostname}\ | |
| \ in {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ Server is disconnected from a master HA agent running on host {hostname}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.HA.VcDisconnectedFromMasterEvent|vCenter\ | |
| \ Server is disconnected from a master HA agent running on host {hostname}\ | |
| \ in {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.VcDisconnectedFromMasterEvent">\ | |
| \ <description> This event is logged when vCenter Server\ | |
| \ is disconnected from a master vSphere HA agent. </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine failed to become vSphere HA Protected</description><category>error</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} in cluster {computeResource.name} failed to become vSphere\ | |
| \ HA Protected and HA may not attempt to restart it after a failure.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} failed to become vSphere HA Protected and HA may not attempt\ | |
| \ to restart it after a failure.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} failed to become vSphere HA Protected and HA may not attempt\ | |
| \ to restart it after a failure.</formatOnHost><formatOnVm>This virtual machine\ | |
| \ failed to become vSphere HA Protected and HA may not attempt to restart\ | |
| \ it after a failure.</formatOnVm><fullFormat>com.vmware.vc.HA.VmNotProtectedEvent|Virtual\ | |
| \ machine {vm.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ failed to become vSphere HA Protected and HA may not attempt to restart\ | |
| \ it after a failure.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.VmNotProtectedEvent"> <description>\ | |
| \ The virtual machine successfully powered on in a vSphere HA\ | |
| \ cluster after a user-initiated power operation but the VM has\ | |
| \ not transitioned to vSphere HA Protected in the time period\ | |
| \ expected. This condition exists because the master vSphere\ | |
| \ HA agent has not yet persisted that the VM successfully powered on or \ | |
| \ vCenter is unaware that it did. Consequently, vSphere HA may\ | |
| \ not restart the VM after a failure. </description>\ | |
| \ <action> There are a number of reasons why\ | |
| \ a VM may remain not protected for a period of time. First,\ | |
| \ the system may be heavily loaded, in which case the transition will just\ | |
| \ take longer. Second, vCenter may be unable to communicate with\ | |
| \ the vSphere HA master agent. Examine the inventory to see if\ | |
| \ any hosts in the cluster are not responding. Third, the the\ | |
| \ management network may be partitioned, which is preventing the master that\ | |
| \ owns the VM from protecting it or reporting this information\ | |
| \ to vCenter. The cluster summary page may report a config issue\ | |
| \ in this case or hosts in the VM inventory will be reported as \ | |
| \ not responding. Finally, the vSphere HA master election is taking too\ | |
| \ long to complete. The cluster summary page will report if this\ | |
| \ situation exists. See the product documentation for additional\ | |
| \ troubleshooting tips. </action> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine is vSphere HA protected</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} in cluster {computeResource.name} is vSphere HA Protected\ | |
| \ and HA will attempt to restart it after a failure.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} is vSphere HA Protected and HA will attempt to restart\ | |
| \ it after a failure.</formatOnComputeResource><formatOnHost>Virtual machine\ | |
| \ {vm.name} is vSphere HA Protected and HA will attempt to restart it after\ | |
| \ a failure.</formatOnHost><formatOnVm>This virtual machine is vSphere HA\ | |
| \ Protected and HA will attempt to restart it after a failure.</formatOnVm><fullFormat>com.vmware.vc.HA.VmProtectedEvent|Virtual\ | |
| \ machine {vm.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ is vSphere HA Protected and HA will attempt to restart it after a failure.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.HA.VmProtectedEvent">\ | |
| \ <description> The virtual machine successfully\ | |
| \ powered on in a vSphere HA cluster after a user-initiated power\ | |
| \ operation and vSphere HA has persisted this fact. Consequently, \ | |
| \ vSphere HA will attempt to restart the VM after a failure. \ | |
| \ </description> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine is not vSphere HA Protected</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} in cluster {computeResource.name} is not vSphere HA Protected.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} is not vSphere HA Protected.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} is not vSphere HA Protected.</formatOnHost><formatOnVm>This\ | |
| \ virtual machine is not vSphere HA Protected.</formatOnVm><fullFormat>com.vmware.vc.HA.VmUnprotectedEvent|Virtual\ | |
| \ machine {vm.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ is not vSphere HA Protected.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.HA.VmUnprotectedEvent"> <description>\ | |
| \ The virtual machine transitioned from the vSphere HA protected\ | |
| \ to unprotected state. This transition is a result of a user\ | |
| \ powering off the virtual machine, disabling vSphere HA, disconnecting\ | |
| \ the host on which the virtual machine is running, or destroying\ | |
| \ the cluster in which the virtual machine resides. </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA has unprotected out-of-disk-space VM</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA has unprotected virtual machine {vm.name} in cluster {computeResource.name}\ | |
| \ because it ran out of disk space</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA has unprotected virtual machine {vm.name} because it ran out of disk\ | |
| \ space</formatOnComputeResource><formatOnHost>vSphere HA has unprotected\ | |
| \ virtual machine {vm.name} because it ran out of disk space</formatOnHost><formatOnVm>vSphere\ | |
| \ HA has unprotected this virtual machine because it ran out of disk space</formatOnVm><fullFormat>com.vmware.vc.HA.VmUnprotectedOnDiskSpaceFull|vSphere\ | |
| \ HA has unprotected virtual machine {vm.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} because it ran out of disk space</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>vCenter\ | |
| \ Service Overall Health Changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.VCHealthStateChangedEvent|vCenter\ | |
| \ Service overall health changed from '{oldState}' to '{newState}'</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.VCHealthStateChangedEvent">\ | |
| \ <description> This event is logged when the overall health\ | |
| \ of vCenter Service has changed or become unavailable. </description>\ | |
| \ <cause> <description> The vCenter Service\ | |
| \ overall health state has changed or become unavailable </description>\ | |
| \ <action> Examine the vCenter Service health state\ | |
| \ and make sure the VimWebServices service is up and running on the vCenter\ | |
| \ Server </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>Cannot\ | |
| \ complete virtual machine clone.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.VmCloneFailedInvalidDestinationEvent|Cannot\ | |
| \ clone {vm.name} as {destVmName} to invalid or non-existent destination with\ | |
| \ ID {invalidMoRef}: {fault}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Cannot\ | |
| \ complete virtual machine clone.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.VmCloneToResourcePoolFailedEvent|Cannot\ | |
| \ clone {vm.name} as {destVmName} to resource pool {destResourcePool}: {fault}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine disks consolidation succeeded.</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} disks consolidatation succeeded on {host.name} in cluster\ | |
| \ {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} disks consolidation succeeded on {host.name}.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} disks consolidation succeeded.</formatOnHost><formatOnVm>Virtual\ | |
| \ machine disks consolidation succeeded.</formatOnVm><fullFormat>com.vmware.vc.VmDiskConsolidatedEvent|Virtual\ | |
| \ machine {vm.name} disks consolidated successfully on {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine disks consolidation needed.</description><category>warning</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} disks consolidatation is needed on {host.name} in cluster\ | |
| \ {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} disks consolidation is needed on {host.name}.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} disks consolidation is needed.</formatOnHost><formatOnVm>Virtual\ | |
| \ machine disks consolidation is needed.</formatOnVm><fullFormat>com.vmware.vc.VmDiskConsolidationNeeded|Virtual\ | |
| \ machine {vm.name} disks consolidation is needed on {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine disks consolidation no longer needed.</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} disks consolidatation is no longer needed on {host.name}\ | |
| \ in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} disks consolidation is no longer needed on {host.name}.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} disks consolidation is no longer needed.</formatOnHost><formatOnVm>Virtual\ | |
| \ machine disks consolidation is no longer needed.</formatOnVm><fullFormat>com.vmware.vc.VmDiskConsolidationNoLongerNeeded|Virtual\ | |
| \ machine {vm.name} disks consolidation is no longer needed on {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine disks consolidation failed.</description><category>warning</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} disks consolidation failed on {host.name} in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} disks consolidation failed on {host.name}.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} disks consolidation failed.</formatOnHost><formatOnVm>Virtual\ | |
| \ machine disks consolidation failed.</formatOnVm><fullFormat>com.vmware.vc.VmDiskFailedToConsolidateEvent|Virtual\ | |
| \ machine {vm.name} disks consolidation failed on {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Hardware\ | |
| \ Health Changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.cim.CIMGroupHealthStateChanged|Health\ | |
| \ of {data.group} changed from {data.oldState} to {data.newState}. {data.cause}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id=" com.vmware.vc.cim.CIMGroupHealthStateChanged">\ | |
| \ <description> This event is\ | |
| \ logged when the health of a group of hardware sensors changes for a particular\ | |
| \ host. </description> <cause> \ | |
| \ <description> \ | |
| \ A change in the health of a group of hardware sensors. \ | |
| \ </description> <action> \ | |
| \ Check the "Hardware Status" tab for the\ | |
| \ host and find the group in question. Examine\ | |
| \ any sensors in the Warning or Alert states and examine the relevant hardware.\ | |
| \ </action> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to update VM files</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Failed\ | |
| \ to update VM files on datastore {ds.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.datastore.UpdateVmFilesFailedEvent|Failed\ | |
| \ to update VM files on datastore {ds.name} using host {hostName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Updated\ | |
| \ VM files</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Updated\ | |
| \ VM files on datastore {ds.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.datastore.UpdatedVmFilesEvent|Updated\ | |
| \ VM files on datastore {ds.name} using host {hostName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Updating\ | |
| \ VM Files</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Updating\ | |
| \ VM files on datastore {ds.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.datastore.UpdatingVmFilesEvent|Updating\ | |
| \ VM files on datastore {ds.name} using host {hostName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Link\ | |
| \ Aggregation Control Protocol configuration is inconsistent</description><category>info</category><formatOnDatacenter>Single\ | |
| \ Link Aggregation Control Group is enabled on Uplink Port Groups while enhanced\ | |
| \ LACP support is enabled.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.dvs.LacpConfigInconsistentEvent|Single\ | |
| \ Link Aggregation Control Group is enabled on Uplink Port Groups while enhanced\ | |
| \ LACP support is enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Fault\ | |
| \ Tolerance VM restart disabled</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA has been disabled in cluster {computeResource.name}. vSphere HA will\ | |
| \ not restart VM {vm.name} or its Secondary VM after a failure.</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA has been disabled. vSphere HA will not restart VM {vm.name} or its Secondary\ | |
| \ VM after a failure.</formatOnComputeResource><formatOnHost>vSphere HA has\ | |
| \ been disabled. vSphere HA will not restart VM {vm.name} or its Secondary\ | |
| \ VM after a failure.</formatOnHost><formatOnVm>vSphere HA has been disabled.\ | |
| \ vSphere HA will not restart this VM or its Secondary VM after a failure.</formatOnVm><fullFormat>com.vmware.vc.ft.VmAffectedByDasDisabledEvent|vSphere\ | |
| \ HA has been disabled in cluster {computeResource.name} of datacenter {datacenter.name}.\ | |
| \ vSphere HA will not restart VM {vm.name} or its Secondary VM after a failure.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.ft.VmAffectedByDasDisabledEvent">\ | |
| \ <description> When vSphere HA is disabled in a cluster,\ | |
| \ you cannot restart a Primary VM or its Secondary VM after a failure.\ | |
| \ This event is issued when vSphere HA is disabled and a Fault Tolerant virtual\ | |
| \ machine is powered on. The event alerts you of the risk to the\ | |
| \ Fault Tolerant virtual machine that results \t from disabling vSphere HA.\ | |
| \ </description> <cause> <description>vSphere\ | |
| \ HA was disabled when a Fault Tolerant virtual machine was powered on</description>\ | |
| \ <action>Re-enable vSphere HA</action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>Guest\ | |
| \ operation</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Guest\ | |
| \ operation {[email protected]} performed.</formatOnVm><fullFormat>com.vmware.vc.guestOperations.GuestOperation|Guest\ | |
| \ operation {[email protected]} performed on Virtual\ | |
| \ machine {vm.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Guest\ | |
| \ operation authentication failure</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Guest\ | |
| \ operation authentication failed for operation {[email protected]}.</formatOnVm><fullFormat>com.vmware.vc.guestOperations.GuestOperationAuthFailure|Guest\ | |
| \ operation authentication failed for operation {[email protected]}\ | |
| \ on Virtual machine {vm.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA restarted a virtual machine</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA restarted virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA restarted virtual machine {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA restarted virtual machine {vm.name}</formatOnHost><formatOnVm>vSphere\ | |
| \ HA restarted this virtual machine</formatOnVm><fullFormat>com.vmware.vc.ha.VmRestartedByHAEvent|vSphere\ | |
| \ HA restarted virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.ha.VmRestartedByHAEvent">\ | |
| \ <description> The virtual machine\ | |
| \ was restarted automatically by vSphere HA on this\ | |
| \ host. This response may be triggered by a failure \ | |
| \ of the host the virtual machine was originally running on \ | |
| \ or by an unclean power-off of the virtual machine (eg. if the\ | |
| \ vmx process was killed). </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Autostart\ | |
| \ rules reconfigure failed</description><category>error</category><formatOnDatacenter>Reconfiguring\ | |
| \ autostart rules for virtual machines on host {host.name} failed</formatOnDatacenter><formatOnComputeResource>Reconfiguring\ | |
| \ autostart rules for virtual machines on host {host.name} failed</formatOnComputeResource><formatOnHost>Reconfiguring\ | |
| \ autostart rules for virtual machines on this host failed</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.AutoStartReconfigureFailedEvent|Reconfiguring\ | |
| \ autostart rules for virtual machines on {host.name} in datacenter {datacenter.name}\ | |
| \ failed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host's\ | |
| \ virtual flash resource is accessible.</description><category>info</category><formatOnDatacenter>Host's\ | |
| \ virtual flash resource is restored to be accessible.</formatOnDatacenter><formatOnComputeResource>Host's\ | |
| \ virtual flash resource is restored to be accessible.</formatOnComputeResource><formatOnHost>Host's\ | |
| \ virtual flash resource is restored to be accessible.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.clear.vFlashResource.inaccessible|Host's\ | |
| \ virtual flash resource is restored to be accessible.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Host's\ | |
| \ virtual flash resource usage dropped below the threshold.</description><category>info</category><formatOnDatacenter>Host's\ | |
| \ virtual flash resource usage dropped below {1}%.</formatOnDatacenter><formatOnComputeResource>Host's\ | |
| \ virtual flash resource usage dropped below {1}%.</formatOnComputeResource><formatOnHost>Host's\ | |
| \ virtual flash resource usage dropped below {1}%.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.clear.vFlashResource.reachthreshold|Host's\ | |
| \ virtual flash resource usage dropped below {1}%.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host's\ | |
| \ virtual flash resource is inaccessible.</description><category>warning</category><formatOnDatacenter>Host's\ | |
| \ virtual flash resource is inaccessible.</formatOnDatacenter><formatOnComputeResource>Host's\ | |
| \ virtual flash resource is inaccessible.</formatOnComputeResource><formatOnHost>Host's\ | |
| \ virtual flash resource is inaccessible.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.problem.vFlashResource.inaccessible|Host's\ | |
| \ virtual flash resource is inaccessible.</fullFormat><longDescription> \ | |
| \ <EventLongDescription id="com.vmware.vc.host.problem.vFlashResource.inaccessible">\ | |
| \ <description> Inaccessible host virtual flash resource\ | |
| \ indicates that its backing VFFS volume is inaccessible. Due to inaccessible\ | |
| \ host virtual flash resource, virtual machines with virtual Flash Read Cache\ | |
| \ configured cannot be powered on or might experience unpredicted\ | |
| \ behavior if powered on. </description> <cause> \ | |
| \ <description> This might be caused by an unmounted VFFS\ | |
| \ volume or an APD/PDL on the VFFS volume. </description> \ | |
| \ <action> Check the backing VFFS volume connection status.\ | |
| \ For example, mount the unmounted volume or resolve the APD/PDL\ | |
| \ issues. The host virtual flash resource is accessible as long as the \ | |
| \ backing VFFS volume is accessible. </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>Host's\ | |
| \ virtual flash resource usage exceeds the threshold.</description><category>warning</category><formatOnDatacenter>Host's\ | |
| \ virtual flash resource usage is more than {1}%.</formatOnDatacenter><formatOnComputeResource>Host's\ | |
| \ virtual flash resource usage is more than {1}%.</formatOnComputeResource><formatOnHost>Host's\ | |
| \ virtual flash resource usage is more than {1}%.</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.problem.vFlashResource.reachthreshold|Host's\ | |
| \ virtual flash resource usage is more than {1}%.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ flash resource capacity is extended</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ flash resource capacity is extended</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ flash resource capacity is extended</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ flash resource capacity is extended</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.vFlash.VFlashResourceCapacityExtendedEvent|Virtual\ | |
| \ flash resource capacity is extended</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ flash resource is configured on the host</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ flash resource is configured on the host</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ flash resource is configured on the host</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ flash resource is configured on the host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.vFlash.VFlashResourceConfiguredEvent|Virtual\ | |
| \ flash resource is configured on the host</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ flash resource is removed from the host</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ flash resource is removed from the host</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ flash resource is removed from the host</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ flash resource is removed from the host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.vFlash.VFlashResourceRemovedEvent|Virtual\ | |
| \ flash resource is removed from the host</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Default\ | |
| \ virtual flash module is changed to {vFlashModule} on the host</description><category>info</category><formatOnDatacenter>Default\ | |
| \ virtual flash module is changed to {vFlashModule} on the host</formatOnDatacenter><formatOnComputeResource>Default\ | |
| \ virtual flash module is changed to {vFlashModule} on the host</formatOnComputeResource><formatOnHost>Default\ | |
| \ virtual flash module is changed to {vFlashModule} on the host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.vFlash.defaultModuleChangedEvent|Any\ | |
| \ new virtual Flash Read Cache configuration request will use {vFlashModule}\ | |
| \ as default virtual flash module. All existing virtual Flash Read Cache configurations\ | |
| \ remain unchanged.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.host.vFlash.defaultModuleChangedEvent"> \ | |
| \ <description> The default virtual flash module has been\ | |
| \ changed. Any new virtual Flash Read Cache configuration uses the\ | |
| \ new default virtual flash module if undefined in configuration. \ | |
| \ All existing configurations will remain unchanged. </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ flash modules are loaded or reloaded on the host</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ flash modules are loaded or reloaded on the host</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ flash modules are loaded or reloaded on the host</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ flash modules are loaded or reloaded on the host</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.host.vFlash.modulesLoadedEvent|Virtual\ | |
| \ flash modules are loaded or reloaded on the host</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ NIC entered passthrough mode</description><category>info</category><formatOnDatacenter>Network\ | |
| \ passthrough is active on adapter {deviceLabel} of virtual machine {vm.name}\ | |
| \ on host {host.name}</formatOnDatacenter><formatOnComputeResource>Network\ | |
| \ passthrough is active on adapter {deviceLabel} of virtual machine {vm.name}\ | |
| \ on host {host.name}</formatOnComputeResource><formatOnHost>Network passthrough\ | |
| \ is active on adapter {deviceLabel} of virtual machine {vm.name}</formatOnHost><formatOnVm>Network\ | |
| \ passthrough is active on adapter {deviceLabel}</formatOnVm><fullFormat>com.vmware.vc.npt.VmAdapterEnteredPassthroughEvent|Network\ | |
| \ passthrough is active on adapter {deviceLabel} of virtual machine {vm.name}\ | |
| \ on host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ NIC exited passthrough mode</description><category>info</category><formatOnDatacenter>Network\ | |
| \ passthrough is inactive on adapter {deviceLabel} of virtual machine {vm.name}\ | |
| \ on host {host.name}</formatOnDatacenter><formatOnComputeResource>Network\ | |
| \ passthrough is inactive on adapter {deviceLabel} of virtual machine {vm.name}\ | |
| \ on host {host.name}</formatOnComputeResource><formatOnHost>Network passthrough\ | |
| \ is inactive on adapter {deviceLabel} of virtual machine {vm.name}</formatOnHost><formatOnVm>Network\ | |
| \ passthrough is inactive on adapter {deviceLabel}</formatOnVm><fullFormat>com.vmware.vc.npt.VmAdapterExitedPassthroughEvent|Network\ | |
| \ passthrough is inactive on adapter {deviceLabel} of virtual machine {vm.name}\ | |
| \ on host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to clone state for entity on extension</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Failed\ | |
| \ to clone state on extension {extensionName}</formatOnVm><fullFormat>com.vmware.vc.ovfconsumers.CloneOvfConsumerStateErrorEvent|Failed\ | |
| \ to clone state for the entity '{entityName}' on extension {extensionName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to retrieve OVF environment sections for VM on extension</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Failed\ | |
| \ to retrieve OVF environment sections from extension {extensionName}</formatOnVm><fullFormat>com.vmware.vc.ovfconsumers.GetOvfEnvironmentSectionsErrorEvent|Failed\ | |
| \ to retrieve OVF environment sections for VM '{vm.name}' from extension\ | |
| \ {extensionName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Unable\ | |
| \ to power on VM after cloning</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Powering\ | |
| \ on after cloning was blocked by an extension. Message: {description}</formatOnVm><fullFormat>com.vmware.vc.ovfconsumers.PowerOnAfterCloneErrorEvent|Powering\ | |
| \ on VM '{vm.name}' after cloning was blocked by an extension. Message:\ | |
| \ {description}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to register entity on extension</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.ovfconsumers.RegisterEntityErrorEvent|Failed\ | |
| \ to register entity '{entityName}' on extension {extensionName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to unregister entities on extension</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.ovfconsumers.UnregisterEntitiesErrorEvent|Failed\ | |
| \ to unregister entities on extension {extensionName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to validate OVF descriptor on extension</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.ovfconsumers.ValidateOstErrorEvent|Failed\ | |
| \ to validate OVF descriptor on extension {extensionName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Answer\ | |
| \ file exported</description><category>info</category><formatOnDatacenter>Answer\ | |
| \ file for host {host.name} has been exported</formatOnDatacenter><formatOnComputeResource>Answer\ | |
| \ file for host {host.name} has been exported</formatOnComputeResource><formatOnHost>Answer\ | |
| \ file exported</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.profile.AnswerFileExportedEvent|Answer\ | |
| \ file for host {host.name} in datacenter {datacenter.name} has been exported</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ customization settings updated</description><category>info</category><formatOnDatacenter>Host\ | |
| \ customization settings for host {host.name} has been updated</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ customization settings for host {host.name} has been updated</formatOnComputeResource><formatOnHost>Host\ | |
| \ customization settings updated</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.profile.AnswerFileUpdatedEvent|Host\ | |
| \ customization settings for host {host.name} in datacenter {datacenter.name}\ | |
| \ has been updated</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Resource\ | |
| \ pool renamed</description><category>info</category><formatOnDatacenter>Resource\ | |
| \ pool '{oldName}' has been renamed to '{newName}'</formatOnDatacenter><formatOnComputeResource>Resource\ | |
| \ pool '{oldName}' has been renamed to '{newName}'</formatOnComputeResource><formatOnHost>Resource\ | |
| \ pool '{oldName}' has been renamed to '{newName}'</formatOnHost><formatOnVm>Resource\ | |
| \ pool '{oldName}' has been renamed to '{newName}'</formatOnVm><fullFormat>com.vmware.vc.rp.ResourcePoolRenamedEvent|Resource\ | |
| \ pool '{oldName}' has been renamed to '{newName}'</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Datastore\ | |
| \ maintenance mode operation canceled</description><category>info</category><formatOnDatacenter>The\ | |
| \ datastore maintenance mode operation has been canceled</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ datastore maintenance mode operation has been canceled</formatOnComputeResource><formatOnHost>The\ | |
| \ datastore maintenance mode operation has been canceled</formatOnHost><formatOnVm>The\ | |
| \ datastore maintenance mode operation has been canceled</formatOnVm><fullFormat>com.vmware.vc.sdrs.CanceledDatastoreMaintenanceModeEvent|The\ | |
| \ datastore maintenance mode operation has been canceled</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Configured\ | |
| \ storage DRS</description><category>info</category><formatOnDatacenter>Configured\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnDatacenter><formatOnComputeResource>Configured\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnComputeResource><formatOnHost>Configured\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnHost><formatOnVm>Configured\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnVm><fullFormat>com.vmware.vc.sdrs.ConfiguredStorageDrsOnPodEvent|Configured\ | |
| \ storage DRS on datastore cluster {objectName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Datastore\ | |
| \ cluster has datastores that belong to different SRM Consistency Groups</description><category>warning</category><formatOnDatacenter>Datastore\ | |
| \ cluster {objectName} has datastores that belong to different SRM Consistency\ | |
| \ Groups</formatOnDatacenter><formatOnComputeResource>Datastore cluster {objectName}\ | |
| \ has datastores that belong to different SRM Consistency Groups</formatOnComputeResource><formatOnHost>Datastore\ | |
| \ cluster {objectName} has datastores that belong to different SRM Consistency\ | |
| \ Groups</formatOnHost><formatOnVm>Datastore cluster {objectName} has datastores\ | |
| \ that belong to different SRM Consistency Groups</formatOnVm><fullFormat>com.vmware.vc.sdrs.ConsistencyGroupViolationEvent|Datastore\ | |
| \ cluster {objectName} has datastores that belong to different SRM Consistency\ | |
| \ Groups</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Datastore\ | |
| \ entered maintenance mode</description><category>info</category><formatOnDatacenter>Datastore\ | |
| \ {ds.name} has entered maintenance mode</formatOnDatacenter><formatOnComputeResource>Datastore\ | |
| \ {ds.name} has entered maintenance mode</formatOnComputeResource><formatOnHost>Datastore\ | |
| \ {ds.name} has entered maintenance mode</formatOnHost><formatOnVm>Datastore\ | |
| \ {ds.name} has entered maintenance mode</formatOnVm><fullFormat>com.vmware.vc.sdrs.DatastoreEnteredMaintenanceModeEvent|Datastore\ | |
| \ {ds.name} has entered maintenance mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Datastore\ | |
| \ is entering maintenance mode</description><category>info</category><formatOnDatacenter>Datastore\ | |
| \ {ds.name} is entering maintenance mode</formatOnDatacenter><formatOnComputeResource>Datastore\ | |
| \ {ds.name} is entering maintenance mode</formatOnComputeResource><formatOnHost>Datastore\ | |
| \ {ds.name} is entering maintenance mode</formatOnHost><formatOnVm>Datastore\ | |
| \ {ds.name} is entering maintenance mode</formatOnVm><fullFormat>com.vmware.vc.sdrs.DatastoreEnteringMaintenanceModeEvent|Datastore\ | |
| \ {ds.name} is entering maintenance mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Datastore\ | |
| \ exited maintenance mode</description><category>info</category><formatOnDatacenter>Datastore\ | |
| \ {ds.name} has exited maintenance mode</formatOnDatacenter><formatOnComputeResource>Datastore\ | |
| \ {ds.name} has exited maintenance mode</formatOnComputeResource><formatOnHost>Datastore\ | |
| \ {ds.name} has exited maintenance mode</formatOnHost><formatOnVm>Datastore\ | |
| \ {ds.name} has exited maintenance mode</formatOnVm><fullFormat>com.vmware.vc.sdrs.DatastoreExitedMaintenanceModeEvent|Datastore\ | |
| \ {ds.name} has exited maintenance mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Datastore\ | |
| \ cluster has datastores shared across multiple datacenters</description><category>warning</category><formatOnDatacenter>Datastore\ | |
| \ cluster {objectName} has one or more datastores {datastore} shared across\ | |
| \ multiple datacenters</formatOnDatacenter><formatOnComputeResource>Datastore\ | |
| \ cluster {objectName} has one or more datastores {datastore} shared across\ | |
| \ multiple datacenters</formatOnComputeResource><formatOnHost>Datastore cluster\ | |
| \ {objectName} has one or more datastores {datastore} shared across multiple\ | |
| \ datacenters</formatOnHost><formatOnVm>Datastore cluster {objectName} has\ | |
| \ one or more datastores {datastore} shared across multiple datacenters</formatOnVm><fullFormat>com.vmware.vc.sdrs.DatastoreInMultipleDatacentersEvent|Datastore\ | |
| \ cluster {objectName} has one or more datastores {datastore} shared across\ | |
| \ multiple datacenters</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Errors\ | |
| \ encountered while datastore entering into maintenance mode</description><category>error</category><formatOnDatacenter>Datastore\ | |
| \ {ds.name} encountered errors while entering maintenance mode</formatOnDatacenter><formatOnComputeResource>Datastore\ | |
| \ {ds.name} encountered errors while entering maintenance mode</formatOnComputeResource><formatOnHost>Datastore\ | |
| \ {ds.name} encountered errors while entering maintenance mode</formatOnHost><formatOnVm>Datastore\ | |
| \ {ds.name} encountered errors while entering maintenance mode</formatOnVm><fullFormat>com.vmware.vc.sdrs.DatastoreMaintenanceModeErrorsEvent|Datastore\ | |
| \ {ds.name} encountered errors while entering maintenance mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Storage\ | |
| \ DRS disabled</description><category>info</category><formatOnDatacenter>Disabled\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnDatacenter><formatOnComputeResource>Disabled\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnComputeResource><formatOnHost>Disabled\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnHost><formatOnVm>Disabled\ | |
| \ storage DRS on datastore cluster {objectName}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsDisabledEvent|Disabled\ | |
| \ storage DRS on datastore cluster {objectName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Storage\ | |
| \ DRS enabled</description><category>info</category><formatOnDatacenter>Enabled\ | |
| \ storage DRS on datastore cluster {objectName} with automation level {[email protected]}</formatOnDatacenter><formatOnComputeResource>Enabled\ | |
| \ storage DRS on datastore cluster {objectName} with automation level {[email protected]}</formatOnComputeResource><formatOnHost>Enabled\ | |
| \ storage DRS on datastore cluster {objectName} with automation level {[email protected]}</formatOnHost><formatOnVm>Enabled\ | |
| \ storage DRS on datastore cluster {objectName} with automation level {[email protected]}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsEnabledEvent|Enabled\ | |
| \ storage DRS on datastore cluster {objectName} with automation level {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Storage\ | |
| \ DRS invocation failed</description><category>error</category><formatOnDatacenter>Storage\ | |
| \ DRS invocation failed on datastore cluster {objectName}</formatOnDatacenter><formatOnComputeResource>Storage\ | |
| \ DRS invocation failed on datastore cluster {objectName}</formatOnComputeResource><formatOnHost>Storage\ | |
| \ DRS invocation failed on datastore cluster {objectName}</formatOnHost><formatOnVm>Storage\ | |
| \ DRS invocation failed on datastore cluster {objectName}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsInvocationFailedEvent|Storage\ | |
| \ DRS invocation failed on datastore cluster {objectName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>New\ | |
| \ storage DRS recommendation generated</description><category>info</category><formatOnDatacenter>A\ | |
| \ new storage DRS recommendation has been generated on datastore cluster {objectName}</formatOnDatacenter><formatOnComputeResource>A\ | |
| \ new storage DRS recommendation has been generated on datastore cluster {objectName}</formatOnComputeResource><formatOnHost>A\ | |
| \ new storage DRS recommendation has been generated on datastore cluster {objectName}</formatOnHost><formatOnVm>A\ | |
| \ new storage DRS recommendation has been generated on datastore cluster {objectName}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsNewRecommendationPendingEvent|A\ | |
| \ new storage DRS recommendation has been generated on datastore cluster {objectName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Datastore\ | |
| \ cluster connected to host(s) that do not support storage DRS</description><category>warning</category><formatOnDatacenter>Datastore\ | |
| \ cluster {objectName} is connected to one or more hosts {host} that do not\ | |
| \ support storage DRS</formatOnDatacenter><formatOnComputeResource>Datastore\ | |
| \ cluster {objectName} is connected to one or more hosts {host} that do not\ | |
| \ support storage DRS</formatOnComputeResource><formatOnHost>Datastore cluster\ | |
| \ {objectName} is connected to one or more hosts {host} that do not support\ | |
| \ storage DRS</formatOnHost><formatOnVm>Datastore cluster {objectName} is\ | |
| \ connected to one or more hosts {host} that do not support storage DRS</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsNotSupportedHostConnectedToPodEvent|Datastore\ | |
| \ cluster {objectName} is connected to one or more hosts {host} that do not\ | |
| \ support storage DRS</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Pending\ | |
| \ storage recommendations were applied</description><category>info</category><formatOnDatacenter>All\ | |
| \ pending recommendations on datastore cluster {objectName} were applied</formatOnDatacenter><formatOnComputeResource>All\ | |
| \ pending recommendations on datastore cluster {objectName} were applied</formatOnComputeResource><formatOnHost>All\ | |
| \ pending recommendations on datastore cluster {objectName} were applied</formatOnHost><formatOnVm>All\ | |
| \ pending recommendations on datastore cluster {objectName} were applied</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsRecommendationApplied|All\ | |
| \ pending recommendations on datastore cluster {objectName} were applied</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Storage\ | |
| \ DRS migrated VM disks</description><category>info</category><formatOnDatacenter>Storage\ | |
| \ DRS migrated disks of VM {vm.name} to datastore {ds.name}</formatOnDatacenter><formatOnComputeResource>Storage\ | |
| \ DRS migrated disks of VM {vm.name} to datastore {ds.name}</formatOnComputeResource><formatOnHost>Storage\ | |
| \ DRS migrated disks of VM {vm.name} to datastore {ds.name}</formatOnHost><formatOnVm>Storage\ | |
| \ DRS migrated disks of VM {vm.name} to datastore {ds.name}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsStorageMigrationEvent|Storage\ | |
| \ DRS migrated disks of VM {vm.name} to datastore {ds.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Storage\ | |
| \ DRS placed VM disks</description><category>info</category><formatOnDatacenter>Storage\ | |
| \ DRS placed disks of VM {vm.name} on datastore {ds.name}</formatOnDatacenter><formatOnComputeResource>Storage\ | |
| \ DRS placed disks of VM {vm.name} on datastore {ds.name}</formatOnComputeResource><formatOnHost>Storage\ | |
| \ DRS placed disks of VM {vm.name} on datastore {ds.name}</formatOnHost><formatOnVm>Storage\ | |
| \ DRS placed disks of VM {vm.name} on datastore {ds.name}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StorageDrsStoragePlacementEvent|Storage\ | |
| \ DRS placed disks of VM {vm.name} on datastore {ds.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Datastore\ | |
| \ cluster created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ datastore cluster {objectName}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ datastore cluster {objectName}</formatOnComputeResource><formatOnHost>Created\ | |
| \ datastore cluster {objectName}</formatOnHost><formatOnVm>Created datastore\ | |
| \ cluster {objectName}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StoragePodCreatedEvent|Created\ | |
| \ datastore cluster {objectName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Datastore\ | |
| \ cluster deleted</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ datastore cluster {objectName}</formatOnDatacenter><formatOnComputeResource>Removed\ | |
| \ datastore cluster {objectName}</formatOnComputeResource><formatOnHost>Removed\ | |
| \ datastore cluster {objectName}</formatOnHost><formatOnVm>Removed datastore\ | |
| \ cluster {objectName}</formatOnVm><fullFormat>com.vmware.vc.sdrs.StoragePodDestroyedEvent|Removed\ | |
| \ datastore cluster {objectName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>SIOC:\ | |
| \ pre-4.1 host {host} connected to SIOC-enabled datastore {objectName}</description><category>warning</category><formatOnDatacenter>SIOC\ | |
| \ has detected that a host: {host} connected to a SIOC-enabled datastore:\ | |
| \ {objectName} is running an older version of ESX that does not support SIOC.\ | |
| \ This is an unsupported configuration.</formatOnDatacenter><formatOnComputeResource>SIOC\ | |
| \ has detected that a host: {host} connected to a SIOC-enabled datastore:\ | |
| \ {objectName} is running an older version of ESX that does not support SIOC.\ | |
| \ This is an unsupported configuration.</formatOnComputeResource><formatOnHost>SIOC\ | |
| \ has detected that a host: {host} connected to a SIOC-enabled datastore:\ | |
| \ {objectName} is running an older version of ESX that does not support SIOC.\ | |
| \ This is an unsupported configuration.</formatOnHost><formatOnVm>SIOC has\ | |
| \ detected that a host: {host} connected to a SIOC-enabled datastore: {objectName}\ | |
| \ is running an older version of ESX that does not support SIOC. This is an\ | |
| \ unsupported configuration.</formatOnVm><fullFormat>com.vmware.vc.sioc.NotSupportedHostConnectedToDatastoreEvent|SIOC\ | |
| \ has detected that a host: {host} connected to a SIOC-enabled datastore:\ | |
| \ {objectName} is running an older version of ESX that does not support SIOC.\ | |
| \ This is an unsupported configuration.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>System\ | |
| \ capability warning from storage provider</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.LunCapabilityInitEvent|Storage\ | |
| \ provider [{providerName}] : system capability warning for {eventSubjectId}\ | |
| \ : {msgTxt}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>System\ | |
| \ capability normal event from storage provider</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.LunCapabilityMetEvent|Storage\ | |
| \ provider [{providerName}] : system capability normal for {eventSubjectId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>System\ | |
| \ capability alert from storage provider</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.LunCapabilityNotMetEvent|Storage\ | |
| \ provider [{providerName}] : system capability alert for {eventSubjectId}\ | |
| \ : {msgTxt}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Thin\ | |
| \ provisioning capacity threshold normal event from storage provider</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.ThinProvisionedLunThresholdClearedEvent|Storage\ | |
| \ provider [{providerName}] : thin provisioning capacity threshold normal\ | |
| \ for {eventSubjectId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Thin\ | |
| \ provisioning capacity threshold alert from storage provider</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.ThinProvisionedLunThresholdCrossedEvent|Storage\ | |
| \ provider [{providerName}] : thin provisioning capacity threshold alert for\ | |
| \ {eventSubjectId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Thin\ | |
| \ provisioning capacity threshold warning from storage provider</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.ThinProvisionedLunThresholdInitEvent|Storage\ | |
| \ provider [{providerName}] : thin provisioning capacity threshold warning\ | |
| \ for {eventSubjectId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ disk bound to a policy profile is compliant backing object based storage.</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ disk {diskKey} on {vmName} connected to {datastore.name} is compliant from\ | |
| \ storage provider {providerName}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.datastore.ComplianceStatusCompliantEvent|Virtual\ | |
| \ disk {diskKey} on {vmName} connected to datastore {datastore.name} in {datacenter.name}\ | |
| \ is compliant from storage provider {providerName}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ disk bound to a policy profile is non compliant backing object based storage.</description><category>error</category><formatOnDatacenter>Virtual\ | |
| \ disk {diskKey} on {vmName} connected to {datastore.name} is not compliant\ | |
| \ {operationalStatus] from storage provider {providerName}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.datastore.ComplianceStatusNonCompliantEvent|Virtual\ | |
| \ disk {diskKey} on {vmName} connected to {datastore.name} in {datacenter.name}\ | |
| \ is not compliant {operationalStatus] from storage provider {providerName}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ disk bound to a policy profile is unknown compliance status backing object\ | |
| \ based storage.</description><category>warning</category><formatOnDatacenter>Virtual\ | |
| \ disk {diskKey} on {vmName} connected to {datastore.name} compliance status\ | |
| \ is unknown from storage provider {providerName}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.datastore.ComplianceStatusUnknownEvent|Virtual\ | |
| \ disk {diskKey} on {vmName} connected to {datastore.name} in {datacenter.name}\ | |
| \ compliance status is unknown from storage provider {providerName}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Health\ | |
| \ event from storage provider</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.provider.health.event|Storage\ | |
| \ provider [{providerName}] : health event for {eventSubjectId} : {msgTxt}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>System\ | |
| \ event from storage provider</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.sms.provider.system.event|Storage\ | |
| \ provider [{providerName}] : system event : {msgTxt}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ disk bound to a policy profile is complianct backing object based storage.</description><category>info</category><formatOnDatacenter>Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} in {computeResource.name} is\ | |
| \ compliant from storage provider {providerName}.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} is compliant from storage provider\ | |
| \ {providerName}.</formatOnComputeResource><formatOnHost>Virtual disk {diskKey}\ | |
| \ on {vm.name} is compliant from storage provider {providerName}.</formatOnHost><formatOnVm>Virtual\ | |
| \ disk {diskKey} on {vm.name} is compliant from storage provider {providerName}.</formatOnVm><fullFormat>com.vmware.vc.sms.vm.ComplianceStatusCompliantEvent|Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} and {computeResource.name} in\ | |
| \ {datacenter.name} is compliant from storage provider {providerName}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ disk bound to a policy profile is non complianct backing object based storage.</description><category>error</category><formatOnDatacenter>Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} in {computeResource.name} is\ | |
| \ not compliant {operationalStatus] from storage provider {providerName}.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} is not compliant {operationalStatus]\ | |
| \ from storage provider {providerName}.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ disk {diskKey} on {vm.name} is not compliant {operationalStatus] from storage\ | |
| \ provider {providerName}.</formatOnHost><formatOnVm>Virtual disk {diskKey}\ | |
| \ on {vm.name} is not compliant {operationalStatus] from storage provider\ | |
| \ {providerName}.</formatOnVm><fullFormat>com.vmware.vc.sms.vm.ComplianceStatusNonCompliantEvent|Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} and {computeResource.name} in\ | |
| \ {datacenter.name} is not compliant {operationalStatus] from storage provider\ | |
| \ {providerName}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ disk bound to a policy profile is unknown compliance status backing object\ | |
| \ based storage.</description><category>warning</category><formatOnDatacenter>Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} in {computeResource.name} compliance\ | |
| \ status is unknown from storage provider {providerName}.</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} compliance status is unknown\ | |
| \ from storage provider {providerName}.</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ disk {diskKey} on {vm.name} compliance status is unknown from storage provider\ | |
| \ {providerName}.</formatOnHost><formatOnVm>Virtual disk {diskKey} on {vm.name}\ | |
| \ compliance status is unknown from storage provider {providerName}.</formatOnVm><fullFormat>com.vmware.vc.sms.vm.ComplianceStatusUnknownEvent|Virtual\ | |
| \ disk {diskKey} on {vm.name} on {host.name} and {computeResource.name} in\ | |
| \ {datacenter.name} compliance status is unknown from storage provider {providerName}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Profile\ | |
| \ association/dissociation failed</description><category>error</category><formatOnDatacenter>Profile\ | |
| \ association/dissociation failed for {entityName}</formatOnDatacenter><formatOnComputeResource>Profile\ | |
| \ association/dissociation failed for {entityName}</formatOnComputeResource><formatOnHost>Profile\ | |
| \ association/dissociation failed for {entityName}</formatOnHost><formatOnVm>Profile\ | |
| \ association/dissociation failed for {entityName}</formatOnVm><fullFormat>com.vmware.vc.spbm.ProfileAssociationFailedEvent|Profile\ | |
| \ association/dissociation failed for {entityName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Quick\ | |
| \ stats is not up-to-date</description><category>info</category><formatOnDatacenter>Quick\ | |
| \ stats on {host.name} in {computeResource.name} is not up-to-date</formatOnDatacenter><formatOnComputeResource>Quick\ | |
| \ stats on {host.name} is not up-to-date</formatOnComputeResource><formatOnHost>Quick\ | |
| \ stats on {host.name} is not up-to-date</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.stats.HostQuickStatesNotUpToDateEvent|Quick\ | |
| \ stats on {host.name} in {computeResource.name} in {datacenter.name} is not\ | |
| \ up-to-date</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="com.vmware.vc.stats.HostQuickStatesNotUpToDateEvent"> \ | |
| \ <description> "Quick stats on the host is\ | |
| \ not up-to-date. </description> <cause>\ | |
| \ <description> Quickstats on the host\ | |
| \ are not up-to-date. This is expected if the host was recently \ | |
| \ added or reconnected or VC just started up. </description>\ | |
| \ <action> No specific action needs to\ | |
| \ be taken. </action> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>FT\ | |
| \ Disabled VM protected as non-FT VM</description><category>info</category><formatOnDatacenter>HA\ | |
| \ VM Component Protection protects virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} as non-FT virtual machine because the\ | |
| \ FT state is disabled</formatOnDatacenter><formatOnComputeResource>HA VM\ | |
| \ Component Protection protects virtual machine {vm.name} on host {host.name}\ | |
| \ as non-FT virtual machine because the FT state is disabled</formatOnComputeResource><formatOnHost>HA\ | |
| \ VM Component Protection protects virtual machine {vm.name} as non-FT virtual\ | |
| \ machine because the FT state is disabled</formatOnHost><formatOnVm>HA VM\ | |
| \ Component Protection will protect this virtul machine as non-FT virtual\ | |
| \ machine because the FT state is disabled</formatOnVm><fullFormat>com.vmware.vc.vcp.FtDisabledVmTreatAsNonFtEvent|HA\ | |
| \ VM Component Protection protects virtual machine {vm.name} on {host.name}\ | |
| \ in cluster {computeResource.name} in datacenter {datacenter.name} as non-FT\ | |
| \ virtual machine because the FT state is disabled</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Failover\ | |
| \ FT VM due to component failure</description><category>info</category><formatOnDatacenter>FT\ | |
| \ Primary VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ is going to fail over to Secondary VM due to component failure</formatOnDatacenter><formatOnComputeResource>FT\ | |
| \ Primary VM {vm.name} on host {host.name} is going to fail over to Secondary\ | |
| \ VM due to component failure</formatOnComputeResource><formatOnHost>FT Primary\ | |
| \ VM {vm.name} is going to fail over to Secondary VM due to component failure</formatOnHost><formatOnVm>FT\ | |
| \ Primary VM is going to fail over to Secondary VM due to component failure</formatOnVm><fullFormat>com.vmware.vc.vcp.FtFailoverEvent|FT\ | |
| \ Primary VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} is going to fail over to Secondary VM due\ | |
| \ to component failure </fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>FT\ | |
| \ VM failover failed</description><category>error</category><formatOnDatacenter>FT\ | |
| \ virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ failed to failover to secondary</formatOnDatacenter><formatOnComputeResource>FT\ | |
| \ virtual machine {vm.name} on host {host.name} failed to failover to secondary</formatOnComputeResource><formatOnHost>FT\ | |
| \ virtual machine {vm.name} failed to failover to secondary</formatOnHost><formatOnVm>FT\ | |
| \ virtual machine failed to failover to secondary</formatOnVm><fullFormat>com.vmware.vc.vcp.FtFailoverFailedEvent|FT\ | |
| \ virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} failed to failover to secondary</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Restarting\ | |
| \ FT secondary due to component failure</description><category>info</category><formatOnDatacenter>HA\ | |
| \ VM Component Protection is restarting FT secondary virtual machine {vm.name}\ | |
| \ on host {host.name} in cluster {computeResource.name} due to component failure</formatOnDatacenter><formatOnComputeResource>HA\ | |
| \ VM Component Protection is restarting FT secondary virtual machine {vm.name}\ | |
| \ on host {host.name} due to component failure</formatOnComputeResource><formatOnHost>HA\ | |
| \ VM Component Protection is restarting FT secondary virtual machine {vm.name}\ | |
| \ due to component failure</formatOnHost><formatOnVm>HA VM Component Protection\ | |
| \ is restarting FT secondary virtual machine due to component failure</formatOnVm><fullFormat>com.vmware.vc.vcp.FtSecondaryRestartEvent|HA\ | |
| \ VM Component Protection is restarting FT secondary virtual machine {vm.name}\ | |
| \ on host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name}\ | |
| \ due to component failure</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>FT\ | |
| \ secondary VM restart failed</description><category>error</category><formatOnDatacenter>FT\ | |
| \ Secondary VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ failed to restart</formatOnDatacenter><formatOnComputeResource>FT Secondary\ | |
| \ VM {vm.name} on host {host.name} failed to restart</formatOnComputeResource><formatOnHost>FT\ | |
| \ Secondary VM {vm.name} failed to restart</formatOnHost><formatOnVm>FT Secondary\ | |
| \ VM failed to restart</formatOnVm><fullFormat>com.vmware.vc.vcp.FtSecondaryRestartFailedEvent|FT\ | |
| \ Secondary VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} failed to restart</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Need\ | |
| \ secondary VM protected as non-FT VM</description><category>info</category><formatOnDatacenter>HA\ | |
| \ VM Component Protection protects virtual machine {vm.name} on host {host.name}\ | |
| \ as non-FT virtual machine since it has been in the needSecondary state too\ | |
| \ long</formatOnDatacenter><formatOnComputeResource>HA VM Component Protection\ | |
| \ protects virtual machine {vm.name} on host {host.name} as non-FT virtual\ | |
| \ machine because it has been in the needSecondary state too long</formatOnComputeResource><formatOnHost>HA\ | |
| \ VM Component Protection protects virtual machine {vm.name} as non-FT virtual\ | |
| \ machine because it has been in the needSecondary state too long</formatOnHost><formatOnVm>HA\ | |
| \ VM Component Protection protects this virtul machine as non-FT virtual machine\ | |
| \ because it has been in the needSecondary state too long</formatOnVm><fullFormat>com.vmware.vc.vcp.NeedSecondaryFtVmTreatAsNonFtEvent|HA\ | |
| \ VM Component Protection protects virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in datacenter {datacenter.name} as non-FT\ | |
| \ virtual machine because it has been in the needSecondary state too long</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VM\ | |
| \ Component Protection test ends</description><category>info</category><formatOnDatacenter>VM\ | |
| \ Component Protection test ends on host {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>VM\ | |
| \ Component Protection test ends on host {host.name}</formatOnComputeResource><formatOnHost>VM\ | |
| \ Component Protection test ends</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vcp.TestEndEvent|VM\ | |
| \ Component Protection test ends on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VM\ | |
| \ Component Protection test starts</description><category>info</category><formatOnDatacenter>VM\ | |
| \ Component Protection test starts on host {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>VM\ | |
| \ Component Protection test starts on host {host.name}</formatOnComputeResource><formatOnHost>VM\ | |
| \ Component Protection test starts</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vcp.TestStartEvent|VM\ | |
| \ Component Protection test starts on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>No\ | |
| \ action on VM</description><category>info</category><formatOnDatacenter>HA\ | |
| \ VM Component Protection did not take action on virtual machine {vm.name}\ | |
| \ on host {host.name} in cluster {computeResource.name} due to the feature\ | |
| \ configuration setting</formatOnDatacenter><formatOnComputeResource>HA VM\ | |
| \ Component Protection did not take action on virtual machine {vm.name} on\ | |
| \ host {host.name} due to the feature configuration setting</formatOnComputeResource><formatOnHost>HA\ | |
| \ VM Component Protection did not take action on virtual machine {vm.name}\ | |
| \ due to the feature configuration setting</formatOnHost><formatOnVm>HA VM\ | |
| \ Component Protection did not take action due to the feature configuration\ | |
| \ setting</formatOnVm><fullFormat>com.vmware.vc.vcp.VcpNoActionEvent|HA VM\ | |
| \ Component Protection did not take action on virtual machine {vm.name} on\ | |
| \ host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name}\ | |
| \ due to the feature configuration setting</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ machine lost datastore access</description><category>error</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ lost access to {datastore}</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} on host {host.name} lost access to {datastore}</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} lost access to {datastore}</formatOnHost><formatOnVm>Virtual\ | |
| \ machine lost access to {datastore}</formatOnVm><fullFormat>com.vmware.vc.vcp.VmDatastoreFailedEvent|Virtual\ | |
| \ machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} lost access to {datastore}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ machine lost VM network accessibility</description><category>error</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ lost access to {network}</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} on host {host.name} lost access to {network}</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} lost access to {network}</formatOnHost><formatOnVm>Virtual\ | |
| \ machine lost access to {network}</formatOnVm><fullFormat>com.vmware.vc.vcp.VmNetworkFailedEvent|Virtual\ | |
| \ machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} lost access to {network}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VM\ | |
| \ power off hang</description><category>error</category><formatOnDatacenter>HA\ | |
| \ VM Component Protection could not power off virtual machine {vm.name} on\ | |
| \ host {host.name} in cluster {computeResource.name} successfully after trying\ | |
| \ {numTimes} times and will keep trying</formatOnDatacenter><formatOnComputeResource>HA\ | |
| \ VM Component Protection could not power off virtual machine {vm.name} on\ | |
| \ host {host.name} successfully after trying {numTimes} times and will keep\ | |
| \ trying</formatOnComputeResource><formatOnHost>HA VM Component Protection\ | |
| \ could not power off virtual machine {vm.name} successfully after trying\ | |
| \ {numTimes} times and will keep trying</formatOnHost><formatOnVm>HA VM Component\ | |
| \ Protection could not power off virtual machine successfully after trying\ | |
| \ {numTimes} times and will keep trying</formatOnVm><fullFormat>com.vmware.vc.vcp.VmPowerOffHangEvent|HA\ | |
| \ VM Component Protection could not power off virtual machine {vm.name} on\ | |
| \ host {host.name} in cluster {computeResource.name} in datacenter {datacenter.name}\ | |
| \ successfully after trying {numTimes} times and will keep trying</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Restarting\ | |
| \ VM due to component failure</description><category>info</category><formatOnDatacenter>HA\ | |
| \ VM Component Protection is restarting virtual machine {vm.name} due to component\ | |
| \ failure on host {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>HA\ | |
| \ VM Component Protection is restarting virtual machine {vm.name} due to component\ | |
| \ failure on host {host.name}</formatOnComputeResource><formatOnHost>HA VM\ | |
| \ Component Protection is restarting virtual machine {vm.name} due to component\ | |
| \ failure</formatOnHost><formatOnVm>HA VM Component Protection is restarting\ | |
| \ virtual machine due to component failure</formatOnVm><fullFormat>com.vmware.vc.vcp.VmRestartEvent|HA\ | |
| \ VM Component Protection is restarting virtual machine {vm.name} due to component\ | |
| \ failure on host {host.name} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine affected by component failure failed to restart</description><category>error</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} affected by component failure on host {host.name} in cluster\ | |
| \ {computeResource.name} failed to restart</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} affected by component failure on host {host.name} failed\ | |
| \ to restart</formatOnComputeResource><formatOnHost>Virtual machine {vm.name}\ | |
| \ affected by component failure failed to restart</formatOnHost><formatOnVm>Virtual\ | |
| \ machine affected by component failure failed to restart</formatOnVm><fullFormat>com.vmware.vc.vcp.VmRestartFailedEvent|Virtual\ | |
| \ machine {vm.name} affected by component failure on host {host.name} in cluster\ | |
| \ {computeResource.name} in datacenter {datacenter.name} failed to restart</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>No\ | |
| \ candidate host to restart</description><category>error</category><formatOnDatacenter>HA\ | |
| \ VM Component Protection could not find a destination host for virtual machine\ | |
| \ {vm.name} on host {host.name} in cluster {computeResource.name} after waiting\ | |
| \ {numSecWait} seconds and will keep trying</formatOnDatacenter><formatOnComputeResource>HA\ | |
| \ VM Component Protection could not find a destination host for virtual machine\ | |
| \ {vm.name} on host {host.name} after waiting {numSecWait} seconds and will\ | |
| \ keep trying</formatOnComputeResource><formatOnHost>HA VM Component Protection\ | |
| \ could not find a destination host for virtual machine {vm.name} after waiting\ | |
| \ {numSecWait} seconds and will keep trying</formatOnHost><formatOnVm>HA VM\ | |
| \ Component Protection could not find a destination host for this virtual\ | |
| \ machine after waiting {numSecWait} seconds and will keep trying</formatOnVm><fullFormat>com.vmware.vc.vcp.VmWaitForCandidateHostEvent|HA\ | |
| \ VM Component Protection could not find a destination host for virtual machine\ | |
| \ {vm.name} on host {host.name} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name} after waiting {numSecWait} seconds and will keep trying</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Operation\ | |
| \ on the SSD device failed</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Configuration\ | |
| \ on disk {disk.path} failed. Reason : {fault.msg}</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vflash.SsdConfigurationFailedEvent|Configuration\ | |
| \ on disk {disk.path} failed. Reason : {fault.msg}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Virtual\ | |
| \ machine register failed</description><category>error</category><formatOnDatacenter>Virtual\ | |
| \ machine {vm.name} registration on host {host.name} failed</formatOnDatacenter><formatOnComputeResource>Virtual\ | |
| \ machine {vm.name} registration on host {host.name} failed</formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} registration on this host failed</formatOnHost><formatOnVm>Virtual\ | |
| \ machine registration failed</formatOnVm><fullFormat>com.vmware.vc.vm.VmRegisterFailedEvent|Virtual\ | |
| \ machine {vm.name} registration on {host.name} in datacenter {datacenter.name}\ | |
| \ failed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to revert the virtual machine state to a snapshot</description><category>error</category><formatOnDatacenter>Failed\ | |
| \ to revert the execution state of the virtual machine {vm.name} on host {host.name},\ | |
| \ in compute resource {computeResource.name} to snapshot {snapshotName}, with\ | |
| \ ID {snapshotId}</formatOnDatacenter><formatOnComputeResource>Failed to revert\ | |
| \ the execution state of the virtual machine {vm.name} on host {host.name}\ | |
| \ to snapshot {snapshotName}, with ID {snapshotId}</formatOnComputeResource><formatOnHost>Failed\ | |
| \ to revert the execution state of the virtual machine {vm.name} to snapshot\ | |
| \ {snapshotName}, with ID {snapshotId}</formatOnHost><formatOnVm>Failed to\ | |
| \ revert the execution state of the virtual machine to snapshot {snapshotName},\ | |
| \ with ID {snapshotId}</formatOnVm><fullFormat>com.vmware.vc.vm.VmStateFailedToRevertToSnapshot|Failed\ | |
| \ to revert the execution state of the virtual machine {vm.name} on host {host.name},\ | |
| \ in compute resource {computeResource.name} to snapshot {snapshotName}, with\ | |
| \ ID {snapshotId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>The\ | |
| \ virtual machine state has been reverted to a snapshot</description><category>info</category><formatOnDatacenter>The\ | |
| \ execution state of the virtual machine {vm.name} on host {host.name}, in\ | |
| \ compute resource {computeResource.name} has been reverted to the state of\ | |
| \ snapshot {snapshotName}, with ID {snapshotId}</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ execution state of the virtual machine {vm.name} on host {host.name} has\ | |
| \ been reverted to the state of snapshot {snapshotName}, with ID {snapshotId}</formatOnComputeResource><formatOnHost>The\ | |
| \ execution state of the virtual machine {vm.name} has been reverted to the\ | |
| \ state of snapshot {snapshotName}, with ID {snapshotId}</formatOnHost><formatOnVm>The\ | |
| \ execution state of the virtual machine has been reverted to the state of\ | |
| \ snapshot {snapshotName}, with ID {snapshotId}</formatOnVm><fullFormat>com.vmware.vc.vm.VmStateRevertedToSnapshot|The\ | |
| \ execution state of the virtual machine {vm.name} on host {host.name}, in\ | |
| \ compute resource {computeResource.name} has been reverted to the state of\ | |
| \ snapshot {snapshotName}, with ID {snapshotId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Application\ | |
| \ Monitoring Is Not Supported</description><category>warning</category><formatOnDatacenter>Application\ | |
| \ monitoring is not supported on {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Application\ | |
| \ monitoring is not supported on {host.name}</formatOnComputeResource><formatOnHost>Application\ | |
| \ monitoring is not supported</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vmam.AppMonitoringNotSupported|Application\ | |
| \ monitoring is not supported on {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA detected application heartbeat status change</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA detected that the application heartbeat status changed to {[email protected]}\ | |
| \ for {vm.name} on {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA detected that the application heartbeat status changed to {[email protected]}\ | |
| \ for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected that the application heartbeat status changed to {[email protected]}\ | |
| \ for {vm.name}</formatOnHost><formatOnVm>vSphere HA detected that the application\ | |
| \ heartbeat status changed to {[email protected]}\ | |
| \ for this virtual machine</formatOnVm><fullFormat>com.vmware.vc.vmam.VmAppHealthMonitoringStateChangedEvent|vSphere\ | |
| \ HA detected that the application heartbeat status changed to {[email protected]}\ | |
| \ for {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.vmam.VmAppHealthMonitoringStateChangedEvent">\ | |
| \ <description> Application monitoring state changes indicate\ | |
| \ a change in the health of the application being monitored or in\ | |
| \ the application-monitoring process. A transition from gray to green\ | |
| \ occurs when application heartbeat is being enabled from within the\ | |
| \ guest. A transition to red occurs after vSphere HA didn't receive\ | |
| \ any heartbeats within 30 seconds. A transition from red to green\ | |
| \ occurs if heartbeats begin again before vSphere HA can react. A\ | |
| \ transition to gray occurs after application heartbeating is disabled from\ | |
| \ within the guest. </description> <cause>\ | |
| \ <description> Either the user initiated action from\ | |
| \ inside the guest or vSphere HA did not receive application heartbeats\ | |
| \ from the application-monitoring agent within a 30-second interval. \ | |
| \ </description> <action> If the state transitions\ | |
| \ to red, investigate why the application-monitoring agent stopped \ | |
| \ heartbeating. Missing heartbeats may be a result of the application failing\ | |
| \ or a problem with the application-monitoring agent. Frequent state\ | |
| \ transitions to or from gray may indicate a problem with the application-monitoring\ | |
| \ agent. If they occur, investigate whether the enabling/disabling\ | |
| \ of monitoring is expected. </action> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>vSphere\ | |
| \ HA detected application state change</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA detected that the application state changed to {[email protected]}\ | |
| \ for {vm.name} on {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA detected that the application state changed to {[email protected]}\ | |
| \ for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected that the application state changed to {[email protected]}\ | |
| \ for {vm.name}</formatOnHost><formatOnVm>vSphere HA detected that the application\ | |
| \ state changed to {[email protected]} for this virtual\ | |
| \ machine</formatOnVm><fullFormat>com.vmware.vc.vmam.VmAppHealthStateChangedEvent|vSphere\ | |
| \ HA detected that the application state changed to {[email protected]}\ | |
| \ for {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.vmam.VmAppHealthStateChangedEvent">\ | |
| \ <description> Application state changes indicate that\ | |
| \ an in-guest application has posted one of the two allowed values\ | |
| \ - appStateOk or appStateNeedReset. The former indicates that the\ | |
| \ monitored application is fine, the latter causes an immediate reset\ | |
| \ if Application Monitoring is enabled for this virtual machine. </description>\ | |
| \ <cause> <description> This is an in-guest\ | |
| \ initated action. </description> <action> \ | |
| \ If vSphere HA and Application Monitoring are enabled for this virtual\ | |
| \ machine, it is reset if the state is appStateNeedReset. If the\ | |
| \ virtual machine is being migrated using vMotion the reset will be delayed\ | |
| \ until the virtual machine has reached its destination. Also, the\ | |
| \ reset will be delayed until the datastore connectivity issues are resolved.\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vSphere\ | |
| \ HA detected application heartbeat failure</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA detected application heartbeat failure for {vm.name} on {host.name} in\ | |
| \ cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA detected application heartbeat failure for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA detected application heartbeat failure for {vm.name}</formatOnHost><formatOnVm>vSphere\ | |
| \ HA detected application heartbeat failure for this virtual machine</formatOnVm><fullFormat>com.vmware.vc.vmam.VmDasAppHeartbeatFailedEvent|vSphere\ | |
| \ HA detected application heartbeat failure for {vm.name} on {host.name} in\ | |
| \ cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.vmam.VmDasAppHeartbeatFailedEvent">\ | |
| \ <description> vSphere HA has detected a heartbeat failure\ | |
| \ from the application-monitoring agent inside the guest. If application\ | |
| \ monitoring is enabled in vSphere the virtual machine will be reset. \ | |
| \ </description> <cause> <description> \ | |
| \ vSphere HA did not receive application heartbeats from the application-monitoring\ | |
| \ agent within a 30-second interval. </description>\ | |
| \ <action> Investigate why the application-monitoring\ | |
| \ agent stopped heartbeating. Missing heartbeats may be a result\ | |
| \ of the application failing or a problem with the application-monitoring\ | |
| \ agent. </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>EventEx</key><description>VSAN\ | |
| \ datastore {datastoreName} does not have capacity</description><category>error</category><formatOnDatacenter>VSAN\ | |
| \ datastore {datastoreName} in cluster {computeResource.name} does not have\ | |
| \ capacity</formatOnDatacenter><formatOnComputeResource>VSAN datastore {datastoreName}\ | |
| \ does not have capacity</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>VSAN\ | |
| \ datastore {datastoreName} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name} does not have capacity</formatOnVm><fullFormat>com.vmware.vc.vsan.DatastoreNoCapacityEvent|VSAN\ | |
| \ datastore {datastoreName} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name} does not have capacity</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Host\ | |
| \ cannot communicate with all other nodes in the VSAN enabled cluster</description><category>error</category><formatOnDatacenter>Host\ | |
| \ {host.name} in cluster {computeResource.name} cannot communicate with all\ | |
| \ other nodes in the VSAN enabled cluster</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} cannot communicate with all other nodes in the VSAN enabled\ | |
| \ cluster</formatOnComputeResource><formatOnHost>Host cannot communicate with\ | |
| \ all other nodes in the VSAN enabled cluster</formatOnHost><formatOnVm>Host\ | |
| \ {host.name} in cluster {computeResource.name} in datacenter {datacenter.name}\ | |
| \ cannot communicate with all other nodes in the VSAN enabled cluster</formatOnVm><fullFormat>com.vmware.vc.vsan.HostCommunicationErrorEvent|Host\ | |
| \ {host.name} in cluster {computeResource.name} in datacenter {datacenter.name}\ | |
| \ cannot communicate with all other nodes in the VSAN enabled cluster</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ with the VSAN service enabled is not in the vCenter cluster</description><category>error</category><formatOnDatacenter>{host.name}\ | |
| \ with the VSAN service enabled is not in the vCenter cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>{host.name}\ | |
| \ with the VSAN service enabled is not in the vCenter cluster</formatOnComputeResource><formatOnHost>Host\ | |
| \ with the VSAN service enabled is not in the vCenter cluster</formatOnHost><formatOnVm>{host.name}\ | |
| \ with the VSAN service enabled is not in the vCenter cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</formatOnVm><fullFormat>com.vmware.vc.vsan.HostNotInClusterEvent|{host.name}\ | |
| \ with the VSAN service enabled is not in the vCenter cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ is in a VSAN enabled cluster but does not have VSAN service enabled</description><category>error</category><formatOnDatacenter>{host.name}\ | |
| \ is in a VSAN enabled cluster {computeResource.name} but does not have VSAN\ | |
| \ service enabled</formatOnDatacenter><formatOnComputeResource>{host.name}\ | |
| \ is in a VSAN enabled cluster but does not have VSAN service enabled</formatOnComputeResource><formatOnHost>Host\ | |
| \ is in a VSAN enabled cluster but does not have VSAN service enabled</formatOnHost><formatOnVm>{host.name}\ | |
| \ is in a VSAN enabled cluster {computeResource.name} in datacenter {datacenter.name}\ | |
| \ but does not have VSAN service enabled</formatOnVm><fullFormat>com.vmware.vc.vsan.HostNotInVsanClusterEvent|{host.name}\ | |
| \ is in a VSAN enabled cluster {computeResource.name} in datacenter {datacenter.name}\ | |
| \ but does not have VSAN service enabled</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ vendor provider deregistration failed</description><category>error</category><formatOnDatacenter>Vendor\ | |
| \ provider {host.name} deregistration failed</formatOnDatacenter><formatOnComputeResource>Vendor\ | |
| \ provider {host.name} deregistration failed</formatOnComputeResource><formatOnHost>Vendor\ | |
| \ provider {host.name} deregistration failed</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vsan.HostVendorProviderDeregistrationFailedEvent|Vendor\ | |
| \ provider {host.name} deregistration failed</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.vsan.HostVendorProviderDeregistrationFailedEvent">\ | |
| \ <description> Cannot deregister host vendor provider\ | |
| \ in Storage management service </description> <cause>\ | |
| \ <description>Host vendor provider deregistration failed</description>\ | |
| \ <action>Check if Storage management service is running</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ vendor provider deregistration succeeded</description><category>info</category><formatOnDatacenter>Vendor\ | |
| \ provider {host.name} deregistration succeeded</formatOnDatacenter><formatOnComputeResource>Vendor\ | |
| \ provider {host.name} deregistration succeeded</formatOnComputeResource><formatOnHost>Vendor\ | |
| \ provider {host.name} deregistration succeeded</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vsan.HostVendorProviderDeregistrationSuccessEvent|Vendor\ | |
| \ provider {host.name} deregistration succeeded</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.vsan.HostVendorProviderDeregistrationSuccessEvent">\ | |
| \ <description> Deregistered host vendor provider from\ | |
| \ Storage management service </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ vendor provider registration failed</description><category>error</category><formatOnDatacenter>Vendor\ | |
| \ provider {host.name} registration failed</formatOnDatacenter><formatOnComputeResource>Vendor\ | |
| \ provider {host.name} registration failed</formatOnComputeResource><formatOnHost>Vendor\ | |
| \ provider {host.name} registration failed</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vsan.HostVendorProviderRegistrationFailedEvent|Vendor\ | |
| \ provider {host.name} registration failed</fullFormat><longDescription> \ | |
| \ <EventLongDescription id="com.vmware.vc.vsan.HostVendorProviderRegistrationFailedEvent">\ | |
| \ <description> Cannot register host vendor provider\ | |
| \ in Storage management service </description> <cause>\ | |
| \ <description>Host vendor provider registration failed</description>\ | |
| \ <action>Check if Storage management service is running</action>\ | |
| \ <action>Check if the vendor provider on host is running</action>\ | |
| \ <action>Check if there are network connectivity issues\ | |
| \ between host and VC</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ vendor provider registration succeeded</description><category>info</category><formatOnDatacenter>Vendor\ | |
| \ provider {host.name} registration succeeded</formatOnDatacenter><formatOnComputeResource>Vendor\ | |
| \ provider {host.name} registration succeeded</formatOnComputeResource><formatOnHost>Vendor\ | |
| \ provider {host.name} registration succeeded</formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vc.vsan.HostVendorProviderRegistrationSuccessEvent|Vendor\ | |
| \ provider {host.name} registration succeeded</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="com.vmware.vc.vsan.HostVendorProviderRegistrationSuccessEvent">\ | |
| \ <description> Registered host vendor provider in Storage\ | |
| \ management service </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>VSAN\ | |
| \ network is not configured</description><category>error</category><formatOnDatacenter>VSAN\ | |
| \ network is not configured on {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>VSAN\ | |
| \ network is not configured on {host.name}</formatOnComputeResource><formatOnHost>VSAN\ | |
| \ network is not configured</formatOnHost><formatOnVm>VSAN network is not\ | |
| \ configured on {host.name} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name}</formatOnVm><fullFormat>com.vmware.vc.vsan.NetworkMisConfiguredEvent|VSAN\ | |
| \ network is not configured on {host.name} in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Found\ | |
| \ another host participating in the VSAN service which is not a member of\ | |
| \ this host's vCenter cluster</description><category>error</category><formatOnDatacenter>Found\ | |
| \ another host {hostString} participating in the VSAN service in cluster {computeResource.name}\ | |
| \ which is not a member of this host's vCenter cluster</formatOnDatacenter><formatOnComputeResource>Found\ | |
| \ another host {hostString} participating in the VSAN service which is not\ | |
| \ a member of this host's vCenter cluster</formatOnComputeResource><formatOnHost>Found\ | |
| \ another host {hostString} participating in the VSAN service which is not\ | |
| \ a member of this host's vCenter cluster</formatOnHost><formatOnVm>Found\ | |
| \ another host {hostString} participating in the VSAN service in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} which is not a member of this host's\ | |
| \ vCenter cluster</formatOnVm><fullFormat>com.vmware.vc.vsan.RogueHostFoundEvent|Found\ | |
| \ another host {hostString} participating in the VSAN service in cluster {computeResource.name}\ | |
| \ in datacenter {datacenter.name} which is not a member of this host's\ | |
| \ vCenter cluster</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agency\ | |
| \ created</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agency.create|{agencyName}\ | |
| \ created by {ownerName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agency\ | |
| \ destroyed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agency.destroyed|{agencyName}\ | |
| \ removed from the vSphere ESX Agent Manager</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agency\ | |
| \ state changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agency.goalstate|{agencyName}\ | |
| \ changed goal state from {oldGoalState} to {newGoalState}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agency\ | |
| \ status changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agency.statusChanged|Agency\ | |
| \ status changed from {oldStatus} to {newStatus}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agency\ | |
| \ reconfigured</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agency.updated|Configuration\ | |
| \ updated {agencyName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ added to host</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.created|Agent\ | |
| \ added to host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ removed from host</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.destroyed|Agent\ | |
| \ removed from host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ removed from host</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.destroyedNoHost|Agent\ | |
| \ removed from host ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ VM has been powered on and awaits a mark agent as available</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.markAgentVmAsAvailableAfterPowerOn|Agent\ | |
| \ VM {vm.name} has been powered on. Mark agent as available to proceed agent\ | |
| \ workflow ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ VM has been provisioned and awaits a mark agent as available</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.markAgentVmAsAvailableAfterProvisioning|Agent\ | |
| \ VM {vm.name} has been provisioned. Mark agent as available to proceed agent\ | |
| \ workflow ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ status changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.statusChanged|Agent\ | |
| \ status changed from {oldStatus} to {newStatus}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ VM is deleted</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.task.deleteVm|Agent\ | |
| \ VM {vmName} is deleted on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ VM is provisioned</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.task.deployVm|Agent\ | |
| \ VM {vm.name} is provisioned on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ VM powered off</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.task.powerOffVm|Agent\ | |
| \ VM {vm.name} powered off, on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Agent\ | |
| \ VM powered on</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.task.powerOnVm|Agent\ | |
| \ VM {vm.name} powered on, on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VIB\ | |
| \ installed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.task.vibInstalled|Agent\ | |
| \ installed VIB {vib} on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VIB\ | |
| \ installed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.agent.task.vibUninstalled|Agent\ | |
| \ uninstalled VIB {vib} on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.cannotAccessAgentOVF|Unable\ | |
| \ to access agent OVF package at {url} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.cannotAccessAgentVib|Unable\ | |
| \ to access agent VIB module at {url} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.hostInMaintenanceMode|Agent\ | |
| \ cannot complete an operation since the host {host.name} is in maintenance\ | |
| \ mode ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.hostInStandbyMode|Agent\ | |
| \ cannot complete an operation since the host {host.name} is in standby mode\ | |
| \ ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.hostPoweredOff|Agent\ | |
| \ cannot complete an operation since the host {host.name} is powered off ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.incompatibleHostVersion|Agent\ | |
| \ is not deployed due to incompatible host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.insufficientIpAddresses|Insufficient\ | |
| \ IP addresses in network protocol profile in agent's VM network ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.insufficientResources|Agent\ | |
| \ cannot be provisioned due to insufficient resources on host {host.name}\ | |
| \ ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.insufficientSpace|Agent\ | |
| \ on {host.name} cannot be provisioned due to insufficient space on datastore\ | |
| \ ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.missingAgentIpPool|No\ | |
| \ network protocol profile associated to agent's VM network ({agencyname})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.missingDvFilterSwitch|dvFilter\ | |
| \ switch is not configured on host {host.name} ({agencyname})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.noAgentVmDatastore|No\ | |
| \ agent datastore configuration on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.noAgentVmNetwork|No\ | |
| \ agent network configuration on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.noCustomAgentVmDatastore|Agent\ | |
| \ datastore(s) {customAgentVmDatastoreName} not available on host {host.name}\ | |
| \ ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.noCustomAgentVmNetwork|Agent\ | |
| \ network(s) {customAgentVmNetworkName} not available on host {host.name}\ | |
| \ ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.orphandedDvFilterSwitch|Unused\ | |
| \ dvFilter switch on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.orphanedAgency|Orphaned\ | |
| \ agency found. ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.ovfInvalidFormat|OVF\ | |
| \ used to provision agent on host {host.name} has invalid format ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.ovfInvalidProperty|OVF\ | |
| \ environment used to provision agent on host {host.name} has one or more\ | |
| \ invalid properties ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.resolved|Issue\ | |
| \ {type} resolved (key {key})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.unknownAgentVm|Unknown\ | |
| \ agent VM {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vibCannotPutHostInMaintenanceMode|Cannot\ | |
| \ put host into maintenance mode ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vibInvalidFormat|Invalid\ | |
| \ format for VIB module at {url} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vibNotInstalled|VIB\ | |
| \ module for agent is not installed on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vibRequiresHostInMaintenanceMode|Host\ | |
| \ must be put into maintenance mode to complete agent VIB installation ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vibRequiresHostReboot|Host\ | |
| \ {host.name} must be reboot to complete agent VIB installation ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vibRequiresManualInstallation|VIB\ | |
| \ {vib} requires manual installation on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vibRequiresManualUninstallation|VIB\ | |
| \ {vib} requires manual uninstallation on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmCorrupted|Agent\ | |
| \ VM {vm.name} on host {host.name} is corrupted ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmDeployed|Agent\ | |
| \ VM {vm.name} is provisioned on host {host.name} when it should be removed\ | |
| \ ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmMarkedAsTemplate|Agent\ | |
| \ VM {vm.name} on host {host.name} is marked as template ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmNotDeployed|Agent\ | |
| \ VM is missing on host {host.name} ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmOrphaned|Orphaned\ | |
| \ agent VM {vm.name} on host {host.name} detected ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmPoweredOff|Agent\ | |
| \ VM {vm.name} on host {host.name} is expected to be powered on ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmPoweredOn|Agent\ | |
| \ VM {vm.name} on host {host.name} is expected to be powered off ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmSuspended|Agent\ | |
| \ VM {vm.name} on host {host.name} is expected to be powered on but is suspended\ | |
| \ ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmWrongFolder|Agent\ | |
| \ VM {vm.name} on host {host.name} is in the wrong VM folder ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.issue.vmWrongResourcePool|Agent\ | |
| \ VM {vm.name} on host {host.name} is in the resource pool ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Invalid\ | |
| \ login</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.login.invalid|Failed\ | |
| \ login to vSphere ESX Agent Manager</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Successful\ | |
| \ login to vSphere ESX Agent Manager</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.login.succeeded|Successful\ | |
| \ login by {user} into vSphere ESX Agent Manager</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>User\ | |
| \ logged out of vSphere ESX Agent Manager</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.logout|User\ | |
| \ {user} logged out of vSphere ESX Agent Manager by logging out of the vCenter\ | |
| \ server</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Scan\ | |
| \ for unknown agent VMs completed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.task.scanForUnknownAgentVmsCompleted|Scan\ | |
| \ for unknown agent VMs completed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Scan\ | |
| \ for unknown agent VMs initiated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.task.scanForUnknownAgentVmsInitiated|Scan\ | |
| \ for unknown agent VMs initiated</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Setup\ | |
| \ dvFilter switch</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.task.setupDvFilter|DvFilter\ | |
| \ switch '{switchName}' is setup on host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Tear\ | |
| \ down dvFilter switch</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.task.tearDownDvFilter|DvFilter\ | |
| \ switch '{switchName}' is teared down on host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Unauthorized\ | |
| \ access in vSphere ESX Agent Manager</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.unauthorized.access|Unauthorized\ | |
| \ access by {user} in vSphere ESX Agent Manager</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description></description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>com.vmware.vim.eam.vum.failedtouploadvib|Failed\ | |
| \ to upload {vibUrl} to VMware Update Manager ({agencyName})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.bind.vApp</description><category>event.com.vmware.vim.vsm.dependency.bind.vApp.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.bind.vApp.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.bind.vApp.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.bind.vApp.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.bind.vApp.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.bind.vApp|event.com.vmware.vim.vsm.dependency.bind.vApp.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.bind.vm</description><category>event.com.vmware.vim.vsm.dependency.bind.vm.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.bind.vm.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.bind.vm.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.bind.vm.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.bind.vm.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.bind.vm|event.com.vmware.vim.vsm.dependency.bind.vm.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.create.vApp</description><category>event.com.vmware.vim.vsm.dependency.create.vApp.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.create.vApp.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.create.vApp.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.create.vApp.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.create.vApp.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.create.vApp|event.com.vmware.vim.vsm.dependency.create.vApp.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.create.vm</description><category>event.com.vmware.vim.vsm.dependency.create.vm.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.create.vm.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.create.vm.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.create.vm.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.create.vm.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.create.vm|event.com.vmware.vim.vsm.dependency.create.vm.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.destroy.vApp</description><category>event.com.vmware.vim.vsm.dependency.destroy.vApp.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.destroy.vApp.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.destroy.vApp.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.destroy.vApp.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.destroy.vApp.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.destroy.vApp|event.com.vmware.vim.vsm.dependency.destroy.vApp.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.destroy.vm</description><category>event.com.vmware.vim.vsm.dependency.destroy.vm.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.destroy.vm.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.destroy.vm.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.destroy.vm.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.destroy.vm.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.destroy.vm|event.com.vmware.vim.vsm.dependency.destroy.vm.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.reconfigure.vApp</description><category>event.com.vmware.vim.vsm.dependency.reconfigure.vApp.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.reconfigure.vApp.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.reconfigure.vApp.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.reconfigure.vApp.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.reconfigure.vApp.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.reconfigure.vApp|event.com.vmware.vim.vsm.dependency.reconfigure.vApp.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.reconfigure.vm</description><category>event.com.vmware.vim.vsm.dependency.reconfigure.vm.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.reconfigure.vm.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.reconfigure.vm.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.reconfigure.vm.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.reconfigure.vm.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.reconfigure.vm|event.com.vmware.vim.vsm.dependency.reconfigure.vm.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.unbind.vApp</description><category>event.com.vmware.vim.vsm.dependency.unbind.vApp.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.unbind.vApp.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.unbind.vApp.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.unbind.vApp.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.unbind.vApp.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.unbind.vApp|event.com.vmware.vim.vsm.dependency.unbind.vApp.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.unbind.vm</description><category>event.com.vmware.vim.vsm.dependency.unbind.vm.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.unbind.vm.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.unbind.vm.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.unbind.vm.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.unbind.vm.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.unbind.vm|event.com.vmware.vim.vsm.dependency.unbind.vm.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.update.vApp</description><category>event.com.vmware.vim.vsm.dependency.update.vApp.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.update.vApp.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.update.vApp.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.update.vApp.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.update.vApp.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.update.vApp|event.com.vmware.vim.vsm.dependency.update.vApp.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.dependency.update.vm</description><category>event.com.vmware.vim.vsm.dependency.update.vm.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.dependency.update.vm.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.dependency.update.vm.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.dependency.update.vm.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.dependency.update.vm.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.dependency.update.vm|event.com.vmware.vim.vsm.dependency.update.vm.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.provider.register</description><category>event.com.vmware.vim.vsm.provider.register.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.provider.register.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.provider.register.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.provider.register.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.provider.register.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.provider.register|event.com.vmware.vim.vsm.provider.register.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.provider.unregister</description><category>event.com.vmware.vim.vsm.provider.unregister.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.provider.unregister.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.provider.unregister.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.provider.unregister.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.provider.unregister.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.provider.unregister|event.com.vmware.vim.vsm.provider.unregister.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vim.vsm.provider.update</description><category>event.com.vmware.vim.vsm.provider.update.category</category><formatOnDatacenter>event.com.vmware.vim.vsm.provider.update.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vim.vsm.provider.update.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vim.vsm.provider.update.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vim.vsm.provider.update.fullFormat</formatOnVm><fullFormat>com.vmware.vim.vsm.provider.update|event.com.vmware.vim.vsm.provider.update.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Restoring\ | |
| \ factory defaults through DCUI.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.defaults.factoryrestore|The\ | |
| \ host has been restored to default factory settings. Please consult ESXi\ | |
| \ Embedded and vCenter Server Setup Guide or follow the Ask VMware link for\ | |
| \ more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ DCUI has been disabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.disabled|The\ | |
| \ DCUI has been disabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ DCUI has been enabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.enabled|The\ | |
| \ DCUI has been enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Rebooting\ | |
| \ host through DCUI.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.host.reboot|The\ | |
| \ host is being rebooted through the Direct Console User Interface (DCUI).\ | |
| \ Please consult ESXi Embedded and vCenter Server Setup Guide or follow the\ | |
| \ Ask VMware link for more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Shutting\ | |
| \ down host through DCUI.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.host.shutdown|The\ | |
| \ host is being shut down through the Direct Console User Interface (DCUI).\ | |
| \ Please consult ESXi Embedded and vCenter Server Setup Guide or follow the\ | |
| \ Ask VMware link for more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Restarting\ | |
| \ host agents through DCUI.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.hostagents.restart|The\ | |
| \ management agents on the host are being restarted. Please consult ESXi Embedded\ | |
| \ and vCenter Server Setup Guide or follow the Ask VMware link for more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Login\ | |
| \ authentication on DCUI failed</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.login.failed|Authentication\ | |
| \ of user {1} has failed. Please consult ESXi Embedded and vCenter Server\ | |
| \ Setup Guide or follow the Ask VMware link for more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>DCUI\ | |
| \ login password changed.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.login.passwd.changed|Login\ | |
| \ password for user {1} has been changed. Please consult ESXi Embedded and\ | |
| \ vCenter Server Setup Guide or follow the Ask VMware link for more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Factory\ | |
| \ network settings restored through DCUI.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.network.factoryrestore|The\ | |
| \ host has been restored to factory network settings. Please consult ESXi\ | |
| \ Embedded and vCenter Server Setup Guide or follow the Ask VMware link for\ | |
| \ more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restarting\ | |
| \ network through DCUI.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.dcui.network.restart|A\ | |
| \ management interface {1} has been restarted. Please consult ESXi Embedded\ | |
| \ and vCenter Server Setup Guide or follow the Ask VMware link for more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Powering\ | |
| \ off host through esxcli</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esxcli.host.poweroff|The\ | |
| \ host is being powered off through esxcli. Reason for powering off: {1}.\ | |
| \ Please consult vSphere Documentation Center or follow the Ask VMware link\ | |
| \ for more information.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>esx.audit.esxcli.host.restart</description><category>event.esx.audit.esxcli.host.restart.category</category><formatOnDatacenter>event.esx.audit.esxcli.host.restart.fullFormat</formatOnDatacenter><formatOnComputeResource>event.esx.audit.esxcli.host.restart.fullFormat</formatOnComputeResource><formatOnHost>event.esx.audit.esxcli.host.restart.fullFormat</formatOnHost><formatOnVm>event.esx.audit.esxcli.host.restart.fullFormat</formatOnVm><fullFormat>esx.audit.esxcli.host.restart|event.esx.audit.esxcli.host.restart.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Host\ | |
| \ acceptance level changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esximage.hostacceptance.changed|Host\ | |
| \ acceptance level changed from {1} to {2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Attempting\ | |
| \ to install an image profile with validation disabled.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esximage.install.novalidation|Attempting\ | |
| \ to install an image profile with validation disabled. This may result in\ | |
| \ an image with unsatisfied dependencies, file or package conflicts, and potential\ | |
| \ security violations.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>SECURITY\ | |
| \ ALERT: Installing image profile.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esximage.install.securityalert|SECURITY\ | |
| \ ALERT: Installing image profile '{1}' with {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Successfully\ | |
| \ installed image profile.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esximage.profile.install.successful|Successfully\ | |
| \ installed image profile '{1}'. Installed {2} VIB(s), removed {3}\ | |
| \ VIB(s). Please use 'esxcli software profile get' or see log for\ | |
| \ more detail about the transaction.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Successfully\ | |
| \ updated host to new image profile.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esximage.profile.update.successful|Successfully\ | |
| \ updated host to image profile '{1}'. Installed {2} VIB(s), removed\ | |
| \ {3} VIB(s). Please use 'esxcli software profile get' or see log\ | |
| \ for more detail about the transaction.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Successfully\ | |
| \ installed VIBs.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esximage.vib.install.successful|Successfully\ | |
| \ installed {1} VIB(s), removed {2} VIB(s). Please use 'esxcli software\ | |
| \ profile get' or see log for more detail about the transaction.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Successfully\ | |
| \ removed VIBs</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.esximage.vib.remove.successful|Successfully\ | |
| \ removed {1} VIB(s). Please use 'esxcli software profile get' or\ | |
| \ see log for more detail about the transaction.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ has booted.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.host.boot|Host\ | |
| \ has booted.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>The\ | |
| \ number of virtual machines registered on the host exceeded limit.</description><category>warning</category><formatOnDatacenter>The\ | |
| \ number of virtual machines registered on host {host.name} in cluster {computeResource.name}\ | |
| \ exceeded limit: {current} registered, {limit} is the maximum supported.</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ number of virtual machines registered on host {host.name} exceeded limit:\ | |
| \ {current} registered, {limit} is the maximum supported.</formatOnComputeResource><formatOnHost>The\ | |
| \ number of virtual machines registered on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} exceeded limit: {current} registered, {limit} is the\ | |
| \ maximum supported.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.host.maxRegisteredVMsExceeded|The\ | |
| \ number of virtual machines registered on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} exceeded limit: {current} registered, {limit} is the\ | |
| \ maximum supported.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ is rebooting.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.host.stop.reboot|Host\ | |
| \ is rebooting.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Host\ | |
| \ is shutting down.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.host.stop.shutdown|Host\ | |
| \ is shutting down.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Administrator\ | |
| \ access to the host has been enabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.lockdownmode.disabled|Administrator\ | |
| \ access to the host has been enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Administrator\ | |
| \ access to the host has been disabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.lockdownmode.enabled|Administrator\ | |
| \ access to the host has been disabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ host has canceled entering maintenance mode.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.maintenancemode.canceled|The\ | |
| \ host has canceled entering maintenance mode.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ host has entered maintenance mode.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.maintenancemode.entered|The\ | |
| \ host has entered maintenance mode.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ host has begun entering maintenance mode.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.maintenancemode.entering|The\ | |
| \ host has begun entering maintenance mode.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ host has exited maintenance mode.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.maintenancemode.exited|The\ | |
| \ host has exited maintenance mode.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Firewall\ | |
| \ configuration has changed.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.firewall.config.changed|Firewall\ | |
| \ configuration has changed. Operation '{1}' for rule set {2} succeeded.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Firewall\ | |
| \ has been disabled.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.firewall.disabled|Firewall\ | |
| \ has been disabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Firewall\ | |
| \ has been enabled for port.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.firewall.enabled|Firewall\ | |
| \ has been enabled for port {1}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Port\ | |
| \ is now protected by Firewall.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.firewall.port.hooked|Port\ | |
| \ {1} is now protected by Firewall.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Port\ | |
| \ is no longer protected with Firewall.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.firewall.port.removed|Port\ | |
| \ {1} is no longer protected with Firewall.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>LACP\ | |
| \ disabled</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.lacp.disable|LACP\ | |
| \ for VDS {1} is disabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>LACP\ | |
| \ eabled</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.lacp.enable|LACP\ | |
| \ for VDS {1} is enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ is connected</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.net.lacp.uplink.connected|LACP\ | |
| \ info: uplink {1} on VDS {2} got connected.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ ESXi command line shell has been disabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.shell.disabled|The\ | |
| \ ESXi command line shell has been disabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>The\ | |
| \ ESXi command line shell has been enabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.shell.enabled|The\ | |
| \ ESXi command line shell has been enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>SSH\ | |
| \ access has been disabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.ssh.disabled|SSH\ | |
| \ access has been disabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>SSH\ | |
| \ access has been enabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.ssh.enabled|SSH\ | |
| \ access has been enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>USB\ | |
| \ configuration has changed.</description><category>info</category><formatOnDatacenter>USB\ | |
| \ configuration has changed on host {host.name} in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>USB\ | |
| \ configuration has changed on host {host.name}.</formatOnComputeResource><formatOnHost>USB\ | |
| \ configuration has changed.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.usb.config.changed|USB\ | |
| \ configuration has changed on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Enforcement\ | |
| \ level changed for all security domains.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.uw.secpolicy.alldomains.level.changed|The\ | |
| \ enforcement level for all security domains has been changed to {1}. The\ | |
| \ enforcement level must always be set to enforcing.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Enforcement\ | |
| \ level changed for security domain.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.uw.secpolicy.domain.level.changed|The\ | |
| \ enforcement level for security domain {1} has been changed to {2}. The enforcement\ | |
| \ level must always be set to enforcing.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>LVM\ | |
| \ device discovered.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.vmfs.lvm.device.discovered|One\ | |
| \ or more LVM devices have been discovered on this host.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>File\ | |
| \ system mounted.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.vmfs.volume.mounted|File\ | |
| \ system {1} on volume {2} has been mounted in {3} mode on this host.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>LVM\ | |
| \ volume un-mounted.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.vmfs.volume.umounted|The\ | |
| \ volume {1} has been safely un-mounted. The datastore is no longer accessible\ | |
| \ on this host.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>VSAN\ | |
| \ clustering services have been enabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>VSAN\ | |
| \ clustering and directory services have been enabled.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.audit.vsan.clustering.enabled|VSAN\ | |
| \ clustering and directory services have been enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>A\ | |
| \ vmkcore disk partition is available and/or a network coredump server has\ | |
| \ been configured. Host core dumps will be saved.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>A\ | |
| \ vmkcore disk partition is available and/or a network coredump server has\ | |
| \ been configured. Host core dumps will be saved.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.coredump.configured|A\ | |
| \ vmkcore disk partition is available and/or a network coredump server has\ | |
| \ been configured. Host core dumps will be saved.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ network connectivity to portgroups</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.connectivity.restored|Network\ | |
| \ connectivity restored on virtual switch {1}, portgroups: {2}. Physical NIC\ | |
| \ {3} is up.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ Network Connectivity to DVPorts</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.dvport.connectivity.restored|Network\ | |
| \ connectivity restored on DVPorts: {1}. Physical NIC {2} is up.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ Network Redundancy to DVPorts</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.dvport.redundancy.restored|Uplink\ | |
| \ redundancy restored on DVPorts: {1}. Physical NIC {2} is up.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>lag\ | |
| \ transition up</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.lacp.lag.transition.up|LACP\ | |
| \ info: LAG {1} on VDS {2} is up.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ transition up</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.lacp.uplink.transition.up|LACP\ | |
| \ info: uplink {1} on VDS {2} is moved into link aggregation group.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ is unblocked</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.lacp.uplink.unblocked|LACP\ | |
| \ info: uplink {1} on VDS {2} is unblocked.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ uplink redundancy to portgroups</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.redundancy.restored|Uplink\ | |
| \ redundancy restored on virtual switch {1}, portgroups: {2}. Physical NIC\ | |
| \ {3} is up.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Link\ | |
| \ state up</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.net.vmnic.linkstate.up|Physical\ | |
| \ NIC {1} linkstate is up.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Scsi\ | |
| \ Device I/O Latency has improved</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.scsi.device.io.latency.improved|Device\ | |
| \ {1} performance has improved. I/O latency reduced from {2} microseconds\ | |
| \ to {3} microseconds.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ has been turned on administratively.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.scsi.device.state.on|Device\ | |
| \ {1}, has been turned on administratively.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ that was permanently inaccessible is now online.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.scsi.device.state.permanentloss.deviceonline|Device\ | |
| \ {1}, that was permanently inaccessible is now online. No data consistency\ | |
| \ guarantees.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Exited\ | |
| \ the All Paths Down state</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.storage.apd.exit|Device\ | |
| \ or filesystem with identifier {1} has exited the All Paths Down state.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ connectivity to storage device</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.storage.connectivity.restored|Connectivity\ | |
| \ to storage device {1} (Datastores: {2}) restored. Path {3} is active again.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ path redundancy to storage device</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.storage.redundancy.restored|Path\ | |
| \ redundancy to storage device {1} (Datastores: {2}) restored. Path {3} is\ | |
| \ active again.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>VSAN\ | |
| \ clustering services have now been enabled.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>VSAN\ | |
| \ clustering and directory services have now been enabled.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.clear.vsan.clustering.enabled|VSAN\ | |
| \ clustering and directory services have now been enabled.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>esx.clear.vsan.network.available</description><category>event.esx.clear.vsan.network.available.category</category><formatOnDatacenter>event.esx.clear.vsan.network.available.fullFormat</formatOnDatacenter><formatOnComputeResource>event.esx.clear.vsan.network.available.fullFormat</formatOnComputeResource><formatOnHost>event.esx.clear.vsan.network.available.fullFormat</formatOnHost><formatOnVm>event.esx.clear.vsan.network.available.fullFormat</formatOnVm><fullFormat>esx.clear.vsan.network.available|event.esx.clear.vsan.network.available.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>esx.clear.vsan.vmknic.ready</description><category>event.esx.clear.vsan.vmknic.ready.category</category><formatOnDatacenter>event.esx.clear.vsan.vmknic.ready.fullFormat</formatOnDatacenter><formatOnComputeResource>event.esx.clear.vsan.vmknic.ready.fullFormat</formatOnComputeResource><formatOnHost>event.esx.clear.vsan.vmknic.ready.fullFormat</formatOnHost><formatOnVm>event.esx.clear.vsan.vmknic.ready.fullFormat</formatOnVm><fullFormat>esx.clear.vsan.vmknic.ready|event.esx.clear.vsan.vmknic.ready.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ 3rd party component on ESXi has reported an error.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.3rdParty.error|A\ | |
| \ 3rd party component, {1}, running on ESXi has reported an error. Please\ | |
| \ follow the knowledge base link ({2}) to see the steps to remedy the problem\ | |
| \ as reported by {3}. The message reported is: {4}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>esx.problem.3rdParty.info</description><category>event.esx.problem.3rdParty.info.category</category><formatOnDatacenter>event.esx.problem.3rdParty.info.fullFormat</formatOnDatacenter><formatOnComputeResource>event.esx.problem.3rdParty.info.fullFormat</formatOnComputeResource><formatOnHost>event.esx.problem.3rdParty.info.fullFormat</formatOnHost><formatOnVm>event.esx.problem.3rdParty.info.fullFormat</formatOnVm><fullFormat>esx.problem.3rdParty.info|event.esx.problem.3rdParty.info.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ 3rd party component on ESXi has reported a warning.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.3rdParty.warning|A\ | |
| \ 3rd party component, {1}, running on ESXi has reported a warning related\ | |
| \ to a problem. Please follow the knowledge base link ({2}) to see the steps\ | |
| \ to remedy the problem as reported by {3}. The message reported is: {4}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ corrected memory error occurred</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.apei.bert.memory.error.corrected|A\ | |
| \ corrected memory error occurred in last boot. The following details were\ | |
| \ reported. Physical Addr: {1}, Physical Addr Mask: {2}, Node: {3}, Card:\ | |
| \ {4}, Module: {5}, Bank: {6}, Device: {7}, Row: {8}, Column: {9} Error type:\ | |
| \ {10}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ fatal memory error occurred</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.apei.bert.memory.error.fatal|A\ | |
| \ fatal memory error occurred in the last boot. The following details were\ | |
| \ reported. Physical Addr: {1}, Physical Addr Mask: {2}, Node: {3}, Card:\ | |
| \ {4}, Module: {5}, Bank: {6}, Device: {7}, Row: {8}, Column: {9} Error type:\ | |
| \ {10}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ recoverable memory error occurred</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.apei.bert.memory.error.recoverable|A\ | |
| \ recoverable memory error occurred in last boot. The following details were\ | |
| \ reported. Physical Addr: {1}, Physical Addr Mask: {2}, Node: {3}, Card:\ | |
| \ {4}, Module: {5}, Bank: {6}, Device: {7}, Row: {8}, Column: {9} Error type:\ | |
| \ {10}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ corrected PCIe error occurred</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.apei.bert.pcie.error.corrected|A\ | |
| \ corrected PCIe error occurred in last boot. The following details were reported.\ | |
| \ Port Type: {1}, Device: {2}, Bus #: {3}, Function: {4}, Slot: {5}, Device\ | |
| \ Vendor: {6}, Version: {7}, Command Register: {8}, Status Register: {9}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ fatal PCIe error occurred</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.apei.bert.pcie.error.fatal|Platform\ | |
| \ encounterd a fatal PCIe error in last boot. The following details were reported.\ | |
| \ Port Type: {1}, Device: {2}, Bus #: {3}, Function: {4}, Slot: {5}, Device\ | |
| \ Vendor: {6}, Version: {7}, Command Register: {8}, Status Register: {9}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ recoverable PCIe error occurred</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.apei.bert.pcie.error.recoverable|A\ | |
| \ recoverable PCIe error occurred in last boot. The following details were\ | |
| \ reported. Port Type: {1}, Device: {2}, Bus #: {3}, Function: {4}, Slot:\ | |
| \ {5}, Device Vendor: {6}, Version: {7}, Command Register: {8}, Status Register:\ | |
| \ {9}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>An\ | |
| \ application running on ESXi host has crashed and a core file was created.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.application.core.dumped|An\ | |
| \ application ({1}) running on ESXi host has crashed ({2} time(s) so far).\ | |
| \ A core file might have been created at {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>No\ | |
| \ vmkcore disk partition is available and no network coredump server has been\ | |
| \ configured. Host core dumps cannot be saved.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>No\ | |
| \ vmkcore disk partition is available and no network coredump server has been\ | |
| \ configured. Host core dumps cannot be saved.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.coredump.unconfigured|No\ | |
| \ vmkcore disk partition is available and no network coredump server has been\ | |
| \ configured. Host core dumps cannot be saved.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>DRAM\ | |
| \ ECC not enabled. Please enable it in BIOS.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.cpu.amd.mce.dram.disabled|DRAM\ | |
| \ ECC not enabled. Please enable it in BIOS.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Not\ | |
| \ all IO-APICs are listed in the DMAR. Not enabling interrupt remapping on\ | |
| \ this platform. </description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.cpu.intel.ioapic.listing.error|Not\ | |
| \ all IO-APICs are listed in the DMAR. Not enabling interrupt remapping on\ | |
| \ this platform. </fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>MCE\ | |
| \ monitoring will be disabled as an unsupported CPU was detected. Please consult\ | |
| \ the ESX HCL for information on supported hardware.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.cpu.mce.invalid|MCE\ | |
| \ monitoring will be disabled as an unsupported CPU was detected. Please consult\ | |
| \ the ESX HCL for information on supported hardware.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Disabling\ | |
| \ HyperThreading due to invalid configuration: Number of threads: {1}, Number\ | |
| \ of PCPUs: {2}.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.cpu.smp.ht.invalid|Disabling\ | |
| \ HyperThreading due to invalid configuration: Number of threads: {1}, Number\ | |
| \ of PCPUs: {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Found\ | |
| \ {1} PCPUs, but only using {2} of them due to specified limit.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.cpu.smp.ht.numpcpus.max|Found\ | |
| \ {1} PCPUs, but only using {2} of them due to specified limit.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Disabling\ | |
| \ HyperThreading due to invalid configuration: HT partner {1} is missing from\ | |
| \ PCPU {2}.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.cpu.smp.ht.partner.missing|Disabling\ | |
| \ HyperThreading due to invalid configuration: HT partner {1} is missing from\ | |
| \ PCPU {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Unable\ | |
| \ to obtain a DHCP lease.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.dhclient.lease.none|Unable\ | |
| \ to obtain a DHCP lease on interface {1}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>esx.problem.dhclient.lease.offered.error</description><category>event.esx.problem.dhclient.lease.offered.error.category</category><formatOnDatacenter>event.esx.problem.dhclient.lease.offered.error.fullFormat</formatOnDatacenter><formatOnComputeResource>event.esx.problem.dhclient.lease.offered.error.fullFormat</formatOnComputeResource><formatOnHost>event.esx.problem.dhclient.lease.offered.error.fullFormat</formatOnHost><formatOnVm>event.esx.problem.dhclient.lease.offered.error.fullFormat</formatOnVm><fullFormat>esx.problem.dhclient.lease.offered.error|event.esx.problem.dhclient.lease.offered.error.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Could\ | |
| \ not install image profile.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.esximage.install.error|Could\ | |
| \ not install image profile: {1}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Host\ | |
| \ doesn't meet image profile hardware requirements.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.esximage.install.invalidhardware|Host\ | |
| \ doesn't meet image profile '{1}' hardware requirements: {2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Could\ | |
| \ not stage image profile.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.esximage.install.stage.error|Could\ | |
| \ not stage image profile '{1}': {2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Skipping\ | |
| \ interrupt routing entry with bad device number: {1}. This is a BIOS bug.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.hardware.acpi.interrupt.routing.device.invalid|Skipping\ | |
| \ interrupt routing entry with bad device number: {1}. This is a BIOS bug.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Skipping\ | |
| \ interrupt routing entry with bad device pin: {1}. This is a BIOS bug.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.hardware.acpi.interrupt.routing.pin.invalid|Skipping\ | |
| \ interrupt routing entry with bad device pin: {1}. This is a BIOS bug.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>IOAPIC\ | |
| \ Num {1} is missing. Please check BIOS settings to enable this IOAPIC.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.hardware.ioapic.missing|IOAPIC\ | |
| \ Num {1} is missing. Please check BIOS settings to enable this IOAPIC.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>An\ | |
| \ unread host kernel core dump has been found.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.host.coredump|An\ | |
| \ unread host kernel core dump has been found.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Hostd\ | |
| \ crashed and a core file was created.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.hostd.core.dumped|{1}\ | |
| \ crashed ({2} time(s) so far) and a core file might have been created at\ | |
| \ {3}. This might have caused connections to the host to be dropped.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Storage\ | |
| \ I/O Control version mismatch</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.iorm.badversion|Host\ | |
| \ {1} cannot participate in Storage I/O Control(SIOC) on datastore {2} because\ | |
| \ the version number {3} of the SIOC agent on this host is incompatible with\ | |
| \ number {4} of its counterparts on other hosts connected to this datastore.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Unmanaged\ | |
| \ workload detected on SIOC-enabled datastore</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.iorm.nonviworkload|An\ | |
| \ unmanaged I/O workload is detected on a SIOC-enabled datastore: {1}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Failed\ | |
| \ to create default migration heap</description><category>warning</category><formatOnDatacenter>Failed\ | |
| \ to create default migration heap. This might be the result of severe host\ | |
| \ memory pressure or virtual address space exhaustion. Migration might still\ | |
| \ be possible, but will be unreliable in cases of extreme host memory pressure.</formatOnDatacenter><formatOnComputeResource>Failed\ | |
| \ to create default migration heap. This might be the result of severe host\ | |
| \ memory pressure or virtual address space exhaustion. Migration might still\ | |
| \ be possible, but will be unreliable in cases of extreme host memory pressure.</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.migrate.vmotion.default.heap.create.failed|Failed\ | |
| \ to create default migration heap. This might be the result of severe host\ | |
| \ memory pressure or virtual address space exhaustion. Migration might still\ | |
| \ be possible, but will be unreliable in cases of extreme host memory pressure.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Error\ | |
| \ with migration listen socket</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.migrate.vmotion.server.pending.cnx.listen.socket.shutdown|The\ | |
| \ ESXi host's vMotion network server encountered an error while monitoring\ | |
| \ incoming network connections. Shutting down listener socket. vMotion might\ | |
| \ not be possible with this host until vMotion is manually re-enabled. Failure\ | |
| \ status: {1}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Network Connectivity</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.connectivity.lost|Lost\ | |
| \ network connectivity on virtual switch {1}. Physical NIC {2} is down. Affected\ | |
| \ portgroups:{3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Network Connectivity to DVPorts</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.dvport.connectivity.lost|Lost\ | |
| \ network connectivity on DVPorts: {1}. Physical NIC {2} is down.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Network\ | |
| \ Redundancy Degraded on DVPorts</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.dvport.redundancy.degraded|Uplink\ | |
| \ redundancy degraded on DVPorts: {1}. Physical NIC {2} is down.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Network Redundancy on DVPorts</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.dvport.redundancy.lost|Lost\ | |
| \ uplink redundancy on DVPorts: {1}. Physical NIC {2} is down.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>No\ | |
| \ IPv6 TSO support</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.e1000.tso6.notsupported|Guest-initiated\ | |
| \ IPv6 TCP Segmentation Offload (TSO) packets ignored. Manually disable TSO\ | |
| \ inside the guest operating system in virtual machine {1}, or use a different\ | |
| \ virtual adapter.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Invalid\ | |
| \ fenceId configuration on dvPort</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.fence.port.badfenceid|VMkernel\ | |
| \ failed to set fenceId {1} on distributed virtual port {2} on switch {3}.\ | |
| \ Reason: invalid fenceId.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Maximum\ | |
| \ number of fence networks or ports</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.fence.resource.limited|Vmkernel\ | |
| \ failed to set fenceId {1} on distributed virtual port {2} on switch {3}.\ | |
| \ Reason: maximum number of fence networks or ports have been reached.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Switch\ | |
| \ fence property is not set</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.fence.switch.unavailable|Vmkernel\ | |
| \ failed to set fenceId {1} on distributed virtual port {2} on switch {3}.\ | |
| \ Reason: dvSwitch fence property is not set.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Firewall\ | |
| \ configuration operation failed. The changes were not applied.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.firewall.config.failed|Firewall\ | |
| \ configuration operation '{1}' failed. The changes were not applied\ | |
| \ to rule set {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Adding\ | |
| \ port to Firewall failed.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.firewall.port.hookfailed|Adding\ | |
| \ port {1} to Firewall failed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to set gateway</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.gateway.set.failed|Cannot\ | |
| \ connect to the specified gateway {1}. Failed to set it.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Network\ | |
| \ memory pool threshold</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.heap.belowthreshold|{1}\ | |
| \ free size dropped below {2} percent.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>lag\ | |
| \ transition down</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.lag.transition.down|LACP\ | |
| \ warning: LAG {1} on VDS {2} is down.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>No\ | |
| \ peer response</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.peer.noresponse|LACP\ | |
| \ error: No peer response on uplink {1} for VDS {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Current\ | |
| \ teaming policy is incompatible</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.policy.incompatible|LACP\ | |
| \ error: Current teaming policy on VDS {1} is incompatible, supported is IP\ | |
| \ hash only.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Current\ | |
| \ teaming policy is incompatible</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.policy.linkstatus|LACP\ | |
| \ error: Current teaming policy on VDS {1} is incompatible, supported link\ | |
| \ failover detection is link status only.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ is blocked</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.uplink.blocked|LACP\ | |
| \ warning: uplink {1} on VDS {2} is blocked.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ is disconnected</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.uplink.disconnected|LACP\ | |
| \ warning: uplink {1} on VDS {2} got disconnected.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ duplex mode is different</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.uplink.fail.duplex|LACP\ | |
| \ error: Duplex mode across all uplink ports must be full, VDS {1} uplink\ | |
| \ {2} has different mode.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ speed is different</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.uplink.fail.speed|LACP\ | |
| \ error: Speed across all uplink ports must be same, VDS {1} uplink {2} has\ | |
| \ different speed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>All\ | |
| \ uplinks must be active</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.uplink.inactive|LACP\ | |
| \ error: All uplinks on VDS {1} must be active.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>uplink\ | |
| \ transition down</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.lacp.uplink.transition.down|LACP\ | |
| \ warning: uplink {1} on VDS {2} is moved out of link aggregation group.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Invalid\ | |
| \ vmknic specified in /Migrate/Vmknic</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.migrate.bindtovmk|The\ | |
| \ ESX advanced configuration option /Migrate/Vmknic is set to an invalid vmknic:\ | |
| \ {1}. /Migrate/Vmknic specifies a vmknic that vMotion binds to for improved\ | |
| \ performance. Update the configuration option with a valid vmknic. Alternatively,\ | |
| \ if you do not want vMotion to bind to a specific vmknic, remove the invalid\ | |
| \ vmknic and leave the option blank.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Unsupported\ | |
| \ vMotion network latency detected</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.migrate.unsupported.latency|ESXi\ | |
| \ has detected {1}ms round-trip vMotion network latency between host {2} and\ | |
| \ {3}. High latency vMotion networks are supported only if both ESXi hosts\ | |
| \ have been configured for vMotion latency tolerance.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to apply for free ports</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.portset.port.full|Portset\ | |
| \ {1} has reached the maximum number of ports ({2}). Cannot apply for any\ | |
| \ more free ports.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Vlan\ | |
| \ ID of the port is invalid</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.portset.port.vlan.invalidid|{1}\ | |
| \ VLANID {2} is invalid. VLAN ID must be between 0 and 4095.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ NIC connection to switch failed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.proxyswitch.port.unavailable|Virtual\ | |
| \ NIC with hardware address {1} failed to connect to distributed virtual port\ | |
| \ {2} on switch {3}. There are no more ports available on the host proxy switch.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Network\ | |
| \ Redundancy Degraded</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.redundancy.degraded|Uplink\ | |
| \ redundancy degraded on virtual switch {1}. Physical NIC {2} is down. Affected\ | |
| \ portgroups:{3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Network Redundancy</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.redundancy.lost|Lost\ | |
| \ uplink redundancy on virtual switch {1}. Physical NIC {2} is down. Affected\ | |
| \ portgroups:{3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to set MTU on an uplink</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.uplink.mtu.failed|VMkernel\ | |
| \ failed to set the MTU value {1} on the uplink {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ duplicate IP address was detected on a vmknic interface</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.vmknic.ip.duplicate|A\ | |
| \ duplicate IP address was detected for {1} on the interface {2}. The current\ | |
| \ owner is {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Link\ | |
| \ state down</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.vmnic.linkstate.down|Physical\ | |
| \ NIC {1} linkstate is down.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Link\ | |
| \ state unstable</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.vmnic.linkstate.flapping|Taking\ | |
| \ down physical NIC {1} because the link is unstable.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Nic\ | |
| \ Watchdog Reset</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.net.vmnic.watchdog.reset|Uplink\ | |
| \ {1} has recovered from a transient failure due to watchdog timeout</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>NTP\ | |
| \ daemon stopped. Time correction out of bounds.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.ntpd.clock.correction.error|NTP\ | |
| \ daemon stopped. Time correction {1} > {2} seconds. Manually set the\ | |
| \ time and restart ntpd.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Memory\ | |
| \ page retirement requested by platform firmware.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.pageretire.platform.retire.request|Memory\ | |
| \ page retirement requested by platform firmware. FRU ID: {1}. Refer to System\ | |
| \ Hardware Log: {2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Number\ | |
| \ of host physical memory pages selected for retirement exceeds threshold.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.pageretire.selectedmpnthreshold.host.exceeded|Number\ | |
| \ of host physical memory pages that have been selected for retirement ({1})\ | |
| \ exceeds threshold ({2}).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>No\ | |
| \ memory to allocate APD Event</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.apd.event.descriptor.alloc.failed|No\ | |
| \ memory to allocate APD (All Paths Down) event subsystem.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Scsi\ | |
| \ Device close failed.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.close.failed|"Failed\ | |
| \ to close the device {1} properly, plugin {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ detach failed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.detach.failed|Detach\ | |
| \ failed for device :{1}. Exceeded the number of devices that can be detached,\ | |
| \ please cleanup stale detach entries.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to attach filter to device.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.filter.attach.failed|Failed\ | |
| \ to attach filters to device '%s' during registration. Plugin load\ | |
| \ failed or the filter rules are incorrect.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Plugin\ | |
| \ trying to issue command to device does not have a valid storage plugin type.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.io.bad.plugin.type|Bad\ | |
| \ plugin type for device {1}, plugin {2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to obtain INQUIRY data from the device</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.io.inquiry.failed|Failed\ | |
| \ to get standard inquiry for device {1} from Plugin {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Scsi\ | |
| \ device queue parameters incorrectly set.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.io.invalid.disk.qfull.value|QFullSampleSize\ | |
| \ should be bigger than QFullThreshold. LUN queue depth throttling algorithm\ | |
| \ will not function as expected. Please set the QFullSampleSize and QFullThreshold\ | |
| \ disk configuration values in ESX correctly.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Scsi\ | |
| \ Device I/O Latency going high</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.io.latency.high|Device\ | |
| \ {1} performance has deteriorated. I/O latency increased from average value\ | |
| \ of {2} microseconds to {3} microseconds.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>QErr\ | |
| \ cannot be changed on device. Please change it manually on the device if\ | |
| \ possible.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.io.qerr.change.config|QErr\ | |
| \ set to 0x{1} for device {2}. This may cause unexpected behavior. The system\ | |
| \ is not configured to change the QErr setting of device. The QErr value supported\ | |
| \ by system is 0x{3}. Please check the SCSI ChangeQErrSetting configuration\ | |
| \ value for ESX.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Scsi\ | |
| \ Device QErr setting changed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.io.qerr.changed|QErr\ | |
| \ set to 0x{1} for device {2}. This may cause unexpected behavior. The device\ | |
| \ was originally configured to the supported QErr setting of 0x{3}, but this\ | |
| \ has been changed and could not be changed back.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Plugin's\ | |
| \ isLocal entry point failed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.is.local.failed|Failed\ | |
| \ to verify if the device {1} from plugin {2} is a local - not shared - device</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Plugin's\ | |
| \ isPseudo entry point failed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.is.pseudo.failed|Failed\ | |
| \ to verify if the device {1} from plugin {2} is a pseudo device</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Plugin's\ | |
| \ isSSD entry point failed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.is.ssd.failed|Failed\ | |
| \ to verify if the device {1} from plugin {2} is a Solid State Disk device</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Maximum\ | |
| \ number of storage devices</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.limitreached|The\ | |
| \ maximum number of supported devices of {1} has been reached. A device from\ | |
| \ plugin {2} could not be created.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ has been turned off administratively.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.state.off|Device\ | |
| \ {1}, has been turned off administratively.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ has been removed or is permanently inaccessible.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.state.permanentloss|Device\ | |
| \ {1} has been removed or is permanently inaccessible. Affected datastores\ | |
| \ (if any): {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Permanently\ | |
| \ inaccessible device has no more opens.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.state.permanentloss.noopens|Permanently\ | |
| \ inaccessible device {1} has no more opens. It is now safe to unmount datastores\ | |
| \ (if any) {2} and delete the device.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ has been plugged back in after being marked permanently inaccessible.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.state.permanentloss.pluggedback|Device\ | |
| \ {1} has been plugged back in after being marked permanently inaccessible.\ | |
| \ No data consistency guarantees.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ has been removed or is permanently inaccessible.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.state.permanentloss.withreservationheld|Device\ | |
| \ {1} has been removed or is permanently inaccessible, while holding a reservation.\ | |
| \ Affected datastores (if any): {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Thin\ | |
| \ Provisioned Device Nearing Capacity</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.device.thinprov.atquota|Space\ | |
| \ utilization on thin-provisioned device {1} exceeded configured threshold.\ | |
| \ Affected datastores (if any): {2}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Maximum\ | |
| \ number of storage paths</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.scsipath.limitreached|The\ | |
| \ maximum number of supported paths of {1} has been reached. Path {2} could\ | |
| \ not be added.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Storage\ | |
| \ plugin of unsupported type tried to register.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.scsi.unsupported.plugin.type|Scsi\ | |
| \ Device Allocation not supported for plugin type {1}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>All\ | |
| \ paths are down</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.apd.start|Device\ | |
| \ or filesystem with identifier {1} has entered the All Paths Down state.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>All\ | |
| \ Paths Down timed out, I/Os will be fast failed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.apd.timeout|Device\ | |
| \ or filesystem with identifier {1} has entered the All Paths Down Timeout\ | |
| \ state after being in the All Paths Down state for {2} seconds. I/Os will\ | |
| \ now be fast failed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Frequent\ | |
| \ PowerOn Reset Unit Attention of Storage Path</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.connectivity.devicepor|Frequent\ | |
| \ PowerOn Reset Unit Attentions are occurring on device {1}. This might indicate\ | |
| \ a storage problem. Affected datastores: {2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Storage Connectivity</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.connectivity.lost|Lost\ | |
| \ connectivity to storage device {1}. Path {2} is down. Affected datastores:\ | |
| \ {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Frequent\ | |
| \ PowerOn Reset Unit Attention of Storage Path</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.connectivity.pathpor|Frequent\ | |
| \ PowerOn Reset Unit Attentions are occurring on path {1}. This might indicate\ | |
| \ a storage problem. Affected device: {2}. Affected datastores: {3}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Frequent\ | |
| \ State Changes of Storage Path</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.connectivity.pathstatechanges|Frequent\ | |
| \ path state changes are occurring for path {1}. This might indicate a storage\ | |
| \ problem. Affected device: {2}. Affected datastores: {3}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>iSCSI\ | |
| \ discovery target login connection problem</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.iscsi.discovery.connect.error|iSCSI\ | |
| \ discovery to {1} on {2} failed. The iSCSI Initiator could not establish\ | |
| \ a network connection to the discovery address.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>iSCSI\ | |
| \ Discovery target login error</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.iscsi.discovery.login.error|iSCSI\ | |
| \ discovery to {1} on {2} failed. The Discovery target returned a login error\ | |
| \ of: {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>iSCSI\ | |
| \ Target login connection problem</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.iscsi.target.connect.error|Login\ | |
| \ to iSCSI target {1} on {2} failed. The iSCSI initiator could not establish\ | |
| \ a network connection to the target.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>iSCSI\ | |
| \ Target login error</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.iscsi.target.login.error|Login\ | |
| \ to iSCSI target {1} on {2} failed. Target returned login error of: {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>iSCSI\ | |
| \ target permanently removed</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.iscsi.target.permanently.lost|The\ | |
| \ iSCSI target {2} was permanently removed from {1}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Degraded\ | |
| \ Storage Path Redundancy</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.redundancy.degraded|Path\ | |
| \ redundancy to storage device {1} degraded. Path {2} is down. Affected datastores:\ | |
| \ {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Storage Path Redundancy</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.storage.redundancy.lost|Lost\ | |
| \ path redundancy to storage device {1}. Path {2} is down. Affected datastores:\ | |
| \ {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>System\ | |
| \ logging is not configured.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>System\ | |
| \ logging is not configured on host {host.name}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.syslog.config|System\ | |
| \ logging is not configured on host {host.name}. Please check Syslog options\ | |
| \ for the host under Configuration -> Software -> Advanced Settings\ | |
| \ in vSphere client.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>System\ | |
| \ logs are stored on non-persistent storage.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>System\ | |
| \ logs on host {host.name} are stored on non-persistent storage.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.syslog.nonpersistent|System\ | |
| \ logs on host {host.name} are stored on non-persistent storage. Consult product\ | |
| \ documentation to configure a syslog server or a scratch partition.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ VFAT filesystem is full.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vfat.filesystem.full.other|The\ | |
| \ VFAT filesystem {1} (UUID {2}) is full.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ VFAT filesystem, being used as the host's scratch partition, is full.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vfat.filesystem.full.scratch|The\ | |
| \ host's scratch partition, which is the VFAT filesystem {1} (UUID {2}),\ | |
| \ is full.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>An\ | |
| \ operation on the root filesystem has failed.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.visorfs.failure|An\ | |
| \ operation on the root filesystem has failed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>The\ | |
| \ root filesystem's file table is full.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.visorfs.inodetable.full|The\ | |
| \ root filesystem's file table is full. As a result, the file {1} could\ | |
| \ not be created by the application '{2}'.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ ramdisk is full.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.visorfs.ramdisk.full|The\ | |
| \ ramdisk '{1}' is full. As a result, the file {2} could not be\ | |
| \ written.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ ramdisk's file table is full.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.visorfs.ramdisk.inodetable.full|The\ | |
| \ file table of the ramdisk '{1}' is full. As a result, the file\ | |
| \ {2} could not be created by the application '{3}'.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ VM could not fault in the a page. The VM is terminated as further progress\ | |
| \ is impossible.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vm.kill.unexpected.fault.failure|The\ | |
| \ VM using the config file {1} could not fault in a guest physical page from\ | |
| \ the hypervisor level swap file at {2}. The VM is terminated as further progress\ | |
| \ is impossible.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ VM did not respond to swap actions and is forcefully powered off to prevent\ | |
| \ system instability.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vm.kill.unexpected.forcefulPageRetire|The\ | |
| \ VM using the config file {1} contains the host physical page {2} which was\ | |
| \ scheduled for immediate retirement. To avoid system instability the VM is\ | |
| \ forcefully powered off.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ VM did not respond to swap actions and is forcefully powered off to prevent\ | |
| \ system instability.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vm.kill.unexpected.noSwapResponse|The\ | |
| \ VM using the config file {1} did not respond to {2} swap actions in {3}\ | |
| \ seconds and is forcefully powered off to prevent system instability.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ VM is allocating too many pages while system is critically low in free memory.\ | |
| \ It is forcefully terminated to prevent system instability.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vm.kill.unexpected.vmtrack|The\ | |
| \ VM using the config file {1} is allocating too many pages while system is\ | |
| \ critically low in free memory. It is forcefully terminated to prevent system\ | |
| \ instability.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ Backing VMFS has lost ATS Support</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.ats.support.lost|ATS-Only\ | |
| \ VMFS volume '{1}' not mounted. Host does not support ATS or ATS\ | |
| \ initialization has failed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Locked By Remote Host</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.error.volume.is.locked|Volume\ | |
| \ on device {1} is locked, possibly because some remote host encountered an\ | |
| \ error during a volume operation and could not recover.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ backing an extent of a file system is offline.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.extent.offline|An\ | |
| \ attached device {1} may be offline. The file system {2} is now in a degraded\ | |
| \ state. While the datastore is still available, parts of data that reside\ | |
| \ on the extent that went offline might be inaccessible.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ backing an extent of a file system came online</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.extent.online|Device\ | |
| \ {1} backing file system {2} came online. This extent was previously offline.\ | |
| \ All resources on this device are now available.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Volume Connectivity Restored</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.heartbeat.recovered|Successfully\ | |
| \ restored access to volume {1} ({2}) following connectivity issues.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Volume Connectivity Degraded</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.heartbeat.timedout|Lost\ | |
| \ access to volume {1} ({2}) due to connectivity issues. Recovery attempt\ | |
| \ is in progress and outcome will be reported shortly.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Volume Connectivity Lost</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.heartbeat.unrecoverable|Lost\ | |
| \ connectivity to volume {1} ({2}) and subsequent recovery attempts have failed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>No\ | |
| \ Space To Create VMFS Journal</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.journal.createfailed|No\ | |
| \ space for journal on volume {1} ({2}). Opening volume in read-only metadata\ | |
| \ mode with limited write support.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Lock Corruption Detected</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.lock.corruptondisk|At\ | |
| \ least one corrupt on-disk lock was detected on volume {1} ({2}). Other regions\ | |
| \ of the volume might be damaged too.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ connection to NFS server</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.nfs.server.disconnect|Lost\ | |
| \ connection to server {1} mount point {2} mounted as {3} ({4}).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ connection to NFS server</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.nfs.server.restored|Restored\ | |
| \ connection to server {1} mount point {2} mounted as {3} ({4}).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Resource Corruption Detected</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmfs.resource.corruptondisk|At\ | |
| \ least one corrupt resource metadata region was detected on volume {1} ({2}).\ | |
| \ Other regions of the volume might be damaged too.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Remote\ | |
| \ logging host has become unreachable.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmsyslogd.remote.failure|The\ | |
| \ host "{1}" has become unreachable. Remote logging to this host\ | |
| \ has stopped.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Logging\ | |
| \ to storage has failed.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmsyslogd.storage.failure|Logging\ | |
| \ to storage has failed. Logs are no longer being stored locally on this\ | |
| \ host.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>The\ | |
| \ configured log directory cannot be used. The default directory will be\ | |
| \ used instead.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmsyslogd.storage.logdir.invalid|The\ | |
| \ configured log directory {1} cannot be used. The default directory {2}\ | |
| \ will be used instead.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Log\ | |
| \ daemon has failed for an unexpected reason.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vmsyslogd.unexpected|Log\ | |
| \ daemon has failed for an unexpected reason: {1}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Vpxa\ | |
| \ crashed and a core file was created.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vpxa.core.dumped|{1}\ | |
| \ crashed ({2} time(s) so far) and a core file might have been created at\ | |
| \ {3}. This might have caused connections to the host to be dropped.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>VSAN\ | |
| \ clustering services have been disabled.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>VSAN\ | |
| \ clustering and directory services have been disabled thus will be no longer\ | |
| \ available.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vsan.clustering.disabled|VSAN\ | |
| \ clustering and directory services have been disabled thus will be no longer\ | |
| \ available.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ vmknic added to VSAN network configuration doesn't have valid IP. Network\ | |
| \ is not ready.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>vmknic\ | |
| \ {1} that is currently configured to be used with VSAN doesn't have\ | |
| \ an IP address yet. There are no other active network configuration and therefore\ | |
| \ the VSAN node doesn't have network connectivity.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vsan.net.not.ready|vmknic\ | |
| \ {1} that is currently configured to be used with VSAN doesn't have\ | |
| \ an IP address yet. There are no other active network configuration and therefore\ | |
| \ the VSAN node doesn't have network connectivity.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VSAN\ | |
| \ doesn't haven any redundancy in its network configuration.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>VSAN\ | |
| \ network configuration doesn't have any redundancy. This might be a\ | |
| \ problem if further network configuration is removed.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vsan.net.redundancy.lost|VSAN\ | |
| \ network configuration doesn't have any redundancy. This might be a\ | |
| \ problem if further network configuration is removed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VSAN\ | |
| \ is operating on reduced network redundancy.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>VSAN\ | |
| \ network configuration redundancy has been reduced. This might be a problem\ | |
| \ if further network configuration is removed.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vsan.net.redundancy.reduced|VSAN\ | |
| \ network configuration redundancy has been reduced. This might be a problem\ | |
| \ if further network configuration is removed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VSAN\ | |
| \ doesn't have any networking configuration for use.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>VSAN\ | |
| \ doesn't have any network configuration. This can severely impact several\ | |
| \ objects in the VSAN datastore.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vsan.no.network.connectivity|VSAN\ | |
| \ doesn't have any network configuration. This can severely impact several\ | |
| \ objects in the VSAN datastore.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>A\ | |
| \ vmknic added to VSAN network configuration doesn't have valid IP. It\ | |
| \ will not be in use.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>vmknic\ | |
| \ {1} that is currently configured to be used with VSAN doesn't have\ | |
| \ an IP address yet. However, there are other network configuration which\ | |
| \ are active. If those configurations are removed that may cause problems.</formatOnHost><formatOnVm></formatOnVm><fullFormat>esx.problem.vsan.vmknic.not.ready|vmknic\ | |
| \ {1} that is currently configured to be used with VSAN doesn't have\ | |
| \ an IP address yet. However, there are other network configuration which\ | |
| \ are active. If those configurations are removed that may cause problems.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Application\ | |
| \ consistent delta completed.</description><category>info</category><formatOnDatacenter>Application\ | |
| \ consistent delta completed for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Application\ | |
| \ consistent delta completed for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Application\ | |
| \ consistent delta completed for virtual machine {vm.name}.</formatOnHost><formatOnVm>Application\ | |
| \ consistent delta completed.</formatOnVm><fullFormat>hbr.primary.AppQuiescedDeltaCompletedEvent|Application\ | |
| \ consistent delta completed for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name} ({bytes} bytes transferred)</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Connection\ | |
| \ to VR Server restored.</description><category>info</category><formatOnDatacenter>Connection\ | |
| \ to VR Server restored for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Connection\ | |
| \ to VR Server restored for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Connection\ | |
| \ to VR Server restored for virtual machine {vm.name}.</formatOnHost><formatOnVm>Connection\ | |
| \ to VR Server restored.</formatOnVm><fullFormat>hbr.primary.ConnectionRestoredToHbrServerEvent|Connection\ | |
| \ to VR Server restored for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Delta\ | |
| \ aborted.</description><category>warning</category><formatOnDatacenter>Delta\ | |
| \ aborted for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}:\ | |
| \ {[email protected]}</formatOnDatacenter><formatOnComputeResource>Delta\ | |
| \ aborted for virtual machine {vm.name} on host {host.name}: {[email protected]}</formatOnComputeResource><formatOnHost>Delta\ | |
| \ aborted for virtual machine {vm.name}: {[email protected]}</formatOnHost><formatOnVm>Delta\ | |
| \ aborted: {[email protected]}</formatOnVm><fullFormat>hbr.primary.DeltaAbortedEvent|Delta\ | |
| \ aborted for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}: {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Delta\ | |
| \ completed.</description><category>info</category><formatOnDatacenter>Delta\ | |
| \ completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Delta\ | |
| \ completed for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Delta\ | |
| \ completed for virtual machine {vm.name}.</formatOnHost><formatOnVm>Delta\ | |
| \ completed.</formatOnVm><fullFormat>hbr.primary.DeltaCompletedEvent|Delta\ | |
| \ completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} ({bytes} bytes transferred).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Delta\ | |
| \ started.</description><category>info</category><formatOnDatacenter>Delta\ | |
| \ started by {userName} for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Delta\ | |
| \ started by {userName} for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Delta\ | |
| \ started by {userName} for virtual machine {vm.name}.</formatOnHost><formatOnVm>Delta\ | |
| \ started by {userName}.</formatOnVm><fullFormat>hbr.primary.DeltaStartedEvent|Delta\ | |
| \ started by {userName} for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>File\ | |
| \ system consistent delta completed.</description><category>warning</category><formatOnDatacenter>File\ | |
| \ system consistent delta completed for virtual machine {vm.name} on host\ | |
| \ {host.name} in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>File\ | |
| \ system consistent delta completed for virtual machine {vm.name} on host\ | |
| \ {host.name}.</formatOnComputeResource><formatOnHost>File system consistent\ | |
| \ delta completed for virtual machine {vm.name}.</formatOnHost><formatOnVm>File\ | |
| \ system consistent delta completed.</formatOnVm><fullFormat>hbr.primary.FSQuiescedDeltaCompletedEvent|File\ | |
| \ system consistent delta completed for virtual machine {vm.name} on host\ | |
| \ {host.name} in cluster {computeResource.name} in {datacenter.name} ({bytes}\ | |
| \ bytes transferred)</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to start delta.</description><category>error</category><formatOnDatacenter>Failed\ | |
| \ to start delta for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}: {[email protected]}</formatOnDatacenter><formatOnComputeResource>Failed\ | |
| \ to start delta for virtual machine {vm.name} on host {host.name}: {[email protected]}</formatOnComputeResource><formatOnHost>Failed\ | |
| \ to start delta for virtual machine {vm.name}: {[email protected]}</formatOnHost><formatOnVm>Failed\ | |
| \ to start delta: {[email protected]}</formatOnVm><fullFormat>hbr.primary.FailedToStartDeltaEvent|Failed\ | |
| \ to start delta for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}: {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Failed\ | |
| \ to start full sync.</description><category>error</category><formatOnDatacenter>Failed\ | |
| \ to start full sync for virtual machine {vm.name} on host {host.name} in\ | |
| \ cluster {computeResource.name}: {[email protected]}</formatOnDatacenter><formatOnComputeResource>Failed\ | |
| \ to start full sync for virtual machine {vm.name} on host {host.name}: {[email protected]}</formatOnComputeResource><formatOnHost>Failed\ | |
| \ to start full sync for virtual machine {vm.name}: {[email protected]}</formatOnHost><formatOnVm>Failed\ | |
| \ to start full sync: {[email protected]}</formatOnVm><fullFormat>hbr.primary.FailedToStartSyncEvent|Failed\ | |
| \ to start full sync for virtual machine {vm.name} on host {host.name} in\ | |
| \ cluster {computeResource.name} in {datacenter.name}: {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Disk\ | |
| \ replication configuration is invalid.</description><category>error</category><formatOnDatacenter>Replication\ | |
| \ configuration is invalid for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}, disk {diskKey}: {reasonForFault.@enum.fault.ReplicationDiskConfigFault.ReasonForFault}</formatOnDatacenter><formatOnComputeResource>Replication\ | |
| \ configuration is invalid for virtual machine {vm.name} on host {host.name}\ | |
| \ disk {diskKey}: {reasonForFault.@enum.fault.ReplicationDiskConfigFault.ReasonForFault}</formatOnComputeResource><formatOnHost>Replication\ | |
| \ configuration is invalid for virtual machine {vm.name} disk {diskKey}: {reasonForFault.@enum.fault.ReplicationDiskConfigFault.ReasonForFault}</formatOnHost><formatOnVm>Replication\ | |
| \ configuration is invalid for disk {diskKey}: {reasonForFault.@enum.fault.ReplicationDiskConfigFault.ReasonForFault}</formatOnVm><fullFormat>hbr.primary.InvalidDiskReplicationConfigurationEvent|Replication\ | |
| \ configuration is invalid for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}, disk {diskKey}:\ | |
| \ {reasonForFault.@enum.fault.ReplicationDiskConfigFault.ReasonForFault}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ machine replication configuration is invalid.</description><category>error</category><formatOnDatacenter>Replication\ | |
| \ configuration is invalid for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}: {reasonForFault.@enum.fault.ReplicationVmConfigFault.ReasonForFault}</formatOnDatacenter><formatOnComputeResource>Replication\ | |
| \ configuration is invalid for virtual machine {vm.name} on host {host.name}:\ | |
| \ {reasonForFault.@enum.fault.ReplicationVmConfigFault.ReasonForFault}</formatOnComputeResource><formatOnHost>Replication\ | |
| \ configuration is invalid for virtual machine {vm.name}: {reasonForFault.@enum.fault.ReplicationVmConfigFault.ReasonForFault}</formatOnHost><formatOnVm>Replication\ | |
| \ configuration is invalid: {reasonForFault.@enum.fault.ReplicationVmConfigFault.ReasonForFault}</formatOnVm><fullFormat>hbr.primary.InvalidVmReplicationConfigurationEvent|Replication\ | |
| \ configuration is invalid for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}: {reasonForFault.@enum.fault.ReplicationVmConfigFault.ReasonForFault}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>No\ | |
| \ connection to VR Server.</description><category>warning</category><formatOnDatacenter>No\ | |
| \ connection to VR Server for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}: {[email protected]}</formatOnDatacenter><formatOnComputeResource>No\ | |
| \ connection to VR Server for virtual machine {vm.name} on host {host.name}:\ | |
| \ {[email protected]}</formatOnComputeResource><formatOnHost>No\ | |
| \ connection to VR Server for virtual machine {vm.name}: {[email protected]}</formatOnHost><formatOnVm>No\ | |
| \ connection to VR Server: {[email protected]}</formatOnVm><fullFormat>hbr.primary.NoConnectionToHbrServerEvent|No\ | |
| \ connection to VR Server for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}: {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VR\ | |
| \ Server error: {[email protected]}</description><category>error</category><formatOnDatacenter>VR\ | |
| \ Server error for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}: {[email protected]}</formatOnDatacenter><formatOnComputeResource>VR\ | |
| \ Server error for virtual machine {vm.name} on host {host.name}: {[email protected]}</formatOnComputeResource><formatOnHost>VR\ | |
| \ Server error for virtual machine {vm.name}: {[email protected]}</formatOnHost><formatOnVm>VR\ | |
| \ Server error: {[email protected]}</formatOnVm><fullFormat>hbr.primary.NoProgressWithHbrServerEvent|VR\ | |
| \ Server error for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}: {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Quiescing\ | |
| \ is not supported for this virtual machine.</description><category>warning</category><formatOnDatacenter>Quiescing\ | |
| \ is not supported for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Quiescing\ | |
| \ is not supported for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Quiescing\ | |
| \ is not supported for virtual machine {vm.name}.</formatOnHost><formatOnVm>Quiescing\ | |
| \ is not supported for this virtual machine.</formatOnVm><fullFormat>hbr.primary.QuiesceNotSupported|Quiescing\ | |
| \ is not supported for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Full\ | |
| \ sync completed.</description><category>info</category><formatOnDatacenter>Full\ | |
| \ sync completed for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Full\ | |
| \ sync completed for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Full\ | |
| \ sync completed for virtual machine {vm.name}.</formatOnHost><formatOnVm>Full\ | |
| \ sync completed.</formatOnVm><fullFormat>hbr.primary.SyncCompletedEvent|Full\ | |
| \ sync completed for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name} ({bytes} bytes transferred).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Full\ | |
| \ sync started.</description><category>info</category><formatOnDatacenter>Full\ | |
| \ sync started for virtual machine {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Full\ | |
| \ sync started for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Full\ | |
| \ sync started for virtual machine {vm.name}.</formatOnHost><formatOnVm>Full\ | |
| \ sync started.</formatOnVm><fullFormat>hbr.primary.SyncStartedEvent|Full\ | |
| \ sync started by {userName} for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Delta\ | |
| \ completed.</description><category>warning</category><formatOnDatacenter>Delta\ | |
| \ completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Delta\ | |
| \ completed for virtual machine {vm.name} on host {host.name}.</formatOnComputeResource><formatOnHost>Delta\ | |
| \ completed for virtual machine {vm.name}.</formatOnHost><formatOnVm>Delta\ | |
| \ completed.</formatOnVm><fullFormat>hbr.primary.UnquiescedDeltaCompletedEvent|Delta\ | |
| \ completed for virtual machine {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} ({bytes} bytes transferred).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Replication\ | |
| \ configuration changed.</description><category>info</category><formatOnDatacenter>Replication\ | |
| \ configuration changed for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} ({numDisks} disks, {rpo} minutes RPO,\ | |
| \ VR Server is {vrServerAddress}:{vrServerPort}).</formatOnDatacenter><formatOnComputeResource>Replication\ | |
| \ configuration changed for virtual machine {vm.name} on host {host.name}\ | |
| \ ({numDisks} disks, {rpo} minutes RPO, VR Server is {vrServerAddress}:{vrServerPort}).</formatOnComputeResource><formatOnHost>Replication\ | |
| \ configuration changed for virtual machine {vm.name} ({numDisks} disks, {rpo}\ | |
| \ minutes RPO, VR Server is {vrServerAddress}:{vrServerPort}).</formatOnHost><formatOnVm>Replication\ | |
| \ configuration changed ({numDisks} disks, {rpo} minutes RPO, VR Server is\ | |
| \ {vrServerAddress}:{vrServerPort}).</formatOnVm><fullFormat>hbr.primary.VmReplicationConfigurationChangedEvent|Replication\ | |
| \ configuration changed for virtual machine {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name} ({numDisks} disks,\ | |
| \ {rpo} minutes RPO, VR Server is {vrServerAddress}:{vrServerPort}).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AccountCreatedEvent</key><description>Account\ | |
| \ created</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>An\ | |
| \ account was created</formatOnHost><formatOnVm></formatOnVm><fullFormat>An\ | |
| \ account was created on host {host.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.AccountCreatedEvent">\t\ | |
| \t<description>\t\t\tAn account has been created on the host\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>AccountRemovedEvent</key><description>Account\ | |
| \ removed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Account\ | |
| \ {account} was removed</formatOnHost><formatOnVm></formatOnVm><fullFormat>Account\ | |
| \ {account} was removed on host {host.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.AccountRemovedEvent">\t\ | |
| \t<description>\t\t\tAn account has been removed from the host\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>AccountUpdatedEvent</key><description>Account\ | |
| \ updated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>An\ | |
| \ account was updated</formatOnHost><formatOnVm></formatOnVm><fullFormat>An\ | |
| \ account was updated on host {host.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.AccountUpdatedEvent">\t\ | |
| \t<description>\t\t\tAn account has been updated on the host\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>AdminPasswordNotChangedEvent</key><description>Administrator\ | |
| \ password not changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ default password for the root user has not been changed</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ default password for the root user on the host {host.name} has not been\ | |
| \ changed</fullFormat><longDescription>\t<EventLongDescription id="vim.event.AdminPasswordNotChangedEvent">\t\ | |
| \t<description>\t\t\tThe default password for the Administrator user\ | |
| \ on the host has not been changed\t\t</description>\t\t<cause>\t\ | |
| \t\t<description>\t\t\tYou have not changed the password for the Administrator\ | |
| \ user on the host so the \t\t\tdefault password is still active\t\t\t</description>\t\ | |
| \t\t<action>\t\t\t\tChange the password for the Administrator user on\ | |
| \ the host\t\t\t</action>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>AlarmAcknowledgedEvent</key><description>Alarm\ | |
| \ acknowledged</description><category>info</category><formatOnDatacenter>Acknowledged\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnDatacenter><formatOnComputeResource>Acknowledged\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnComputeResource><formatOnHost>Acknowledged\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnHost><formatOnVm>Acknowledged\ | |
| \ alarm '{alarm.name}'</formatOnVm><fullFormat>Acknowledged alarm\ | |
| \ '{alarm.name}' on {entity.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmActionTriggeredEvent</key><description>Alarm\ | |
| \ action triggered</description><category>info</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}' on {entity.name} triggered an action</formatOnDatacenter><formatOnComputeResource>Alarm\ | |
| \ '{alarm.name}' on {entity.name} triggered an action</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Alarm\ | |
| \ '{alarm.name}' on {entity.name} triggered an action</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmClearedEvent</key><description>Alarm\ | |
| \ cleared</description><category>info</category><formatOnDatacenter>Manually\ | |
| \ cleared alarm '{alarm.name}' on {entity.name} from {[email protected]}</formatOnDatacenter><formatOnComputeResource>Manually\ | |
| \ cleared alarm '{alarm.name}' on {entity.name} from {[email protected]}</formatOnComputeResource><formatOnHost>Manually\ | |
| \ cleared alarm '{alarm.name}' on {entity.name} from {[email protected]}</formatOnHost><formatOnVm>Manually\ | |
| \ cleared alarm '{alarm.name}' from {[email protected]}</formatOnVm><fullFormat>Manually\ | |
| \ cleared alarm '{alarm.name}' on {entity.name} from {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmCreatedEvent</key><description>Alarm\ | |
| \ created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnComputeResource><formatOnHost>Created\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnHost><formatOnVm>Created\ | |
| \ alarm '{alarm.name}'</formatOnVm><fullFormat>Created alarm '{alarm.name}'\ | |
| \ on {entity.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmEmailCompletedEvent</key><description>Alarm\ | |
| \ email sent</description><category>info</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}' on {entity.name} sent email to {to}</formatOnDatacenter><formatOnComputeResource>Alarm\ | |
| \ '{alarm.name}' on {entity.name} sent email to {to}</formatOnComputeResource><formatOnHost>Alarm\ | |
| \ '{alarm.name}' on {entity.name} sent email to {to}</formatOnHost><formatOnVm>Alarm\ | |
| \ '{alarm.name}' sent email to {to}</formatOnVm><fullFormat>Alarm\ | |
| \ '{alarm.name}' on {entity.name} sent email to {to}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmEmailFailedEvent</key><description>Cannot\ | |
| \ send alarm email</description><category>error</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}' on {entity.name} cannot send email to {to}</formatOnDatacenter><formatOnComputeResource>Alarm\ | |
| \ '{alarm.name}' on {entity.name} cannot send email to {to}</formatOnComputeResource><formatOnHost>Alarm\ | |
| \ '{alarm.name}' on {entity.name} cannot send email to {to}</formatOnHost><formatOnVm>Alarm\ | |
| \ '{alarm.name}' cannot send email to {to}</formatOnVm><fullFormat>Alarm\ | |
| \ '{alarm.name}' on {entity.name} cannot send email to {to}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.AlarmEmailFailedEvent">\ | |
| \ <description> An error occurred while sending email\ | |
| \ notification of a triggered alarm </description> <cause>\ | |
| \ <description>Failed to send email for a triggered alarm</description>\ | |
| \ <action>Check the vCenter Server SMTP settings for sending\ | |
| \ email notifications</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>AlarmEvent</key><description><Alarm\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmReconfiguredEvent</key><description>Alarm\ | |
| \ reconfigured</description><category>info</category><formatOnDatacenter>Reconfigured\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnDatacenter><formatOnComputeResource>Reconfigured\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnComputeResource><formatOnHost>Reconfigured\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnHost><formatOnVm>Reconfigured\ | |
| \ alarm '{alarm.name}'</formatOnVm><fullFormat>Reconfigured alarm\ | |
| \ '{alarm.name}' on {entity.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.AlarmReconfiguredEvent">\ | |
| \ <description> An alarm has been reconfigured \ | |
| \ </description> <cause> <description>A\ | |
| \ user has reconfigured an alarm</description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>AlarmRemovedEvent</key><description>Alarm\ | |
| \ removed</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnDatacenter><formatOnComputeResource>Removed\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnComputeResource><formatOnHost>Removed\ | |
| \ alarm '{alarm.name}' on {entity.name}</formatOnHost><formatOnVm>Removed\ | |
| \ alarm '{alarm.name}'</formatOnVm><fullFormat>Removed alarm '{alarm.name}'\ | |
| \ on {entity.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmScriptCompleteEvent</key><description>Alarm\ | |
| \ script completed</description><category>info</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}' on {entity.name} ran script {script}</formatOnDatacenter><formatOnComputeResource>Alarm\ | |
| \ '{alarm.name}' on {entity.name} ran script {script}</formatOnComputeResource><formatOnHost>Alarm\ | |
| \ '{alarm.name}' on {entity.name} ran script {script}</formatOnHost><formatOnVm>Alarm\ | |
| \ '{alarm.name}' ran script {script}</formatOnVm><fullFormat>Alarm\ | |
| \ '{alarm.name}' on {entity.name} ran script {script}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmScriptFailedEvent</key><description>Alarm\ | |
| \ script not completed</description><category>error</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}' on {entity.name} did not complete script: {reason.msg}</formatOnDatacenter><formatOnComputeResource>Alarm\ | |
| \ '{alarm.name}' on {entity.name} did not complete script: {reason.msg}</formatOnComputeResource><formatOnHost>Alarm\ | |
| \ '{alarm.name}' on {entity.name} did not complete script: {reason.msg}</formatOnHost><formatOnVm>Alarm\ | |
| \ '{alarm.name}' did not complete script: {reason.msg}</formatOnVm><fullFormat>Alarm\ | |
| \ '{alarm.name}' on {entity.name} did not complete script: {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.AlarmScriptFailedEvent">\ | |
| \ <description> The vCenter Server logs this event if\ | |
| \ an error occurs while running a script after an alarm triggers. </description>\ | |
| \ <cause> <description>There was an error running\ | |
| \ the script</description> <action>Fix the script or\ | |
| \ failure condition</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>AlarmSnmpCompletedEvent</key><description>Alarm\ | |
| \ SNMP trap sent</description><category>info</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}': an SNMP trap for entity {entity.name} was sent</formatOnDatacenter><formatOnComputeResource>Alarm\ | |
| \ '{alarm.name}': an SNMP trap for entity {entity.name} was sent</formatOnComputeResource><formatOnHost>Alarm\ | |
| \ '{alarm.name}': an SNMP trap for entity {entity.name} was sent</formatOnHost><formatOnVm>Alarm\ | |
| \ '{alarm.name}': an SNMP trap was sent</formatOnVm><fullFormat>Alarm\ | |
| \ '{alarm.name}': an SNMP trap for entity {entity.name} was sent</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlarmSnmpFailedEvent</key><description>Alarm\ | |
| \ SNMP trap not sent</description><category>error</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}' on entity {entity.name} did not send SNMP trap:\ | |
| \ {reason.msg}</formatOnDatacenter><formatOnComputeResource>Alarm '{alarm.name}'\ | |
| \ on entity {entity.name} did not send SNMP trap: {reason.msg}</formatOnComputeResource><formatOnHost>Alarm\ | |
| \ '{alarm.name}' on entity {entity.name} did not send SNMP trap:\ | |
| \ {reason.msg}</formatOnHost><formatOnVm>Alarm '{alarm.name}' did\ | |
| \ not send SNMP trap: {reason.msg}</formatOnVm><fullFormat>Alarm '{alarm.name}'\ | |
| \ on entity {entity.name} did not send SNMP trap: {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.AlarmSnmpFailedEvent">\ | |
| \ <description> The vCenter Server logs this event if\ | |
| \ an error occurs while sending an SNMP trap when an alarm triggers. \ | |
| \ </description> <cause> <description>An\ | |
| \ SNMP trap could not be sent for a triggered alarm</description> \ | |
| \ <action>Check the vCenter Server SNMP settings. Make sure\ | |
| \ that the vCenter Server network can handle SNMP packets.</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>AlarmStatusChangedEvent</key><description>Alarm\ | |
| \ status changed</description><category>info</category><formatOnDatacenter>Alarm\ | |
| \ '{alarm.name}' on {entity.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnDatacenter><formatOnComputeResource>Alarm\ | |
| \ '{alarm.name}' on {entity.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnComputeResource><formatOnHost>Alarm\ | |
| \ '{alarm.name}' on {entity.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnHost><formatOnVm>Alarm '{alarm.name}'\ | |
| \ changed from {[email protected]} to {[email protected]}</formatOnVm><fullFormat>Alarm\ | |
| \ '{alarm.name}' on {entity.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AllVirtualMachinesLicensedEvent</key><description>All\ | |
| \ virtual machines are licensed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>All\ | |
| \ running virtual machines are licensed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AlreadyAuthenticatedSessionEvent</key><description>Already\ | |
| \ authenticated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>User\ | |
| \ cannot logon since the user is already logged on</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>AuthorizationEvent</key><description><Authorization\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>BadUsernameSessionEvent</key><description>Invalid\ | |
| \ user name</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ login {userName}@{ipAddress}</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.BadUsernameSessionEvent">\t\t<description>\t\ | |
| \t\tA user attempted to log in with an unknown or invalid username\t\t</description>\t\ | |
| \t<cause> <description>\t\t\tThe username is unknown\ | |
| \ to the system\t\t\t</description> <action>\t\t\t\ | |
| Use a username that is included in the system user directory\t\t\t</action>\ | |
| \ <action>\t\t\tOn Linux, verify that the user directory\ | |
| \ is correctly configured\t\t\t</action> <action>\t\ | |
| \t\tIf you are using Active Directory, check the health of the domain controller\t\ | |
| \t\t</action> </cause> <cause> \ | |
| \ <description>\t\t\tThe user provided an invalid password\t\t\t</description>\ | |
| \ <action>\t\t\tSupply the correct password\t\t\t</action>\ | |
| \ </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>CanceledHostOperationEvent</key><description>Canceled\ | |
| \ host operation</description><category>info</category><formatOnDatacenter>The\ | |
| \ operation performed on host {host.name} was canceled</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ operation performed on host {host.name} was canceled</formatOnComputeResource><formatOnHost>The\ | |
| \ operation was canceled</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ operation performed on host {host.name} in {datacenter.name} was canceled</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.CanceledHostOperationEvent">\t\ | |
| \t<description>\t\t\tAn operation performed on the host was canceled\t\ | |
| \t</description>\t\t<cause>\t\t\t<description>\t\tA previous\ | |
| \ event in the sequence of events will provide more information about the\ | |
| \ \t\tcause of this cancellation\t\t\t</description>\t\t</cause>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ChangeOwnerOfFileEvent</key><description>Change\ | |
| \ owner of file</description><category>info</category><formatOnDatacenter>Changed\ | |
| \ ownership of file name {filename} from {oldOwner} to {newOwner} on {host.name}\ | |
| \ in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Changed\ | |
| \ ownership of file name {filename} from {oldOwner} to {newOwner} on {host.name}.</formatOnComputeResource><formatOnHost>Changed\ | |
| \ ownership of file name {filename} from {oldOwner} to {newOwner}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Changed\ | |
| \ ownership of file name {filename} from {oldOwner} to {newOwner} on {host.name}\ | |
| \ in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ChangeOwnerOfFileFailedEvent</key><description>Cannot\ | |
| \ change owner of file name</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ change ownership of file name {filename} from {owner} to {attemptedOwner}\ | |
| \ on {host.name} in cluster {computeResource.name}.</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ change ownership of file name {filename} from {owner} to {attemptedOwner}\ | |
| \ on {host.name}.</formatOnComputeResource><formatOnHost>Cannot change ownership\ | |
| \ of file name {filename} from {owner} to {attemptedOwner}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ change ownership of file name {filename} from {owner} to {attemptedOwner}\ | |
| \ on {host.name} in {datacenter.name}.</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.ChangeOwnerOfFileFailedEvent">\t\t<description>\t\ | |
| \t\tChanging the owner of a file or folder failed\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ClusterComplianceCheckedEvent</key><description>Checked\ | |
| \ cluster for compliance</description><category>info</category><formatOnDatacenter>Checked\ | |
| \ cluster {computeResource.name} for compliance</formatOnDatacenter><formatOnComputeResource>Cluster\ | |
| \ was checked for compliance with profile {profile.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Checked\ | |
| \ cluster for compliance</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.ClusterComplianceCheckedEvent"> <description>\ | |
| \ The cluster was checked for compliance with a cluster profile \ | |
| \ </description> <cause> <description> \ | |
| \ The user initiated a compliance check on the cluster against a\ | |
| \ cluster profile </description> </cause> <cause>\ | |
| \ <description> A scheduled has initiated a compliance\ | |
| \ check for the cluster against a cluster profile </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ClusterCreatedEvent</key><description>Cluster\ | |
| \ created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ in folder {parent.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Created\ | |
| \ cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ClusterDestroyedEvent</key><description>Cluster\ | |
| \ deleted</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Removed\ | |
| \ cluster</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Removed\ | |
| \ cluster {computeResource.name} in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ClusterEvent</key><description><Cluster\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ClusterOvercommittedEvent</key><description>Cluster\ | |
| \ overcommitted</description><category>error</category><formatOnDatacenter>Insufficient\ | |
| \ capacity in cluster {computeResource.name} to satisfy resource configuration</formatOnDatacenter><formatOnComputeResource>Insufficient\ | |
| \ capacity to satisfy resource configuration</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Insufficient\ | |
| \ capacity in cluster {computeResource.name} to satisfy resource configuration\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.ClusterOvercommittedEvent"> <description>\ | |
| \ The cumulative CPU and/or memory resources of all hosts in the\ | |
| \ cluster are not adequate to satisfy the resource reservations\ | |
| \ of all virtual machines in the cluster </description>\ | |
| \ <cause> <description>You attempted to power\ | |
| \ on a virtual machine bypassing vCenter Server. \t\tThis condition occurs\ | |
| \ when you attempt the power on using the vSphere Client directly \t\tconnected\ | |
| \ to the host.</description> <action>In a DRS cluster,\ | |
| \ do not power on virtual machines bypassing vCenter Server</action>\ | |
| \ </cause> <cause> <description>A\ | |
| \ host was placed in Maintenance, Standby, or Disconnected Mode</description>\ | |
| \ <action>Bring any host in Maintenance, Standby, or Disconnected\ | |
| \ mode out of these modes</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ClusterReconfiguredEvent</key><description>Cluster\ | |
| \ reconfigured</description><category>info</category><formatOnDatacenter>Reconfigured\ | |
| \ cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Cluster\ | |
| \ reconfigured</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Reconfigured\ | |
| \ cluster {computeResource.name} in datacenter {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.ClusterReconfiguredEvent">\ | |
| \ <description> The cluster configuration was changed.\ | |
| \ The cluster configuration includes information about the DRS,\ | |
| \ DPM, EVC and vSphere HA settings of the cluster. All DRS rules\ | |
| \ are also stored in the cluster configuration. Editing the cluster \ | |
| \ configuration may trigger an invocation of DRS and/or enabling/disabling\ | |
| \ of vSphere HA on each host in the cluster. </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ClusterStatusChangedEvent</key><description>Cluster\ | |
| \ status changed</description><category>info</category><formatOnDatacenter>Configuration\ | |
| \ status on cluster {computeResource.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnDatacenter><formatOnComputeResource>Configuration\ | |
| \ status changed from {[email protected]} to {[email protected]}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Configuration\ | |
| \ status on cluster {computeResource.name} changed from {[email protected]}\ | |
| \ to {[email protected]} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.ClusterStatusChangedEvent">\ | |
| \ <description> The cluster status has changed. This\ | |
| \ status is the status of the root resource pool that encompasses\ | |
| \ the entire cluster. A cluster status change may be accompanied\ | |
| \ by the removal of a configuration issue if one was previously detected.\ | |
| \ A cluster status of green indicates that everything is fine. A\ | |
| \ yellow status indicates that the root resource pool does not have\ | |
| \ the resources to meet the reservations of its children. A red\ | |
| \ status means that a node in the resource pool has children whose\ | |
| \ reservations exceed the configuration of the node. </description>\ | |
| \ <cause> <description>The cluster status changed\ | |
| \ to yellow</description> <action>Add more resources\ | |
| \ (more hosts), or reduce the reservation of the resource pools\ | |
| \ directly under the root to match the new capacity</action> \ | |
| \ </cause> <cause> <description>The cluster\ | |
| \ status changed to red</description> <action>Change\ | |
| \ the resource settings on the resource pools that are red so \ | |
| \ that they can accommodate their child virtual machines. If this is not possible,\ | |
| \ lower the virtual machine reservations. If this is not possible\ | |
| \ either, power off some \t virtual machines.</action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>CustomFieldDefAddedEvent</key><description>Custom\ | |
| \ field definition added</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Created\ | |
| \ new custom field definition {name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomFieldDefEvent</key><description><Custom\ | |
| \ Field Definition Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomFieldDefRemovedEvent</key><description>Custom\ | |
| \ field definition removed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Removed\ | |
| \ field definition {name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomFieldDefRenamedEvent</key><description>Custom\ | |
| \ field definition renamed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Renamed\ | |
| \ field definition from {name} to {newName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomFieldEvent</key><description><Custom\ | |
| \ Field Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomFieldValueChangedEvent</key><description>Custom\ | |
| \ field value changed</description><category>info</category><formatOnDatacenter>Changed\ | |
| \ custom field {name} on {entity.name} to {value}</formatOnDatacenter><formatOnComputeResource>Changed\ | |
| \ custom field {name} on {entity.name} to {value}</formatOnComputeResource><formatOnHost>Changed\ | |
| \ custom field {name} on {entity.name} to {value}</formatOnHost><formatOnVm>Changed\ | |
| \ custom field {name} to {value}</formatOnVm><fullFormat>Changed custom field\ | |
| \ {name} on {entity.name} in {datacenter.name} to {value}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomizationEvent</key><description><Customization\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomizationFailed</key><description><Customization\ | |
| \ failed></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Cannot\ | |
| \ complete customization</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ complete customization of VM {vm.name}. See customization log at {logLocation}\ | |
| \ on the guest OS for details.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomizationLinuxIdentityFailed</key><description>Customization\ | |
| \ Linux Identity Failed</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>An\ | |
| \ error occurred while setting up Linux identity. See log file '{logLocation}'\ | |
| \ on guest OS for details.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.CustomizationLinuxIdentityFailed"> <description>\ | |
| \ The guest operating system Linux distribution is not supported by\ | |
| \ the customization scripts. Please refer to the VMware vSphere Compatibility\ | |
| \ Matrix for the list of the supported Linux distributions. \ | |
| \ </description> <cause> <description> \ | |
| \ Customization of the target guest operating system Linux distribution\ | |
| \ is not supported. </description> <action> \ | |
| \ Consult with VMware on when the specific Linux distribution will\ | |
| \ be supported. If the Linux distribution is already supported\ | |
| \ in a newer release, consider upgrading. </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>CustomizationNetworkSetupFailed</key><description>Cannot\ | |
| \ complete customization network setup</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>An\ | |
| \ error occurred while setting up network properties of the guest OS. See\ | |
| \ the log file {logLocation} in the guest OS for details.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.CustomizationNetworkSetupFailed">\ | |
| \ <description> The customization scripts failed to set\ | |
| \ the parameters in the corresponding configuration files for Linux\ | |
| \ or in the Windows registry </description> <cause>\ | |
| \ <description> The Customization Specification contains\ | |
| \ an invalid host name or domain name </description> \ | |
| \ <action> Review the guest operating system log files for\ | |
| \ this event for more details </action> <action>\ | |
| \ Provide a valid host name for the target guest operating system.\ | |
| \ The name must comply with the host name and domain name definitions\ | |
| \ in RFC 952, 1035, 1123, 2181. </action> </cause>\ | |
| \ <cause> <description> Could not find\ | |
| \ a NIC with the MAC address specified in the Customization Package \ | |
| \ </description> <action> Review the guest\ | |
| \ operating system log files for this event for more details </action>\ | |
| \ <action> Confirm that there was no change in the\ | |
| \ virtual NIC MAC address between the creation of the Customization\ | |
| \ Package and its deployment. Deployment occurs during the first\ | |
| \ boot of the virtual machine after customization has been scheduled. \ | |
| \ </action> </cause> <cause> <description>\ | |
| \ The customization code needs read/write permissions for certain\ | |
| \ configuration files. \t These permissions were not granted to the 'root'\ | |
| \ account on Linux or to the account used by the VMware Tools\ | |
| \ Service on the Windows guest operating system. </description>\ | |
| \ <action> Review the guest operating system log\ | |
| \ files for this event for more details </action> <action>\ | |
| \ Grant read/write permissions to the 'root' account\ | |
| \ for Linux or to the account used by the VMware Tools Service\ | |
| \ on the Windows guest operating system and the registry keys\ | |
| \ that need to be modified by the customization code </action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>CustomizationStartedEvent</key><description>Started\ | |
| \ customization</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Started\ | |
| \ customization of VM {vm.name}. Customization log located at {logLocation}\ | |
| \ in the guest OS.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomizationSucceeded</key><description>Customization\ | |
| \ succeeded</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Customization\ | |
| \ of VM {vm.name} succeeded. Customization log located at {logLocation} in\ | |
| \ the guest OS.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>CustomizationSysprepFailed</key><description>Cannot\ | |
| \ complete customization Sysprep</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ version of Sysprep {sysprepVersion} provided for customizing VM {vm.name}\ | |
| \ does not match the version of guest OS {systemVersion}. See the log file\ | |
| \ {logLocation} in the guest OS for more information.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.CustomizationSysprepFailed">\ | |
| \ <description> The sysprep files in the folder corresponding\ | |
| \ to the selected target guest operating system are not compatible\ | |
| \ with the actual version of the guest operation system </description>\ | |
| \ <cause> <description> The sysprep files\ | |
| \ in the folder corresponding to the target guest operating system \t (for\ | |
| \ example Windows XP) are for a different guest operating system (for example\ | |
| \ Windows 2003) </description> <action> \ | |
| \ On the machine running vCenter Server, place the correct sysprep files\ | |
| \ in the folder \t corresponding to the target guest operating system \ | |
| \ </action> </cause> <cause> <description>\ | |
| \ The sysprep files in the folder corresponding to the guest operating\ | |
| \ system are for a different Service Pack, for example the guest\ | |
| \ operating system is Windows XP SP2 and but the sysprep files\ | |
| \ are for Windows XP SP1. </description> <action>\ | |
| \ On the machine running vCenter Server, place the correct sysprep\ | |
| \ files in the folder \t corresponding to the target guest operating system\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>CustomizationUnknownFailure</key><description>Unknown\ | |
| \ customization error</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>An\ | |
| \ error occurred while customizing VM {vm.name}. For details reference the\ | |
| \ log file {logLocation} in the guest OS.</fullFormat><longDescription> \ | |
| \ <EventLongDescription id="vim.event.CustomizationUnknownFailure">\ | |
| \ <description> The customization component failed to set\ | |
| \ the required parameters inside the guest operating system \ | |
| \ </description> <cause> <description> \ | |
| \ On Windows, the user account under which the customization code runs\ | |
| \ has no \t read/write permissions for the registry keys used by the customization\ | |
| \ code. \t Customization code is usually run under the 'Local\ | |
| \ System' account but you can change this by selecting a\ | |
| \ different account for VMware Tools Service execution. </description>\ | |
| \ <action> Review the guest operating system log\ | |
| \ files for this event for more details </action> <action>\ | |
| \ Determine which user account is selected for VMware Tools Service\ | |
| \ execution and confirm that this account has read/write permissions\ | |
| \ on registry keys used by the customization code </action>\ | |
| \ </cause> <cause> <description> \ | |
| \ On Windows, the user account under which the customization code runs\ | |
| \ has no read/write \t permissions for the files and folders used by the\ | |
| \ customization code. Customization code is usually run \ | |
| \ under the 'Local System' account but you can change this\ | |
| \ by selecting a different account for VMware Tools Service execution.\ | |
| \ </description> <action> Review the\ | |
| \ guest operating system log files for this event for more details \ | |
| \ </action> <action> Determine which user\ | |
| \ account is selected for VMware Tools Service execution and confirm\ | |
| \ that this account has read/write permissions on the files and folders \ | |
| \ used by the customization code </action> </cause>\ | |
| \ <cause> <description> On Linux, an invalid\ | |
| \ or unsupported time zone is passed to the customization scripts\ | |
| \ and the time zone configuration failed as a result </description>\ | |
| \ <action> Review the guest operating system log\ | |
| \ files for this event for more details </action> <action>\ | |
| \ Confirm that a supported time zone is passed in Customization\ | |
| \ Specification. </action> </cause> <cause>\ | |
| \ <description> On Linux, the guest operating system\ | |
| \ 'root' account does not have read/write \t permissions for\ | |
| \ the configuration files that the customization scripts need to modify \t\ | |
| \ ('/etc/hosts') </description> <action>\ | |
| \ Grant read/write permissions for the configuration files to the\ | |
| \ guest operating system 'root' \t account </action>\ | |
| \ </cause> <cause> <description> \ | |
| \ Customization of the target guest operating system is not supported\ | |
| \ </description> <action> Consult with\ | |
| \ VMware on when the specific Linux distribution will be supported.\ | |
| \ If the Linux distribution is already supported in a newer release, consider\ | |
| \ upgrading. </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>DVPortgroupCreatedEvent</key><description>dvPort\ | |
| \ group created</description><category>info</category><formatOnDatacenter>dvPort\ | |
| \ group {net.name} was added to switch {dvs}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>dvPort\ | |
| \ group {net.name} in {datacenter.name} was added to switch {dvs.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DVPortgroupDestroyedEvent</key><description>dvPort\ | |
| \ group deleted</description><category>info</category><formatOnDatacenter>dvPort\ | |
| \ group {net.name} was deleted.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>dvPort\ | |
| \ group {net.name} in {datacenter.name} was deleted.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DVPortgroupEvent</key><description>dvPort\ | |
| \ group event</description><category>info</category><formatOnDatacenter>dvPort\ | |
| \ group event</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>dvPort\ | |
| \ group event</formatOnVm><fullFormat></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DVPortgroupReconfiguredEvent</key><description>dvPort\ | |
| \ group reconfigured</description><category>info</category><formatOnDatacenter>dvPort\ | |
| \ group {net.name} was reconfigured.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>dvPort\ | |
| \ group {net.name} in {datacenter.name} was reconfigured.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DVPortgroupRenamedEvent</key><description>dvPort\ | |
| \ group renamed</description><category>info</category><formatOnDatacenter>dvPort\ | |
| \ group {oldName} was renamed to {newName}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>dvPort\ | |
| \ group {oldName} in {datacenter.name} was renamed to {newName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DasAdmissionControlDisabledEvent</key><description>vSphere\ | |
| \ HA admission control disabled</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA admission control disabled for cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA admission control disabled</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA admission control disabled for cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DasAdmissionControlEnabledEvent</key><description>vSphere\ | |
| \ HA admission control enabled</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA admission control enabled for cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA admission control enabled</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA admission control enabled for cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DasAgentFoundEvent</key><description>vSphere\ | |
| \ HA agent found</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Re-established\ | |
| \ contact with a primary host in this vSphere HA cluster</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DasAgentUnavailableEvent</key><description>vSphere\ | |
| \ HA agent unavailable</description><category>error</category><formatOnDatacenter>Unable\ | |
| \ to contact a primary vSphere HA agent in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Unable\ | |
| \ to contact a primary vSphere HA agent</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Unable\ | |
| \ to contact a primary vSphere HA agent in cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.DasAgentUnavailableEvent"> <description>\ | |
| \ vCenter Server is not able to contact any good primary hosts in\ | |
| \ the vSphere HA cluster. vSphere HA protection may not be available\ | |
| \ for virtual machines running in the cluster. In addition, you\ | |
| \ cannot enable or reconfigure vSphere HA on hosts in the cluster\ | |
| \ until contact between vCenter Server and a good primary host is restored.\ | |
| \ </description> <cause> <description>\ | |
| \ There was a network outage, and all hosts show up in the inventory\ | |
| \ as "not responding" </description> <action>Restore\ | |
| \ the network</action> </cause> <cause> \ | |
| \ <description>All the primary hosts in the cluster failed</description>\ | |
| \ <action> If the failed primary hosts cannot be\ | |
| \ restored, disable vSphere HA on the cluster, wait for the Unconfigure\ | |
| \ vSphere HA tasks to complete on all hosts, and re-enable vSphere\ | |
| \ HA on the cluster </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>DasClusterIsolatedEvent</key><description>All\ | |
| \ vSphere HA hosts isolated</description><category>error</category><formatOnDatacenter>All\ | |
| \ hosts in the vSphere HA cluster {computeResource.name} were isolated from\ | |
| \ the network. Check the network configuration for proper network redundancy\ | |
| \ in the management network</formatOnDatacenter><formatOnComputeResource>All\ | |
| \ hosts in the vSphere HA cluster were isolated from the network. Check the\ | |
| \ network configuration for proper network redundancy in the management network</formatOnComputeResource><formatOnHost>All\ | |
| \ hosts in the vSphere HA cluster were isolated from the network. Check the\ | |
| \ network configuration for proper network redundancy in the management network</formatOnHost><formatOnVm></formatOnVm><fullFormat>All\ | |
| \ hosts in the vSphere HA cluster {computeResource.name} in {datacenter.name}\ | |
| \ were isolated from the network. Check the network configuration for proper\ | |
| \ network redundancy in the management network.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DasDisabledEvent</key><description>vSphere\ | |
| \ HA disabled for cluster</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA disabled for cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA disabled for this cluster</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA disabled for cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.DasDisabledEvent">\ | |
| \ <description> vSphere HA has been disabled on this\ | |
| \ host due to a user action. vSphere HA is disabled when a host \ | |
| \ is disconnected from vCenter Server or placed into maintenance or standby\ | |
| \ mode. Virtual machines on other hosts in the cluster will not\ | |
| \ be failed over to this host in the event of a host failure. In\ | |
| \ addition, if the host is disconnected, any virtual machines running\ | |
| \ on this host will not be failed if the host fails. Further, no\ | |
| \ attempt will be made by vSphere HA VM and Application Monitoring to reset\ | |
| \ VMs. </description> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>DasEnabledEvent</key><description>vSphere\ | |
| \ HA enabled for cluster</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA enabled for cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA enabled for this cluster</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA enabled for cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.DasEnabledEvent">\ | |
| \ <description> vSphere HA has been enabled on this host\ | |
| \ due to a user action. vSphere HA is enabled when a host is added\ | |
| \ to or moved into an vSphere HA cluster or when vSphere HA is enabled on\ | |
| \ a cluster. If the host was already in an vSphere HA cluster, vSphere\ | |
| \ HA will be enabled when the host is reconnected to vCenter Server\ | |
| \ or brought out of maintenance or standby mode. vSphere HA will attempt to\ | |
| \ protect any VMs that are running on the host at the time that\ | |
| \ HA is enabled on it. </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>DasHostFailedEvent</key><description>vSphere\ | |
| \ HA host failed</description><category>error</category><formatOnDatacenter>A\ | |
| \ possible host failure has been detected by vSphere HA on {failedHost.name}</formatOnDatacenter><formatOnComputeResource>A\ | |
| \ possible host failure has been detected by vSphere HA on {failedHost.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ possible host failure has been detected by vSphere HA on {failedHost.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DasHostIsolatedEvent</key><description>vSphere\ | |
| \ HA host isolated</description><category>warning</category><formatOnDatacenter>Host\ | |
| \ {isolatedHost.name} has been isolated from cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {isolatedHost.name} has been isolated</formatOnComputeResource><formatOnHost>Host\ | |
| \ has been isolated from cluster</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {isolatedHost.name} has been isolated from cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.DasHostIsolatedEvent"> <description>\ | |
| \ vSphere HA detected that the host is network isolated. When a host\ | |
| \ is in this state, vSphere HA applies the power-off or shutdown\ | |
| \ host isolation response to virtual machines running on the host, and \ | |
| \ continues to monitor the virtual machines that are left powered on.\ | |
| \ While a host is in this state, vSphere HA's ability to restart\ | |
| \ virtual machines after a failure is impacted. vSphere HA only powers off\ | |
| \ or shuts down a virtual machine if the agent on the host determines\ | |
| \ that a master agent is responsible for the virtual machine. </description>\ | |
| \ <cause> <description> A host\ | |
| \ is network isolated if both of the following conditions are met: (1) \ | |
| \ isolation addresses have been configured and the host is unable\ | |
| \ to ping them; (2) the vSphere HA agent on the host is unable\ | |
| \ to access any of the agents running on the other cluster hosts. \ | |
| \ </description> <action> Resolve\ | |
| \ the networking problem that is preventing the host from pinging its isolation\ | |
| \ addresses and communicating with other hosts. Ensure that\ | |
| \ there is redundancy in the management networks used by\ | |
| \ vSphere HA. With redundancy, vSphere HA is able to communicate \ | |
| \ over more than one path thus reducing the chance of a host becoming\ | |
| \ isolated. </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>DatacenterCreatedEvent</key><description>Datacenter\ | |
| \ created</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ in folder {parent.name}</formatOnComputeResource><formatOnHost>Created datacenter\ | |
| \ {datacenter.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Created\ | |
| \ datacenter {datacenter.name} in folder {parent.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatacenterEvent</key><description><Datacenter\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatacenterRenamedEvent</key><description>Datacenter\ | |
| \ renamed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource>Renamed\ | |
| \ datacenter</formatOnComputeResource><formatOnHost>Renamed datacenter from\ | |
| \ {oldName} to {newName}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Renamed\ | |
| \ datacenter from {oldName} to {newName}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreCapacityIncreasedEvent</key><description>Datastore\ | |
| \ capacity increased</description><category>info</category><formatOnDatacenter>Datastore\ | |
| \ {datastore.name} increased in capacity from {oldCapacity} bytes to {newCapacity}\ | |
| \ bytes</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Datastore\ | |
| \ {datastore.name} increased in capacity from {oldCapacity} bytes to {newCapacity}\ | |
| \ bytes in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreDestroyedEvent</key><description>Datastore\ | |
| \ deleted</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ unconfigured datastore {datastore.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Removed\ | |
| \ unconfigured datastore {datastore.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreDiscoveredEvent</key><description>Datastore\ | |
| \ discovered</description><category>info</category><formatOnDatacenter>Discovered\ | |
| \ datastore {datastore.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Discovered\ | |
| \ datastore {datastore.name} on {host.name}</formatOnComputeResource><formatOnHost>Discovered\ | |
| \ datastore {datastore.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Discovered\ | |
| \ datastore {datastore.name} on {host.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.DatastoreDiscoveredEvent">\ | |
| \ <description> A datastore was discovered on a host \ | |
| \ </description> <cause> <description> \ | |
| \ A host that has access to this datastore was added to the datacenter\ | |
| \ </description> </cause> <cause> \ | |
| \ <description> The storage backing this datastore was unmasked\ | |
| \ to a host in the datacenter </description> </cause>\ | |
| \ <cause> <description> A user or system\ | |
| \ action caused this datastore to be created on a host </description>\ | |
| \ </cause> <cause> <description> \ | |
| \ A user or system action caused this datastore to be created on a host\ | |
| \ and the datastore was visible on at least one other host in the\ | |
| \ datacenter prior to this operation. </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>DatastoreDuplicatedEvent</key><description>Datastore\ | |
| \ duplicated</description><category>error</category><formatOnDatacenter>Multiple\ | |
| \ datastores named {datastore} detected on host {host.name}</formatOnDatacenter><formatOnComputeResource>Multiple\ | |
| \ datastores named {datastore} detected on host {host.name}</formatOnComputeResource><formatOnHost>Multiple\ | |
| \ datastores named {datastore} detected</formatOnHost><formatOnVm></formatOnVm><fullFormat>Multiple\ | |
| \ datastores named {datastore} detected on host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreEvent</key><description><Datastore\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreFileCopiedEvent</key><description>File\ | |
| \ or directory copied to datastore</description><category>info</category><formatOnDatacenter>File\ | |
| \ or directory {sourceFile} copied from {sourceDatastore.name} to {datastore.name}\ | |
| \ as {targetFile}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>File\ | |
| \ or directory {sourceFile} copied from {sourceDatastore.name} to {datastore.name}\ | |
| \ as {targetFile}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreFileDeletedEvent</key><description>File\ | |
| \ or directory deleted</description><category>info</category><formatOnDatacenter>File\ | |
| \ or directory {targetFile} deleted from {datastore.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>File\ | |
| \ or directory {targetFile} deleted from {datastore.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreFileEvent</key><description><Datastore\ | |
| \ File Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreFileMovedEvent</key><description>File\ | |
| \ or directory moved to datastore</description><category>info</category><formatOnDatacenter>File\ | |
| \ or directory {sourceFile} moved from {sourceDatastore.name} to {datastore.name}\ | |
| \ as {targetFile}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>File\ | |
| \ or directory {sourceFile} moved from {sourceDatastore.name} to {datastore.name}\ | |
| \ as {targetFile}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreIORMReconfiguredEvent</key><description>Reconfigured\ | |
| \ Storage I/O Control on datastore</description><category>info</category><formatOnDatacenter>Reconfigured\ | |
| \ Storage I/O Control on datastore {datastore.name}</formatOnDatacenter><formatOnComputeResource>Reconfigured\ | |
| \ Storage I/O Control on datastore {datastore.name}</formatOnComputeResource><formatOnHost>Reconfigured\ | |
| \ Storage I/O Control on datastore {datastore.name}</formatOnHost><formatOnVm>Reconfigured\ | |
| \ Storage I/O Control on datastore {datastore.name}</formatOnVm><fullFormat>Reconfigured\ | |
| \ Storage I/O Control on datastore {datastore.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastorePrincipalConfigured</key><description>Datastore\ | |
| \ principal configured</description><category>info</category><formatOnDatacenter>Configured\ | |
| \ datastore principal {datastorePrincipal} on host {host.name}</formatOnDatacenter><formatOnComputeResource>Configured\ | |
| \ datastore principal {datastorePrincipal} on host {host.name}</formatOnComputeResource><formatOnHost>Configured\ | |
| \ datastore principal {datastorePrincipal}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Configured\ | |
| \ datastore principal {datastorePrincipal} on host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreRemovedOnHostEvent</key><description>Datastore\ | |
| \ removed from host</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ datastore {datastore.name} from {host.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Removed\ | |
| \ datastore {datastore.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Removed\ | |
| \ datastore {datastore.name} from {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreRenamedEvent</key><description>Datastore\ | |
| \ renamed</description><category>info</category><formatOnDatacenter>Renamed\ | |
| \ datastore from {oldName} to {newName}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Renamed\ | |
| \ datastore from {oldName} to {newName} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DatastoreRenamedOnHostEvent</key><description>Datastore\ | |
| \ renamed from host</description><category>info</category><formatOnDatacenter>Renamed\ | |
| \ datastore from {oldName} to {newName}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Renamed\ | |
| \ datastore from {oldName} to {newName} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.DatastoreRenamedOnHostEvent">\ | |
| \ <description> A datastore was renamed on a host managed\ | |
| \ by vCenter Server </description> <cause> <description>\ | |
| \ vCenter Server discovered datastore on a host and renamed the\ | |
| \ datastore because it already exists in the vCenter Server inventory\ | |
| \ under a different name. vCenter Server might also have renamed\ | |
| \ the datastore because the name conflicts with \t another datastore in\ | |
| \ the same datacenter. </description> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>DrsDisabledEvent</key><description>DRS\ | |
| \ disabled</description><category>info</category><formatOnDatacenter>Disabled\ | |
| \ DRS on cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Disabled\ | |
| \ DRS</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Disabled\ | |
| \ DRS on cluster {computeResource.name} in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsEnabledEvent</key><description>DRS\ | |
| \ enabled</description><category>info</category><formatOnDatacenter>Enabled\ | |
| \ DRS on cluster {computeResource.name} with automation level {behavior}</formatOnDatacenter><formatOnComputeResource>Enabled\ | |
| \ DRS with automation level {behavior}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Enabled\ | |
| \ DRS on {computeResource.name} with automation level {behavior} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsEnteredStandbyModeEvent</key><description>DRS\ | |
| \ entered standby mode</description><category>info</category><formatOnDatacenter>DRS\ | |
| \ put {host.name} into standby mode</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ put {host.name} into standby mode</formatOnComputeResource><formatOnHost>DRS\ | |
| \ put the host into standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>DRS\ | |
| \ put {host.name} into standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsEnteringStandbyModeEvent</key><description>DRS\ | |
| \ entering standby mode</description><category>info</category><formatOnDatacenter>DRS\ | |
| \ is putting {host.name} into standby mode</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ is putting {host.name} into standby mode</formatOnComputeResource><formatOnHost>DRS\ | |
| \ is putting the host into standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>DRS\ | |
| \ is putting {host.name} into standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsExitStandbyModeFailedEvent</key><description>DRS\ | |
| \ cannot exit the host out of standby mode</description><category>error</category><formatOnDatacenter>DRS\ | |
| \ cannot move {host.name} out of standby mode</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ cannot move {host.name} out of standby mode</formatOnComputeResource><formatOnHost>DRS\ | |
| \ cannot move the host out of standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>DRS\ | |
| \ cannot move {host.name} out of standby mode</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.DrsExitStandbyModeFailedEvent">\ | |
| \ <description> DPM failed to power on a host in standby\ | |
| \ mode. \t\t DPM tried to power on a host using IPMI, iLO or Wake-on-LAN\ | |
| \ protocol, but the host \t\t did not power on. </description>\ | |
| \ <cause> <description>DPM could not communicate\ | |
| \ with the BMC on the host</description> <action>Verify\ | |
| \ the IPMI/iLO credentials entered in vCenter Server</action> \ | |
| \ <action>Verify that LAN access is enabled in the BMC</action>\ | |
| \ </cause> <cause> <description>The\ | |
| \ vMotion NIC on the host does not support Wake-on-LAN</description>\ | |
| \ <action>Select a vMotion NIC that supports Wake-on-LAN</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>DrsExitedStandbyModeEvent</key><description>DRS\ | |
| \ exited standby mode</description><category>info</category><formatOnDatacenter>DRS\ | |
| \ moved {host.name} out of standby mode</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ moved {host.name} out of standby mode</formatOnComputeResource><formatOnHost>DRS\ | |
| \ moved the host out of standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>DRS\ | |
| \ moved {host.name} out of standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsExitingStandbyModeEvent</key><description>DRS\ | |
| \ exiting standby mode</description><category>info</category><formatOnDatacenter>DRS\ | |
| \ is moving {host.name} out of standby mode</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ is moving {host.name} out of standby mode</formatOnComputeResource><formatOnHost>DRS\ | |
| \ is moving the host out of standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>DRS\ | |
| \ is moving {host.name} out of standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsInvocationFailedEvent</key><description>DRS\ | |
| \ invocation not completed</description><category>error</category><formatOnDatacenter>DRS\ | |
| \ invocation not completed</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ invocation not completed</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>DRS\ | |
| \ invocation not completed</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.DrsInvocationFailedEvent"> <description>\ | |
| \ A DRS invocation failed to complete successfully. This condition\ | |
| \ can occur for a \t\t variety of reasons, some of which may be transient.\ | |
| \ </description> <cause> <description>An\ | |
| \ error was encountered during a DRS invocation</description> \ | |
| \ <action>Disable and re-enable DRS</action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>DrsRecoveredFromFailureEvent</key><description>DRS\ | |
| \ has recovered from the failure</description><category>info</category><formatOnDatacenter>DRS\ | |
| \ has recovered from the failure</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ has recovered from the failure</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>DRS\ | |
| \ has recovered from the failure</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsResourceConfigureFailedEvent</key><description>Cannot\ | |
| \ complete DRS resource configuration</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Unable\ | |
| \ to apply DRS resource settings on host. {reason.msg}. This can significantly\ | |
| \ reduce the effectiveness of DRS.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Unable\ | |
| \ to apply DRS resource settings on host {host.name} in {datacenter.name}.\ | |
| \ {reason.msg}. This can significantly reduce the effectiveness of DRS.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.DrsResourceConfigureFailedEvent">\ | |
| \ <description> The DRS resource settings could not be\ | |
| \ successfully applied to a host in the \t\t cluster. This condition is typically\ | |
| \ transient. </description> <cause> <description>DRS\ | |
| \ resource settings could not be applied to a host.</description> \ | |
| \ <action>DRS generates resource settings that map the cluster\ | |
| \ values to the host. However, in this case, the values could\ | |
| \ not be successfully applied to the host. This is typically a\ | |
| \ transient error caused by delayed synchronization from DRS to\ | |
| \ the host. If this condition persists, enable debug logging in vpxa and \t\ | |
| \ contact VMware Support. </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>DrsResourceConfigureSyncedEvent</key><description>DRS\ | |
| \ resource configuration synchronized</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Resource\ | |
| \ configuration specification returns to synchronization from previous failure</formatOnHost><formatOnVm></formatOnVm><fullFormat>Resource\ | |
| \ configuration specification returns to synchronization from previous failure\ | |
| \ on host '{host.name}' in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsRuleComplianceEvent</key><description>VM\ | |
| \ is now compliant with DRS VM-Host affinity rules</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is now compliant with DRS VM-Host affinity rules</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is now compliant with DRS VM-Host affinity rules</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is now compliant with DRS VM-Host affinity rules</formatOnHost><formatOnVm>virtual\ | |
| \ machine on {host.name} is now compliant with DRS VM-Host affinity rules</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is now compliant with DRS VM-Host affinity\ | |
| \ rules</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsRuleViolationEvent</key><description>VM\ | |
| \ is violating a DRS VM-Host affinity rule</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is violating a DRS VM-Host affinity rule</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is violating a DRS VM-Host affinity rule</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is violating a DRS VM-Host affinity rule</formatOnHost><formatOnVm>virtual\ | |
| \ machine on {host.name} is violating a DRS VM-Host affinity rule</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is violating a DRS VM-Host affinity\ | |
| \ rule</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DrsVmMigratedEvent</key><description>DRS\ | |
| \ VM migrated</description><category>info</category><formatOnDatacenter>DRS\ | |
| \ migrated {vm.name} from {sourceHost.name} to {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ migrated {vm.name} from {sourceHost.name} to {host.name}</formatOnComputeResource><formatOnHost>DRS\ | |
| \ migrated {vm.name} from {sourceHost.name}</formatOnHost><formatOnVm>Migrated\ | |
| \ from {sourceHost.name} to {host.name} by DRS</formatOnVm><fullFormat>DRS\ | |
| \ migrated {vm.name} from {sourceHost.name} to {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.DrsVmMigratedEvent"> <description>\ | |
| \ A virtual machine was migrated based on a DRS recommendation. The\ | |
| \ recommendation might \t\t have been made be to achieve better load balancing\ | |
| \ in the cluster or to evacuate \t\t a host in the cluster that is being\ | |
| \ put into Standby or Maintenance Mode. </description> \ | |
| \ <cause> <description>DRS recommended the migration\ | |
| \ of a virtual machine</description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>DrsVmPoweredOnEvent</key><description>DRS\ | |
| \ VM powered on</description><category>info</category><formatOnDatacenter>DRS\ | |
| \ powered On {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>DRS\ | |
| \ powered On {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>DRS\ | |
| \ powered On {vm.name}</formatOnHost><formatOnVm>DRS powered On the virtual\ | |
| \ machine on {host.name}</formatOnVm><fullFormat>DRS powered On {vm.name}\ | |
| \ on {host.name} in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.DrsVmPoweredOnEvent"> <description>\ | |
| \ A virtual machine was powered on by the user and DRS choose a host\ | |
| \ for the virtual machine based on \t the current cluster load distribution\ | |
| \ combined with the virtual machine's resource requirements </description>\ | |
| \ <cause> <description>DRS chose a host for a\ | |
| \ virtual machine that was being powered on</description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>DuplicateIpDetectedEvent</key><description>Duplicate\ | |
| \ IP detected</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {macAddress} has a duplicate IP {duplicateIP}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Virtual\ | |
| \ machine {macAddress} on host {host.name} has a duplicate IP {duplicateIP}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvpgImportEvent</key><description>Import\ | |
| \ Operation event</description><category>info</category><formatOnDatacenter>Import\ | |
| \ operation with type {importType} was performed on {net.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Import\ | |
| \ operation with type {importType} was performed on {net.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvpgRestoreEvent</key><description>Restore\ | |
| \ Operation event</description><category>info</category><formatOnDatacenter>Restore\ | |
| \ operation was performed on {net.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Restore\ | |
| \ operation was performed on {net.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsCreatedEvent</key><description>vSphere\ | |
| \ Distributed Switch created</description><category>info</category><formatOnDatacenter>A\ | |
| \ vSphere Distributed Switch {dvs.name} was created</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ vSphere Distributed Switch {dvs.name} was created in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsDestroyedEvent</key><description>vSphere\ | |
| \ Distributed Switch deleted</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ Distributed Switch {dvs.name} was deleted.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ Distributed Switch {dvs.name} in {datacenter.name} was deleted.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsEvent</key><description>vSphere\ | |
| \ Distributed Switch event</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ Distributed Switch event</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ Distributed Switch event</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsHealthStatusChangeEvent</key><description>Health\ | |
| \ check status of the switch changed.</description><category>info</category><formatOnDatacenter>Health\ | |
| \ check status changed in vSphere Distributed Switch {dvs.name} on host {host.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Health\ | |
| \ check status changed in vSphere Distributed Switch {dvs.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Health\ | |
| \ check status was changed in vSphere Distributed Switch {dvs.name} on host\ | |
| \ {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsHostBackInSyncEvent</key><description>The\ | |
| \ vSphere Distributed Switch configuration on the host was synchronized with\ | |
| \ that of the vCenter Server.</description><category>info</category><formatOnDatacenter>The\ | |
| \ vSphere Distributed Switch {dvs.name} configuration on the host was synchronized\ | |
| \ with that of the vCenter Server.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ vSphere Distributed Switch {dvs.name} configuration on the host was synchronized\ | |
| \ with that of the vCenter Server.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsHostJoinedEvent</key><description>Host\ | |
| \ joined the vSphere Distributed Switch</description><category>info</category><formatOnDatacenter>The\ | |
| \ host {hostJoined.name} joined the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ host {hostJoined.name} joined the vSphere Distributed Switch {dvs.name}\ | |
| \ in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsHostLeftEvent</key><description>\ | |
| \ Host left vSphere Distributed Switch</description><category>info</category><formatOnDatacenter>The\ | |
| \ host {hostLeft.name} left the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ host {hostLeft.name} left the vSphere Distributed Switch {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsHostStatusUpdated</key><description>Host\ | |
| \ status changed on the vSphere Distributed Switch</description><category>info</category><formatOnDatacenter>The\ | |
| \ host {hostMember.name} changed status on the vSphere Distributed Switch\ | |
| \ {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ host {hostMember.name} changed status on the vSphere Distributed Switch\ | |
| \ {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsHostWentOutOfSyncEvent</key><description>The\ | |
| \ vSphere Distributed Switch configuration on the host differed from that\ | |
| \ of the vCenter Server.</description><category>warning</category><formatOnDatacenter>The\ | |
| \ vSphere Distributed Switch {dvs.name} configuration on the host differed\ | |
| \ from that of the vCenter Server.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ vSphere Distributed Switch {dvs.name} configuration on the host differed\ | |
| \ from that of the vCenter Server.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.DvsHostWentOutOfSyncEvent"> <description>\ | |
| \ The vSphere Distributed Switch configuration on the host differed from that\ | |
| \ of the vCenter Server </description> <cause> \ | |
| \ <description> The host was not connected to the vCenter Server\ | |
| \ when updates were sent </description> </cause> <cause>\ | |
| \ <description> vCenter Server failed to push the vSphere Distributed\ | |
| \ Switch configuration to the host in the past</description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>DvsImportEvent</key><description>Import\ | |
| \ Operation event</description><category>info</category><formatOnDatacenter>Import\ | |
| \ operation with type {importType} was performed on {dvs.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Import\ | |
| \ operation with type {importType} was performed on {dvs.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsMergedEvent</key><description>vSphere\ | |
| \ Distributed Switch merged</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ Distributed Switch {srcDvs.name} was merged into {dstDvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ Distributed Switch {srcDvs.name} was merged into {dstDvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortBlockedEvent</key><description>dvPort\ | |
| \ blocked</description><category>info</category><formatOnDatacenter>The dvPort\ | |
| \ {portKey} was blocked in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was blocked in the vSphere Distributed Switch {dvs.name}\ | |
| \ in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortConnectedEvent</key><description>dvPort\ | |
| \ connected</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} was connected in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was connected in the vSphere Distributed Switch {dvs.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortCreatedEvent</key><description>dvPort\ | |
| \ created</description><category>info</category><formatOnDatacenter>New ports\ | |
| \ were created in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>New\ | |
| \ ports were created in the vSphere Distributed Switch {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortDeletedEvent</key><description>dvPort\ | |
| \ deleted</description><category>info</category><formatOnDatacenter>Ports\ | |
| \ were deleted in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Deleted\ | |
| \ ports in the vSphere Distributed Switch {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortDisconnectedEvent</key><description>dvPort\ | |
| \ disconnected</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} was disconnected in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was disconnected in the vSphere Distributed Switch {dvs.name}\ | |
| \ in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortEnteredPassthruEvent</key><description>dvPort\ | |
| \ in passthrough mode</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} was in passthrough mode in the vSphere Distributed Switch\ | |
| \ {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was in passthrough mode in the vSphere Distributed Switch\ | |
| \ {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortExitedPassthruEvent</key><description>dvPort\ | |
| \ not in passthrough mode</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} was not in passthrough mode in the vSphere Distributed\ | |
| \ Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was not in passthrough mode in the vSphere Distributed\ | |
| \ Switch {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortJoinPortgroupEvent</key><description>A\ | |
| \ dvPort was moved into the dvPort group.</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} was moved into the dvPort group {portgroupName}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was moved into the dvPort group {portgroupName} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortLeavePortgroupEvent</key><description>A\ | |
| \ dvPort was moved out of the dvPort group.</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} was moved out of the dvPort group {portgroupName}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was moved out of the dvPort group {portgroupName} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortLinkDownEvent</key><description>dvPort\ | |
| \ link was down</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} link was down in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} link was down in the vSphere Distributed Switch {dvs.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortLinkUpEvent</key><description>dvPort\ | |
| \ link was up</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} link was up in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} link was up in the vSphere Distributed Switch {dvs.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortReconfiguredEvent</key><description>dvPort\ | |
| \ reconfigured</description><category>info</category><formatOnDatacenter>Ports\ | |
| \ were reconfigured in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Reconfigured\ | |
| \ ports in the vSphere Distributed Switch {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortRuntimeChangeEvent</key><description>dvPort\ | |
| \ runtime information changed.</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} runtime information changed in the vSphere Distributed\ | |
| \ Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} runtime information changed in the vSphere Distributed\ | |
| \ Switch {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortUnblockedEvent</key><description>dvPort\ | |
| \ unblocked</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} was unblocked in the vSphere Distributed Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} was unblocked in the vSphere Distributed Switch {dvs.name}\ | |
| \ in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsPortVendorSpecificStateChangeEvent</key><description>dvPort\ | |
| \ vendor specific state changed.</description><category>info</category><formatOnDatacenter>The\ | |
| \ dvPort {portKey} vendor specific state changed in the vSphere Distributed\ | |
| \ Switch {dvs.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ dvPort {portKey} vendor specific state changed in the vSphere Distributed\ | |
| \ Switch {dvs.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsReconfiguredEvent</key><description>vSphere\ | |
| \ Distributed Switch reconfigured</description><category>info</category><formatOnDatacenter>The\ | |
| \ vSphere Distributed Switch {dvs.name} was reconfigured.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ vSphere Distributed Switch {dvs.name} in {datacenter.name} was reconfigured.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsRenamedEvent</key><description>vSphere\ | |
| \ Distributed Switch renamed</description><category>info</category><formatOnDatacenter>The\ | |
| \ vSphere Distributed Switch {oldName} was renamed to {newName}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ vSphere Distributed Switch {oldName} in {datacenter.name} was renamed to\ | |
| \ {newName}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsRestoreEvent</key><description>Restore\ | |
| \ Operation event</description><category>info</category><formatOnDatacenter>Restore\ | |
| \ operation was performed on {dvs.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Restore\ | |
| \ operation was performed on {dvs.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsUpgradeAvailableEvent</key><description>An\ | |
| \ upgrade for the vSphere Distributed Switch is available.</description><category>info</category><formatOnDatacenter>An\ | |
| \ upgrade for vSphere Distributed Switch {dvs.name} is available.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>\ | |
| \ An upgrade for the vSphere Distributed Switch {dvs.name} in datacenter {datacenter.name}\ | |
| \ is available.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsUpgradeInProgressEvent</key><description>An\ | |
| \ upgrade for the vSphere Distributed Switch is in progress.</description><category>info</category><formatOnDatacenter>An\ | |
| \ upgrade for vSphere Distributed Switch {dvs.name} is in progress.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>An\ | |
| \ upgrade for the vSphere Distributed Switch {dvs.name} in datacenter {datacenter.name}\ | |
| \ is in progress.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsUpgradeRejectedEvent</key><description>Cannot\ | |
| \ complete the upgrade for the vSphere Distributed Switch</description><category>info</category><formatOnDatacenter>An\ | |
| \ upgrade for vSphere Distributed Switch {dvs.name} was rejected.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ complete an upgrade for the vSphere Distributed Switch {dvs.name} in datacenter\ | |
| \ {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>DvsUpgradedEvent</key><description>The\ | |
| \ vSphere Distributed Switch was upgraded.</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ Distributed Switch {dvs.name} was upgraded.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ Distributed Switch {dvs.name} in datacenter {datacenter.name} was upgraded.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EnteredMaintenanceModeEvent</key><description>Entered\ | |
| \ maintenance mode</description><category>info</category><formatOnDatacenter>Host\ | |
| \ {host.name} in {datacenter.name} has entered maintenance mode</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} in {datacenter.name} has entered maintenance mode</formatOnComputeResource><formatOnHost>Enter\ | |
| \ maintenance mode completed. All virtual machine operations are disabled</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} in {datacenter.name} has entered maintenance mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EnteredStandbyModeEvent</key><description>Entered\ | |
| \ standby mode</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Entered\ | |
| \ standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>The host\ | |
| \ {host.name} is in standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EnteringMaintenanceModeEvent</key><description>Entering\ | |
| \ maintenance mode</description><category>info</category><formatOnDatacenter>Host\ | |
| \ {host.name} has started to enter maintenance mode</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} has started to enter maintenance mode</formatOnComputeResource><formatOnHost>Started\ | |
| \ to enter maintenance mode. Waiting for virtual machines to shut down, suspend,\ | |
| \ or migrate</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host {host.name}\ | |
| \ in {datacenter.name} has started to enter maintenance mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EnteringStandbyModeEvent</key><description>Entering\ | |
| \ standby mode</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Entering\ | |
| \ standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>The host\ | |
| \ {host.name} is entering standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ErrorUpgradeEvent</key><description>Upgrade\ | |
| \ error</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>{message}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.ErrorUpgradeEvent">\t\t\ | |
| <description>\t\t\tAn error occurred during agent upgrade\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>Event</key><description><Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExitMaintenanceModeEvent</key><description>Exit\ | |
| \ maintenance mode</description><category>info</category><formatOnDatacenter>Host\ | |
| \ {host.name} has exited maintenance mode</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} has exited maintenance mode</formatOnComputeResource><formatOnHost>Exited\ | |
| \ maintenance mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} in {datacenter.name} has exited maintenance mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExitStandbyModeFailedEvent</key><description>Cannot\ | |
| \ exit standby mode</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Could\ | |
| \ not exit standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ host {host.name} could not exit standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExitedStandbyModeEvent</key><description>Exited\ | |
| \ standby mode</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Exited\ | |
| \ standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>The host\ | |
| \ {host.name} is no longer in standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExitingStandbyModeEvent</key><description>Exiting\ | |
| \ standby mode</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Exiting\ | |
| \ standby mode</formatOnHost><formatOnVm></formatOnVm><fullFormat>The host\ | |
| \ {host.name} is exiting standby mode</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>FailoverLevelRestored</key><description>vSphere\ | |
| \ HA failover resources are sufficient</description><category>info</category><formatOnDatacenter>Sufficient\ | |
| \ resources are available to satisfy vSphere HA failover level in cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Sufficient\ | |
| \ resources are available to satisfy vSphere HA failover level</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Sufficient\ | |
| \ resources are available to satisfy vSphere HA failover level in cluster\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>GeneralEvent</key><description>General\ | |
| \ event</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>General\ | |
| \ event: {message}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>GeneralHostErrorEvent</key><description>Host\ | |
| \ error</description><category>error</category><formatOnDatacenter>Error detected\ | |
| \ on {host.name}: {message}</formatOnDatacenter><formatOnComputeResource>Error\ | |
| \ detected on {host.name}: {message}</formatOnComputeResource><formatOnHost>{message}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Error\ | |
| \ detected on {host.name} in {datacenter.name}: {message}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.GeneralHostErrorEvent">\t\ | |
| \t<description>\t\t\tAn error occurred on the host\t\t</description>\t\ | |
| \t<cause> <description>\t\t\tThe agent cannot send\ | |
| \ heartbeats because of a networking related failure on host\t\t\t</description>\ | |
| \ </cause> <cause> <description>\t\ | |
| \t\tThe agent failed to update the configuration file on host\t\t\t</description>\ | |
| \ </cause> <cause> <description>\t\ | |
| \t\tThe agent failed to save the configuration file to disk on host\t\t\t\ | |
| </description> </cause> <cause> \ | |
| \ <description>\t\t\tThe provisioning module failed to load. As a result,\ | |
| \ all provisioning operations will \t\t\tfail on host.\t\t\t</description>\ | |
| \ </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>GeneralHostInfoEvent</key><description>Host\ | |
| \ information</description><category>info</category><formatOnDatacenter>Issue\ | |
| \ detected on {host.name}: {message}</formatOnDatacenter><formatOnComputeResource>Issue\ | |
| \ detected on {host.name}: {message}</formatOnComputeResource><formatOnHost>{message}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Issue\ | |
| \ detected on {host.name} in {datacenter.name}: {message}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.GeneralHostInfoEvent">\t\ | |
| \t<description>\t\t\tA general information event occurred on the host\t\ | |
| \t</description>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>GeneralHostWarningEvent</key><description>Host\ | |
| \ warning</description><category>warning</category><formatOnDatacenter>Issue\ | |
| \ detected on {host.name}: {message}</formatOnDatacenter><formatOnComputeResource>Issue\ | |
| \ detected on {host.name}: {message}</formatOnComputeResource><formatOnHost>{message}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Issue\ | |
| \ detected on {host.name} in {datacenter.name}: {message}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.GeneralHostWarningEvent">\t\ | |
| \t<description>\t\t\tA general warning event occurred on the host\t\t\ | |
| </description>\t\t<cause> <description>\t\t\t\ | |
| Virtual machine creation might fail because the agent was unable to retrieve\ | |
| \ virtual \t\t\tmachine creation options from the host\t\t\t</description>\ | |
| \ </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>GeneralUserEvent</key><description>User\ | |
| \ event</description><category>user</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>User\ | |
| \ logged event: {message}</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.GeneralUserEvent">\t\t<description>\t\t\ | |
| \tA general user event occurred on the host\t\t</description>\t\t<cause>\ | |
| \ <description>\t\t\tA user initiated an action on the host\t\ | |
| \t\t</description> </cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>GeneralVmErrorEvent</key><description>VM\ | |
| \ error</description><category>error</category><formatOnDatacenter>Error detected\ | |
| \ for {vm.name} on {host.name} in {datacenter.name}: {message}</formatOnDatacenter><formatOnComputeResource>Error\ | |
| \ detected for {vm.name} on {host.name} in {datacenter.name}: {message}</formatOnComputeResource><formatOnHost>Error\ | |
| \ detected for {vm.name}: {message}</formatOnHost><formatOnVm>{message} on\ | |
| \ {host.name}</formatOnVm><fullFormat>Error detected for {vm.name} on {host.name}\ | |
| \ in {datacenter.name}: {message}</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.GeneralVmErrorEvent">\t\t<description>\t\ | |
| \t\tAn error occurred on the virtual machine\t\t</description>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>GeneralVmInfoEvent</key><description>VM\ | |
| \ information</description><category>info</category><formatOnDatacenter>Issue\ | |
| \ detected for {vm.name} on {host.name} in {datacenter.name}: {message}</formatOnDatacenter><formatOnComputeResource>Issue\ | |
| \ detected for {vm.name} on {host.name} in {datacenter.name}: {message}</formatOnComputeResource><formatOnHost>Issue\ | |
| \ detected for {vm.name}: {message}</formatOnHost><formatOnVm>{message} on\ | |
| \ {host.name}</formatOnVm><fullFormat>Issue detected for {vm.name} on {host.name}\ | |
| \ in {datacenter.name}: {message}</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.GeneralVmInfoEvent">\t\t<description>\t\ | |
| \t\tA general information event occurred on the virtual machine\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>GeneralVmWarningEvent</key><description>VM\ | |
| \ warning</description><category>warning</category><formatOnDatacenter>Issue\ | |
| \ detected for {vm.name} on {host.name} in {datacenter.name}: {message}</formatOnDatacenter><formatOnComputeResource>Issue\ | |
| \ detected for {vm.name} on {host.name} in {datacenter.name}: {message}</formatOnComputeResource><formatOnHost>Issue\ | |
| \ detected for {vm.name}: {message}</formatOnHost><formatOnVm>{message} on\ | |
| \ {host.name}</formatOnVm><fullFormat>Issue detected for {vm.name} on {host.name}\ | |
| \ in {datacenter.name}: {message}</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.GeneralVmWarningEvent">\t\t<description>\t\ | |
| \t\tA general warning event occurred on the virtual machine\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>GhostDvsProxySwitchDetectedEvent</key><description>The\ | |
| \ vSphere Distributed Switch corresponding to the proxy switches on the host\ | |
| \ does not exist in vCenter Server or does not contain this host.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ vSphere Distributed Switch corresponding to the proxy switches {switchUuid}\ | |
| \ on the host does not exist in vCenter Server or does not contain this host.</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ vSphere Distributed Switch corresponding to the proxy switches {switchUuid}\ | |
| \ on the host {host.name} does not exist in vCenter Server or does not contain\ | |
| \ this host.</fullFormat><longDescription> <EventLongDescription id="vim.event.GhostDvsProxySwitchDetectedEvent">\ | |
| \ <description> vCenter Server found a vSphere Distributed\ | |
| \ Switch proxy switch on the host that does not match any vSphere\ | |
| \ Distributed Switch in vCenter Server </description> <cause>\ | |
| \ <description> The vSphere Distributed Switch\ | |
| \ corresponding to the vSphere Distributed Switch proxy switch \ | |
| \ on the host was deleted while host was disconnected from the vCenter\ | |
| \ Server </description> </cause> <cause>\ | |
| \ <description> The host is no longer a member\ | |
| \ of the vSphere Distributed Switch that the proxy switch in the host \ | |
| \ corresponds to </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>GhostDvsProxySwitchRemovedEvent</key><description>A\ | |
| \ ghost proxy switch on the host was resolved.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>A\ | |
| \ ghost proxy switch {switchUuid} on the host was resolved.</formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ ghost proxy switch {switchUuid} on the host {host.name} was resolved.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>GlobalMessageChangedEvent</key><description>Message\ | |
| \ changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ message changed: {message}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HealthStatusChangedEvent</key><description>Status\ | |
| \ change</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>{componentName}\ | |
| \ status changed from {oldStatus} to {newStatus}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostAddFailedEvent</key><description>Cannot\ | |
| \ add host</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ add host {hostname}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ add host {hostname}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ add host {hostname} to datacenter {datacenter.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostAddFailedEvent">\t\ | |
| \t<description>\t\t\tAdding a host failed\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostAddedEvent</key><description>Host\ | |
| \ Added</description><category>info</category><formatOnDatacenter>Added host\ | |
| \ {host.name}</formatOnDatacenter><formatOnComputeResource>Added host {host.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Added\ | |
| \ host {host.name} to datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostAdminDisableEvent</key><description>Host\ | |
| \ administrator access disabled</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Administrator\ | |
| \ access to the host is disabled</formatOnHost><formatOnVm></formatOnVm><fullFormat>Administrator\ | |
| \ access to the host {host.name} is disabled</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostAdminDisableEvent">\t\ | |
| \t<description>\t\t\tHost permissions have been changed so that only\ | |
| \ the account used for \t\t\tvCenter Server operations has Administrator permissions\t\ | |
| \t</description>\t\t<cause>\t\t\t<description>\t\t\t\tThis\ | |
| \ condition occurs when vCenter Server removes all other Administrator \t\t\ | |
| \t\taccess to the host because the host has been placed in Lockdown Mode.\ | |
| \ \t\t\t\tThe host can be managed by vCenter Server only and \t\t\t\tOnly\ | |
| \ vCenter Server can re-enable Administrator access for other accounts.\t\t\ | |
| \t</description>\t\t</cause>\t</EventLongDescription> \ | |
| \ </longDescription></eventInfo><eventInfo><key>HostAdminEnableEvent</key><description>Host\ | |
| \ administrator access enabled</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Administrator\ | |
| \ access to the host has been restored</formatOnHost><formatOnVm></formatOnVm><fullFormat>Administrator\ | |
| \ access to the host {host.name} has been restored</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostAdminEnableEvent">\t\ | |
| \t<description>\t\t\tvCenter Server has restored Administrator permissions\ | |
| \ for host user accounts \t\t\t whose permissions were disabled by Lockdown\ | |
| \ Mode\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tThis condition occurs when vCenter Server restores Administrator access\ | |
| \ to \t\t\t\thost user accounts that lost their Administrator permissions\ | |
| \ when the host was \t\t\t\tplaced in Lockdown Mode\t\t\t</description>\t\ | |
| \t</cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedAccountFailedEvent</key><description>Cannot\ | |
| \ connect host and configure management account</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: cannot configure management account</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: cannot configure management account</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: cannot configure management account</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: cannot configure management account</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedAccountFailedEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host because setting up\ | |
| \ a management account failed\t\t</description>\t\t<cause>\t\t\ | |
| \t<description>\t\t\tThe account used by vCenter Server to manage the\ | |
| \ host could not be configured\t\t\t</description>\t\t</cause>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedAlreadyManagedEvent</key><description>Cannot\ | |
| \ connect host - already managed</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: already managed by {serverName}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: already managed by {serverName}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: already managed by {serverName}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: already managed by {serverName}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedAlreadyManagedEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host because it is already\ | |
| \ being managed by a \t\t\tdifferent vCenter Server instance.\t\t</description>\t\ | |
| \t<cause>\t\t\t<description>\t\t\t\tThe host is already being\ | |
| \ managed by a different vCenter Server instance\t\t\t</description>\t\ | |
| \t\t<action>\t\t\t\tRemove the host from the inventory for the other\ | |
| \ vCenter Server instance\t\t\t</action>\t\t\t<action>\t\t\t\t\ | |
| Force the addition of the host to the current vCenter Server instance\t\t\t\ | |
| </action>\t\t</cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedBadCcagentEvent</key><description>Cannot\ | |
| \ connect host - incorrect Ccagent</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name} : server agent is not responding</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name} : server agent is not responding</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: server agent is not responding</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect host {host.name} in {datacenter.name} : server agent is not responding</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedBadCcagentEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host because the host\ | |
| \ agent did not respond\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tNo response was received from the host agent\t\t\t</description>\t\ | |
| \t\t<action>\t\t\t\tRestart the host agent on the ESX/ESXi host\t\t\t\ | |
| </action>\t\t</cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedBadUsernameEvent</key><description>Cannot\ | |
| \ connect host - incorrect user name</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: incorrect user name or password</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: incorrect user name or password</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: incorrect user name or password</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: incorrect user name or password</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedBadUsernameEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host due to an invalid\ | |
| \ username and password combination\t\t</description>\t\t<cause>\t\ | |
| \t\t<description>\t\t\t\tInvalid username and password combination\t\ | |
| \t\t</description>\t\t\t<action>\t\t\t\tUse the correct username\ | |
| \ and password\t\t\t</action>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostCnxFailedBadVersionEvent</key><description>Cannot\ | |
| \ connect host - incompatible version</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: incompatible version</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: incompatible version</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: incompatible version</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: incompatible version</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedBadVersionEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host due to an incompatible\ | |
| \ vSphere Client version\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\tThe version of the vSphere Client is incompatible with the ESX/ESXi host\ | |
| \ so \t\t\tthe connection attempt failed\t\t\t</description>\t\t\t<action>\t\ | |
| \t\tDownload and use a compatible vSphere Client version to connect to the\ | |
| \ host\t\t\t</action>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostCnxFailedCcagentUpgradeEvent</key><description>Cannot\ | |
| \ connect host - Ccagent upgrade</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect host {host.name}: did not install or upgrade vCenter agent service</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect host {host.name}: did not install or upgrade vCenter agent service</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: did not install or upgrade vCenter agent service</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect host {host.name} in {datacenter.name}. Did not install or upgrade\ | |
| \ vCenter agent service.</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.HostCnxFailedCcagentUpgradeEvent">\t\t<description>\t\ | |
| \t\tCould not connect to the host because a host agent upgrade or \t\t\tinstallation\ | |
| \ is in process\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tThe host agent is being upgraded or installed on the host\t\t\t</description>\t\ | |
| \t\t<action>\t\t\t\tWait for the host agent upgrade or installation\ | |
| \ to complete\t\t\t</action>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostCnxFailedEvent</key><description>Cannot\ | |
| \ connect host</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect host {host.name}: error connecting to host</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect host {host.name}: error connecting to host</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: error connecting to host</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: error connecting to host</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host due to an unspecified\ | |
| \ condition\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tUnknown cause of failure\t\t\t</description>\t\t</cause>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedNetworkErrorEvent</key><description>Cannot\ | |
| \ connect host - network error</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: network error</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: network error</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: network error</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: network error</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedNetworkErrorEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host due to a network\ | |
| \ error\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tA Network error occurred while connecting to the host\t\t\t</description>\t\ | |
| \t\t<action>\t\t\t\tVerify that host networking is configured correctly\t\ | |
| \t\t</action>\t\t</cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedNoAccessEvent</key><description>Cannot\ | |
| \ connect host - no access</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: account has insufficient privileges</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: account has insufficient privileges</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: account has insufficient privileges</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect host {host.name} in {datacenter.name}: account has insufficient\ | |
| \ privileges</fullFormat><longDescription>\t<EventLongDescription id="vim.event.HostCnxFailedNoAccessEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host due to insufficient\ | |
| \ account privileges\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tThe account used to connect to host does not have host access privileges\t\ | |
| \t\t</description>\t\t\t<action>\t\t\t\tUse an account that has\ | |
| \ sufficient privileges to connect to the host\t\t\t</action>\t\t</cause>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedNoConnectionEvent</key><description>Cannot\ | |
| \ connect host - no connection</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}</formatOnComputeResource><formatOnHost>Cannot connect\ | |
| \ to host</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot connect\ | |
| \ host {host.name} in {datacenter.name}</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.HostCnxFailedNoConnectionEvent">\t\t<description>\t\ | |
| \t\tCould not connect to the host because the host is not in the network\t\ | |
| \t</description>\t\t<cause>\t\t\t<description>\t\t\t\tThe\ | |
| \ host that you are attempting to connect to is not present in the network\t\ | |
| \t\t</description>\t\t\t<action>\t\t\t\tVerify that host networking\ | |
| \ is configured correctly and the host is connected \t\t\t\tto the same network\ | |
| \ as vCenter Server\t\t\t</action>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostCnxFailedNoLicenseEvent</key><description>Cannot\ | |
| \ connect host - no license</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: not enough CPU licenses</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: not enough CPU licenses</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: not enough CPU licenses</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: not enough CPU licenses</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedNoLicenseEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host due to a licensing\ | |
| \ issue\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tThere are not enough licenses to add the host to the vCenter Server\ | |
| \ inventory. \t\t\t\tThis event is accompanied by a fault that specifies the\ | |
| \ missing licenses \t\t\t\trequired to add the host successfully.\t\t\t</description>\t\ | |
| \t\t<action>\t\t\t\tAdd the necessary licenses to vCenter Server and\ | |
| \ try adding the host again\t\t\t</action>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostCnxFailedNotFoundEvent</key><description>Cannot\ | |
| \ connect host - host not found</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: incorrect host name</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: incorrect host name</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: incorrect host name</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: incorrect host name</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedNotFoundEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host because vCenter Server\ | |
| \ could not resolve the host name\t\t</description>\t\t<cause>\t\ | |
| \t\t<description>\t\t\t\tUnable to resolve the host name of the host\t\ | |
| \t\t</description>\t\t\t<action>\t\t\t\tVerify that the correct\ | |
| \ host name has been supplied for the host\t\t\t</action>\t\t\t<action>\t\ | |
| \t\t\tConfigure the host to use a known-good (resolvable) host name\t\t\t\ | |
| </action>\t\t\t<action>\t\t\t\tAdd the host name to the DNS server\t\ | |
| \t\t</action>\t\t</cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCnxFailedTimeoutEvent</key><description>Cannot\ | |
| \ connect host - time-out</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ connect {host.name}: time-out waiting for host response</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ connect {host.name}: time-out waiting for host response</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ connect: time-out waiting for host response</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ connect {host.name} in {datacenter.name}: time-out waiting for host response</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostCnxFailedTimeoutEvent">\t\ | |
| \t<description>\t\t\tCould not connect to the host because the connection\ | |
| \ attempt timed out\t\t</description>\t\t<cause>\t\t\t<description>\t\ | |
| \t\t\tA timeout occurred while attempting to connect to the host\t\t\t</description>\t\ | |
| \t</cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostComplianceCheckedEvent</key><description>Checked\ | |
| \ host for compliance</description><category>info</category><formatOnDatacenter>Host\ | |
| \ {host.name} checked for compliance with profile {profile.name}</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} checked for compliance with profile {profile.name}</formatOnComputeResource><formatOnHost>Checked\ | |
| \ host for compliance with profile {profile.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} checked for compliance.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.HostComplianceCheckedEvent"> <description>\ | |
| \ The host was checked for compliance with a host profile </description>\ | |
| \ <cause> <description> The user initiated\ | |
| \ a compliance check on the host against a host profile </description>\ | |
| \ </cause> <cause> <description> \ | |
| \ A scheduled task initiated a compliance check for the\ | |
| \ host against a host profile </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostCompliantEvent</key><description>Host\ | |
| \ compliant with profile</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Host\ | |
| \ is in compliance with the attached profile.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} is in compliance with the attached profile</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostConfigAppliedEvent</key><description>Host\ | |
| \ configuration changes applied to host</description><category>info</category><formatOnDatacenter>Host\ | |
| \ configuration changes applied to {host.name}</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ configuration changes applied to {host.name}</formatOnComputeResource><formatOnHost>Host\ | |
| \ configuration changes applied.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ configuration changes applied.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostConnectedEvent</key><description>Host\ | |
| \ connected</description><category>info</category><formatOnDatacenter>Connected\ | |
| \ to {host.name}</formatOnDatacenter><formatOnComputeResource>Connected to\ | |
| \ {host.name}</formatOnComputeResource><formatOnHost>Established a connection</formatOnHost><formatOnVm></formatOnVm><fullFormat>Connected\ | |
| \ to {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostConnectionLostEvent</key><description>Host\ | |
| \ connection lost</description><category>error</category><formatOnDatacenter>Host\ | |
| \ {host.name} is not responding</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} is not responding</formatOnComputeResource><formatOnHost>Host\ | |
| \ is not responding</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} in {datacenter.name} is not responding</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostConnectionLostEvent">\t\ | |
| \t<description>\t\t\tConnection to the host has been lost\t\t</description>\t\ | |
| \t<cause>\t\t\t<description>\t\t\t\tThe host is not in a state\ | |
| \ where it can respond\t\t\t</description>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostDVPortEvent</key><description>dvPort\ | |
| \ connected to host changed status</description><category>info</category><formatOnDatacenter>dvPort\ | |
| \ connected to host {host.name} changed status</formatOnDatacenter><formatOnComputeResource>dvPort\ | |
| \ connected to host {host.name} changed status</formatOnComputeResource><formatOnHost>dvPort\ | |
| \ changed status</formatOnHost><formatOnVm></formatOnVm><fullFormat>dvPort\ | |
| \ connected to host {host.name} in {datacenter.name} changed status</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostDasDisabledEvent</key><description>vSphere\ | |
| \ HA agent disabled on host</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on {host.name} in cluster {computeResource.name} is disabled</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on {host.name} is disabled</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent on this host is disabled</formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ is disabled</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostDasDisablingEvent</key><description>Disabling\ | |
| \ vSphere HA</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA is being disabled on {host.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA is being disabled on {host.name}</formatOnComputeResource><formatOnHost>Disabling\ | |
| \ vSphere HA</formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere HA\ | |
| \ is being disabled on {host.name} in cluster {computeResource.name} in datacenter\ | |
| \ {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostDasEnabledEvent</key><description>vSphere\ | |
| \ HA agent enabled on host</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on {host.name} in cluster {computeResource.name} is enabled</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on {host.name} is enabled</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent on this host is enabled</formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ is enabled</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostDasEnablingEvent</key><description>Enabling\ | |
| \ host vSphere HA agent</description><category>warning</category><formatOnDatacenter>Enabling\ | |
| \ vSphere HA agent on {host.name}</formatOnDatacenter><formatOnComputeResource>Enabling\ | |
| \ vSphere HA agent on {host.name}</formatOnComputeResource><formatOnHost>Enabling\ | |
| \ vSphere HA agent</formatOnHost><formatOnVm></formatOnVm><fullFormat>Enabling\ | |
| \ vSphere HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostDasEnablingEvent">\ | |
| \ <description> vSphere HA is being enabled on this\ | |
| \ host. </description> </EventLongDescription> \ | |
| \ </longDescription></eventInfo><eventInfo><key>HostDasErrorEvent</key><description>vSphere\ | |
| \ HA agent error</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on host {host.name} has an error {message} : {[email protected]}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on host {host.name} has an error {message} : {[email protected]}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent has an error {message} : {[email protected]}</formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA agent on {host.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ has an error {message}: {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostDasEvent</key><description><Host\ | |
| \ vSphere HA Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostDasOkEvent</key><description>vSphere\ | |
| \ HA agent configured</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on host {host.name} is configured correctly</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on host {host.name} is configured correctly</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent is configured correctly</formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA agent on host {host.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ is configured correctly</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostDisconnectedEvent</key><description>Host\ | |
| \ disconnected</description><category>info</category><formatOnDatacenter>Disconnected\ | |
| \ from {host.name}. Reason: {[email protected]}</formatOnDatacenter><formatOnComputeResource>Disconnected\ | |
| \ from {host.name}. Reason: {[email protected]}</formatOnComputeResource><formatOnHost>Disconnected\ | |
| \ from host. Reason: {[email protected]}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Disconnected\ | |
| \ from {host.name} in {datacenter.name}. Reason: {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostEnableAdminFailedEvent</key><description>Cannot\ | |
| \ restore administrator permissions to host</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Cannot\ | |
| \ restore some administrator permissions to the host</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ restore some administrator permissions to the host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostEvent</key><description><Host\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostExtraNetworksEvent</key><description>Host\ | |
| \ has extra vSphere HA networks</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Host\ | |
| \ {host.name} has the following extra networks not used by other hosts for\ | |
| \ vSphere HA communication:{ips}. Consider using vSphere HA advanced option\ | |
| \ das.allowNetwork to control network usage</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} has the following extra networks not used by other hosts for\ | |
| \ vSphere HA communication:{ips}. Consider using vSphere HA advanced option\ | |
| \ das.allowNetwork to control network usage</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostExtraNetworksEvent">\ | |
| \ <description> The host being added to the vSphere HA cluster\ | |
| \ has more management networks than existing hosts in the cluster.\ | |
| \ When vSphere HA is being configured for a host, an existing host\ | |
| \ in the cluster is examined for the networks used by vSphere HA for heartbeats\ | |
| \ and other vSphere HA communication. The joining host is expected\ | |
| \ to have the same number of management networks, and optimally, be\ | |
| \ on the same subnets. This helps to facilitate the pairing up of source/destination\ | |
| \ pairs for heartbeats. If common subnets are not detected (using the\ | |
| \ IP address/subnet mask) between the member being added and the existing\ | |
| \ members of the cluster, this event is generated and the configuration\ | |
| \ task fails. The event details report the subnet of the joining member\ | |
| \ that are not present on the existing member. </description>\ | |
| \ <cause> <description> The host\ | |
| \ has extra networks missing on an existing cluster member </description>\ | |
| \ <action> Change the host's network\ | |
| \ configuration to enable vSphere HA traffic on the same subnets\ | |
| \ as existing hosts in the cluster. vSphere HA will use the Service Console\ | |
| \ port groups on ESX and, on ESXi hosts, the port groups with\ | |
| \ the "Management Traffic" checkbox selected. \ | |
| \ </action> <action> Use advanced\ | |
| \ options to override the default port group selection for vSphere HA \ | |
| \ cluster communication. You can use the das.allowNetwork[X]\ | |
| \ advanced option to tell vSphere HA to use the port group\ | |
| \ specified in this option. For each port group name that\ | |
| \ should be used, specify one das.allowNetwork[X] advanced\ | |
| \ option. The vSphere HA configuration examines the host being added for\ | |
| \ port groups that match the name specified. The configuration\ | |
| \ task also examines an existing member whose port groups\ | |
| \ match the name specified. The number of matched port group\ | |
| \ names must be the same on each host. After setting the advanced \ | |
| \ options, re-enable vSphere HA for the cluster. </action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostGetShortNameFailedEvent</key><description>Cannot\ | |
| \ get short host name</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Cannot\ | |
| \ complete command 'hostname -s' or returned incorrect name format</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ complete command 'hostname -s' on host {host.name} or returned\ | |
| \ incorrect name format</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.HostGetShortNameFailedEvent"> <description>\ | |
| \ The hostname -s command has failed on the host </description>\ | |
| \ <cause> <description>The host network is improperly\ | |
| \ configured</description> <action>Check the host network\ | |
| \ configuration</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostInAuditModeEvent</key><description>Host\ | |
| \ is in audit mode.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Host\ | |
| \ is running in audit mode.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} is running in audit mode. The host's configuration will\ | |
| \ not be persistent across reboots.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostInventoryFullEvent</key><description>Host\ | |
| \ inventory full</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Maximum\ | |
| \ ({capacity}) number of hosts allowed for this edition of vCenter Server\ | |
| \ has been reached</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.HostInventoryFullEvent"> <description>\ | |
| \ The vCenter Server Foundation license key currently allows only\ | |
| \ three hosts to be added to the inventory. Adding extra hosts results\ | |
| \ in errors and the logging of this event. </description>\ | |
| \ <cause> <description>Attempting to add more\ | |
| \ hosts than the number allowed by the license key\ | |
| \ assigned to vCenter Server</description> <action>Assign\ | |
| \ vCenter Server a license key that allows more hosts or has \ | |
| \ no host limit</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostInventoryUnreadableEvent</key><description>Host\ | |
| \ Inventory Unreadable</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ virtual machine inventory file is damaged or unreadable.</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ virtual machine inventory file on host {host.name} is damaged or unreadable.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostIpChangedEvent</key><description>Host\ | |
| \ IP changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>IP\ | |
| \ address changed from {oldIP} to {newIP}</formatOnHost><formatOnVm></formatOnVm><fullFormat>IP\ | |
| \ address of the host {host.name} changed from {oldIP} to {newIP}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostIpChangedEvent">\ | |
| \ <description> The IP address of the host was changed\ | |
| \ </description> <cause> <description>\ | |
| \ The IP address of the host was changed through vCenter Server </description>\ | |
| \ </cause> <cause> <description>\ | |
| \ The IP address of the host was changed through the host </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostIpInconsistentEvent</key><description>Host\ | |
| \ IP inconsistent</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Configuration\ | |
| \ of host IP address is inconsistent: address resolved to {ipAddress} and\ | |
| \ {ipAddress2}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Configuration\ | |
| \ of host IP address is inconsistent on host {host.name}: address resolved\ | |
| \ to {ipAddress} and {ipAddress2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostIpToShortNameFailedEvent</key><description>Host\ | |
| \ IP to short name not completed</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Cannot\ | |
| \ resolve IP address to short name</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ resolve IP address to short name on host {host.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostIpToShortNameFailedEvent">\ | |
| \ <description> The host's IP address could not\ | |
| \ be resolved to a short name </description> <cause>\ | |
| \ <description>The host or DNS records are improperly configured</description>\ | |
| \ <action>Check the host network configuration</action>\ | |
| \ <action>Check the DNS configuration</action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostIsolationIpPingFailedEvent</key><description>vSphere\ | |
| \ HA isolation address unreachable</description><category>error</category><formatOnDatacenter>vSphere\ | |
| \ HA agent on host {host.name} in cluster {computeResource.name} could not\ | |
| \ reach isolation address: {isolationIp}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA agent on host {host.name} could not reach isolation address: {isolationIp}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA agent on this host could not reach isolation address: {isolationIp}</formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA agent on host {host.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ could not reach isolation address: {isolationIp}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostIsolationIpPingFailedEvent">\ | |
| \ <description> vSphere HA was unable to ping one or more\ | |
| \ of the isolation IP addresses. The inabilikty to ping the addresses \ | |
| \ may cause HA to incorrectly declare the host as network isolated. A\ | |
| \ host is declared as isolated if it cannot ping the configured isolation\ | |
| \ addresses and the vSphere HA agent on the host is unable to access any of\ | |
| \ the agents running on the other cluster hosts. </description>\ | |
| \ <cause> <description>Could not ping the isolation\ | |
| \ address</description> <action>Correct the cause of\ | |
| \ the failure to ping the address</action> <action>\ | |
| \ Use advanced options to change the addresses used by vSphere\ | |
| \ HA for determining if a host is network isolated. By default,\ | |
| \ the isolation address is the default \tgateway of the management\ | |
| \ network. You can override the default using advanced options, or \ | |
| \ \tspecify additional addresses to use for determining if a host is network\ | |
| \ isolated. \tSet the das.useDefaultIsolationAddress advanced option\ | |
| \ to "false" if you prefer that \t\tvSphere HA not use the default\ | |
| \ gateway as the isolation address. Specify the das.isolationAddress[X] \t\ | |
| \tadvanced option for each isolation address that you want to specify. The\ | |
| \ new values \ttake effect when vSphere HA is reconfigured for each\ | |
| \ host. </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostLicenseExpiredEvent</key><description>Host\ | |
| \ license expired</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ host license for {host.name} has expired</fullFormat><longDescription> \ | |
| \ <EventLongDescription id="vim.event.HostLicenseExpiredEvent">\ | |
| \ <description> vCenter Server tracks the expiration\ | |
| \ times of host licenses on the license server and uses this event\ | |
| \ to notify you of any host licenses that are about to expire </description>\ | |
| \ <cause> <description>Host licenses on the license\ | |
| \ server are about to expire</description> <action>Update\ | |
| \ the license server to get a new version of the host license</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostLocalPortCreatedEvent</key><description>A\ | |
| \ host local port is created to recover from management network connectivity\ | |
| \ loss.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>A\ | |
| \ host local port {hostLocalPort.portKey} is created on vSphere Distributed\ | |
| \ Switch {hostLocalPort.switchUuid} to recover from management network connectivity\ | |
| \ loss on virtual NIC device {hostLocalPort.vnic}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ host local port {hostLocalPort.portKey} is created on vSphere Distributed\ | |
| \ Switch {hostLocalPort.switchUuid} to recover from management network connectivity\ | |
| \ loss on virtual NIC device {hostLocalPort.vnic} on the host {host.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostMissingNetworksEvent</key><description>Host\ | |
| \ is missing vSphere HA networks</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Host\ | |
| \ {host.name} does not have the following networks used by other hosts for\ | |
| \ vSphere HA communication:{ips}. Consider using vSphere HA advanced option\ | |
| \ das.allowNetwork to control network usage</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} does not have the following networks used by other hosts for\ | |
| \ vSphere HA communication:{ips}. Consider using vSphere HA advanced option\ | |
| \ das.allowNetwork to control network usage</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostMissingNetworksEvent">\ | |
| \ <description> The host being added to the vSphere HA cluster\ | |
| \ has fewer management networks than existing hosts in the cluster.\ | |
| \ When vSphere HA is being configured for a host, an existing host\ | |
| \ in the cluster is examined for the networks used by vSphere HA for heartbeats\ | |
| \ and other vSphere HA communication. The joining host is expected\ | |
| \ to have the same number of management networks, and optimally, have\ | |
| \ common subnets. This helps facilitate the pairing of source/destination\ | |
| \ pairs for heartbeats. If common subnets are not detected (using the\ | |
| \ IP address/subnet mask) between the member being added and the existing\ | |
| \ members of the cluster, this event is generated and the configuration\ | |
| \ task fails. The event details report the subnets of the existing\ | |
| \ member that are not present on the joining member. </description>\ | |
| \ <cause> <description> The host\ | |
| \ does not have networks compatible with an existing cluster member \ | |
| \ </description> <action> Change\ | |
| \ the host's network configuration to enable vSphere HA traffic on the\ | |
| \ same subnets as existing hosts in the cluster. vSphere HA\ | |
| \ will use the Service Console port groups on ESX and, on\ | |
| \ ESXi hosts, the port groups with the "Management Traffic" \ | |
| \ checkbox selected. After you change the host's network\ | |
| \ configuration, reconfigure \t\tvSphere HA for this host. </action>\ | |
| \ <action> Use advanced options to override\ | |
| \ the default port group selection for vSphere HA cluster\ | |
| \ communication. You can use the das.allowNetwork[X] advanced option \ | |
| \ to tell vSphere HA to use the port group specified in this option.\ | |
| \ For each port group name that should be used, specify one\ | |
| \ das.allowNetwork[X] advanced option. The vSphere HA configuration\ | |
| \ examines the host being added for port groups that match\ | |
| \ the name specified. The configuration task also examines \ | |
| \ an existing member whose port groups match the name specified. The number\ | |
| \ of matched port group names must be the same on each host.\ | |
| \ After setting the advanced \t\toptions, re-enable vSphere HA for this cluster.\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostMonitoringStateChangedEvent</key><description>vSphere\ | |
| \ HA host monitoring state changed</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA host monitoring state in {computeResource.name} changed to {[email protected]}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA host monitoring state changed to {[email protected]}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA host monitoring state in {computeResource.name} in {datacenter.name}\ | |
| \ changed to {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostNoAvailableNetworksEvent</key><description>Host\ | |
| \ has no available networks for vSphere HA communication</description><category>error</category><formatOnDatacenter>Host\ | |
| \ {host.name} in cluster {computeResource.name} currently has no available\ | |
| \ networks for vSphere HA Communication. The following networks are currently\ | |
| \ used by HA: {ips}</formatOnDatacenter><formatOnComputeResource>Host {host.name}\ | |
| \ currently has no available networks for vSphere HA Communication. The following\ | |
| \ networks are currently used by HA: {ips}</formatOnComputeResource><formatOnHost>This\ | |
| \ host currently has no available networks for vSphere HA Communication. The\ | |
| \ following networks are currently used by HA: {ips}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} in cluster {computeResource.name} in {datacenter.name} currently\ | |
| \ has no available networks for vSphere HA Communication. The following networks\ | |
| \ are currently used by HA: {ips}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.HostNoAvailableNetworksEvent"> <description>\ | |
| \ The host being added to the vSphere HA cluster has no management networks\ | |
| \ available for vSphere HA cluster communication. The advanced option\ | |
| \ das.allowNetwork[X] is set, but no port group names match the advanced\ | |
| \ option for this host. </description> <cause> \ | |
| \ <description> The host has no port groups that\ | |
| \ match the names used by the advanced options to control\ | |
| \ which port groups vSphere HA uses </description> \ | |
| \ <action> Delete the advanced options das.allowNetwork[X]\ | |
| \ to allow vSphere HA to select the default management port\ | |
| \ groups </action> <action> \ | |
| \ Correct the names of the port groups specified in the advanced options\ | |
| \ to match those to be used by vSphere HA for this host \ | |
| \ </action> <action> Specify\ | |
| \ additional das.allowNetwork[X] advanced options to match the port group\ | |
| \ names for this host </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostNoHAEnabledPortGroupsEvent</key><description>Host\ | |
| \ has no port groups enabled for vSphere HA</description><category>error</category><formatOnDatacenter>Host\ | |
| \ {host.name} in cluster {computeResource.name} has no port groups enabled\ | |
| \ for vSphere HA communication.</formatOnDatacenter><formatOnComputeResource>Host\ | |
| \ {host.name} has no port groups enabled for vSphere HA communication.</formatOnComputeResource><formatOnHost>This\ | |
| \ host has no port groups enabled for vSphere HA communication.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} in cluster {computeResource.name} in {datacenter.name} has no\ | |
| \ port groups enabled for vSphere HA communication.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostNoHAEnabledPortGroupsEvent">\ | |
| \ <description> vSphere HA has determined that there are\ | |
| \ no management networks available on the host for vSphere HA inter-agent\ | |
| \ communication. </description> <cause> \ | |
| \ <description> The host has no vSphere HA management\ | |
| \ networks available </description> <action>\ | |
| \ If this event is observed when the host is being added to\ | |
| \ a vSphere HA cluster, change the host's network configuration\ | |
| \ to enable vSphere HA traffic on one or more port groups.\ | |
| \ By default, vSphere HA will use the Service Console port groups \ | |
| \ on ESX and ESXi hosts, the port groups with the Management Traffic\ | |
| \ checkbox selected. If vSphere HA was already configured\ | |
| \ on the host, it is possible that the host's network\ | |
| \ settings have changed and invalidated the management network \ | |
| \ configuration. Review the settings to make sure the port groups configured\ | |
| \ for managment network still exist on the host and for ESXi\ | |
| \ the Management Traffic option is enabled. Reconfigure vSphere\ | |
| \ HA on the host after fixing any configuration issues. </action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostNoRedundantManagementNetworkEvent</key><description>No\ | |
| \ redundant management network for host</description><category>warning</category><formatOnDatacenter>Host\ | |
| \ {host.name} in cluster {computeResource.name} currently has no management\ | |
| \ network redundancy</formatOnDatacenter><formatOnComputeResource>Host {host.name}\ | |
| \ currently has no management network redundancy</formatOnComputeResource><formatOnHost>This\ | |
| \ host currently has no management network redundancy</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} in cluster {computeResource.name} in {datacenter.name} currently\ | |
| \ has no management network redundancy</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.HostNoRedundantManagementNetworkEvent"> \ | |
| \ <description> vSphere HA has determined that there is only\ | |
| \ one path for vSphere HA management traffic, resulting in a single\ | |
| \ point of failure. Best practices require more than one path for\ | |
| \ vSphere HA to use for heartbeats and cluster communication. A host\ | |
| \ with a single path is more likely to be declared dead, network partitioned\ | |
| \ or isolated after a network failure. If declared dead, vSphere HA\ | |
| \ will not respond if the host subsequently actually fails, while if\ | |
| \ declared isolated, vSphere HA may apply the isolation response thus\ | |
| \ impacting the uptime of the virtual machines running on it. </description>\ | |
| \ <cause> <description>There is only one port\ | |
| \ group available for vSphere HA communication</description> \ | |
| \ <action>Configure another Service Console port group on the ESX\ | |
| \ host</action> <action> Configure another\ | |
| \ port group on the ESXi host by selecting the "Management Traffic"\ | |
| \ \t\tcheck box </action> <action> \ | |
| \ Use NIC teaming on the management port group to allow ESX or ESXi\ | |
| \ to direct management traffic out of more than one physical\ | |
| \ NIC in case of a path failure </action> <action>\ | |
| \ If you accept the risk of not having redundancy for vSphere\ | |
| \ HA communication, you can eliminate the configuration issue\ | |
| \ by setting the das.ignoreRedundantNetWarning advanced option\ | |
| \ to "true" </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostNonCompliantEvent</key><description>Host\ | |
| \ non-compliant with profile</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Host\ | |
| \ is not in compliance with the attached profile.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} is not in compliance with the attached profile</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostNonCompliantEvent">\ | |
| \ <description> The host does not comply with the host\ | |
| \ profile </description> <cause> <description>\ | |
| \ The host is not in compliance with the attached profile \ | |
| \ </description> <action> Check the Summary\ | |
| \ tab for the host in the vSphere Client to determine the possible\ | |
| \ cause(s) of noncompliance </action> </cause></EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostNotInClusterEvent</key><description>Host\ | |
| \ not in cluster</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Not\ | |
| \ a cluster member in {datacenter.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} is not a cluster member in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostOvercommittedEvent</key><description>Host\ | |
| \ resource overcommitted</description><category>error</category><formatOnDatacenter>Insufficient\ | |
| \ capacity in host {computeResource.name} to satisfy resource configuration</formatOnDatacenter><formatOnComputeResource>Insufficient\ | |
| \ capacity to satisfy resource configuration</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Insufficient\ | |
| \ capacity in host {computeResource.name} to satisfy resource configuration\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.HostOvercommittedEvent"> <description>\ | |
| \ A host does not have sufficient CPU and/or memory capacity to \t\ | |
| \t satisfy its resource configuration. The host has its own admission control,\ | |
| \ so this \t\t condition should never occur. </description> \ | |
| \ <cause> <description>A host has insufficient\ | |
| \ capacity for its resource configuration</description> <action>If\ | |
| \ you encounter this condition, contact VMware Support </action>\t</cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostPrimaryAgentNotShortNameEvent</key><description>Host\ | |
| \ primary agent not specified as short name</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Primary\ | |
| \ agent {primaryAgent} was not specified as a short name</formatOnHost><formatOnVm></formatOnVm><fullFormat>Primary\ | |
| \ agent {primaryAgent} was not specified as a short name to host {host.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostPrimaryAgentNotShortNameEvent">\ | |
| \ <description> The primary agent is not specified in\ | |
| \ short name format </description> <cause> \ | |
| \ <description>The host network is improperly configured</description>\ | |
| \ <action>Check the host network configuration</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostProfileAppliedEvent</key><description>Host\ | |
| \ profile applied</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Profile\ | |
| \ configuration applied to the host</formatOnHost><formatOnVm></formatOnVm><fullFormat>Profile\ | |
| \ is applied on the host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostReconnectionFailedEvent</key><description>Cannot\ | |
| \ reconnect host</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ reconnect to {host.name}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ reconnect to {host.name}</formatOnComputeResource><formatOnHost>Cannot reconnect</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ reconnect to {host.name} in {datacenter.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostReconnectionFailedEvent">\t\ | |
| \t<description>\t\t\tCould not reestablish a connection to the host\t\ | |
| \t</description>\t\t<cause>\t\t\t<description>\t\t\t\tThe\ | |
| \ host is not in a state where it can respond\t\t\t</description>\t\t\ | |
| </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostRemovedEvent</key><description>Host\ | |
| \ removed</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ host {host.name}</formatOnDatacenter><formatOnComputeResource>Removed host\ | |
| \ {host.name}</formatOnComputeResource><formatOnHost>Removed from inventory</formatOnHost><formatOnVm></formatOnVm><fullFormat>Removed\ | |
| \ host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostShortNameInconsistentEvent</key><description>Host\ | |
| \ short name inconsistent</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ names {shortName} and {shortName2} both resolved to the same IP address.\ | |
| \ Check the host's network configuration and DNS entries</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostShortNameInconsistentEvent">\ | |
| \ <description> The name resolution check on the host\ | |
| \ returns different names for the host </description> <cause>\ | |
| \ <description>The host network is improperly configured</description>\ | |
| \ <action>Check the host network configuration</action>\ | |
| \ <action>Check the DNS configuration</action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostShortNameToIpFailedEvent</key><description>Host\ | |
| \ short name to IP not completed</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Cannot\ | |
| \ resolve short name {shortName} to IP address</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ resolve short name {shortName} to IP address on host {host.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostShortNameToIpFailedEvent">\ | |
| \ <description> The short name of the host can not be\ | |
| \ resolved to an IP address </description> <cause>\ | |
| \ <description>The host network is improperly configured</description>\ | |
| \ <action>Check the host network configuration</action>\ | |
| \ <action>Check the DNS configuration</action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostShutdownEvent</key><description>Host\ | |
| \ shut down</description><category>info</category><formatOnDatacenter>Shut\ | |
| \ down of {host.name}: {reason}</formatOnDatacenter><formatOnComputeResource>Shut\ | |
| \ down of {host.name}: {reason}</formatOnComputeResource><formatOnHost>Shut\ | |
| \ down of host: {reason}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Shut\ | |
| \ down of {host.name} in {datacenter.name}: {reason}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostStatusChangedEvent</key><description>Host\ | |
| \ status changed</description><category>info</category><formatOnDatacenter>Configuration\ | |
| \ status on host {computeResource.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnDatacenter><formatOnComputeResource>Configuration\ | |
| \ status changed from {[email protected]} to {[email protected]}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Configuration\ | |
| \ status on host {computeResource.name} changed from {[email protected]}\ | |
| \ to {[email protected]} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostStatusChangedEvent">\ | |
| \ <description> The host status has changed. This status\ | |
| \ is the status of the root resource pool that encompasses the entire\ | |
| \ host. A host status change may be accompanied by the removal of\ | |
| \ a configuration issue if one was previously detected. A host status\ | |
| \ of green indicates that everything is fine. A yellow status indicates\ | |
| \ that the root resource pool does not have the resources to meet the \ | |
| \ reservations of its children. A red status means that a node in the\ | |
| \ resource pool has children whose reservations exceed the configuration\ | |
| \ of the node. </description> <cause> \ | |
| \ <description>The host status changed to yellow</description>\ | |
| \ <action>Reduce the reservation of the resource pools directly\ | |
| \ under the root to match the new capacity</action> \ | |
| \ </cause> <cause> <description>The\ | |
| \ host status changed to red</description> <action>Change\ | |
| \ the resource settings on the resource pools that are red so \ | |
| \ that they can accommodate their child virtual machines. If this is not possible,\ | |
| \ lower the virtual machine reservations. If this is not possible\ | |
| \ either, power off some \t virtual machines.</action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostSyncFailedEvent</key><description>Cannot\ | |
| \ synchronize host</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ synchronize host {host.name}. {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ synchronize host {host.name}. {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ synchronize host {host.name}. {reason.msg}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ synchronize host {host.name}. {reason.msg}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostSyncFailedEvent">\t\ | |
| \t<description>\t\t\tFailed to sync with the vCenter Agent on the host\t\ | |
| \t</description>\t\t<cause> <description>\t\t\ | |
| \t\tThe event contains details on why this failure occurred\t\t\t</description>\ | |
| \ </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>HostUpgradeFailedEvent</key><description>Host\ | |
| \ upgrade failed</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ install or upgrade vCenter agent service on {host.name}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ install or upgrade vCenter agent service on {host.name}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ install or upgrade vCenter agent service on {host.name} in {datacenter.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ install or upgrade vCenter agent service on {host.name} in {datacenter.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.HostUpgradeFailedEvent">\t\ | |
| \t<description>\t\t\tFailed to connect to the host due to an installation\ | |
| \ or upgrade issue\t\t</description>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>HostUserWorldSwapNotEnabledEvent</key><description>The\ | |
| \ userworld swap is not enabled on the host</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ userworld swap is not enabled on the host</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ userworld swap is not enabled on the host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostVnicConnectedToCustomizedDVPortEvent</key><description>Some\ | |
| \ host vNICs were reconfigured to use dvPorts with port level configuration,\ | |
| \ which might be different from the dvPort group.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Host\ | |
| \ vNIC {vnic.vnic} was reconfigured to use dvPort {vnic.port.portKey} with\ | |
| \ port level configuration, which might be different from the dvPort group.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} vNIC {vnic.vnic} was reconfigured to use dvPort {vnic.port.portKey}\ | |
| \ with port level configuration, which might be different from the dvPort\ | |
| \ group.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostWwnChangedEvent</key><description>Host\ | |
| \ WWN changed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>WWNs\ | |
| \ are changed</formatOnHost><formatOnVm></formatOnVm><fullFormat>WWNs are\ | |
| \ changed for {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>HostWwnConflictEvent</key><description>Host\ | |
| \ WWN conflict</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ WWN ({wwn}) conflicts with the currently registered WWN</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ WWN ({wwn}) of {host.name} conflicts with the currently registered WWN</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.HostWwnConflictEvent">\ | |
| \ <description> The WWN (World Wide Name) of this host\ | |
| \ conflicts with the WWN of another host or virtual machine </description>\ | |
| \ <cause> <description> The WWN of this host\ | |
| \ conflicts with WWN of another host </description> </cause>\ | |
| \ <cause> <description> The WWN of this host\ | |
| \ conflicts with WWN of another virtual machine</description> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>IncorrectHostInformationEvent</key><description>Incorrect\ | |
| \ host information</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Information\ | |
| \ needed to acquire the correct set of licenses not provided</formatOnHost><formatOnVm></formatOnVm><fullFormat>Host\ | |
| \ {host.name} did not provide the information needed to acquire the correct\ | |
| \ set of licenses</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.IncorrectHostInformationEvent"> <description>\ | |
| \ The host did not provide the information needed to acquire the correct set\ | |
| \ of licenses </description> <cause> <description>\ | |
| \ The cpuCores, cpuPackages or hostType information on the host is not valid\ | |
| \ </description> </cause> <cause> \ | |
| \ <description> The host information is not available because host\ | |
| \ was added as disconnected </description> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>InfoUpgradeEvent</key><description>Information\ | |
| \ upgrade</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>{message}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>InsufficientFailoverResourcesEvent</key><description>vSphere\ | |
| \ HA failover resources are insufficient</description><category>error</category><formatOnDatacenter>Insufficient\ | |
| \ resources to satisfy vSphere HA failover level on cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Insufficient\ | |
| \ resources to satisfy vSphere HA failover level</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Insufficient\ | |
| \ resources to satisfy vSphere HA failover level on cluster {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.InsufficientFailoverResourcesEvent"> <description>\ | |
| \ The cluster does not have enough unreserved capacity to satisfy\ | |
| \ the level configured for vSphere HA admission control. Failovers\ | |
| \ may still be performed by vSphere HA but will be on a best effort\ | |
| \ basis. </description> <cause> <description>\ | |
| \ If the "number of host failures to tolerate" policy\ | |
| \ is configured and a few virtual machines have a much higher\ | |
| \ CPU or memory reservation than the other virtual machines,\ | |
| \ vSphere HA admission control can be excessively conservative \ | |
| \ to ensure that there are enough unfragmented resources if a host fails.\ | |
| \ </description> <action> \ | |
| \ Use similar CPU and memory reservations for all virtual machines \ | |
| \ in the cluster. If this is not possible, consider using a different\ | |
| \ vSphere HA admission control policy, such as reserving a\ | |
| \ percentage of cluster resource for failover. Alternatively,\ | |
| \ you can use advanced options to specify a cap for the slot size. \ | |
| \ See the vSphere Availability Guide for details. </action>\ | |
| \ </cause> <cause> <description>\ | |
| \ Hosts with vSphere HA agent errors are not good candidates\ | |
| \ for providing failover capacity in the cluster, and their\ | |
| \ resources are not considered for vSphere HA admission control\ | |
| \ purposes. If many hosts have an vSphere HA agent error,\ | |
| \ vCenter Server generates this event. </description> \ | |
| \ <action> Check the event log of the hosts to\ | |
| \ determine the cause of the vSphere HA agent errors. After\ | |
| \ addressing any configuration issues, reconfigure vSphere HA on the affected\ | |
| \ hosts or on the cluster. </action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>InvalidEditionEvent</key><description>Invalid\ | |
| \ edition</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ license edition '{feature}' is invalid</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.InvalidEditionEvent">\ | |
| \ <description> vCenter Server attempted to acquire an\ | |
| \ undefined feature from the license server </description> \ | |
| \ <cause> <description>Any operation that requires\ | |
| \ a feature license such as vMotion, DRS, vSphere HA might result \ | |
| \ in this event if that feature is not defined on the license\ | |
| \ server</description> <action>Verify that the feature\ | |
| \ in question is present on the license server</action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>EventEx</key><description>License\ | |
| \ downgrade</description><category>warning</category><formatOnDatacenter>License\ | |
| \ downgrade</formatOnDatacenter><formatOnComputeResource>License downgrade</formatOnComputeResource><formatOnHost>License\ | |
| \ downgrade</formatOnHost><formatOnVm></formatOnVm><fullFormat>vim.event.LicenseDowngradedEvent|License\ | |
| \ downgrade: {licenseKey} removes the following features: {lostFeatures}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.LicenseDowngradedEvent">\ | |
| \ <description> The installed license reduces the set\ | |
| \ of available features. Some of the features, previously available,\ | |
| \ will not be accessible with the new license. </description>\ | |
| \ <cause> <description>The license has been replaced.</description>\ | |
| \ <action>Revert to the license previously installed if it\ | |
| \ is not already expired.</action> <action>Contact\ | |
| \ VMware in order to obtain new license with the required features.</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>LicenseEvent</key><description><License\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>LicenseExpiredEvent</key><description>License\ | |
| \ expired</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>License\ | |
| \ {feature.featureName} has expired</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>LicenseNonComplianceEvent</key><description>Insufficient\ | |
| \ licenses.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>License\ | |
| \ inventory is not compliant. Licenses are overused</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.LicenseNonComplianceEvent">\ | |
| \ <description> vCenter Server does not strictly enforce\ | |
| \ license usage. Instead, it checks for license overuse periodically.\ | |
| \ If vCenter Server detects overuse, it logs this event and \t triggers an\ | |
| \ alarm. </description> <cause> <description>Overuse\ | |
| \ of licenses</description> <action>Check the license\ | |
| \ reports through the vSphere Client and reduce the number of \ | |
| \ entities using the license key or add a new license key with a greater\ | |
| \ capacity</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>LicenseRestrictedEvent</key><description>Unable\ | |
| \ to acquire licenses due to a restriction on the license server</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Unable\ | |
| \ to acquire licenses due to a restriction in the option file on the license\ | |
| \ server.</fullFormat><longDescription> <EventLongDescription id="vim.event.LicenseRestrictedEvent">\ | |
| \ <description> vCenter Server logs this event if it\ | |
| \ is unable to check out a license from the license server due to\ | |
| \ restrictions in the license file </description> <cause>\ | |
| \ <description>License file in the license server has restrictions\ | |
| \ that prevent check out</description> <action>Check\ | |
| \ the license file and remove any restrictions that you can</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>LicenseServerAvailableEvent</key><description>License\ | |
| \ server available</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>License\ | |
| \ server {licenseServer} is available</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>LicenseServerUnavailableEvent</key><description>License\ | |
| \ server unavailable</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>License\ | |
| \ server {licenseServer} is unavailable</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.LicenseServerUnavailableEvent"> <description>\ | |
| \ vCenter Server tracks the license server state and logs this event\ | |
| \ if the license server has stopped responding. </description>\ | |
| \ <cause> <description>License server is not\ | |
| \ responding and not available to vCenter Server</description> \ | |
| \ <action>Verify that the license server is running. If it is,\ | |
| \ check the connectivity between vCenter Server and the\ | |
| \ license server.</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>LocalDatastoreCreatedEvent</key><description>Local\ | |
| \ datastore created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ local datastore {datastore.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ local datastore {datastore.name} on {host.name}</formatOnComputeResource><formatOnHost>Created\ | |
| \ local datastore {datastore.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Created\ | |
| \ local datastore {datastore.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>LocalTSMEnabledEvent</key><description>ESXi\ | |
| \ Shell is enabled</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>ESXi\ | |
| \ Shell for the host has been enabled</formatOnHost><formatOnVm></formatOnVm><fullFormat>ESXi\ | |
| \ Shell for the host {host.name} has been enabled</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>LockerMisconfiguredEvent</key><description>Locker\ | |
| \ misconfigured</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Datastore\ | |
| \ {datastore} which is configured to back the locker does not exist</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>LockerReconfiguredEvent</key><description>Locker\ | |
| \ reconfigured</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Locker\ | |
| \ was reconfigured from {oldDatastore} to {newDatastore} datastore</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>MigrationErrorEvent</key><description>Migration\ | |
| \ error</description><category>error</category><formatOnDatacenter>Unable\ | |
| \ to migrate {vm.name} from {host.name}: {fault.msg}</formatOnDatacenter><formatOnComputeResource>Unable\ | |
| \ to migrate {vm.name}: {fault.msg}</formatOnComputeResource><formatOnHost>Unable\ | |
| \ to migrate {vm.name}: {fault.msg}</formatOnHost><formatOnVm>Unable to migrate\ | |
| \ from {host.name}: {fault.msg}</formatOnVm><fullFormat>Unable to migrate\ | |
| \ {vm.name} from {host.name} in {datacenter.name}: {fault.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.MigrationErrorEvent">\ | |
| \ <description> A virtual machine failed to\ | |
| \ migrate because it did not meet all compatibility criteria </description>\ | |
| \ <cause> <description> Migrating a\ | |
| \ virtual machine from the source host failed because \ | |
| \ the virtual machine did not meet all the compatibility criteria \ | |
| \ </description> <action> \ | |
| \ Use the VSphere Client to check for errors at the time \ | |
| \ of the failure so that you can identify possible reasons\ | |
| \ for this problem </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>MigrationEvent</key><description><Migration\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>MigrationHostErrorEvent</key><description>Migration\ | |
| \ host error</description><category>error</category><formatOnDatacenter>Unable\ | |
| \ to migrate {vm.name} from {host.name} to {dstHost.name}: {fault.msg}</formatOnDatacenter><formatOnComputeResource>Unable\ | |
| \ to migrate {vm.name} to host {dstHost.name}: {fault.msg}</formatOnComputeResource><formatOnHost>Unable\ | |
| \ to migrate {vm.name} to {dstHost.name}: {fault.msg}</formatOnHost><formatOnVm>Unable\ | |
| \ to migrate from {host.name} to {dstHost.name}: {fault.msg}</formatOnVm><fullFormat>Unable\ | |
| \ to migrate {vm.name} from {host.name} to {dstHost.name} in {datacenter.name}:\ | |
| \ {fault.msg}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.MigrationHostErrorEvent"> <description>\ | |
| \ A virtual machine failed to migrate because it did not meet\ | |
| \ all compatibility criteria </description> <cause>\ | |
| \ <description> Migrating a virtual machine to the destination\ | |
| \ host or datastore failed because the virtual machine\ | |
| \ did not meet all the compatibility criteria </description>\ | |
| \ <action> Use the VSphere Client\ | |
| \ to check for errors at the time of the failure so\ | |
| \ that you can identify possible reasons for this problem \ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>MigrationHostWarningEvent</key><description>Migration\ | |
| \ host warning</description><category>warning</category><formatOnDatacenter>Migration\ | |
| \ of {vm.name} from {host.name} to {dstHost.name}: {fault.msg}</formatOnDatacenter><formatOnComputeResource>Migration\ | |
| \ of {vm.name} to {dstHost.name}: {fault.msg}</formatOnComputeResource><formatOnHost>Migration\ | |
| \ of {vm.name} to {dstHost.name}: {fault.msg}</formatOnHost><formatOnVm>Migration\ | |
| \ from {host.name} to {dstHost.name}: {fault.msg}</formatOnVm><fullFormat>Migration\ | |
| \ of {vm.name} from {host.name} to {dstHost.name} in {datacenter.name}: {fault.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.MigrationHostWarningEvent">\ | |
| \ <description> The virtual machine can be\ | |
| \ migrated but might lose some functionality after migration \t is complete\ | |
| \ </description> <cause> \ | |
| \ <description> Migrating the virtual machine to the destination host\ | |
| \ or datastore is likely to succeed but some functionality\ | |
| \ might not work correctly afterward because the virtual \t \t machine\ | |
| \ did not meet all the compatibility criteria. </description>\ | |
| \ <action> Use the VSphere Client\ | |
| \ to check for warnings at the time of the failure so\ | |
| \ that you can identify possible reasons for this problem \ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>MigrationResourceErrorEvent</key><description>Migration\ | |
| \ resource error</description><category>error</category><formatOnDatacenter>Unable\ | |
| \ to migrate {vm.name} from {host.name} to {dstHost.name} and resource pool\ | |
| \ {dstPool.name}: {fault.msg}</formatOnDatacenter><formatOnComputeResource>Unable\ | |
| \ to migrate {vm.name} to {dstHost.name} and resource pool {dstPool.name}:\ | |
| \ {fault.msg}</formatOnComputeResource><formatOnHost>Unable to migrate {vm.name}\ | |
| \ to {dstHost.name} and resource pool {dstPool.name}: {fault.msg}</formatOnHost><formatOnVm>Unable\ | |
| \ to migrate from {host.name} to {dstHost.name} and resource pool {dstPool.name}:\ | |
| \ {fault.msg}</formatOnVm><fullFormat>Cannot migrate {vm.name} from {host.name}\ | |
| \ to {dstHost.name} and resource pool {dstPool.name} in {datacenter.name}:\ | |
| \ {fault.msg}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.MigrationResourceErrorEvent"> <description>\ | |
| \ A virtual machine failed to migrate due to incompatibilities\ | |
| \ with target resource pool </description> <cause>\ | |
| \ <description>Migrating a virtual machine to the destination\ | |
| \ host or datastore is not possible due to incompatibilities\ | |
| \ with the target resource pool. </description> \ | |
| \ <action> Use the VSphere Client\ | |
| \ to check for errors at the time of the failure so\ | |
| \ that you can identify possible reasons for this problem \ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>MigrationResourceWarningEvent</key><description>Migration\ | |
| \ resource warning</description><category>warning</category><formatOnDatacenter>Migration\ | |
| \ of {vm.name} from {host.name} to {dstHost.name} and resource pool {dstPool.name}:\ | |
| \ {fault.msg}</formatOnDatacenter><formatOnComputeResource>Migration of {vm.name}\ | |
| \ to {dstHost.name} and resource pool {dstPool.name}: {fault.msg}</formatOnComputeResource><formatOnHost>Migration\ | |
| \ of {vm.name} to {dstHost.name} and resource pool {dstPool.name}: {fault.msg}</formatOnHost><formatOnVm>Migration\ | |
| \ from {host.name} to {dstHost.name} and resource pool {dstPool.name}: {fault.msg}</formatOnVm><fullFormat>Migration\ | |
| \ of {vm.name} from {host.name} to {dstHost.name} and resource pool {dstPool.name}\ | |
| \ in {datacenter.name}: {fault.msg}</fullFormat><longDescription> \ | |
| \ <EventLongDescription id="vim.event.MigrationResourceWarningEvent">\ | |
| \ <description> The virtual machine can be\ | |
| \ migrated but might lose some functionality after migration \t is complete\ | |
| \ </description> <cause> \ | |
| \ <description> Migrating the virtual machine to the destination resource\ | |
| \ pool is likely to succeed but some functionality might\ | |
| \ not work correctly afterward because the virtual \t \t machine did\ | |
| \ not meet all the compatibility criteria. </description>\ | |
| \ <action> Use the VSphere Client\ | |
| \ to check for warnings at the time of the failure so\ | |
| \ that you can identify possible reasons for this problem \ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>MigrationWarningEvent</key><description>Migration\ | |
| \ warning</description><category>warning</category><formatOnDatacenter>Migration\ | |
| \ of {vm.name} from {host.name}: {fault.msg}</formatOnDatacenter><formatOnComputeResource>Migration\ | |
| \ of {vm.name}: {fault.msg}</formatOnComputeResource><formatOnHost>Migration\ | |
| \ of {vm.name}: {fault.msg}</formatOnHost><formatOnVm>Migration from {host.name}:\ | |
| \ {fault.msg}</formatOnVm><fullFormat>Migration of {vm.name} from {host.name}\ | |
| \ in {datacenter.name}: {fault.msg}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>MtuMatchEvent</key><description>The\ | |
| \ MTU configured in the vSphere Distributed Switch matches the physical switch\ | |
| \ connected to the physical NIC.</description><category>info</category><formatOnDatacenter>The\ | |
| \ MTU configured in the vSphere Distributed Switch matches the physical switch\ | |
| \ connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed\ | |
| \ Switch {dvs.name} on host {host.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ MTU configured in the vSphere Distributed Switch matches the physical switch\ | |
| \ connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed\ | |
| \ Switch {dvs.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ MTU configured in the vSphere Distributed Switch matches the physical switch\ | |
| \ connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed\ | |
| \ Switch {dvs.name} on host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>MtuMismatchEvent</key><description>The\ | |
| \ MTU configured in the vSphere Distributed Switch does not match the physical\ | |
| \ switch connected to the physical NIC.</description><category>error</category><formatOnDatacenter>The\ | |
| \ MTU configured in the vSphere Distributed Switch does not match the physical\ | |
| \ switch connected to uplink port {healthResult.uplinkPortKey} in vSphere\ | |
| \ Distributed Switch {dvs.name} on host {host.name}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ MTU configured in the vSphere Distributed Switch does not match the physical\ | |
| \ switch connected to uplink port {healthResult.uplinkPortKey} in vSphere\ | |
| \ Distributed Switch {dvs.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ MTU configured in the vSphere Distributed Switch does not match the physical\ | |
| \ switch connected to uplink port {healthResult.uplinkPortKey} in vSphere\ | |
| \ Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>NASDatastoreCreatedEvent</key><description>NAS\ | |
| \ datastore created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ NAS datastore {datastore.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ NAS datastore {datastore.name} on {host.name}</formatOnComputeResource><formatOnHost>Created\ | |
| \ NAS datastore {datastore.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Created\ | |
| \ NAS datastore {datastore.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>NetworkRollbackEvent</key><description>Network\ | |
| \ configuration on the host {host.name} is rolled back as it disconnects the\ | |
| \ host from vCenter server.</description><category>error</category><formatOnDatacenter>Network\ | |
| \ configuration on the host {host.name} is rolled back as it disconnects the\ | |
| \ host from vCenter server.</formatOnDatacenter><formatOnComputeResource>Network\ | |
| \ configuration on the host {host.name} is rolled back as it disconnects the\ | |
| \ host from vCenter server.</formatOnComputeResource><formatOnHost>Network\ | |
| \ configuration on the host {host.name} is rolled back as it disconnects the\ | |
| \ host from vCenter server.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Network\ | |
| \ configuration on the host {host.name} is rolled back as it disconnects the\ | |
| \ host from vCenter server.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>NoAccessUserEvent</key><description>No\ | |
| \ access for user</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ login user {userName}@{ipAddress}: no permission</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.NoAccessUserEvent">\t\t\ | |
| <description>\t\t\tA user could not log in due to insufficient access\ | |
| \ permission\t\t</description>\t\t<cause> <description>\t\ | |
| \t\tThe user account has insufficient access permission\t\t\t</description>\t\ | |
| \t\t<action>\t\t\tLog in with a user account that has the necessary\ | |
| \ access permissions or grant \t\t\tadditional access permissions to the current\ | |
| \ user\t\t\t</action> </cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>NoDatastoresConfiguredEvent</key><description>No\ | |
| \ datastores configured</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>No\ | |
| \ datastores have been configured</formatOnHost><formatOnVm></formatOnVm><fullFormat>No\ | |
| \ datastores have been configured on the host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>NoLicenseEvent</key><description>No\ | |
| \ license</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ required license {feature.featureName} is not reserved</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.NoLicenseEvent">\ | |
| \ <description> vCenter Server logs this event if it\ | |
| \ fails to acquire a feature from the license server for an unknown\ | |
| \ reason. </description> <cause> <description>Acquiring\ | |
| \ a feature license fails for an unknown reason</description> \ | |
| \ <action>Verify that the license server has the license for the\ | |
| \ feature</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>NoMaintenanceModeDrsRecommendationForVM</key><description>No\ | |
| \ maintenance mode DRS recommendation for the VM</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Unable\ | |
| \ to automatically migrate {vm.name}</formatOnHost><formatOnVm>Unable to automatically\ | |
| \ migrate from {host.name}</formatOnVm><fullFormat>Unable to automatically\ | |
| \ migrate {vm.name} from {host.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.NoMaintenanceModeDrsRecommendationForVM"> \ | |
| \ <description> DRS failed to generate a vMotion recommendation\ | |
| \ for a virtual machine on a host entering \t\t Maintenance Mode. This condition\ | |
| \ typically occurs because no other host \t\t in the DRS cluster is compatible\ | |
| \ with the virtual machine. Unless you manually \t\t migrate or power off\ | |
| \ this virtual machine, the host will be unable to enter \t\t Maintenance\ | |
| \ Mode. </description> <cause> <description>DRS\ | |
| \ failed to evacuate a powered on virtual machine</description> \ | |
| \ <action>Manually migrate the virtual machine to another host\ | |
| \ in the cluster</action> <action>Power off the virtual\ | |
| \ machine</action> <action>Bring any hosts in Maintenance\ | |
| \ Mode out of that mode</action> <action>Cancel the\ | |
| \ task that is making the host enter Maintenance Mode </action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>NonVIWorkloadDetectedOnDatastoreEvent</key><description>Unmanaged\ | |
| \ workload detected on SIOC-enabled datastore</description><category>info</category><formatOnDatacenter>An\ | |
| \ unmanaged I/O workload is detected on a SIOC-enabled datastore: {datastore.name}.</formatOnDatacenter><formatOnComputeResource>An\ | |
| \ unmanaged I/O workload is detected on a SIOC-enabled datastore: {datastore.name}.</formatOnComputeResource><formatOnHost>An\ | |
| \ unmanaged I/O workload is detected on a SIOC-enabled datastore: {datastore.name}.</formatOnHost><formatOnVm>An\ | |
| \ unmanaged I/O workload is detected on a SIOC-enabled datastore: {datastore.name}.</formatOnVm><fullFormat>An\ | |
| \ unmanaged I/O workload is detected on a SIOC-enabled datastore: {datastore.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>NotEnoughResourcesToStartVmEvent</key><description>Not\ | |
| \ enough resources for vSphere HA to start VM</description><category>warning</category><formatOnDatacenter>Not\ | |
| \ enough resources to failover {vm.name} in {computeResource.name}. vSphere\ | |
| \ HA will retry when resources become available.</formatOnDatacenter><formatOnComputeResource>Not\ | |
| \ enough resources to failover {vm.name}. vSphere HA will retry when resources\ | |
| \ become available.</formatOnComputeResource><formatOnHost>Not enough resources\ | |
| \ to failover {vm.name}. vSphere HA will retry when resources become available.</formatOnHost><formatOnVm>Not\ | |
| \ enough resources to failover this virtual machine. vSphere HA will retry\ | |
| \ when resources become available.</formatOnVm><fullFormat>Not enough resources\ | |
| \ to failover {vm.name} in {computeResource.name} in {datacenter.name}. vSphere\ | |
| \ HA will retry when resources become available.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.NotEnoughResourcesToStartVmEvent</longDescription></eventInfo><eventInfo><key>OutOfSyncDvsHost</key><description>The\ | |
| \ vSphere Distributed Switch configuration on some hosts differed from that\ | |
| \ of the vCenter Server.</description><category>warning</category><formatOnDatacenter>The\ | |
| \ vSphere Distributed Switch configuration on some hosts differed from that\ | |
| \ of the vCenter Server.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ vSphere Distributed Switch configuration on some hosts differed from that\ | |
| \ of the vCenter Server.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>PermissionAddedEvent</key><description>Permission\ | |
| \ added</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Permission\ | |
| \ created for {principal} on {entity.name}, role is {role.name}, propagation\ | |
| \ is {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>PermissionEvent</key><description><Permission\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>PermissionRemovedEvent</key><description>Permission\ | |
| \ removed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Permission\ | |
| \ rule removed for {principal} on {entity.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>PermissionUpdatedEvent</key><description>Permission\ | |
| \ updated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Permission\ | |
| \ changed for {principal} on {entity.name}, role is {role.name}, propagation\ | |
| \ is {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ProfileAssociatedEvent</key><description>Profile\ | |
| \ attached to host</description><category>info</category><formatOnDatacenter>Profile\ | |
| \ {profile.name} has been attached.</formatOnDatacenter><formatOnComputeResource>Profile\ | |
| \ {profile.name} has been attached.</formatOnComputeResource><formatOnHost>Profile\ | |
| \ {profile.name} has been attached with the host.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Profile\ | |
| \ {profile.name} attached.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ProfileChangedEvent</key><description>Profile\ | |
| \ was changed</description><category>info</category><formatOnDatacenter>Profile\ | |
| \ {profile.name} was changed.</formatOnDatacenter><formatOnComputeResource>Profile\ | |
| \ {profile.name} was changed.</formatOnComputeResource><formatOnHost>Profile\ | |
| \ {profile.name} was changed.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Profile\ | |
| \ {profile.name} was changed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ProfileCreatedEvent</key><description>Profile\ | |
| \ created</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Profile\ | |
| \ is created.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ProfileDissociatedEvent</key><description>Profile\ | |
| \ detached from host</description><category>info</category><formatOnDatacenter>Profile\ | |
| \ {profile.name} has been detached.</formatOnDatacenter><formatOnComputeResource>Profile\ | |
| \ {profile.name} has been detached. </formatOnComputeResource><formatOnHost>Profile\ | |
| \ {profile.name} has been detached from the host.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Profile\ | |
| \ {profile.name} detached.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ProfileEvent</key><description></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ProfileReferenceHostChangedEvent</key><description>The\ | |
| \ profile reference host was changed</description><category>info</category><formatOnDatacenter>The\ | |
| \ profile {profile.name} reference host was changed to {referenceHost.name}.</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ profile {profile.name} reference host was changed to {referenceHost.name}.</formatOnComputeResource><formatOnHost>The\ | |
| \ profile {profile.name} reference host was changed to {referenceHost.name}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Profile\ | |
| \ {profile.name} reference host changed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ProfileRemovedEvent</key><description>Profile\ | |
| \ removed</description><category>info</category><formatOnDatacenter>Profile\ | |
| \ {profile.name} was removed.</formatOnDatacenter><formatOnComputeResource>Profile\ | |
| \ {profile.name} was removed.</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Profile\ | |
| \ was removed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>RecoveryEvent</key><description>Recovery\ | |
| \ completed on the host.</description><category>info</category><formatOnDatacenter>The\ | |
| \ host {hostName} network connectivity was recovered on the virtual management\ | |
| \ NIC {vnic}. A new port {portKey} was created on vSphere Distributed Switch\ | |
| \ {dvsUuid}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ host {hostName} network connectivity was recovered on the virtual management\ | |
| \ NIC {vnic}. A new port {portKey} was created on vSphere Distributed Switch\ | |
| \ {dvsUuid}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>The host {hostName}\ | |
| \ network connectivity was recovered on the management virtual NIC {vnic}\ | |
| \ by connecting to a new port {portKey} on the vSphere Distributed Switch\ | |
| \ {dvsUuid}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>RemoteTSMEnabledEvent</key><description>SSH\ | |
| \ is enabled</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>SSH\ | |
| \ for the host has been enabled</formatOnHost><formatOnVm></formatOnVm><fullFormat>SSH\ | |
| \ for the host {host.name} has been enabled</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ResourcePoolCreatedEvent</key><description>Resource\ | |
| \ pool created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ resource pool {resourcePool.name} in compute-resource {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ resource pool {resourcePool.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Created\ | |
| \ resource pool {resourcePool.name} in compute-resource {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ResourcePoolDestroyedEvent</key><description>Resource\ | |
| \ pool deleted</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ resource pool {resourcePool.name} on {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Removed\ | |
| \ resource pool {resourcePool.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Removed\ | |
| \ resource pool {resourcePool.name} on {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ResourcePoolEvent</key><description><Resource\ | |
| \ Pool Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ResourcePoolMovedEvent</key><description>Resource\ | |
| \ pool moved</description><category>info</category><formatOnDatacenter>Moved\ | |
| \ resource pool {resourcePool.name} from {oldParent.name} to {newParent.name}\ | |
| \ on {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Moved\ | |
| \ resource pool {resourcePool.name} from {oldParent.name} to {newParent.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Moved\ | |
| \ resource pool {resourcePool.name} from {oldParent.name} to {newParent.name}\ | |
| \ on {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ResourcePoolReconfiguredEvent</key><description>Resource\ | |
| \ pool reconfigured</description><category>info</category><formatOnDatacenter>Updated\ | |
| \ configuration for {resourcePool.name} in compute-resource {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Updated\ | |
| \ configuration on {resourcePool.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Updated\ | |
| \ configuration for {resourcePool.name} in compute-resource {computeResource.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.ResourcePoolReconfiguredEvent"> <description>\ | |
| \ The resource pool configuration changed. The resource pool configuration\ | |
| \ includes information about the resource reservations of the resource\ | |
| \ pool and the resource reservations of its children. </description>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ResourceViolatedEvent</key><description>Resource\ | |
| \ usage exceeds configuration</description><category>error</category><formatOnDatacenter>Resource\ | |
| \ usage exceeds configuration for resource pool {resourcePool.name} in compute-resource\ | |
| \ {computeResource.name}'</formatOnDatacenter><formatOnComputeResource>Resource\ | |
| \ usage exceeds configuration on resource pool {resourcePool.name}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Resource\ | |
| \ usage exceeds configuration for resource pool {resourcePool.name} in compute-resource\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.ResourceViolatedEvent">\ | |
| \ <description> The cumulative CPU and/or memory resource\ | |
| \ consumption of all virtual machines in the resource pool exceeds\ | |
| \ the resource pool configuration </description> <cause>\ | |
| \ <description>You attempted to move a virtual machine from\ | |
| \ one resource pool into another \t\tbypassing vCenter Server. This condition\ | |
| \ occurs when you attempt the move using \t\tthe vSphere Client directly connected\ | |
| \ to the host.\t\t</description> <action>In a DRS cluster,\ | |
| \ do not move and power on a virtual machine bypassing vCenter Server</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>RoleAddedEvent</key><description>Role\ | |
| \ added</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>New\ | |
| \ role {role.name} created</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>RoleEvent</key><description><Role\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>RoleRemovedEvent</key><description>Role\ | |
| \ removed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Role\ | |
| \ {role.name} removed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>RoleUpdatedEvent</key><description>Role\ | |
| \ updated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Modified\ | |
| \ role {role.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>RollbackEvent</key><description>Host\ | |
| \ Network operation rolled back</description><category>info</category><formatOnDatacenter>The\ | |
| \ Network API {methodName} on this entity caused the host {hostName} to be\ | |
| \ disconnected from the vCenter Server. The configuration change was rolled\ | |
| \ back on the host.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ operation {methodName} on the host {hostName} disconnected the host and\ | |
| \ was rolled back .</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ Network API {methodName} on this entity caused the host {hostName} to be\ | |
| \ disconnected from the vCenter Server. The configuration change was rolled\ | |
| \ back on the host.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ScheduledTaskCompletedEvent</key><description>Scheduled\ | |
| \ task completed</description><category>info</category><formatOnDatacenter>Task\ | |
| \ {scheduledTask.name} on {entity.name} completed successfully</formatOnDatacenter><formatOnComputeResource>Task\ | |
| \ {scheduledTask.name} on {entity.name} completed successfully</formatOnComputeResource><formatOnHost>Task\ | |
| \ {scheduledTask.name} on {entity.name} completed successfully</formatOnHost><formatOnVm>Task\ | |
| \ {scheduledTask.name} completed successfully</formatOnVm><fullFormat>Task\ | |
| \ {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} completed\ | |
| \ successfully</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ScheduledTaskCreatedEvent</key><description>Scheduled\ | |
| \ task created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnComputeResource><formatOnHost>Created\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnHost><formatOnVm>Created\ | |
| \ task {scheduledTask.name}</formatOnVm><fullFormat>Created task {scheduledTask.name}\ | |
| \ on {entity.name} in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ScheduledTaskEmailCompletedEvent</key><description>Sent\ | |
| \ scheduled task email</description><category>info</category><formatOnDatacenter>Task\ | |
| \ {scheduledTask.name} on {entity.name} sent email to {to}</formatOnDatacenter><formatOnComputeResource>Task\ | |
| \ {scheduledTask.name} on {entity.name} sent email to {to}</formatOnComputeResource><formatOnHost>Task\ | |
| \ {scheduledTask.name} on {entity.name} sent email to {to}</formatOnHost><formatOnVm>Task\ | |
| \ {scheduledTask.name} on {entity.name} sent email to {to}</formatOnVm><fullFormat>Task\ | |
| \ {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} sent\ | |
| \ email to {to}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ScheduledTaskEmailFailedEvent</key><description>Scheduled\ | |
| \ task email not sent</description><category>error</category><formatOnDatacenter>Task\ | |
| \ {scheduledTask.name} on {entity.name} cannot send email to {to}: {reason.msg}</formatOnDatacenter><formatOnComputeResource>Task\ | |
| \ {scheduledTask.name} on {entity.name} cannot send email to {to}: {reason.msg}</formatOnComputeResource><formatOnHost>Task\ | |
| \ {scheduledTask.name} on {entity.name} cannot send email to {to}: {reason.msg}</formatOnHost><formatOnVm>Task\ | |
| \ {scheduledTask.name} cannot send email to {to}: {reason.msg}</formatOnVm><fullFormat>Task\ | |
| \ {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} cannot\ | |
| \ send email to {to}: {reason.msg}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.ScheduledTaskEmailFailedEvent"> <description>\ | |
| \ An error occurred while sending email notification that a scheduled\ | |
| \ task is running </description> <cause> \ | |
| \ <description>Failed to send email for the scheduled task</description>\ | |
| \ <action>Check the vCenter Server SMTP settings for sending\ | |
| \ emails</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>ScheduledTaskEvent</key><description><Scheduled\ | |
| \ Task Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ScheduledTaskFailedEvent</key><description>Cannot\ | |
| \ complete scheduled task</description><category>error</category><formatOnDatacenter>Task\ | |
| \ {scheduledTask.name} on {entity.name} cannot be completed: {reason.msg}</formatOnDatacenter><formatOnComputeResource>Task\ | |
| \ {scheduledTask.name} on {entity.name} cannot be completed: {reason.msg}</formatOnComputeResource><formatOnHost>Task\ | |
| \ {scheduledTask.name} on {entity.name} cannot be completed: {reason.msg}</formatOnHost><formatOnVm>Task\ | |
| \ {scheduledTask.name} cannot be completed: {reason.msg}</formatOnVm><fullFormat>Task\ | |
| \ {scheduledTask.name} on {entity.name} in datacenter {datacenter.name} cannot\ | |
| \ be completed: {reason.msg}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.ScheduledTaskFailedEvent"> <description>\ | |
| \ An error occurred while running a scheduled task </description>\ | |
| \ <cause> <description>Failed to run a scheduled\ | |
| \ task</description> <action>Correct the failure condition</action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>ScheduledTaskReconfiguredEvent</key><description>Scheduled\ | |
| \ task reconfigured</description><category>info</category><formatOnDatacenter>Reconfigured\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnDatacenter><formatOnComputeResource>Reconfigured\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnComputeResource><formatOnHost>Reconfigured\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnHost><formatOnVm>Reconfigured\ | |
| \ task {scheduledTask.name}</formatOnVm><fullFormat>Reconfigured task {scheduledTask.name}\ | |
| \ on {entity.name} in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ScheduledTaskRemovedEvent</key><description>Scheduled\ | |
| \ task removed</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnDatacenter><formatOnComputeResource>Removed\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnComputeResource><formatOnHost>Removed\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnHost><formatOnVm>Removed\ | |
| \ task {scheduledTask.name}</formatOnVm><fullFormat>Removed task {scheduledTask.name}\ | |
| \ on {entity.name} in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ScheduledTaskStartedEvent</key><description>Scheduled\ | |
| \ task started</description><category>info</category><formatOnDatacenter>Running\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnDatacenter><formatOnComputeResource>Running\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnComputeResource><formatOnHost>Running\ | |
| \ task {scheduledTask.name} on {entity.name}</formatOnHost><formatOnVm>Running\ | |
| \ task {scheduledTask.name}</formatOnVm><fullFormat>Running task {scheduledTask.name}\ | |
| \ on {entity.name} in datacenter {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ServerLicenseExpiredEvent</key><description>Server\ | |
| \ license expired</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ vCenter Server license has expired</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ServerStartedSessionEvent</key><description>Server\ | |
| \ started session</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vCenter\ | |
| \ started</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>SessionEvent</key><description><Session\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>SessionTerminatedEvent</key><description>Session\ | |
| \ stopped</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ session for user '{terminatedUsername}' has stopped</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.SessionTerminatedEvent">\t\ | |
| \t<description>\t\t\tA session has been terminated\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>TaskEvent</key><description>Task\ | |
| \ event</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Task:\ | |
| \ {info.descriptionId}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>TaskTimeoutEvent</key><description>Task\ | |
| \ time-out</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Task:\ | |
| \ {info.descriptionId} time-out</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.TaskTimeoutEvent">\t\t<description>\t\t\ | |
| \tA task has been cleaned up because it timed out\t\t</description>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>TeamingMatchEvent</key><description>Teaming\ | |
| \ configuration in the vSphere Distributed Switch matches the physical switch\ | |
| \ configuration</description><category>info</category><formatOnDatacenter>Teaming\ | |
| \ configuration in the vSphere Distributed Switch {dvs.name} on host {host.name}\ | |
| \ matches the physical switch configuration. Detail: {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Teaming\ | |
| \ configuration in the vSphere Distributed Switch {dvs.name} matches the physical\ | |
| \ switch configuration. Detail: {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Teaming\ | |
| \ configuration in the vSphere Distributed Switch {dvs.name} on host {host.name}\ | |
| \ matches the physical switch configuration in {datacenter.name}. Detail:\ | |
| \ {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>TeamingMisMatchEvent</key><description>Teaming\ | |
| \ configuration in the vSphere Distributed Switch does not match the physical\ | |
| \ switch configuration</description><category>error</category><formatOnDatacenter>Teaming\ | |
| \ configuration in the vSphere Distributed Switch {dvs.name} on host {host.name}\ | |
| \ does not match the physical switch configuration. Detail: {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus}</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Teaming\ | |
| \ configuration in the vSphere Distributed Switch {dvs.name} does not match\ | |
| \ the physical switch configuration. Detail: {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Teaming\ | |
| \ configuration in the vSphere Distributed Switch {dvs.name} on host {host.name}\ | |
| \ does not match the physical switch configuration in {datacenter.name}. Detail:\ | |
| \ {healthResult.summary.@enum.dvs.VmwareDistributedVirtualSwitch.TeamingMatchStatus}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>TemplateBeingUpgradedEvent</key><description>Upgrading\ | |
| \ template</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Upgrading\ | |
| \ template {legacyTemplate}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>TemplateUpgradeEvent</key><description><Template\ | |
| \ Upgrade Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>TemplateUpgradeFailedEvent</key><description>Cannot\ | |
| \ upgrade template</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ upgrade template {legacyTemplate} due to: {reason.msg}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>TemplateUpgradedEvent</key><description>Template\ | |
| \ upgraded</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Template\ | |
| \ {legacyTemplate} upgrade completed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>TimedOutHostOperationEvent</key><description>Host\ | |
| \ operation timed out</description><category>warning</category><formatOnDatacenter>The\ | |
| \ operation performed on host {host.name} timed out</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ operation performed on host {host.name} timed out</formatOnComputeResource><formatOnHost>The\ | |
| \ operation timed out</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ operation performed on {host.name} in {datacenter.name} timed out</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.TimedOutHostOperationEvent">\t\ | |
| \t<description>\t\t\tAn operation performed on the host has timed out\t\ | |
| \t</description>\t\t<cause>\t\t\t<description>\t\t\tA previous\ | |
| \ event in the sequence of events will provide information on the \t\t\treason\ | |
| \ for the timeout\t\t\t</description>\t\t</cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>UnlicensedVirtualMachinesEvent</key><description>Unlicensed\ | |
| \ virtual machines</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>There\ | |
| \ are {unlicensed} unlicensed virtual machines on host {host} - there are\ | |
| \ only {available} licenses available</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UnlicensedVirtualMachinesFoundEvent</key><description>Unlicensed\ | |
| \ virtual machines found</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>{unlicensed}\ | |
| \ unlicensed virtual machines found on host {host}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UpdatedAgentBeingRestartedEvent</key><description>Restarting\ | |
| \ updated agent</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ agent is updated and will soon restart</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ agent on host {host.name} is updated and will soon restart</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UpgradeEvent</key><description><Upgrade\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UplinkPortMtuNotSupportEvent</key><description>Not\ | |
| \ all VLAN MTU settings on the external physical switch allow the vSphere\ | |
| \ Distributed Switch maximum MTU size packets to pass.</description><category>error</category><formatOnDatacenter>Not\ | |
| \ all VLAN MTU settings on the external physical switch allow the vSphere\ | |
| \ Distributed Switch maximum MTU size packets to pass on the uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name} on host {host.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Not\ | |
| \ all VLAN MTU settings on the external physical switch allow the vSphere\ | |
| \ Distributed Switch maximum MTU size packets to pass on the uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Not\ | |
| \ all VLAN MTU settings on the external physical switch allow the vSphere\ | |
| \ Distributed Switch maximum MTU size packets to pass on the uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UplinkPortMtuSupportEvent</key><description>All\ | |
| \ VLAN MTU settings on the external physical switch allow the vSphere Distributed\ | |
| \ Switch maximum MTU size packets to pass.</description><category>info</category><formatOnDatacenter>All\ | |
| \ VLAN MTU settings on the external physical switch allow the vSphere Distributed\ | |
| \ Switch maximum MTU size packets to pass on uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name} on host {host.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>All\ | |
| \ VLAN MTU setting on the external physical switch allows the vSphere Distributed\ | |
| \ Switch max MTU size packets passing on uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>All\ | |
| \ VLAN MTU settings on the external physical switch allow the vSphere Distributed\ | |
| \ Switch maximum MTU size packets to pass on the uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UplinkPortVlanTrunkedEvent</key><description>The\ | |
| \ configured VLAN in the vSphere Distributed Switch was trunked by the physical\ | |
| \ switch.</description><category>info</category><formatOnDatacenter>The configured\ | |
| \ VLAN in the vSphere Distributed Switch was trunked by the physical switch\ | |
| \ connected to uplink port {healthResult.uplinkPortKey} in vSphere Distributed\ | |
| \ Switch {dvs.name} on host {host.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>The\ | |
| \ configured VLAN in the vSphere Distributed Switch was trunked by the physical\ | |
| \ switch connected to uplink port {healthResult.uplinkPortKey} in vSphere\ | |
| \ Distributed Switch {dvs.name}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ configured VLAN in the vSphere Distributed Switch was trunked by the physical\ | |
| \ switch connected to uplink port {healthResult.uplinkPortKey} in vSphere\ | |
| \ Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UplinkPortVlanUntrunkedEvent</key><description>Not\ | |
| \ all the configured VLANs in the vSphere Distributed Switch were trunked\ | |
| \ by the physical switch.</description><category>error</category><formatOnDatacenter>Not\ | |
| \ all the configured VLANs in the vSphere Distributed Switch were trunked\ | |
| \ by the physical switch connected to uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name} on host {host.name}.</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Not\ | |
| \ all the configured VLANs in the vSphere Distributed Switch were trunked\ | |
| \ by the physical switch connected to uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name}.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Not\ | |
| \ all the configured VLANs in the vSphere Distributed Switch were trunked\ | |
| \ by the physical switch connected to uplink port {healthResult.uplinkPortKey}\ | |
| \ in vSphere Distributed Switch {dvs.name} on host {host.name} in {datacenter.name}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UserAssignedToGroup</key><description>User\ | |
| \ assigned to group</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>User\ | |
| \ {userLogin} was added to group {group}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UserLoginSessionEvent</key><description>User\ | |
| \ login</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>User\ | |
| \ {userName}@{ipAddress} logged in as {userAgent}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UserLogoutSessionEvent</key><description>User\ | |
| \ logout</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>User\ | |
| \ {userName}@{ipAddress} logged out (login time: {loginTime}, number of API\ | |
| \ invocations: {callCount}, user agent: {userAgent})</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UserPasswordChanged</key><description>User\ | |
| \ password changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Password\ | |
| \ was changed for account {userLogin}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Password\ | |
| \ was changed for account {userLogin} on host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UserUnassignedFromGroup</key><description>User\ | |
| \ removed from group</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>User\ | |
| \ {userLogin} removed from group {group}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>UserUpgradeEvent</key><description>User\ | |
| \ upgrade</description><category>user</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>{message}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.UserUpgradeEvent">\t\t\ | |
| <description>\t\t\tA general user event occurred due to an upgrade\t\ | |
| \t</description>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>VMFSDatastoreCreatedEvent</key><description>VMFS\ | |
| \ datastore created</description><category>info</category><formatOnDatacenter>Created\ | |
| \ VMFS datastore {datastore.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ VMFS datastore {datastore.name} on {host.name}</formatOnComputeResource><formatOnHost>Created\ | |
| \ VMFS datastore {datastore.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Created\ | |
| \ VMFS datastore {datastore.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VMFSDatastoreExpandedEvent</key><description>VMFS\ | |
| \ datastore expanded</description><category>info</category><formatOnDatacenter>Expanded\ | |
| \ VMFS datastore {datastore.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Expanded\ | |
| \ VMFS datastore {datastore.name} on {host.name}</formatOnComputeResource><formatOnHost>Expanded\ | |
| \ VMFS datastore {datastore.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Expanded\ | |
| \ VMFS datastore {datastore.name} on {host.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VMFSDatastoreExpandedEvent">\ | |
| \ <description> An existing extent in a VMFS volume was\ | |
| \ grown to increase its capacity </description> <cause>\ | |
| \ <description> A user or system action caused an\ | |
| \ extent of an existing VMFS datastore to be grown. Only extents\ | |
| \ with free space immediately after them are expandable. As a\ | |
| \ result, the action filled the available adjacent capacity on\ | |
| \ the LUN. </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VMFSDatastoreExtendedEvent</key><description>VMFS\ | |
| \ datastore extended</description><category>info</category><formatOnDatacenter>Extended\ | |
| \ VMFS datastore {datastore.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Extended\ | |
| \ VMFS datastore {datastore.name} on {host.name}</formatOnComputeResource><formatOnHost>Extended\ | |
| \ VMFS datastore {datastore.name}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Extended\ | |
| \ VMFS datastore {datastore.name} on {host.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VMFSDatastoreExtendedEvent">\ | |
| \ <description> An existing VMFS volume was extended to\ | |
| \ increase its capacity </description> <cause> \ | |
| \ <description> A user or system action caused the datastore\ | |
| \ to be extended with a partition on a LUN to increase its capacity.\ | |
| \ </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VMotionLicenseExpiredEvent</key><description>vMotion\ | |
| \ license expired</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>A\ | |
| \ vMotion license for {host.name} has expired</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VMotionLicenseExpiredEvent">\ | |
| \ <description> vCenter Server tracks the expiration\ | |
| \ times of vMotion licenses on the license server and uses this\ | |
| \ event to notify you of any vMotion licenses that are about to expire \ | |
| \ </description> <cause> <description>vMotion\ | |
| \ licenses on the license server are about to expire</description> \ | |
| \ <action>Update the license server to get a fresher version\ | |
| \ of the vMotion license</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VcAgentUninstallFailedEvent</key><description>Cannot\ | |
| \ uninstall vCenter agent</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ uninstall vCenter agent from {host.name}. {[email protected]}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ uninstall vCenter agent from {host.name}. {[email protected]}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ uninstall vCenter agent. {[email protected]}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ uninstall vCenter agent from {host.name} in {datacenter.name}. {[email protected]}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.VcAgentUninstallFailedEvent">\t\ | |
| \t<description>\t\t\tAn attempt to uninstall the vCenter Agent failed\ | |
| \ on the host\t\t</description>\t\t<cause> <description>\t\ | |
| \t\t\tThe event contains details on why this failure occurred\t\t\t</description>\ | |
| \ </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>VcAgentUninstalledEvent</key><description>vCenter\ | |
| \ agent uninstalled</description><category>info</category><formatOnDatacenter>vCenter\ | |
| \ agent has been uninstalled from {host.name}</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ agent has been uninstalled from {host.name}</formatOnComputeResource><formatOnHost>vCenter\ | |
| \ agent has been uninstalled</formatOnHost><formatOnVm></formatOnVm><fullFormat>vCenter\ | |
| \ agent has been uninstalled from {host.name} in {datacenter.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.VcAgentUninstalledEvent">\t\ | |
| \t<description>\t\t\tThe vCenter Agent has been uninstalled from host\t\ | |
| \t</description>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>VcAgentUpgradeFailedEvent</key><description>Cannot\ | |
| \ complete vCenter agent upgrade</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ upgrade vCenter agent on {host.name}. {[email protected]}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ upgrade vCenter agent on {host.name}. {[email protected]}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ upgrade vCenter agent. {[email protected]}</formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ upgrade vCenter agent on {host.name} in {datacenter.name}. {[email protected]}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.VcAgentUpgradeFailedEvent">\t\ | |
| \t<description>\t\t\tA vCenter Agent upgrade attempt failed on the host\t\ | |
| \t</description>\t\t<cause> <description>\t\t\ | |
| \t\tThe event contains details on why this failure occurred\t\t\t</description>\ | |
| \ </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>VcAgentUpgradedEvent</key><description>vCenter\ | |
| \ agent upgraded</description><category>info</category><formatOnDatacenter>vCenter\ | |
| \ agent has been upgraded on {host.name}</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ agent has been upgraded on {host.name}</formatOnComputeResource><formatOnHost>vCenter\ | |
| \ agent has been upgraded</formatOnHost><formatOnVm></formatOnVm><fullFormat>vCenter\ | |
| \ agent has been upgraded on {host.name} in {datacenter.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.VcAgentUpgradedEvent">\t\ | |
| \t<description>\t\t\tThe vCenter Agent has been upgraded on the host\t\ | |
| \t</description>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>VimAccountPasswordChangedEvent</key><description>VIM\ | |
| \ account password changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>VIM\ | |
| \ account password changed</formatOnHost><formatOnVm></formatOnVm><fullFormat>VIM\ | |
| \ account password was changed on host {host.name}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.VimAccountPasswordChangedEvent">\t\ | |
| \t<description>\t\t\tThe password for the Vim account user on the host\ | |
| \ has been changed. \t\t\tThis account is created by vCenter Server and used\ | |
| \ to manage the host.\t\t</description>\t\t<cause> \ | |
| \ <description>\t\t\tvCenter Server periodically changes the password\ | |
| \ of the Vim account that it \t\t\tuses to manage the host\t\t\t</description>\ | |
| \ </cause>\t</EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmAcquiredMksTicketEvent</key><description>VM\ | |
| \ acquired MKS ticket</description><category>info</category><formatOnDatacenter>Remote\ | |
| \ console to {vm.name} on {host.name} has been opened</formatOnDatacenter><formatOnComputeResource>Remote\ | |
| \ console to {vm.name} on {host.name} has been opened</formatOnComputeResource><formatOnHost>Remote\ | |
| \ console to {vm.name} has been opened</formatOnHost><formatOnVm>Remote console\ | |
| \ has been opened for this virtual machine on {host.name}</formatOnVm><fullFormat>Remote\ | |
| \ console to {vm.name} on {host.name} in {datacenter.name} has been opened</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmAcquiredMksTicketEvent">\ | |
| \ <description> Successfully acquired MKS Ticket for\ | |
| \ the virtual machine </description> <cause> \ | |
| \ <description> The MKS Ticket used to connect\ | |
| \ to the virtual machine remote console has been successfully\ | |
| \ acquired. </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmAcquiredTicketEvent</key><description>VM\ | |
| \ acquired ticket</description><category>info</category><formatOnDatacenter>A\ | |
| \ ticket for {vm.name} of type {ticketType} has been acquired</formatOnDatacenter><formatOnComputeResource>A\ | |
| \ ticket for {vm.name} of type {ticketType} has been acquired</formatOnComputeResource><formatOnHost>A\ | |
| \ ticket for {vm.name} of type {ticketType} has been acquired</formatOnHost><formatOnVm>A\ | |
| \ ticket of type {ticketType} has been acquired.</formatOnVm><fullFormat>A\ | |
| \ ticket for {vm.name} of type {ticketType} on {host.name} in {datacenter.name}\ | |
| \ has been acquired</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmAutoRenameEvent</key><description>VM\ | |
| \ auto rename</description><category>info</category><formatOnDatacenter>Invalid\ | |
| \ name for {vm.name} on {host.name}. Renamed from {oldName} to {newName}</formatOnDatacenter><formatOnComputeResource>Invalid\ | |
| \ name for {vm.name} on {host.name}. Renamed from {oldName} to {newName}</formatOnComputeResource><formatOnHost>Invalid\ | |
| \ name for {vm.name}. Renamed from {oldName} to {newName}</formatOnHost><formatOnVm>Conflicting\ | |
| \ or invalid virtual machine name detected. Renamed from {oldName} to {newName}</formatOnVm><fullFormat>Invalid\ | |
| \ name for {vm.name} on {host.name} in {datacenter.name}. Renamed from {oldName}\ | |
| \ to {newName}</fullFormat><longDescription> <EventLongDescription id="vim.event.VmAutoRenameEvent">\ | |
| \ <description> The virtual machine was renamed because\ | |
| \ of possible name conflicts with another virtual machine </description>\ | |
| \ <cause> <description>The virtual machine might\ | |
| \ have been added to the vCenter Server inventory while scanning\ | |
| \ the datastores of hosts added to the inventory. During such an action, the\ | |
| \ \t newly-added virtual machine's name might have been found to\ | |
| \ be in conflict with a virtual machine name already in the\ | |
| \ inventory. To resolve this, vCenter Server renames the \t newly-added\ | |
| \ virtual machine. </description> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmBeingClonedEvent</key><description>VM\ | |
| \ being cloned</description><category>info</category><formatOnDatacenter>Cloning\ | |
| \ {vm.name} on {host.name} to {destName} on {destHost.name}</formatOnDatacenter><formatOnComputeResource>Cloning\ | |
| \ {vm.name} on {host.name} to {destName} on {destHost.name}</formatOnComputeResource><formatOnHost>Cloning\ | |
| \ {vm.name} to {destName} on {destHost.name}</formatOnHost><formatOnVm>Being\ | |
| \ cloned to {destName} on {destHost.name}</formatOnVm><fullFormat>Cloning\ | |
| \ {vm.name} on host {host.name} in {datacenter.name} to {destName} on host\ | |
| \ {destHost.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmBeingClonedNoFolderEvent</key><description>VM\ | |
| \ being cloned to a vApp</description><category>info</category><formatOnDatacenter>Cloning\ | |
| \ {vm.name} on {host.name} to {destName} on {destHost.name} to a vApp</formatOnDatacenter><formatOnComputeResource>Cloning\ | |
| \ {vm.name} on {host.name} to {destName} on {destHost.name} to a vApp</formatOnComputeResource><formatOnHost>Cloning\ | |
| \ {vm.name} to {destName} on {destHost.name} to a vApp</formatOnHost><formatOnVm>Being\ | |
| \ cloned to {destName} on {destHost.name} to a vApp</formatOnVm><fullFormat>Cloning\ | |
| \ {vm.name} on host {host.name} in {datacenter.name} to {destName} on host\ | |
| \ {destHost.name} to a vApp</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmBeingCreatedEvent</key><description>Creating\ | |
| \ VM</description><category>info</category><formatOnDatacenter>Creating {vm.name}\ | |
| \ on host {host.name}</formatOnDatacenter><formatOnComputeResource>Creating\ | |
| \ {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>Creating\ | |
| \ {vm.name}</formatOnHost><formatOnVm>Creating VM on host {host.name}</formatOnVm><fullFormat>Creating\ | |
| \ {vm.name} on host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmBeingDeployedEvent</key><description>Deploying\ | |
| \ VM</description><category>info</category><formatOnDatacenter>Deploying {vm.name}\ | |
| \ on host {host.name} from template {srcTemplate.name}</formatOnDatacenter><formatOnComputeResource>Deploying\ | |
| \ {vm.name} on host {host.name} from template {srcTemplate.name}</formatOnComputeResource><formatOnHost>Deploying\ | |
| \ {vm.name} from template {srcTemplate.name}</formatOnHost><formatOnVm>Deploying\ | |
| \ VM on host {host.name} from template {srcTemplate.name}</formatOnVm><fullFormat>Deploying\ | |
| \ {vm.name} on host {host.name} in {datacenter.name} from template {srcTemplate.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmBeingDeployedEvent">\ | |
| \ <description> A virtual machine is being created\ | |
| \ from a template </description> <cause> \ | |
| \ <description> A user action prompted a virtual\ | |
| \ machine to be created from this template. </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmBeingHotMigratedEvent</key><description>VM\ | |
| \ is hot migrating</description><category>info</category><formatOnDatacenter>Migrating\ | |
| \ {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnDatacenter><formatOnComputeResource>Migrating\ | |
| \ {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnComputeResource><formatOnHost>Migrating\ | |
| \ {vm.name} from {ds.name} to {destHost.name}, {destDatastore.name}</formatOnHost><formatOnVm>Migrating\ | |
| \ VM from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnVm><fullFormat>Migrating\ | |
| \ {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmBeingHotMigratedEvent"> <description>\ | |
| \ A powered-on virtual machine is being migrated with vMotion </description>\ | |
| \ <cause> <description> A user action might have\ | |
| \ caused a powered-on virtual machine to be migrated with vMotion \ | |
| \ </description> </cause> <cause> <description>\ | |
| \ A DRS recommendation might have caused a powered-on virtual machine to be\ | |
| \ migrated \t with vMotion </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmBeingMigratedEvent</key><description>VM\ | |
| \ migrating</description><category>info</category><formatOnDatacenter>Relocating\ | |
| \ {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnDatacenter><formatOnComputeResource>Relocating\ | |
| \ {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnComputeResource><formatOnHost>Relocating\ | |
| \ {vm.name} to {destHost.name}</formatOnHost><formatOnVm>Relocating VM from\ | |
| \ {host.name} to {destHost.name}</formatOnVm><fullFormat>Relocating {vm.name}\ | |
| \ from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name} in\ | |
| \ {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmBeingMigratedEvent"> <description>\ | |
| \ Changing the host on which the virtual machine is executing \ | |
| \ </description> <cause> <description>\ | |
| \ A user action caused the virtual machine to be migrated to\ | |
| \ a different host </description> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmBeingRelocatedEvent</key><description>VM\ | |
| \ relocating</description><category>info</category><formatOnDatacenter>Relocating\ | |
| \ {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnDatacenter><formatOnComputeResource>Relocating\ | |
| \ {vm.name} from {host.name}, {ds.name} to {destHost.name}, {desDatastore.name}</formatOnComputeResource><formatOnHost>Relocating\ | |
| \ {vm.name} from {ds.name} to {destHost.name}, {destDatastore.name}</formatOnHost><formatOnVm>Relocating\ | |
| \ from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnVm><fullFormat>Relocating\ | |
| \ {vm.name} in {datacenter.name} from {host.name}, {ds.name} to {destHost.name},\ | |
| \ {destDatastore.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmBeingRelocatedEvent"> <description>\ | |
| \ The virtual machine execution and/or storage is being relocated\ | |
| \ </description> <cause> <description>\ | |
| \ A user action might have caused the virtual machine's\ | |
| \ execution and/or storage to be changed </description> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmCloneEvent</key><description><VM\ | |
| \ Clone Event></description><category>info</category><formatOnDatacenter><internal></formatOnDatacenter><formatOnComputeResource><internal></formatOnComputeResource><formatOnHost><internal></formatOnHost><formatOnVm><internal></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmCloneFailedEvent</key><description>Cannot\ | |
| \ complete VM clone</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ clone {vm.name}: {reason.msg}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmCloneFailedEvent"> <description>\ | |
| \ Cloning a virtual machine failed </description> \ | |
| \ <cause> <description> An error\ | |
| \ prevented the virtual machine from being cloned </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmClonedEvent</key><description>VM\ | |
| \ cloned</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Clone\ | |
| \ of {sourceVm.name} completed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmConfigMissingEvent</key><description>VM\ | |
| \ configuration missing</description><category>info</category><formatOnDatacenter>Configuration\ | |
| \ file for {vm.name} on {host.name} cannot be found</formatOnDatacenter><formatOnComputeResource>Configuration\ | |
| \ file for {vm.name} on {host.name} cannot be found</formatOnComputeResource><formatOnHost>Configuration\ | |
| \ file for {vm.name} cannot be found</formatOnHost><formatOnVm>Configuration\ | |
| \ file cannot be found</formatOnVm><fullFormat>Configuration file for {vm.name}\ | |
| \ on {host.name} in {datacenter.name} cannot be found</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmConfigMissingEvent">\ | |
| \ <description> One or more configuration files for the\ | |
| \ virtual machine cannot be found </description> <cause>\ | |
| \ <description> The datastore on which this virtual\ | |
| \ machine resides may be inaccessible </description> \ | |
| \ <action> Check the connectivity of the datastore on which\ | |
| \ this virtual machine resides. If the datastore has a backing\ | |
| \ LUN, check to see if there are any transient disk failures.\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmConnectedEvent</key><description>VM\ | |
| \ connected</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Host\ | |
| \ is connected</formatOnVm><fullFormat>Virtual machine {vm.name} is connected</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmConnectedEvent">\ | |
| \ <description> The virtual machine is in a connected state\ | |
| \ in the inventory and vCenter Server can access it </description>\ | |
| \ <cause> <description> A user or system\ | |
| \ action that resulted in operations such as creating, registering,\ | |
| \ cloning or deploying a virtual machine gave vCenter Server access to \ | |
| \ the virtual machine </description> </cause>\ | |
| \ <cause> <description> A user or system\ | |
| \ action that resulted in operations such as adding or reconnecting\ | |
| \ a host gave vCenter Server access to the virtual machine </description>\ | |
| \ </cause> <cause> <description> \ | |
| \ The state of the virtual machine's host changed from \ | |
| \ Not Responding to Connected and the host gave vCenter Server access\ | |
| \ to the virtual machine </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmCreatedEvent</key><description>VM\ | |
| \ created</description><category>info</category><formatOnDatacenter>New virtual\ | |
| \ machine {vm.name} created on {host.name}</formatOnDatacenter><formatOnComputeResource>New\ | |
| \ virtual machine {vm.name} created on {host.name}</formatOnComputeResource><formatOnHost>New\ | |
| \ virtual machine {vm.name} created</formatOnHost><formatOnVm>Virtual machine\ | |
| \ created</formatOnVm><fullFormat>Created virtual machine {vm.name} on {host.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmDVPortEvent</key><description>dvPort\ | |
| \ connected to VM changed status</description><category>info</category><formatOnDatacenter>dvPort\ | |
| \ connected to VM {vm.name} on {host.name} changed status</formatOnDatacenter><formatOnComputeResource>dvPort\ | |
| \ connected to VM {vm.name} changed status</formatOnComputeResource><formatOnHost>dvPort\ | |
| \ connected to VM {vm.name} changed status</formatOnHost><formatOnVm>dvPort\ | |
| \ changed status</formatOnVm><fullFormat>dvPort connected to VM {vm.name}\ | |
| \ on {host.name} in {datacenter.name} changed status</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmDasBeingResetEvent</key><description>vSphere\ | |
| \ HA is resetting VM</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} in cluster {computeResource.name} reset by vSphere HA. Reason:\ | |
| \ {[email protected]}</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} reset by vSphere HA. Reason: {[email protected]}.</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ reset by vSphere HA. Reason: {[email protected]}.</formatOnHost><formatOnVm>This\ | |
| \ virtual machine reset by vSphere HA. Reason: {[email protected]}.\ | |
| \ </formatOnVm><fullFormat>{vm.name} on {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} reset by vSphere HA. Reason: {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmDasBeingResetEvent">\ | |
| \ <description> The virtual machine was reset by vSphere\ | |
| \ HA. Depending on how vSphere HA has been configured, the virtual\ | |
| \ machine might be reset because the VMware Tools heartbeat or application\ | |
| \ heartbeat status turned red. </description> \ | |
| \ <cause> <description> The VMware\ | |
| \ Tools heartbeat turned red. This condition can occur if the operating \ | |
| \ system failed with a blue screen or becomes unresponsive.\ | |
| \ It also can occur because \t\tVMware Tools failed or was shut down. \ | |
| \ </description> <action> If the\ | |
| \ virtual machine is reset frequently, check for a persistent\ | |
| \ problem with the operating system that requires attention. \ | |
| \ Consider configuring the cluster so that vSphere HA waits for a longer\ | |
| \ period after heartbeats are lost before taking action. Specifying\ | |
| \ a longer period helps avoid triggering resets for transient\ | |
| \ problems. You can force a longer period by decreasing \t\tthe "monitoring\ | |
| \ sensitivity" in the VM Monitoring section of the Edit Cluster wizard.\ | |
| \ </action> </cause> <cause> \ | |
| \ <description> The application heartbeat turned\ | |
| \ red. This condition can occur if the application that is\ | |
| \ configured to send heartbeats failed or became unresponsive. </description>\ | |
| \ <action> Determine if the application stopped\ | |
| \ sending heartbeats because of a configuration error and\ | |
| \ remediate the problem. </action> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmDasBeingResetWithScreenshotEvent</key><description>vSphere\ | |
| \ HA enabled VM reset with screenshot</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} in cluster {computeResource.name} reset by vSphere HA. Reason:\ | |
| \ {[email protected]}. A screenshot is saved at\ | |
| \ {screenshotFilePath}.</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} reset by vSphere HA. Reason: {[email protected]}.\ | |
| \ A screenshot is saved at {screenshotFilePath}.</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ reset by vSphere HA. Reason: {[email protected]}.\ | |
| \ A screenshot is saved at {screenshotFilePath}</formatOnHost><formatOnVm>This\ | |
| \ virtual machine reset by vSphere HA. Reason: {[email protected]}.\ | |
| \ A screenshot is saved at {screenshotFilePath}</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in cluster {computeResource.name} in {datacenter.name} reset\ | |
| \ by vSphere HA. Reason: {[email protected]}. A\ | |
| \ screenshot is saved at {screenshotFilePath}.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmDasBeingResetWithScreenshotEvent">\ | |
| \ <description> The virtual machine was reset by vSphere\ | |
| \ HA. Depending on how vSphere HA is configured, this condition\ | |
| \ can occur because the VMware Tools heartbeat or the application \ | |
| \ heartbeat status turned red. The event contains the location of the screenshot\ | |
| \ taken of the guest console before it was reset. You can use this\ | |
| \ information to determine the cause of the heartbeat failure. \ | |
| \ </description> <cause> <description>\ | |
| \ The VMware Tools heartbeat turned red. This condition can\ | |
| \ occur if the operating system failed with a blue screen\ | |
| \ or becomes unresponsive. It also can occur because \t\tVMware Tools failed\ | |
| \ or was shut down. </description> <action>\ | |
| \ Check the screenshot image to see if the cause was a guest\ | |
| \ operating system failure. If the virtual machine is reset\ | |
| \ frequently, check for a persistent problem with the operating\ | |
| \ system that requires attention. Consider configuring the\ | |
| \ cluster so that vSphere HA waits for a longer period after \ | |
| \ heartbeats are lost before taking action. Specifying a longer period\ | |
| \ helps avoid triggering resets for transient problems. You\ | |
| \ can force a longer period by decreasing \t\tthe "monitoring sensitivity"\ | |
| \ in the VM Monitoring section of the Edit Cluster wizard. </action>\ | |
| \ </cause> <cause> <description>\ | |
| \ The application heartbeat turned red. This condition can\ | |
| \ occur if the application that is configured to send heartbeats\ | |
| \ failed or became unresponsive. </description> \ | |
| \ <action> Determine if the application stopped sending\ | |
| \ heartbeats because of a configuration error and remediate\ | |
| \ the problem. </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmDasResetFailedEvent</key><description>vSphere\ | |
| \ HA cannot reset VM</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA cannot reset {vm.name} on {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA cannot reset {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA cannot reset {vm.name}</formatOnHost><formatOnVm>vSphere HA cannot reset\ | |
| \ this virtual machine</formatOnVm><fullFormat>vSphere HA cannot reset {vm.name}\ | |
| \ on {host.name} in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmDasResetFailedEvent">\ | |
| \ <description> vSphere HA attempted to reset the\ | |
| \ virtual machine because of a heartbeat failure from VMware Tools\ | |
| \ or a guest application, depending on how vSphere HA was configured.\ | |
| \ However, the reset operation failed. </description> \ | |
| \ <cause> <description> The most likely\ | |
| \ reason for the reset failure is that the virtual machine was running \ | |
| \ another task at the time the reset was initiated. \ | |
| \ </description> <action>Check to see whether the\ | |
| \ virtual machine requires attention and reset it manually\ | |
| \ if necessary.</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmDasUpdateErrorEvent</key><description>VM\ | |
| \ vSphere HA update error</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Unable\ | |
| \ to update vSphere HA agents given the state of {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmDasUpdateOkEvent</key><description>Completed\ | |
| \ VM DAS update</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA agents have been updated with the current state of the virtual machine</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmDateRolledBackEvent</key><description>VM\ | |
| \ date rolled back</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Disconnecting\ | |
| \ all hosts as the date of virtual machine {vm.name} has been rolled back</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmDeployFailedEvent</key><description>Cannot\ | |
| \ deploy VM</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ deploy template: {reason.msg}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmDeployFailedEvent"> <description>\ | |
| \ Failed to deploy a virtual machine for reasons described in the\ | |
| \ event message </description> <cause> \ | |
| \ <description> The virtual machine failed to deploy.\ | |
| \ This condition can occur if there is not enough disk space,\ | |
| \ the host or virtual machine loses \t\tits network connection, the host is\ | |
| \ disconnected, and so on. </description> <action>\t\ | |
| \ Check the reason in the event message to find the cause of the failure\ | |
| \ and correct the problem. </action>\t</cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmDeployedEvent</key><description>VM\ | |
| \ deployed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Template\ | |
| \ {srcTemplate.name} deployed</formatOnHost><formatOnVm></formatOnVm><fullFormat>Template\ | |
| \ {srcTemplate.name} deployed on host {host.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmDeployedEvent">\ | |
| \ <description> A virtual machine has been created from\ | |
| \ the specified template </description> <cause> \ | |
| \ <description> A user action caused a virtual machine\ | |
| \ to be created from the template </description> </cause>\ | |
| \ <cause> <description> A scheduled task\ | |
| \ caused a virtual machine to be created from the template </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmDisconnectedEvent</key><description>VM\ | |
| \ disconnected</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on host {host.name} is disconnected</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on host {host.name} is disconnected</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is disconnected</formatOnHost><formatOnVm>{host.name} is disconnected</formatOnVm><fullFormat>{vm.name}\ | |
| \ on host {host.name} in {datacenter.name} is disconnected</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmDiscoveredEvent</key><description>VM\ | |
| \ discovered</description><category>info</category><formatOnDatacenter>Discovered\ | |
| \ {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Discovered\ | |
| \ {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Discovered\ | |
| \ {vm.name}</formatOnHost><formatOnVm>Discovered on {host.name}</formatOnVm><fullFormat>Discovered\ | |
| \ {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmDiskFailedEvent</key><description>Cannot\ | |
| \ create VM disk</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ create virtual disk {disk}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmDiskFailedEvent"> <description>\ | |
| \ Failed to create a virtual disk for the virtual machine for reasons\ | |
| \ described in the event message </description> <cause>\ | |
| \ <description> A virtual disk was not created\ | |
| \ for the virtual machine. This condition can occur if the \t\toperation failed\ | |
| \ to access the disk, the disk did not have enough space, you do not have\ | |
| \ \t\tpermission for the operation, and so on. </description>\ | |
| \ <action> Check the reason in the event message\ | |
| \ to find the cause of the failure. Ensure that disk is \t\taccessible, has\ | |
| \ enough space, and that the permission settings allow the operation. \ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmEmigratingEvent</key><description>VM\ | |
| \ emigrating</description><category>info</category><formatOnDatacenter>Migrating\ | |
| \ {vm.name} off host {host.name}</formatOnDatacenter><formatOnComputeResource>Migrating\ | |
| \ {vm.name} off host {host.name}</formatOnComputeResource><formatOnHost>Migrating\ | |
| \ {vm.name} off host</formatOnHost><formatOnVm>Migrating off host {host.name}</formatOnVm><fullFormat>Migrating\ | |
| \ {vm.name} off host {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmEndRecordingEvent</key><description>End\ | |
| \ a recording session</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>End\ | |
| \ a recording session</formatOnVm><fullFormat>End a recording session on {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmEndReplayingEvent</key><description>End\ | |
| \ a replay session</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>End\ | |
| \ a replay session</formatOnVm><fullFormat>End a replay session on {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmEvent</key><description><VM\ | |
| \ Event></description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmFailedMigrateEvent</key><description>Cannot\ | |
| \ migrate VM</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ migrate {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ migrate {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ migrate {vm.name} to {destHost.name}, {destDatastore.name}</formatOnHost><formatOnVm>Cannot\ | |
| \ migrate to {destHost.name}, {destDatastore.name}</formatOnVm><fullFormat>Cannot\ | |
| \ migrate {vm.name} from {host.name}, {ds.name} to {destHost.name}, {destDatastore.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmFailedMigrateEvent"> <description>\ | |
| \ Failed to migrate the virtual machine for reasons described in\ | |
| \ the event message </description> <cause> \ | |
| \ <description> The virtual machine did not migrate.\ | |
| \ This condition can occur if vMotion IPs are not configured, \ | |
| \ the source and destination hosts are not accessible, and so on. \ | |
| \ </description> <action> Check\ | |
| \ the reason in the event message to find the cause of the failure. Ensure\ | |
| \ that the \t\tvMotion IPs are configured on source and destination hosts,\ | |
| \ the hosts are accessible, and so on. </action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmFailedRelayoutEvent</key><description>Cannot\ | |
| \ complete VM relayout.</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ complete relayout {vm.name} on {host.name}: {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ complete relayout {vm.name} on {host.name}: {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ complete relayout {vm.name}: {reason.msg}</formatOnHost><formatOnVm>Cannot\ | |
| \ complete relayout for this virtual machine on {host.name}: {reason.msg}</formatOnVm><fullFormat>Cannot\ | |
| \ complete relayout {vm.name} on {host.name} in {datacenter.name}: {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedRelayoutEvent">\ | |
| \ <description> Failed to lay out a virtual machine\ | |
| \ </description> <cause> <description>\ | |
| \ An attempt to lay out a virtual machine on disk failed for\ | |
| \ reasons described in the event message. This condition can\ | |
| \ occur for any of several reasons, for example inability to access the disk.\ | |
| \ </description> <action>\t Check the reason\ | |
| \ in the event message to find the cause of the failure and correct the problem.\ | |
| \ </action>\t</cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmFailedRelayoutOnVmfs2DatastoreEvent</key><description>Cannot\ | |
| \ complete VM relayout on Vmfs2 datastore</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Cannot\ | |
| \ complete relayout due to disks on a VMFS2 volume</formatOnVm><fullFormat>Cannot\ | |
| \ complete relayout for virtual machine {vm.name} which has disks on a VMFS2\ | |
| \ volume.</fullFormat><longDescription> <EventLongDescription id="vim.event.VmFailedRelayoutOnVmfs2DatastoreEvent">\ | |
| \ <description> Failed to migrate a virtual machine\ | |
| \ on VMFS2 datastore </description> <cause> \ | |
| \ <description> An attempt to migrate a virtual\ | |
| \ machine failed because the virtual machine still has disk(s)\ | |
| \ on a VMFS2 datastore. VMFS2 datastores are read-only for \ | |
| \ ESX 3.0 and later hosts. </description> <action>\ | |
| \ Upgrade the datastore(s) from VMFS2 to VMFS3 </action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmFailedStartingSecondaryEvent</key><description>vCenter\ | |
| \ cannot start the Fault Tolerance secondary VM</description><category>error</category><formatOnDatacenter>vCenter\ | |
| \ cannot start the Fault Tolerance secondary VM for {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}. Reason: {[email protected]}</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ cannot start the Fault Tolerance secondary VM for {vm.name} on host {host.name}.\ | |
| \ Reason: {[email protected]}</formatOnComputeResource><formatOnHost>vCenter\ | |
| \ cannot start the Fault Tolerance secondary VM for {vm.name}. Reason: {[email protected]}</formatOnHost><formatOnVm>vCenter\ | |
| \ cannot start the Fault Tolerance secondary VM. Reason: {[email protected]}</formatOnVm><fullFormat>vCenter\ | |
| \ cannot start the Fault Tolerance secondary VM for {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}. Reason: {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedStartingSecondaryEvent">\ | |
| \ <description> vCenter Server could not start the Secondary\ | |
| \ VM because of an error </description> <cause> \ | |
| \ <description> The remote host is incompatible\ | |
| \ for Secondary VM. For instance, this condition can occur when\ | |
| \ the host does not have access to the virtual machine's\ | |
| \ network or datastore. </description> <action>Ensure\ | |
| \ that the hosts in the cluster are compatible for FT</action> \ | |
| \ </cause> <cause> <description>Login\ | |
| \ to a remote host failed. If the host has been newly added to the inventory\ | |
| \ \t\tor just rebooted, it might take some time for SSL thumbprints to be\ | |
| \ propagated to the hosts. </description> <action>If\ | |
| \ the problem persists, disconnect and re-connect the host.</action>\ | |
| \ </cause> <cause> <description>Registration\ | |
| \ of the Secondary VM on the remote host failed</description> \ | |
| \ <action>Determine whether the remote host has access to the datastore\ | |
| \ that the FT \t\tvirtual machine resides on</action> </cause>\ | |
| \ <cause> <description>An error occurred while\ | |
| \ starting the Secondary VM</description> <action>Determine\ | |
| \ the cause of the migration error. vCenter Server will try to restart the\ | |
| \ \t\tSecondary VM if it can.</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmFailedToPowerOffEvent</key><description>Cannot\ | |
| \ power off the VM.</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ power off {vm.name} on {host.name}. {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ power off {vm.name} on {host.name}. {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ power off {vm.name}. {reason.msg}</formatOnHost><formatOnVm>Cannot power\ | |
| \ off: {reason.msg}</formatOnVm><fullFormat>Cannot power off {vm.name} on\ | |
| \ {host.name} in {datacenter.name}: {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedToPowerOffEvent">\ | |
| \ <description>The virtual machine failed to power off</description>\ | |
| \ <cause> <description> The virtual machine\ | |
| \ might be performing concurrent operations </description> \ | |
| \ <action>Complete the concurrent operations and retry the power-off\ | |
| \ operation</action> </cause> <cause> \ | |
| \ <description>The virtual machine is in an invalid state. Virtual machines\ | |
| \ can enter an \t invalid state for many reasons, for example datastore\ | |
| \ inaccessibility. </description> <action> \ | |
| \ Identify the reason that the virtual machine entered an invalid state, correct\ | |
| \ the problem, \t and retry the operation. </action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmFailedToPowerOnEvent</key><description>Cannot\ | |
| \ power on the VM.</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ power on {vm.name} on {host.name}. {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ power on {vm.name} on {host.name}. {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ power on {vm.name}. {reason.msg}</formatOnHost><formatOnVm>Cannot power\ | |
| \ on {vm.name} on {host.name}. {reason.msg}</formatOnVm><fullFormat>Cannot\ | |
| \ power on {vm.name} on {host.name} in {datacenter.name}. {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedToPowerOnEvent">\ | |
| \ <description> The virtual machine failed to power on\ | |
| \ </description> <cause> <description>\ | |
| \ Virtual machine power-on attempts can fail because the virtual\ | |
| \ machine is already in a powered-on state, concurrent operations\ | |
| \ are running on the virtual machine, and so on. </description> \ | |
| \ <action>\t Check the reason in the event message to find the\ | |
| \ cause of the power-on failure and fix the problem. </action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmFailedToRebootGuestEvent</key><description>VM\ | |
| \ cannot reboot the guest OS.</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ reboot Guest OS. {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ reboot Guest OS. {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ reboot Guest OS. {reason.msg}</formatOnHost><formatOnVm>Cannot reboot Guest\ | |
| \ OS. {reason.msg}</formatOnVm><fullFormat>Cannot reboot the guest OS for\ | |
| \ {vm.name} on {host.name} in {datacenter.name}. {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedToRebootGuestEvent">\ | |
| \ <description> The guest operating system on the virtual\ | |
| \ machine failed to reboot. </description> <cause> \ | |
| \ <description> Guest operating system reboot failures\ | |
| \ can occur because the virtual machine is not in a \t powered-on state,\ | |
| \ concurrent operations are running on the virtual machine, and so on. \ | |
| \ </description>\t <action>\t Check the reason in the event\ | |
| \ message to find the cause of the reboot failure and fix the problem. \ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmFailedToResetEvent</key><description>Cannot\ | |
| \ reset VM</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ suspend {vm.name} on {host.name}: {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ suspend {vm.name} on {host.name}: {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ suspend {vm.name}: {reason.msg}</formatOnHost><formatOnVm>Cannot suspend\ | |
| \ {host.name}: {reason.msg}</formatOnVm><fullFormat>Cannot suspend {vm.name}\ | |
| \ on {host.name} in {datacenter.name}: {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedToResetEvent">\ | |
| \ <description>The virtual machine failed to reset</description>\ | |
| \ <cause> <description>\t\tThe virtual machine might\ | |
| \ be waiting for a response to a question or prompt</description> \ | |
| \ <action> Go to the Summary tab for the virtual machine\ | |
| \ in vSphere client and respond to the question \t or prompt </action>\ | |
| \ </cause> <cause> <description>There\ | |
| \ might not be enough available licenses to perform this operation.</description>\ | |
| \ <action> Obtain the required licenses and retry\ | |
| \ the reset operation </action> </cause> <cause>\ | |
| \ <description> Concurrent operations might be executing\ | |
| \ on the virtual machine </description> <action>Complete\ | |
| \ the concurrent operations and retry the reset operation</action> \ | |
| \ </cause> <cause> <description> \ | |
| \ The host on which the virtual machine is running is entering maintenance\ | |
| \ mode </description> <action> Wait\ | |
| \ until the host exits maintenance mode and retry the operation </action>\ | |
| \ </cause> <cause> <description>The virtual\ | |
| \ machine is in an invalid state. Virtual machines can enter an \t invalid\ | |
| \ state for many reasons, for example datastore inaccessibility.</description>\ | |
| \ <action> Identify the reason that the virtual machine\ | |
| \ entered an invalid state, correct the problem, \t and retry the operation.\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmFailedToShutdownGuestEvent</key><description>Cannot\ | |
| \ shut down the guest OS</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ shut down the guest OS. {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ shut down the guest OS. {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ shut down the guest OS. {reason.msg}</formatOnHost><formatOnVm>Cannot shut\ | |
| \ down the guest OS. {reason.msg}</formatOnVm><fullFormat>{vm.name} cannot\ | |
| \ shut down the guest OS on {host.name} in {datacenter.name}: {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedToShutdownGuestEvent">\ | |
| \ <description> Guest operating system shutdown failed\ | |
| \ for the virtual machine </description> <cause> \ | |
| \ <description> Guest operating system shutdown can\ | |
| \ fail if VMware Tools is not installed in the virtual machine.\ | |
| \ </description> <action>Install VMware Tools.</action>\ | |
| \ </cause> <cause> <description>\t \ | |
| \ The virtual machine might be waiting for a response to a question or prompt</description>\ | |
| \ <action> Go to the Summary tab for the virtual\ | |
| \ machine in vSphere Client and respond to the question \t or prompt \ | |
| \ </action> </cause> <cause> <description>\ | |
| \ Concurrent operations might be running on the virtual machine\ | |
| \ </description> <action>Complete the concurrent\ | |
| \ operations and retry the shutdown operation</action> </cause>\ | |
| \ <cause> <description>The virtual machine is in\ | |
| \ an invalid state. Virtual machines can enter an \t invalid state for\ | |
| \ many reasons, for example datastore inaccessibility.</description>\ | |
| \ <action> Identify the reason that the virtual machine\ | |
| \ entered an invalid state, correct the problem, \t and retry the operation.\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmFailedToStandbyGuestEvent</key><description>VM\ | |
| \ cannot standby the guest OS</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ standby the guest OS. {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ standby the guest OS. {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ standby the guest OS. {reason.msg}</formatOnHost><formatOnVm>Cannot standby\ | |
| \ the guest OS. {reason.msg}</formatOnVm><fullFormat>{vm.name} cannot standby\ | |
| \ the guest OS on {host.name} in {datacenter.name}: {reason.msg}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmFailedToSuspendEvent</key><description>Cannot\ | |
| \ suspend VM</description><category>error</category><formatOnDatacenter>Cannot\ | |
| \ suspend {vm.name} on {host.name}: {reason.msg}</formatOnDatacenter><formatOnComputeResource>Cannot\ | |
| \ suspend {vm.name} on {host.name}: {reason.msg}</formatOnComputeResource><formatOnHost>Cannot\ | |
| \ suspend {vm.name}: {reason.msg}</formatOnHost><formatOnVm>Cannot suspend\ | |
| \ {host.name}: {reason.msg}</formatOnVm><fullFormat>Cannot suspend {vm.name}\ | |
| \ on {host.name} in {datacenter.name}: {reason.msg}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmFailedUpdatingSecondaryConfig</key><description>vCenter\ | |
| \ cannot update the Fault Tolerance secondary VM configuration</description><category>error</category><formatOnDatacenter>vCenter\ | |
| \ cannot update the Fault Tolerance secondary VM configuration for {vm.name}\ | |
| \ on host {host.name} in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ cannot update the Fault Tolerance secondary VM configuration for {vm.name}\ | |
| \ on host {host.name}</formatOnComputeResource><formatOnHost>vCenter cannot\ | |
| \ update the Fault Tolerance secondary VM configuration for {vm.name}</formatOnHost><formatOnVm>vCenter\ | |
| \ cannot update the Fault Tolerance secondary VM configuration</formatOnVm><fullFormat>vCenter\ | |
| \ cannot update the Fault Tolerance secondary VM configuration for {vm.name}\ | |
| \ on host {host.name} in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailedUpdatingSecondaryConfig">\ | |
| \ <description> After a failover, the new Primary VM\ | |
| \ failed to update the configuration of the Secondary VM </description>\ | |
| \ <cause> <description> </description>\ | |
| \ <action></action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmFailoverFailed</key><description>vSphere\ | |
| \ HA virtual machine failover unsuccessful</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA unsuccessfully failed over {vm.name} on {host.name} in cluster {computeResource.name}.\ | |
| \ vSphere HA will retry if the maximum number of attempts has not been exceeded.\ | |
| \ Reason: {reason.msg}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA unsuccessfully failed over {vm.name} on {host.name}. vSphere HA will\ | |
| \ retry if the maximum number of attempts has not been exceeded. Reason: {reason.msg}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA unsuccessfully failed over {vm.name}. vSphere HA will retry if the maximum\ | |
| \ number of attempts has not been exceeded. Reason: {reason.msg}</formatOnHost><formatOnVm>vSphere\ | |
| \ HA unsuccessfully failed over this virtual machine. vSphere HA will retry\ | |
| \ if the maximum number of attempts has not been exceeded. Reason: {reason.msg}</formatOnVm><fullFormat>vSphere\ | |
| \ HA unsuccessfully failed over {vm.name} on {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}. vSphere HA will retry if the maximum number of attempts\ | |
| \ has not been exceeded. Reason: {reason.msg}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFailoverFailed">\ | |
| \ <description> vSphere HA did not failover this virtual\ | |
| \ machine. The event includes the details of the fault that was\ | |
| \ generated when vSphere HA attempted the failover. vSphere HA will retry\ | |
| \ the failover on another host unless the maximum number of failover\ | |
| \ attempts have been exceeded. In many cases, the retry will succeeed.\ | |
| \ </description> <cause> <description>\ | |
| \ The failover did not succeed because a problem occurred while\ | |
| \ vSphere HA was trying to restart the virtual machine. Possible\ | |
| \ problems include the inability to register or reconfigure the virtual \ | |
| \ machine on the new host because another operation on the same\ | |
| \ virtual machine is already in progress, or because the virtual\ | |
| \ machine is still powered on. It may also occur if the configuration\ | |
| \ file of the virtual machine is corrupt. </description> \ | |
| \ <action> If vSphere HA is unable to failover\ | |
| \ the virtual machine after repeated attempts, investigate the \ | |
| \ error reported by each occurences of this event, or trying powering\ | |
| \ on the virtual machine and investigate any returned errors.\ | |
| \ </action> <action> If\ | |
| \ the error reports that a file is locked, the VM may be\ | |
| \ powered on a host that the vSphere HA master agent can no longer \ | |
| \ monitor using the management network or heartbeat datastores,\ | |
| \ or it may have been powered on by a user on a host outside\ | |
| \ of the cluster. If any hosts have been declared dead, investigate whether\ | |
| \ a networking/storage issue may be the cause. \ | |
| \ </action> <action> If, however,\ | |
| \ the error reports that the virtual machine is in an invalid\ | |
| \ state, there may be an inprogress operation that is preventing access to\ | |
| \ the virtual machine's files. Investigate whether\ | |
| \ there are in-progress operations, such as a clone operation \ | |
| \ that is taking a long time to complete. </action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmFaultToleranceStateChangedEvent</key><description>VM\ | |
| \ Fault Tolerance state changed</description><category>info</category><formatOnDatacenter>Fault\ | |
| \ Tolerance state of {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ changed from {[email protected]} to {[email protected]}</formatOnDatacenter><formatOnComputeResource>Fault\ | |
| \ Tolerance state on {vm.name} on host {host.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnComputeResource><formatOnHost>Fault\ | |
| \ Tolerance state of {vm.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnHost><formatOnVm>Fault\ | |
| \ Tolerance state changed from {[email protected]}\ | |
| \ to {[email protected]}</formatOnVm><fullFormat>Fault\ | |
| \ Tolerance state of {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} changed from {[email protected]}\ | |
| \ to {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFaultToleranceStateChangedEvent">\ | |
| \ <description> The Fault Tolerance state of the virtual\ | |
| \ machine changed </description> <cause> \ | |
| \ <description> </description> <action></action>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmFaultToleranceTurnedOffEvent</key><description>VM\ | |
| \ Fault Tolerance turned off</description><category>info</category><formatOnDatacenter>Fault\ | |
| \ Tolerance protection has been turned off for {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Fault\ | |
| \ Tolerance protection has been turned off for {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>Fault\ | |
| \ Tolerance protection has been turned off for {vm.name}</formatOnHost><formatOnVm>Fault\ | |
| \ Tolerance protection has been turned off for this virtual machine</formatOnVm><fullFormat>Fault\ | |
| \ Tolerance protection has been turned off for {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFaultToleranceTurnedOffEvent">\ | |
| \ <description> All Secondary VMs have been removed and\ | |
| \ Fault Tolerance protection \t is turned off for this virtual machine. \ | |
| \ </description> <cause> <description>\ | |
| \ </description> <action></action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmFaultToleranceVmTerminatedEvent</key><description>Fault\ | |
| \ Tolerance VM terminated</description><category>info</category><formatOnDatacenter>The\ | |
| \ Fault Tolerance VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ has been terminated. {[email protected]}</formatOnDatacenter><formatOnComputeResource>The\ | |
| \ Fault Tolerance VM {vm.name} on host {host.name} has been terminated. {[email protected]}</formatOnComputeResource><formatOnHost>The\ | |
| \ Fault Tolerance VM {vm.name} has been terminated. {[email protected]}</formatOnHost><formatOnVm>The\ | |
| \ Fault Tolerance VM has been terminated. {[email protected]}</formatOnVm><fullFormat>The\ | |
| \ Fault Tolerance VM {vm.name} on host {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} has been terminated. {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmFaultToleranceVmTerminatedEvent">\ | |
| \ <description> A Primary VM or Secondary VM became inactive\ | |
| \ </description> <cause> <description>\ | |
| \ The Secondary VM became inactive because its operations are no longer \t\ | |
| \tsynchronized with those of the Primary VM</description> \ | |
| \ <action>vSphere HA will attempt to restart the Secondary VM</action>\ | |
| \ </cause> <cause> <description>\ | |
| \ The Secondary VM became inactive because a hardware or network failure \t\ | |
| \tcaused the Primary VM to lose the Primary-to-Secondary connection</description>\ | |
| \ <action>vSphere HA will attempt to restart the Secondary\ | |
| \ VM</action> </cause> <cause> <description>\ | |
| \ The Fault Tolerant VM became inactive due to a partial hardware failure\ | |
| \ on \t\tthe physical host</description> <action>vSphere\ | |
| \ HA will attempt to restart the Secondary VM</action> </cause>\ | |
| \ <cause> <description> A user stopped the Fault\ | |
| \ Tolerant VM</description> <action>The remaining Fault\ | |
| \ Tolerant VM takes over as the Primary VM. vSphere HA \t\twill attempt to\ | |
| \ restart the Secondary VM.</action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmGuestRebootEvent</key><description>Guest\ | |
| \ reboot</description><category>info</category><formatOnDatacenter>Guest OS\ | |
| \ reboot for {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Guest\ | |
| \ OS reboot for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Guest\ | |
| \ OS reboot for {vm.name}</formatOnHost><formatOnVm>Guest OS reboot</formatOnVm><fullFormat>Guest\ | |
| \ OS reboot for {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmGuestShutdownEvent</key><description>Guest\ | |
| \ OS shut down</description><category>info</category><formatOnDatacenter>Guest\ | |
| \ OS shut down for {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Guest\ | |
| \ OS shut down for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Guest\ | |
| \ OS shut down for {vm.name}</formatOnHost><formatOnVm>Guest OS shut down</formatOnVm><fullFormat>Guest\ | |
| \ OS shut down for {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmGuestStandbyEvent</key><description>Guest\ | |
| \ standby</description><category>info</category><formatOnDatacenter>Guest\ | |
| \ OS standby for {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Guest\ | |
| \ OS standby for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Guest\ | |
| \ OS standby for {vm.name}</formatOnHost><formatOnVm>Guest OS standby</formatOnVm><fullFormat>Guest\ | |
| \ OS standby for {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmHealthMonitoringStateChangedEvent</key><description>vSphere\ | |
| \ HA VM monitoring state changed</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA VM monitoring state in {computeResource.name} changed to {[email protected]}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA VM monitoring state changed to {[email protected]}</formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vSphere\ | |
| \ HA VM monitoring state in {computeResource.name} in {datacenter.name} changed\ | |
| \ to {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmInstanceUuidAssignedEvent</key><description>Assign\ | |
| \ a new instance UUID</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Assign\ | |
| \ a new instance UUID ({instanceUuid})</formatOnVm><fullFormat>Assign a new\ | |
| \ instance UUID ({instanceUuid}) to {vm.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmInstanceUuidAssignedEvent">\ | |
| \ <description>The virtual machine was assigned a new vCenter Server-specific\ | |
| \ instance UUID </description> <cause> <description>\ | |
| \ The user who created the virtual machine did not specify a\ | |
| \ vCenter Server-specific instance UUID at creation time. vCenter\ | |
| \ Server generated a new UUID and assigned it to the \t virtual machine.\ | |
| \ </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmInstanceUuidChangedEvent</key><description>Instance\ | |
| \ UUID Changed</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>The\ | |
| \ instance UUID has been changed from ({oldInstanceUuid}) to ({newInstanceUuid})</formatOnVm><fullFormat>The\ | |
| \ instance UUID of {vm.name} has been changed from ({oldInstanceUuid}) to\ | |
| \ ({newInstanceUuid})</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmInstanceUuidChangedEvent"> <description>\ | |
| \ The vCenter Server-specific instance UUID of the virtual machine\ | |
| \ has changed </description> <cause> <description>\ | |
| \ A user action resulted in a change to the vCenter Server-specific\ | |
| \ instance UUID of the virtual machine </description>\ | |
| \ </cause> <cause> <description> \ | |
| \ vCenter Server changed the instance UUID of the virtual machine because\ | |
| \ it detected a conflict </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmInstanceUuidConflictEvent</key><description>Instance\ | |
| \ UUIDs conflict</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>The\ | |
| \ instance UUID ({instanceUuid}) conflicts with the instance UUID assigned\ | |
| \ to {conflictedVm.name}</formatOnVm><fullFormat>The instance UUID ({instanceUuid})\ | |
| \ of {vm.name} conflicts with the instance UUID assigned to {conflictedVm.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmInstanceUuidChangedEvent">\ | |
| \ <description> The vCenter Server-specific instance UUID\ | |
| \ of the virtual machine conflicted with that of another virtual\ | |
| \ machine. </description> <cause> <description>\ | |
| \ Virtual machine instance UUID conflicts can occur if you copy\ | |
| \ virtual machine files manually without using vCenter Server.\ | |
| \ </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmMacAssignedEvent</key><description>VM\ | |
| \ MAC assigned</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>New\ | |
| \ MAC address ({mac}) assigned to adapter {adapter}</formatOnVm><fullFormat>New\ | |
| \ MAC address ({mac}) assigned to adapter {adapter} for {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmMacChangedEvent</key><description>VM\ | |
| \ MAC changed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Changed\ | |
| \ MAC address from {oldMac} to {newMac} for adapter {adapter}</formatOnVm><fullFormat>Changed\ | |
| \ MAC address from {oldMac} to {newMac} for adapter {adapter} for {vm.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmMacChangedEvent">\ | |
| \ <description> The virtual machine MAC address has changed\ | |
| \ </description> <cause> <description>\ | |
| \ A user action changed the virtual machine MAC address \ | |
| \ </description> </cause> <cause> <description>\ | |
| \ vCenter changed the virtual machine MAC address because it detected\ | |
| \ a MAC address conflict </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmMacConflictEvent</key><description>VM\ | |
| \ MAC conflict</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>The\ | |
| \ MAC address ({mac}) conflicts with MAC assigned to {conflictedVm.name}</formatOnVm><fullFormat>The\ | |
| \ MAC address ({mac}) of {vm.name} conflicts with MAC assigned to {conflictedVm.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmMacConflictEvent">\ | |
| \ <description> The virtual machine MAC address conflicts\ | |
| \ with that of another virtual machine </description>\ | |
| \ <cause> <description> This virtual machine's\ | |
| \ MAC address is the same as that of another virtual machine.\ | |
| \ Refer to the event details for more information on the \ | |
| \ virtual machine that caused the conflict. </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmMaxFTRestartCountReached</key><description>vSphere\ | |
| \ HA reached maximum Secondary VM restart count.</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA stopped trying to restart Secondary VM {vm.name} on {host.name} in cluster\ | |
| \ {computeResource.name} because the maximum VM restart count was reached</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA stopped trying to restart Secondary VM {vm.name} on {host.name} because\ | |
| \ the maximum VM restart count was reached</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA stopped trying to restart Secondary VM {vm.name} because the maximum\ | |
| \ VM restart count was reached</formatOnHost><formatOnVm>vSphere HA stopped\ | |
| \ trying to restart Secondary VM because the maximum VM restart count was\ | |
| \ reached</formatOnVm><fullFormat>vSphere HA stopped trying to restart Secondary\ | |
| \ VM {vm.name} on {host.name} in cluster {computeResource.name} in {datacenter.name}\ | |
| \ because the maximum VM restart count was reached</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmMaxFTRestartCountReached">\ | |
| \ <description> The system reached the maximum restart\ | |
| \ limit in its attempt to restart a Secondary VM </description>\ | |
| \ <cause> <description>The system exceeded the\ | |
| \ number of allowed restart attempts for the Secondary VM \t\twhen it tried\ | |
| \ to reestablish Fault Tolerance</description> <action>Check\ | |
| \ the causes for the restart failures and fix them. Then disable and re-enable\ | |
| \ Fault \t\tTolerance protection.</action> </cause> \ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmMaxRestartCountReached</key><description>vSphere\ | |
| \ HA reached maximum VM restart count</description><category>warning</category><formatOnDatacenter>vSphere\ | |
| \ HA stopped trying to restart {vm.name} on {host.name} in cluster {computeResource.name}because\ | |
| \ the maximum VM restart count was reached</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA stopped trying to restart {vm.name} on {host.name} because the maximum\ | |
| \ VM restart count was reached</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA stopped trying to restart {vm.name} because the maximum VM restart count\ | |
| \ was reached</formatOnHost><formatOnVm>vSphere HA stopped trying to restart\ | |
| \ this VM because the maximum VM restart count was reached</formatOnVm><fullFormat>vSphere\ | |
| \ HA stopped trying to restart {vm.name} on {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name} because the maximum VM restart count was reached</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmMaxRestartCountReached">\ | |
| \ <description> vSphere HA has reached the maximum\ | |
| \ number of failover attempts for this virtual machine and has\ | |
| \ not been able to restart it. No further failover attempts will\ | |
| \ be made. By default vSphere HA attempts to failover a virtual machine 5\ | |
| \ times. </description> <cause> <description>\ | |
| \ Failover can fail for a number of reasons including that\ | |
| \ the configuration file of the virtual machine is corrupt\ | |
| \ or one or more of the virtual machines datastores are not\ | |
| \ accessible by any host in the cluster due to an all paths \ | |
| \ down condition. In addition, the VM may be powered on a host that the vSphere\ | |
| \ HA master agent can no longer monitor using the management\ | |
| \ network or heartbeat datastores, or it may have been powered\ | |
| \ on by a user on a host outside of the cluster. </description>\ | |
| \ <action> To determine why previous failover\ | |
| \ attempts failed, search the events that are logged for the\ | |
| \ VM for occurences of the event vSphere HA reports when a failover fails.\ | |
| \ These events will report the reason for the failed failover.\ | |
| \ vSphere HA events can be located by searching for the phrase\ | |
| \ 'vSphere HA'. To determine whether any issues still exist, try\ | |
| \ to manually power on the virtual machine. If power-on fails,\ | |
| \ investigate the error that is returned. But, if the power-on\ | |
| \ remains pending for a long time, investigate whether an\ | |
| \ all paths down condition exists. Also, if any hosts have been declared dead,\ | |
| \ investigate whether a networking or storage issue may be\ | |
| \ the cause. </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmMessageErrorEvent</key><description>VM\ | |
| \ error message</description><category>error</category><formatOnDatacenter>Error\ | |
| \ message on {vm.name} on {host.name}: {message}</formatOnDatacenter><formatOnComputeResource>Error\ | |
| \ message on {vm.name} on {host.name}: {message}</formatOnComputeResource><formatOnHost>Error\ | |
| \ message on {vm.name}: {message}</formatOnHost><formatOnVm>Error message\ | |
| \ from {host.name}: {message}</formatOnVm><fullFormat>Error message on {vm.name}\ | |
| \ on {host.name} in {datacenter.name}: {message}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.VmMessageErrorEvent">\t\ | |
| \t<description>\t\t\tAn error message listing a collection of observations\ | |
| \ has been reported by the virtual machine\t\t</description>\t\t<cause>\ | |
| \ <description>\t\t\t\tThe event contains details on why\ | |
| \ this error occurred\t\t\t</description> </cause>\t</EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmMessageEvent</key><description>VM\ | |
| \ information message</description><category>info</category><formatOnDatacenter>Message\ | |
| \ on {vm.name} on {host.name}: {message}</formatOnDatacenter><formatOnComputeResource>Message\ | |
| \ on {vm.name} on {host.name}: {message}</formatOnComputeResource><formatOnHost>Message\ | |
| \ on {vm.name}: {message}</formatOnHost><formatOnVm>Message from {host.name}:\ | |
| \ {message}</formatOnVm><fullFormat>Message on {vm.name} on {host.name} in\ | |
| \ {datacenter.name}: {message}</fullFormat><longDescription>\t<EventLongDescription\ | |
| \ id="vim.event.VmMessageEvent">\t\t<description>\t\t\t\ | |
| An information message listing a collection of observations has been reported\ | |
| \ by the virtual machine\t\t</description>\t\t<cause> \ | |
| \ <description>\t\t\t\tThe event contains details on the messages\ | |
| \ from the virtual machine\t\t\t</description> </cause>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmMessageWarningEvent</key><description>VM\ | |
| \ warning message</description><category>warning</category><formatOnDatacenter>Warning\ | |
| \ message on {vm.name} on {host.name}: {message}</formatOnDatacenter><formatOnComputeResource>Warning\ | |
| \ message on {vm.name} on {host.name}: {message}</formatOnComputeResource><formatOnHost>Warning\ | |
| \ message on {vm.name}: {message}</formatOnHost><formatOnVm>Warning message\ | |
| \ from {host.name}: {message}</formatOnVm><fullFormat>Warning message on {vm.name}\ | |
| \ on {host.name} in {datacenter.name}: {message}</fullFormat><longDescription>\t\ | |
| <EventLongDescription id="vim.event.VmMessageWarningEvent">\t\ | |
| \t<description>\t\t\tA warning message listing a collection of observations\ | |
| \ has been reported by the virtual machine\t\t</description>\t\t<cause>\ | |
| \ <description>\t\t\t\tThe event contains details on why\ | |
| \ this warning was issued\t\t\t</description> </cause>\t\ | |
| </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmMigratedEvent</key><description>VM\ | |
| \ migrated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Migration\ | |
| \ of virtual machine {vm.name} from host {sourceHost.name}, {sourceDatastore.name}\ | |
| \ completed</formatOnHost><formatOnVm>Migration from host {sourceHost.name},\ | |
| \ {sourceDatastore.name} completed</formatOnVm><fullFormat>Migration of virtual\ | |
| \ machine {vm.name} from {sourceHost.name}, {sourceDatastore.name} to {host.name},\ | |
| \ {ds.name} completed</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmMigratedEvent"> <description> \ | |
| \ The virtual machine's host was changed successfully </description>\ | |
| \ <cause> <description> A user action\ | |
| \ caused the virtual machine to be successfully migrated to a\ | |
| \ different host </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmNoCompatibleHostForSecondaryEvent</key><description>No\ | |
| \ compatible host for the Fault Tolerance secondary VM</description><category>error</category><formatOnDatacenter>No\ | |
| \ compatible host for the Fault Tolerance secondary VM {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>No\ | |
| \ compatible host for the Fault Tolerance secondary VM {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>No\ | |
| \ compatible host for the Fault Tolerance secondary VM {vm.name}</formatOnHost><formatOnVm>No\ | |
| \ compatible host for the Fault Tolerance secondary VM</formatOnVm><fullFormat>No\ | |
| \ compatible host for the Fault Tolerance secondary VM {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmNoCompatibleHostForSecondaryEvent">\ | |
| \ <description> No compatible host was found when trying\ | |
| \ to place a Secondary VM </description> <cause> \ | |
| \ <description>There was no compatible host available to place\ | |
| \ a Secondary VM</description> <action>Resolve the\ | |
| \ incompatibilities and retry the operation</action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmNoNetworkAccessEvent</key><description>VM\ | |
| \ No Network Access</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Not\ | |
| \ all networks are accessible by {destHost.name}</formatOnVm><fullFormat>Not\ | |
| \ all networks for {vm.name} are accessible by {destHost.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmOrphanedEvent</key><description>VM\ | |
| \ orphaned</description><category>warning</category><formatOnDatacenter>{vm.name}\ | |
| \ does not exist on {host.name}</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ does not exist on {host.name}</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ does not exist</formatOnHost><formatOnVm>Virtual machine does not exist\ | |
| \ on {host.name}</formatOnVm><fullFormat>{vm.name} does not exist on {host.name}\ | |
| \ in {datacenter.name}</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmOrphanedEvent"> <description> \ | |
| \ The virtual machine does not exist on the host with which it is associated\ | |
| \ </description> <cause> <description>\ | |
| \ The virtual machine was deleted while its host was disconnected from vCenter\ | |
| \ Server. </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmPowerOffOnIsolationEvent</key><description>vSphere\ | |
| \ HA powered off VM on isolated host</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA powered off {vm.name} on the isolated host {isolatedHost.name} in cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA powered off {vm.name} on the isolated host {isolatedHost.name}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA powered off {vm.name} on the isolated host {isolatedHost.name}</formatOnHost><formatOnVm>vSphere\ | |
| \ HA powered off this virtual machine on the isolated host {isolatedHost.name}</formatOnVm><fullFormat>vSphere\ | |
| \ HA powered off {vm.name} on the isolated host {isolatedHost.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmPowerOffOnIsolationEvent">\ | |
| \ <description> vSphere HA powered off this virtual\ | |
| \ machine because the host it was running on was isolated from\ | |
| \ the management network. </description> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmPoweredOffEvent</key><description>VM\ | |
| \ powered off</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is powered off</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is powered off</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is powered off</formatOnHost><formatOnVm>Virtual machine on {host.name}\ | |
| \ is powered off</formatOnVm><fullFormat>{vm.name} on {host.name} in {datacenter.name}\ | |
| \ is powered off</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmPoweredOnEvent</key><description>VM\ | |
| \ powered on</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is powered on</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is powered on</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is powered on</formatOnHost><formatOnVm>Virtual machine on {host.name} is\ | |
| \ powered on</formatOnVm><fullFormat>{vm.name} on {host.name} in {datacenter.name}\ | |
| \ is powered on</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmPoweringOnWithCustomizedDVPortEvent</key><description>Virtual\ | |
| \ machine powered on with vNICs connected to dvPorts that have a port level\ | |
| \ configuration, which might be different from the dvPort group configuration.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine powered On with vNICs connected to dvPorts that have a port level\ | |
| \ configuration, which might be different from the dvPort group configuration.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Virtual\ | |
| \ machine {vm.name} powered On with vNICs connected to dvPorts that have a\ | |
| \ port level configuration, which might be different from the dvPort group\ | |
| \ configuration.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmPrimaryFailoverEvent</key><description>Fault\ | |
| \ Tolerance VM failover</description><category>error</category><formatOnDatacenter>Fault\ | |
| \ Tolerance VM ({vm.name}) failed over to {host.name} in cluster {computeResource.name}.\ | |
| \ {[email protected]}</formatOnDatacenter><formatOnComputeResource>Fault\ | |
| \ Tolerance VM ({vm.name}) failed over to {host.name}. {[email protected]}</formatOnComputeResource><formatOnHost>Fault\ | |
| \ Tolerance VM ({vm.name}) failed over to {host.name}. {[email protected]}</formatOnHost><formatOnVm>Fault\ | |
| \ Tolerance VM failed over to {host.name}. {[email protected]}</formatOnVm><fullFormat>Fault\ | |
| \ Tolerance VM ({vm.name}) failed over to {host.name} in cluster {computeResource.name}\ | |
| \ in {datacenter.name}. {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmReconfiguredEvent</key><description>VM\ | |
| \ reconfigured</description><category>info</category><formatOnDatacenter>Reconfigured\ | |
| \ {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Reconfigured\ | |
| \ {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Reconfigured\ | |
| \ {vm.name}</formatOnHost><formatOnVm>Reconfigured virtual machine</formatOnVm><fullFormat>Reconfigured\ | |
| \ {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRegisteredEvent</key><description>VM\ | |
| \ registered</description><category>info</category><formatOnDatacenter>Registered\ | |
| \ {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Registered\ | |
| \ {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Registered\ | |
| \ {vm.name}</formatOnHost><formatOnVm>Registered with vCenter</formatOnVm><fullFormat>Registered\ | |
| \ {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRelayoutSuccessfulEvent</key><description>VM\ | |
| \ relayout completed</description><category>info</category><formatOnDatacenter>Relayout\ | |
| \ of {vm.name} on {host.name} completed</formatOnDatacenter><formatOnComputeResource>Relayout\ | |
| \ of {vm.name} on {host.name} completed</formatOnComputeResource><formatOnHost>Relayout\ | |
| \ of {vm.name} completed</formatOnHost><formatOnVm>Relayout of the virtual\ | |
| \ machine completed</formatOnVm><fullFormat>Relayout of {vm.name} on {host.name}\ | |
| \ in {datacenter.name} completed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRelayoutUpToDateEvent</key><description>VM\ | |
| \ relayout up-to-date</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is in the correct format and relayout is not neccessary</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is in the correct format and relayout is not neccessary</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is in the correct format and relayout is not neccessary</formatOnHost><formatOnVm>In\ | |
| \ the correct format and relayout is not neccessary</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is in the correct format and relayout\ | |
| \ is not necessary</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmReloadFromPathEvent</key><description>Virtual\ | |
| \ machine reloaded from path</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} reloaded from new configuration {configPath}.</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} reloaded from new configuration {configPath}.</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ reloaded from new configuration {configPath}.</formatOnHost><formatOnVm>Virtual\ | |
| \ machine reloaded from new configuration {configPath}.</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} reloaded from new configuration {configPath}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmReloadFromPathFailedEvent</key><description>Virtual\ | |
| \ machine not reloaded from path</description><category>error</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} could not be reloaded from {configPath}.</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} could not be reloaded from path {configPath}.</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ could not be reloaded from {configPath}.</formatOnHost><formatOnVm>This\ | |
| \ virtual machine could not be reloaded from {configPath}.</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} could not be reloaded from {configPath}.</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmReloadFromPathFailedEvent">\ | |
| \ <description> Reloading the virtual machine from a new\ | |
| \ datastore path failed </description> <cause> \ | |
| \ <description>The destination datastore path was inaccessible or\ | |
| \ invalid </description> <action>Use\ | |
| \ a valid destination datastore path </action> </cause>\ | |
| \ <cause> <description>The virtual machine is in\ | |
| \ an invalid state </description> <action>Check\ | |
| \ the virtual machine state power state. If the virtual machine is \ | |
| \ powered on, power it off </action> </cause>\ | |
| \ <cause> <description>The virtual machine is enabled\ | |
| \ for Fault Tolerance </description> <action>Disable\ | |
| \ Fault Tolerance for the virtual machine and retry the operation\ | |
| \ </action> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmRelocateFailedEvent</key><description>Cannot\ | |
| \ relocate VM</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Cannot\ | |
| \ relocate virtual machine</formatOnVm><fullFormat>Cannot relocate virtual\ | |
| \ machine '{vm.name}' in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmRelocateFailedEvent">\ | |
| \ <description> Virtual machine relocation to a different\ | |
| \ host or datastore failed </description> <cause> \ | |
| \ <description> Virtual machine relocation can fail\ | |
| \ for a number of reasons, including network outages, insufficient\ | |
| \ disk space, and so on </description> <action>\ | |
| \ Consider the task related to this event, evaluate the failure\ | |
| \ reason, and take action accordingly </action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmRelocateSpecEvent</key><description><VM\ | |
| \ Relocate Spec Event></description><category>info</category><formatOnDatacenter><internal></formatOnDatacenter><formatOnComputeResource><internal></formatOnComputeResource><formatOnHost><internal></formatOnHost><formatOnVm><internal></formatOnVm><fullFormat><internal></fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRelocatedEvent</key><description>VM\ | |
| \ relocated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Completed\ | |
| \ the relocation of the virtual machine</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmRelocatedEvent"> <description> \ | |
| \ The virtual machine execution and/or storage was successfully relocated\ | |
| \ </description> <cause> <description>\ | |
| \ A user action caused the virtual machine's execution \ | |
| \ and/or storage to be successfully changed </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmRemoteConsoleConnectedEvent</key><description>VM\ | |
| \ remote console connected</description><category>info</category><formatOnDatacenter>Remote\ | |
| \ console connected to {vm.name} on host {host.name}</formatOnDatacenter><formatOnComputeResource>Remote\ | |
| \ console connected to {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>Remote\ | |
| \ console connected to {vm.name}</formatOnHost><formatOnVm>Remote console\ | |
| \ connected</formatOnVm><fullFormat>Remote console connected to {vm.name}\ | |
| \ on host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRemoteConsoleDisconnectedEvent</key><description>VM\ | |
| \ remote console disconnected</description><category>info</category><formatOnDatacenter>Remote\ | |
| \ console disconnected from {vm.name} on host {host.name}</formatOnDatacenter><formatOnComputeResource>Remote\ | |
| \ console disconnected from {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>Remote\ | |
| \ console disconnected from {vm.name}</formatOnHost><formatOnVm>Remote console\ | |
| \ connected</formatOnVm><fullFormat>Remote console disconnected from {vm.name}\ | |
| \ on host {host.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRemovedEvent</key><description>VM\ | |
| \ removed</description><category>info</category><formatOnDatacenter>Removed\ | |
| \ {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Removed\ | |
| \ {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Removed\ | |
| \ {vm.name}</formatOnHost><formatOnVm>Removed</formatOnVm><fullFormat>Removed\ | |
| \ {vm.name} on {host.name} from {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRenamedEvent</key><description>VM\ | |
| \ renamed</description><category>warning</category><formatOnDatacenter>Renamed\ | |
| \ {vm.name} from {oldName} to {newName}</formatOnDatacenter><formatOnComputeResource>Renamed\ | |
| \ {vm.name} from {oldName} to {newName}</formatOnComputeResource><formatOnHost>Renamed\ | |
| \ {vm.name} from {oldName} to {newName}</formatOnHost><formatOnVm>Renamed\ | |
| \ from {oldName} to {newName}</formatOnVm><fullFormat>Renamed {vm.name} from\ | |
| \ {oldName} to {newName} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRequirementsExceedCurrentEVCModeEvent</key><description>Virtual\ | |
| \ machine is using features that exceed the capabilities of the host's\ | |
| \ current EVC mode.</description><category>warning</category><formatOnDatacenter>Feature\ | |
| \ requirements of {vm.name} exceed capabilities of {host.name}'s current\ | |
| \ EVC mode.</formatOnDatacenter><formatOnComputeResource>Feature requirements\ | |
| \ of {vm.name} exceed capabilities of {host.name}'s current EVC mode.</formatOnComputeResource><formatOnHost>Feature\ | |
| \ requirements of {vm.name} exceed capabilities of this host's current\ | |
| \ EVC mode.</formatOnHost><formatOnVm>Feature requirements of this virtual\ | |
| \ machine exceed capabilities of this host's current EVC mode.</formatOnVm><fullFormat>Feature\ | |
| \ requirements of {vm.name} exceed capabilities of {host.name}'s current\ | |
| \ EVC mode.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmResettingEvent</key><description>VM\ | |
| \ resetting</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is reset</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is reset</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is reset</formatOnHost><formatOnVm>Virtual machine on {host.name} is reset</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is reset</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmResourcePoolMovedEvent</key><description>VM\ | |
| \ resource pool moved</description><category>info</category><formatOnDatacenter>Moved\ | |
| \ {vm.name} from resource pool {oldParent.name} to {newParent.name}</formatOnDatacenter><formatOnComputeResource>Moved\ | |
| \ {vm.name} from resource pool {oldParent.name}</formatOnComputeResource><formatOnHost>Moved\ | |
| \ {vm.name} from resource pool {oldParent.name} to {newParent.name}</formatOnHost><formatOnVm>Moved\ | |
| \ from resource pool {oldParent.name} to {newParent.name}</formatOnVm><fullFormat>Moved\ | |
| \ {vm.name} from resource pool {oldParent.name} to {newParent.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmResourceReallocatedEvent</key><description>VM\ | |
| \ resource reallocated</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Resource\ | |
| \ allocation changed</formatOnVm><fullFormat>Changed resource allocation for\ | |
| \ {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmRestartedOnAlternateHostEvent</key><description>VM\ | |
| \ restarted on alternate host</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Virtual\ | |
| \ machine {vm.name} was restarted on this host since {sourceHost.name} failed</formatOnHost><formatOnVm>Virtual\ | |
| \ machine was restarted on {host.name} since {sourceHost.name} failed</formatOnVm><fullFormat>Virtual\ | |
| \ machine {vm.name} was restarted on {host.name} since {sourceHost.name} failed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmResumingEvent</key><description>VM\ | |
| \ resuming</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is resumed</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is resumed</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is resumed</formatOnHost><formatOnVm>Virtual machine on {host.name} is resumed</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is resumed</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmSecondaryAddedEvent</key><description>Fault\ | |
| \ Tolerance secondary VM added</description><category>info</category><formatOnDatacenter>A\ | |
| \ Fault Tolerance secondary VM has been added for {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name}</formatOnDatacenter><formatOnComputeResource>A\ | |
| \ Fault Tolerance secondary VM has been added for {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>A\ | |
| \ Fault Tolerance secondary VM has been added for {vm.name}</formatOnHost><formatOnVm>A\ | |
| \ Fault Tolerance secondary VM has been added for this VM</formatOnVm><fullFormat>A\ | |
| \ Fault Tolerance secondary VM has been added for {vm.name} on host {host.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmSecondaryDisabledBySystemEvent</key><description>vCenter\ | |
| \ disabled Fault Tolerance</description><category>error</category><formatOnDatacenter>vCenter\ | |
| \ disabled Fault Tolerance on VM {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} because the Secondary VM could not be powered On.</formatOnDatacenter><formatOnComputeResource>vCenter\ | |
| \ disabled Fault Tolerance on VM {vm.name} on host {host.name} because the\ | |
| \ Secondary VM could not be powered On.</formatOnComputeResource><formatOnHost>vCenter\ | |
| \ disabled Fault Tolerance on VM {vm.name} because the Secondary VM could\ | |
| \ not be powered On.</formatOnHost><formatOnVm>vCenter disabled Fault Tolerance\ | |
| \ because the Secondary VM could not be powered On.</formatOnVm><fullFormat>vCenter\ | |
| \ disabled Fault Tolerance on VM {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name} because the Secondary VM could\ | |
| \ not be powered On.</fullFormat><longDescription> <EventLongDescription\ | |
| \ id="vim.event.VmSecondaryDisabledBySystemEvent"> <description>\ | |
| \ vCenter Server disabled a Secondary VM because it could not power\ | |
| \ on the Secondary VM </description> <cause> \ | |
| \ <description>vCenter Server failed to power on the Secondary\ | |
| \ VM </description> <action>Check the reason in the\ | |
| \ event message for more details, fix the failure, and \t\tre-enable Fault\ | |
| \ Tolerance protection to power on the Secondary VM.</action> \ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmSecondaryDisabledEvent</key><description>Disabled\ | |
| \ Fault Tolerance secondary VM</description><category>info</category><formatOnDatacenter>Disabled\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Disabled\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>Disabled\ | |
| \ Fault Tolerance secondary VM for {vm.name}</formatOnHost><formatOnVm>Disabled\ | |
| \ Fault Tolerance secondary VM for this virtual machine</formatOnVm><fullFormat>Disabled\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmSecondaryEnabledEvent</key><description>Enabled\ | |
| \ Fault Tolerance secondary VM</description><category>info</category><formatOnDatacenter>Enabled\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Enabled\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>Enabled\ | |
| \ Fault Tolerance secondary VM for {vm.name}</formatOnHost><formatOnVm>Enabled\ | |
| \ Fault Tolerance secondary VM for this VM</formatOnVm><fullFormat>Enabled\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmSecondaryStartedEvent</key><description>Started\ | |
| \ Fault Tolerance secondary VM</description><category>info</category><formatOnDatacenter>Started\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Started\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name}</formatOnComputeResource><formatOnHost>Started\ | |
| \ Fault Tolerance secondary VM for {vm.name}</formatOnHost><formatOnVm>Started\ | |
| \ Fault Tolerance secondary VM for this virtual machine</formatOnVm><fullFormat>Started\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmShutdownOnIsolationEvent</key><description>vSphere\ | |
| \ HA shut down VM on isolated host</description><category>info</category><formatOnDatacenter>vSphere\ | |
| \ HA shut down {vm.name} on the isolated host {isolatedHost.name} in cluster\ | |
| \ {computeResource.name}: {[email protected]}</formatOnDatacenter><formatOnComputeResource>vSphere\ | |
| \ HA shut down {vm.name} on the isolated host {isolatedHost.name}: {[email protected]}</formatOnComputeResource><formatOnHost>vSphere\ | |
| \ HA shut down {vm.name} on the isolated host {isolatedHost.name}: {[email protected]}</formatOnHost><formatOnVm>vSphere\ | |
| \ HA shut down this virtual machine on the isolated host {isolatedHost.name}:\ | |
| \ {[email protected]}</formatOnVm><fullFormat>vSphere\ | |
| \ HA shut down {vm.name} was shut down on the isolated host {isolatedHost.name}\ | |
| \ in cluster {computeResource.name} in {datacenter.name}: {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmShutdownOnIsolationEvent">\ | |
| \ <description> vSphere HA shut down this virtual machine\ | |
| \ because the host it was running on was isolated from the management\ | |
| \ network. </description> </EventLongDescription> \ | |
| \ </longDescription></eventInfo><eventInfo><key>VmStartRecordingEvent</key><description>Start\ | |
| \ a recording session</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Start\ | |
| \ a recording session</formatOnVm><fullFormat>Start a recording session on\ | |
| \ {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmStartReplayingEvent</key><description>Start\ | |
| \ a replay session</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>Start\ | |
| \ a replay session</formatOnVm><fullFormat>Start a replay session on {vm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmStartingEvent</key><description>VM\ | |
| \ starting</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on host {host.name} is starting</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on host {host.name} is starting</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is starting</formatOnHost><formatOnVm>Virtual machine is starting</formatOnVm><fullFormat>{vm.name}\ | |
| \ on host {host.name} in {datacenter.name} is starting</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmStartingSecondaryEvent</key><description>Starting\ | |
| \ Fault Tolerance secondary VM</description><category>info</category><formatOnDatacenter>Starting\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name}</formatOnDatacenter><formatOnComputeResource>Starting\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster</formatOnComputeResource><formatOnHost>Starting\ | |
| \ Fault Tolerance secondary VM for {vm.name}</formatOnHost><formatOnVm>Starting\ | |
| \ Fault Tolerance secondary VM for this virtual machine</formatOnVm><fullFormat>Starting\ | |
| \ Fault Tolerance secondary VM for {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmStaticMacConflictEvent</key><description>VM\ | |
| \ static MAC conflict</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>The\ | |
| \ static MAC address ({mac}) conflicts with MAC assigned to {conflictedVm.name}</formatOnVm><fullFormat>The\ | |
| \ static MAC address ({mac}) of {vm.name} conflicts with MAC assigned to \ | |
| \ {conflictedVm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmStoppingEvent</key><description>VM\ | |
| \ stopping</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is stopping</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is stopping</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is stopping</formatOnHost><formatOnVm>Virtual machine is stopping</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is stopping</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmSuspendedEvent</key><description>VM\ | |
| \ suspended</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is suspended</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is suspended</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is suspended</formatOnHost><formatOnVm>Virtual machine is suspended</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is suspended</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmSuspendingEvent</key><description>VM\ | |
| \ being suspended</description><category>info</category><formatOnDatacenter>{vm.name}\ | |
| \ on {host.name} is being suspended</formatOnDatacenter><formatOnComputeResource>{vm.name}\ | |
| \ on {host.name} is being suspended</formatOnComputeResource><formatOnHost>{vm.name}\ | |
| \ is being suspended</formatOnHost><formatOnVm>Virtual machine is being suspended</formatOnVm><fullFormat>{vm.name}\ | |
| \ on {host.name} in {datacenter.name} is being suspended</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmTimedoutStartingSecondaryEvent</key><description>Starting\ | |
| \ the Fault Tolerance secondary VM timed out</description><category>error</category><formatOnDatacenter>Starting\ | |
| \ the Fault Tolerance secondary VM {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} timed out within {timeout} ms</formatOnDatacenter><formatOnComputeResource>Starting\ | |
| \ the Fault Tolerance secondary VM {vm.name} on host {host.name} timed out\ | |
| \ within {timeout} ms</formatOnComputeResource><formatOnHost>Starting the\ | |
| \ Fault Tolerance secondary VM {vm.name} timed out within {timeout} ms</formatOnHost><formatOnVm>Starting\ | |
| \ the Fault Tolerance secondary VM timed out within {timeout} ms</formatOnVm><fullFormat>Starting\ | |
| \ the Fault Tolerance secondary VM {vm.name} on host {host.name} in cluster\ | |
| \ {computeResource.name} in {datacenter.name} timed out within {timeout} ms</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmTimedoutStartingSecondaryEvent">\ | |
| \ <description> An attempt to start a Secondary VM timed out.\ | |
| \ </description> <cause> <description>A\ | |
| \ user attempted to turn on or enable Fault Tolerance, triggering the start\ | |
| \ of \t\tthe Secondary VM. The start operation timed out and, as a result,\ | |
| \ vCenter Server disables \t\tFault Tolerance. </description> \ | |
| \ <action>Fix any problems and re-enable Fault Tolerance protection</action>\ | |
| \ </cause> <cause> <description>The\ | |
| \ secondary VM was started in response to a failure, but the start attempt\ | |
| \ \t\ttimed out</description> <action> vSphere HA will\ | |
| \ attempt to power on the Secondary VM</action> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmUnsupportedStartingEvent</key><description>VM\ | |
| \ unsupported guest OS is starting</description><category>warning</category><formatOnDatacenter>Unsupported\ | |
| \ guest OS {guestId} for {vm.name}</formatOnDatacenter><formatOnComputeResource>Unsupported\ | |
| \ guest OS {guestId} for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Unsupported\ | |
| \ guest OS {guestId} for {vm.name} on {host.name} in {datacenter.name}</formatOnHost><formatOnVm>Unsupported\ | |
| \ guest OS {guestId}</formatOnVm><fullFormat>Unsupported guest OS {guestId}\ | |
| \ for {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmUnsupportedStartingEvent">\ | |
| \ <description> Attempting to power on a virtual machine\ | |
| \ that has an unsupported guest operating system </description>\ | |
| \ <cause> <description> A user action\ | |
| \ initiated a virtual machine power-on operation, but the virtual machine\ | |
| \ has \t an unsupported guest operating system. </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmUpgradeCompleteEvent</key><description>VM\ | |
| \ upgrade complete</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Virtual\ | |
| \ machine compatibility upgraded to {[email protected]}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmUpgradeFailedEvent</key><description>Cannot\ | |
| \ upgrade VM</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>Cannot\ | |
| \ upgrade virtual machine compatibility.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmUpgradingEvent</key><description>Upgrading\ | |
| \ VM</description><category>info</category><formatOnDatacenter>Upgrading virtual\ | |
| \ machine compatibility of {vm.name} to {[email protected]}</formatOnDatacenter><formatOnComputeResource>Upgrading\ | |
| \ virtual machine compatibility of {vm.name} to {[email protected]}</formatOnComputeResource><formatOnHost>Upgrading\ | |
| \ virtual machine compatibility of {vm.name} to {[email protected]}</formatOnHost><formatOnVm>Upgrading\ | |
| \ virtual machine compatibility to {[email protected]}</formatOnVm><fullFormat>Upgrading\ | |
| \ virtual machine compatibility of {vm.name} in {datacenter.name} to {[email protected]}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmUpgradingEvent">\ | |
| \ <description>The virtual hardware on this virtual machine is\ | |
| \ being upgraded</description> <cause> <description>A\ | |
| \ user-initiated action triggered an upgrade of the virtual machine hardware</description>\ | |
| \ </cause> <cause> <description>A scheduled\ | |
| \ task started an upgrade of the virtual machine hardware</description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmUuidAssignedEvent</key><description>VM\ | |
| \ UUID assigned</description><category>info</category><formatOnDatacenter>Assigned\ | |
| \ new BIOS UUID ({uuid}) to {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Assigned\ | |
| \ new BIOS UUID ({uuid}) to {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Assigned\ | |
| \ new BIOS UUID ({uuid}) to {vm.name}</formatOnHost><formatOnVm>Assigned new\ | |
| \ BIOS UUID ({uuid})</formatOnVm><fullFormat>Assigned new BIOS UUID ({uuid})\ | |
| \ to {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmUuidAssignedEvent">\ | |
| \ <description>The virtual machine was assigned a new BIOS UUID</description>\ | |
| \ <cause> <description>The user who created the virtual\ | |
| \ machine did not specify a BIOS UUID at creation time. vCenter\ | |
| \ Server generated a new UUID and assigned it to the virtual machine. \ | |
| \ </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmUuidChangedEvent</key><description>VM\ | |
| \ UUID Changed</description><category>warning</category><formatOnDatacenter>Changed\ | |
| \ BIOS UUID from {oldUuid} to {newUuid} for {vm.name} on {host.name}</formatOnDatacenter><formatOnComputeResource>Changed\ | |
| \ BIOS UUID from {oldUuid} to {newUuid} for {vm.name} on {host.name}</formatOnComputeResource><formatOnHost>Changed\ | |
| \ BIOS UUID from {oldUuid} to {newUuid} for {vm.name}</formatOnHost><formatOnVm>BIOS\ | |
| \ UUID was changed from {oldUuid} to {newUuid}</formatOnVm><fullFormat>Changed\ | |
| \ BIOS UUID from {oldUuid} to {newUuid} for {vm.name} on {host.name} in {datacenter.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmUuidChangedEvent">\ | |
| \ <description>The virtual machine BIOS UUID has changed</description>\ | |
| \ <cause> <description> A user changed\ | |
| \ the virtual machine BIOS UUID directly on the host </description>\ | |
| \ </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmUuidConflictEvent</key><description>VM\ | |
| \ UUID Conflict</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>BIOS\ | |
| \ ID ({uuid}) conflicts with that of {conflictedVm.name}</formatOnVm><fullFormat>BIOS\ | |
| \ ID ({uuid}) of {vm.name} conflicts with that of {conflictedVm.name}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmVnicPoolReservationViolationClearEvent</key><description>Virtual\ | |
| \ NIC Network Resource Pool Reservation Violation Clear event</description><category>info</category><formatOnDatacenter>The\ | |
| \ reservation violation on the virtual NIC network resource pool {vmVnicResourcePoolName}\ | |
| \ with key {vmVnicResourcePoolKey} on {dvs.name} is cleared</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ reservation violation on the virtual NIC network resource pool {vmVnicResourcePoolName}\ | |
| \ with key {vmVnicResourcePoolKey} on {dvs.name} is cleared</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmVnicPoolReservationViolationRaiseEvent</key><description>Virtual\ | |
| \ NIC Network Resource Pool Reservation Violation event</description><category>info</category><formatOnDatacenter>The\ | |
| \ reservation allocated to the virtual NIC network resource pool {vmVnicResourcePoolName}\ | |
| \ with key {vmVnicResourcePoolKey} on {dvs.name} is violated</formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>The\ | |
| \ reservation allocated to the virtual NIC network resource pool {vmVnicResourcePoolName}\ | |
| \ with key {vmVnicResourcePoolKey} on {dvs.name} is violated</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>VmWwnAssignedEvent</key><description>VM\ | |
| \ WWN assigned</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>New\ | |
| \ WWNs assigned</formatOnVm><fullFormat>New WWNs assigned to {vm.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmWwnAssignedEvent">\ | |
| \ <description> The virtual machine was assigned a WWN\ | |
| \ (World Wide Name) </description> <cause> <description>The\ | |
| \ virtual machine was assigned a WWN because it was created with\ | |
| \ an RDM (Raw Device Mappings) disk or was reconfigured to access\ | |
| \ an RDM disk </description> </cause> </EventLongDescription>\ | |
| \ </longDescription></eventInfo><eventInfo><key>VmWwnChangedEvent</key><description>VM\ | |
| \ WWN changed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>WWNs\ | |
| \ are changed</formatOnVm><fullFormat>WWNs are changed for {vm.name}</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmWwnChangedEvent">\ | |
| \ <description> The WWN (World Wide Name) assigned to the\ | |
| \ virtual machine was changed </description> <cause>\ | |
| \ <description>The virtual machine was assigned a new WWN, possibly\ | |
| \ due to a conflict caused by another virtual machine being \ | |
| \ assigned the same WWN </description> </cause>\ | |
| \ </EventLongDescription> </longDescription></eventInfo><eventInfo><key>VmWwnConflictEvent</key><description>VM\ | |
| \ WWN conflict</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm>The\ | |
| \ WWN ({wwn}) conflicts with the currently registered WWN</formatOnVm><fullFormat>The\ | |
| \ WWN ({wwn}) of {vm.name} conflicts with the currently registered WWN</fullFormat><longDescription>\ | |
| \ <EventLongDescription id="vim.event.VmWwnConflictEvent">\ | |
| \ <description> The WWN (World Wide Name) assigned to the\ | |
| \ virtual machine has a conflict </description> \ | |
| \ <cause> <description>The WWN assigned to this virtual\ | |
| \ machine was the same as that of a different virtual machine.\ | |
| \ </description> <action>\t Check the event details\ | |
| \ for more information on the conflict and correct the problem. </action>\\\ | |
| </cause> </EventLongDescription> </longDescription></eventInfo><eventInfo><key>WarningUpgradeEvent</key><description>Warning\ | |
| \ upgrade</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>{message}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>IScsiBootFailureEvent</key><description>Boot\ | |
| \ from iSCSI failed.</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost>Booting\ | |
| \ from iSCSI failed.</formatOnHost><formatOnVm></formatOnVm><fullFormat>Booting\ | |
| \ from iSCSI failed with an error. See the VMware Knowledge Base for information\ | |
| \ on configuring iBFT networking.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Network Connectivity</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.net.connectivity.lost|Lost\ | |
| \ network connectivity on virtual switch {1}. Physical NIC {2} is down. Affected\ | |
| \ portgroups:{3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>No\ | |
| \ IPv6 TSO support</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.net.e1000.tso6.notsupported|Guest-initiated\ | |
| \ IPv6 TCP Segmentation Offload (TSO) packets ignored. Manually disable TSO\ | |
| \ inside the guest operating system in virtual machine {1}, or use a different\ | |
| \ virtual adapter.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Invalid\ | |
| \ vmknic specified in /Migrate/Vmknic</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.net.migrate.bindtovmk|The\ | |
| \ ESX advanced config option /Migrate/Vmknic is set to an invalid vmknic:\ | |
| \ {1}. /Migrate/Vmknic specifies a vmknic that vMotion binds to for improved\ | |
| \ performance. Please update the config option with a valid vmknic or, if\ | |
| \ you do not want vMotion to bind to a specific vmknic, remove the invalid\ | |
| \ vmknic and leave the option blank.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Virtual\ | |
| \ NIC connection to switch failed</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.net.proxyswitch.port.unavailable|Virtual\ | |
| \ NIC with hardware address {1} failed to connect to distributed virtual port\ | |
| \ {2} on switch {3}. There are no more ports available on the host proxy switch.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Network\ | |
| \ Redundancy Degraded</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.net.redundancy.degraded|Uplink\ | |
| \ redundancy degraded on virtual switch {1}. Physical NIC {2} is down. {3}\ | |
| \ uplinks still up. Affected portgroups:{4}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Network Redundancy</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.net.redundancy.lost|Lost\ | |
| \ uplink redundancy on virtual switch {1}. Physical NIC {2} is down. Affected\ | |
| \ portgroups:{3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Thin\ | |
| \ Provisioned Device Nearing Capacity</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.scsi.device.thinprov.atquota|Space\ | |
| \ utilization on thin-provisioned device {1} exceeded configured threshold.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Storage Connectivity</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.storage.connectivity.lost|Lost\ | |
| \ connectivity to storage device {1}. Path {2} is down. Affected datastores:\ | |
| \ {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Degraded\ | |
| \ Storage Path Redundancy</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.storage.redundancy.degraded|Path\ | |
| \ redundancy to storage device {1} degraded. Path {2} is down. {3} remaining\ | |
| \ active paths. Affected datastores: {4}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ Storage Path Redundancy</description><category>warning</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.storage.redundancy.lost|Lost\ | |
| \ path redundancy to storage device {1}. Path {2} is down. Affected datastores:\ | |
| \ {3}.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Locked By Remote Host</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.error.volume.is.locked|Volume\ | |
| \ on device {1} is locked, possibly because some remote host encountered an\ | |
| \ error during a volume operation and could not recover.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ backing an extent of a file system is offline.</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.extent.offline|An\ | |
| \ attached device {1} might be offline. The file system {2} is now in a degraded\ | |
| \ state. While the datastore is still available, parts of data that reside\ | |
| \ on the extent that went offline might be inaccessible.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Device\ | |
| \ backing an extent of a file system is online.</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.extent.online|Device\ | |
| \ {1} backing file system {2} came online. This extent was previously offline.\ | |
| \ All resources on this device are now available.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Volume Connectivity Restored</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.heartbeat.recovered|Successfully\ | |
| \ restored access to volume {1} ({2}) following connectivity issues.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Volume Connectivity Degraded</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.heartbeat.timedout|Lost\ | |
| \ access to volume {1} ({2}) due to connectivity issues. Recovery attempt\ | |
| \ is in progress and outcome will be reported shortly.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Volume Connectivity Lost</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.heartbeat.unrecoverable|Lost\ | |
| \ connectivity to volume {1} ({2}) and subsequent recovery attempts have failed.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>No\ | |
| \ Space To Create VMFS Journal</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.journal.createfailed|No\ | |
| \ space for journal on volume {1} ({2}). Opening volume in read-only metadata\ | |
| \ mode with limited write support.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Lock Corruption Detected</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.lock.corruptondisk|At\ | |
| \ least one corrupt on-disk lock was detected on volume {1} ({2}). Other regions\ | |
| \ of the volume may be damaged too.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Lost\ | |
| \ connection to NFS server</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.nfs.server.disconnect|Lost\ | |
| \ connection to server {1} mount point {2} mounted as {3} ({4}).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>Restored\ | |
| \ connection to NFS server</description><category>info</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.nfs.server.restored|Restored\ | |
| \ connection to server {1} mount point {2} mounted as {3} ({4}).</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>VMFS\ | |
| \ Resource Corruption Detected</description><category>error</category><formatOnDatacenter></formatOnDatacenter><formatOnComputeResource></formatOnComputeResource><formatOnHost></formatOnHost><formatOnVm></formatOnVm><fullFormat>vprob.vmfs.resource.corruptondisk|At\ | |
| \ least one corrupt resource metadata region was detected on volume {1} ({2}).\ | |
| \ Other regions of the volume might be damaged too.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Data Security scan ended on the VirtualMachine.</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Data Security scan ended on the VirtualMachine.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Data Security scan ended on the VirtualMachine.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Data Security scan ended on the VirtualMachine.</formatOnHost><formatOnVm>vShield\ | |
| \ Data Security scan ended on the VirtualMachine.</formatOnVm><fullFormat>com.vmware.vshield.dlp.VmScanEnd|vShield\ | |
| \ Data Security scan ended on the VirtualMachine.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Data Security scan started on the VirtualMachine.</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Data Security scan started on the VirtualMachine.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Data Security scan started on the VirtualMachine.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Data Security scan started on the VirtualMachine.</formatOnHost><formatOnVm>vShield\ | |
| \ Data Security scan started on the VirtualMachine.</formatOnVm><fullFormat>com.vmware.vshield.dlp.VmScanStart|vShield\ | |
| \ Data Security scan started on the VirtualMachine.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Manager found vShield Edge in bad state. Needs a force sync.</description><category>error</category><formatOnDatacenter>vShield\ | |
| \ Manager found vShield Edge in bad state. Needs a force sync.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Manager found vShield Edge in bad state. Needs a force sync.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Manager found vShield Edge in bad state. Needs a force sync.</formatOnHost><formatOnVm>vShield\ | |
| \ Manager found vShield Edge in bad state. Needs a force sync.</formatOnVm><fullFormat>com.vmware.vshield.edge.BadState|vShield\ | |
| \ Manager found vShield Edge in bad state. Needs a force sync.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge Communication Agent not connected to vCenter Server.</description><category>error</category><formatOnDatacenter>vShield\ | |
| \ Edge Communication Agent not connected to vCenter Server.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge Communication Agent not connected to vCenter Server.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge Communication Agent not connected to vCenter Server.</formatOnHost><formatOnVm>vShield\ | |
| \ Edge Communication Agent not connected to vCenter Server.</formatOnVm><fullFormat>com.vmware.vshield.edge.VixagentNotConnectedToVc|vShield\ | |
| \ Edge Communication Agent not connected to vCenter Server.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge is in Bad State. Needs a force sync.</description><category>error</category><formatOnDatacenter>vShield\ | |
| \ Edge is in Bad State. Needs a force sync.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge is in Bad State. Needs a force sync.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge is in Bad State. Needs a force sync.</formatOnHost><formatOnVm>vShield\ | |
| \ Edge is in Bad State. Needs a force sync.</formatOnVm><fullFormat>com.vmware.vshield.edge.appliance.BadState|vShield\ | |
| \ Edge is in Bad State. Needs a force sync.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Failed\ | |
| \ to communicate with the Vshield Edge VM.</description><category>warning</category><formatOnDatacenter>Failed\ | |
| \ to communicate with the Vshield Edge VM.</formatOnDatacenter><formatOnComputeResource>Failed\ | |
| \ to communicate with the Vshield Edge VM.</formatOnComputeResource><formatOnHost>Failed\ | |
| \ to communicate with the Vshield Edge VM.</formatOnHost><formatOnVm>Failed\ | |
| \ to communicate with the Vshield Edge VM.</formatOnVm><fullFormat>com.vmware.vshield.edge.appliance.CommunicationFailed|Failed\ | |
| \ to communicate with the Vshield Edge VM.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge was force synced</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Edge was force synced</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge was force synced</formatOnComputeResource><formatOnHost>vShield Edge\ | |
| \ was force synced</formatOnHost><formatOnVm>vShield Edge was force synced</formatOnVm><fullFormat>com.vmware.vshield.edge.appliance.ForceSynced|vShield\ | |
| \ Edge was force synced</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge appliance moved in the vCenter inventory. This might have been because\ | |
| \ of activities like vMotion, DRS.</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Edge appliance moved in the vCenter inventory. This might have been because\ | |
| \ of activities like vMotion, DRS.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge appliance moved in the vCenter inventory. This might have been because\ | |
| \ of activities like vMotion, DRS.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge appliance moved in the vCenter inventory. This might have been because\ | |
| \ of activities like vMotion, DRS.</formatOnHost><formatOnVm>vShield Edge\ | |
| \ appliance moved in the vCenter inventory. This might have been because of\ | |
| \ activities like vMotion, DRS.</formatOnVm><fullFormat>com.vmware.vshield.edge.appliance.MovedInVcInventory|vShield\ | |
| \ Edge appliance moved in the vCenter inventory. This might have been because\ | |
| \ of activities like vMotion, DRS.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge VM health check failing with critical vix errors. Further health check\ | |
| \ is disabled for vm. Please redeploy or forcesync vm to resume health check.</description><category>error</category><formatOnDatacenter>vShield\ | |
| \ Edge VM health check failing with critical vix errors. Further health check\ | |
| \ is disabled for vm. Please redeploy or forcesync vm to resume health check.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge VM health check failing with critical vix errors. Further health check\ | |
| \ is disabled for vm. Please redeploy or forcesync vm to resume health check.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge VM health check failing with critical vix errors. Further health check\ | |
| \ is disabled for vm. Please redeploy or forcesync vm to resume health check.</formatOnHost><formatOnVm>vShield\ | |
| \ Edge VM health check failing with critical vix errors. Further health check\ | |
| \ is disabled for vm. Please redeploy or forcesync vm to resume health check.</formatOnVm><fullFormat>com.vmware.vshield.edge.appliance.criticalState|vShield\ | |
| \ Edge VM health check failing with critical vix errors. Further health check\ | |
| \ is disabled for vm. Please redeploy or forcesync vm to resume health check.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Created\ | |
| \ vShield Edge Gateway with name : ''{0}'' and id : ''{1}''</description><category>info</category><formatOnDatacenter>Created\ | |
| \ vShield Edge Gateway with name : ''{0}'' and id : ''{1}''</formatOnDatacenter><formatOnComputeResource>Created\ | |
| \ vShield Edge Gateway with name : ''{0}'' and id : ''{1}''</formatOnComputeResource><formatOnHost>Created\ | |
| \ vShield Edge Gateway with name : ''{0}'' and id : ''{1}''</formatOnHost><formatOnVm>Created\ | |
| \ vShield Edge Gateway with name : ''{0}'' and id : ''{1}''</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.Created|Created\ | |
| \ vShield Edge Gateway with name : ''{0}'' and id : ''{1}''</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Deleted\ | |
| \ vShield Edge Gateway with id : {0}</description><category>info</category><formatOnDatacenter>Deleted\ | |
| \ vShield Edge Gateway with id : {0}</formatOnDatacenter><formatOnComputeResource>Deleted\ | |
| \ vShield Edge Gateway with id : {0}</formatOnComputeResource><formatOnHost>Deleted\ | |
| \ vShield Edge Gateway with id : {0}</formatOnHost><formatOnVm>Deleted vShield\ | |
| \ Edge Gateway with id : {0}</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.Deleted|Deleted\ | |
| \ vShield Edge Gateway with id : {0}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>None\ | |
| \ of the VShield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</description><category>error</category><formatOnDatacenter>None\ | |
| \ of the VShield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</formatOnDatacenter><formatOnComputeResource>None of\ | |
| \ the VShield Edge VMs found in serving state. There is a possibility of network\ | |
| \ disruption.</formatOnComputeResource><formatOnHost>None of the VShield Edge\ | |
| \ VMs found in serving state. There is a possibility of network disruption.</formatOnHost><formatOnVm>None\ | |
| \ of the VShield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.HealthcheckNoPulse|None\ | |
| \ of the VShield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge Gateway has recovered and now responding to health check.</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Edge Gateway has recovered and now responding to health check.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge Gateway has recovered and now responding to health check.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge Gateway has recovered and now responding to health check.</formatOnHost><formatOnVm>vShield\ | |
| \ Edge Gateway has recovered and now responding to health check.</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.HealthcheckRecovery|vShield\ | |
| \ Edge Gateway has recovered and now responding to health check.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>None\ | |
| \ of the Vshield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</description><category>error</category><formatOnDatacenter>None\ | |
| \ of the Vshield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</formatOnDatacenter><formatOnComputeResource>None of\ | |
| \ the Vshield Edge VMs found in serving state. There is a possibility of network\ | |
| \ disruption.</formatOnComputeResource><formatOnHost>None of the Vshield Edge\ | |
| \ VMs found in serving state. There is a possibility of network disruption.</formatOnHost><formatOnVm>None\ | |
| \ of the Vshield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.NoActiveOrSelfStateAppliance|None\ | |
| \ of the Vshield Edge VMs found in serving state. There is a possibility of\ | |
| \ network disruption.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Redeployed\ | |
| \ vShield Edge Gateway with id : {0}</description><category>info</category><formatOnDatacenter>Redeployed\ | |
| \ vShield Edge Gateway with id : {0}</formatOnDatacenter><formatOnComputeResource>Redeployed\ | |
| \ vShield Edge Gateway with id : {0}</formatOnComputeResource><formatOnHost>Redeployed\ | |
| \ vShield Edge Gateway with id : {0}</formatOnHost><formatOnVm>Redeployed\ | |
| \ vShield Edge Gateway with id : {0}</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.Redeployed|Redeployed\ | |
| \ vShield Edge Gateway with id : {0}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>Appliance\ | |
| \ size for vShield Edge Gateway with id : {0} changed to {1}</description><category>info</category><formatOnDatacenter>Appliance\ | |
| \ size for vShield Edge Gateway with id : {0} changed to {1}</formatOnDatacenter><formatOnComputeResource>Appliance\ | |
| \ size for vShield Edge Gateway with id : {0} changed to {1}</formatOnComputeResource><formatOnHost>Appliance\ | |
| \ size for vShield Edge Gateway with id : {0} changed to {1}</formatOnHost><formatOnVm>Appliance\ | |
| \ size for vShield Edge Gateway with id : {0} changed to {1}</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.SizeChanged|Appliance\ | |
| \ size for vShield Edge Gateway with id : {0} changed to {1}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge Gateway with id: {1} and name : {0} upgraded to version : {2}</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Edge Gateway with id : {1} and name : {0} upgraded to version : {2}</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge Gateway with id : {1} and name : {0} upgraded to version : {2}</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge Gateway with id : {1} and name : {0} upgraded to version : {2}</formatOnHost><formatOnVm>vShield\ | |
| \ Edge Gateway with id : {1} and name : {0} upgraded to version : {2}</formatOnVm><fullFormat>com.vmware.vshield.edge.gateway.Upgraded|vShield\ | |
| \ Edge Gateway with id : {1} and name : {0} upgraded to version : {2}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge HighAvailability communication channel is disconnected from peer node</description><category>warning</category><formatOnDatacenter>vShield\ | |
| \ Edge HighAvailability communication channel is disconnected from peer node</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge HighAvailability communication channel is disconnected from peer node</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge HighAvailability communication channel is disconnected from peer node</formatOnHost><formatOnVm>vShield\ | |
| \ Edge HighAvailability communication channel is disconnected from peer node</formatOnVm><fullFormat>com.vmware.vshield.edge.highavailability.ChannelDisconnected|vShield\ | |
| \ Edge HighAvailability communication channel is disconnected from peer node</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge HighAvailability is disabled. VM has moved to SELF state.</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Edge HighAvailability is disabled. VM has moved to SELF state.</formatOnDatacenter><formatOnComputeResource>vShield\ | |
| \ Edge HighAvailability is disabled. VM has moved to SELF state.</formatOnComputeResource><formatOnHost>vShield\ | |
| \ Edge HighAvailability is disabled. VM has moved to SELF state.</formatOnHost><formatOnVm>vShield\ | |
| \ Edge HighAvailability is disabled. VM has moved to SELF state.</formatOnVm><fullFormat>com.vmware.vshield.edge.highavailability.StateSelf|vShield\ | |
| \ Edge HighAvailability is disabled. VM has moved to SELF state.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to down</description><category>info</category><formatOnDatacenter>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to down</formatOnDatacenter><formatOnComputeResource>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to down</formatOnComputeResource><formatOnHost>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to down</formatOnHost><formatOnVm>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to down</formatOnVm><fullFormat>com.vmware.vshield.edge.ipsec.ChannelStateDown|IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to down</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to up</description><category>info</category><formatOnDatacenter>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to up</formatOnDatacenter><formatOnComputeResource>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to up</formatOnComputeResource><formatOnHost>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to up</formatOnHost><formatOnVm>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to up</formatOnVm><fullFormat>com.vmware.vshield.edge.ipsec.ChannelStateUp|IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed the state to up</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed to a wrong state</description><category>info</category><formatOnDatacenter>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed to a wrong state</formatOnDatacenter><formatOnComputeResource>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed to a wrong state</formatOnComputeResource><formatOnHost>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed to a wrong state</formatOnHost><formatOnVm>IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed to a wrong state</formatOnVm><fullFormat>com.vmware.vshield.edge.ipsec.ChannelWrongState|IPsec\ | |
| \ Channel from localIp : {0} to peerIp : {1} changed to a wrong state</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to down</description><category>info</category><formatOnDatacenter>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to down</formatOnDatacenter><formatOnComputeResource>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to down</formatOnComputeResource><formatOnHost>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to down</formatOnHost><formatOnVm>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to down</formatOnVm><fullFormat>com.vmware.vshield.edge.ipsec.TunnelStateDown|IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to down</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to up</description><category>info</category><formatOnDatacenter>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to up</formatOnDatacenter><formatOnComputeResource>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to up</formatOnComputeResource><formatOnHost>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to up</formatOnHost><formatOnVm>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to up</formatOnVm><fullFormat>com.vmware.vshield.edge.ipsec.TunnelStateUp|IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed the state to up</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed to a wrong state</description><category>info</category><formatOnDatacenter>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed to a wrong state</formatOnDatacenter><formatOnComputeResource>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed to a wrong state</formatOnComputeResource><formatOnHost>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed to a wrong state</formatOnHost><formatOnVm>IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed to a wrong state</formatOnVm><fullFormat>com.vmware.vshield.edge.ipsec.TunnelWrongState|IPsec\ | |
| \ Tunnel from localSubnet : {0} to peerSubnet : {1} changed to a wrong state</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to down</description><category>info</category><formatOnDatacenter>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to down</formatOnDatacenter><formatOnComputeResource>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to down</formatOnComputeResource><formatOnHost>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to down</formatOnHost><formatOnVm>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to down</formatOnVm><fullFormat>com.vmware.vshield.edge.loadbalancer.BackendServerDown|LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to down</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to up</description><category>info</category><formatOnDatacenter>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to up</formatOnDatacenter><formatOnComputeResource>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to up</formatOnComputeResource><formatOnHost>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to up</formatOnHost><formatOnVm>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to up</formatOnVm><fullFormat>com.vmware.vshield.edge.loadbalancer.BackendServerUp|LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed the state to up</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed to a wrong state</description><category>info</category><formatOnDatacenter>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed to a wrong state</formatOnDatacenter><formatOnComputeResource>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed to a wrong state</formatOnComputeResource><formatOnHost>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed to a wrong state</formatOnHost><formatOnVm>LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed to a wrong state</formatOnVm><fullFormat>com.vmware.vshield.edge.loadbalancer.BackendServerWrongState|LoadBalancer\ | |
| \ virtualServer : {0} Protocol : {1} serverIp : {2} changed to a wrong state</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>vShield\ | |
| \ Edge 5.0 successfully upgraded to vShield Edge 5.5 gateway with id : {0}\ | |
| \ and name : {1}</description><category>info</category><formatOnDatacenter>vShield\ | |
| \ Edge 5.0 successfully upgraded to vShield Edge 5.5 gateway with id : {0}\ | |
| \ and name : {1}</formatOnDatacenter><formatOnComputeResource>vShield Edge\ | |
| \ 5.0 successfully upgraded to vShield Edge 5.5 gateway with id : {0} and\ | |
| \ name : {1}</formatOnComputeResource><formatOnHost>vShield Edge 5.0 successfully\ | |
| \ upgraded to vShield Edge 5.5 gateway with id : {0} and name : {1}</formatOnHost><formatOnVm>vShield\ | |
| \ Edge 5.0 successfully upgraded to vShield Edge 5.5 gateway with id : {0}\ | |
| \ and name : {1}</formatOnVm><fullFormat>com.vmware.vshield.edge.upgrade.ToSpockGateway|vShield\ | |
| \ Edge 5.0 successfully upgraded to vShield Edge 5.5 gateway with id : {0}\ | |
| \ and name : {1}</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>User\ | |
| \ Role assigned for vCenter User on vSM.</description><category>info</category><formatOnDatacenter>User\ | |
| \ Role assigned for vCenter User on vSM.</formatOnDatacenter><formatOnComputeResource>User\ | |
| \ Role assigned for vCenter User on vSM.</formatOnComputeResource><formatOnHost>User\ | |
| \ Role assigned for vCenter User on vSM.</formatOnHost><formatOnVm>User Role\ | |
| \ assigned for vCenter User on vSM.</formatOnVm><fullFormat>com.vmware.vshield.vsm.usermgmt.vcUserRoleAssigned|User\ | |
| \ Role assigned for vCenter User on vSM.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>ExtendedEvent</key><description>User\ | |
| \ Role unassigned for vCenter User on vSM.</description><category>info</category><formatOnDatacenter>User\ | |
| \ Role unassigned for vCenter User on vSM.</formatOnDatacenter><formatOnComputeResource>User\ | |
| \ Role unassigned for vCenter User on vSM.</formatOnComputeResource><formatOnHost>User\ | |
| \ Role unassigned for vCenter User on vSM.</formatOnHost><formatOnVm>User\ | |
| \ Role unassigned for vCenter User on vSM.</formatOnVm><fullFormat>com.vmware.vshield.vsm.usermgmt.vcUserRoleUnassigned|User\ | |
| \ Role unassigned for vCenter User on vSM.</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vxlan.instance.initfail</description><category>event.com.vmware.vxlan.instance.initfail.category</category><formatOnDatacenter>event.com.vmware.vxlan.instance.initfail.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vxlan.instance.initfail.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vxlan.instance.initfail.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vxlan.instance.initfail.fullFormat</formatOnVm><fullFormat>com.vmware.vxlan.instance.initfail|event.com.vmware.vxlan.instance.initfail.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vxlan.instance.notexist</description><category>event.com.vmware.vxlan.instance.notexist.category</category><formatOnDatacenter>event.com.vmware.vxlan.instance.notexist.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vxlan.instance.notexist.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vxlan.instance.notexist.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vxlan.instance.notexist.fullFormat</formatOnVm><fullFormat>com.vmware.vxlan.instance.notexist|event.com.vmware.vxlan.instance.notexist.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vxlan.mcastgroup.joinfail</description><category>event.com.vmware.vxlan.mcastgroup.joinfail.category</category><formatOnDatacenter>event.com.vmware.vxlan.mcastgroup.joinfail.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vxlan.mcastgroup.joinfail.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vxlan.mcastgroup.joinfail.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vxlan.mcastgroup.joinfail.fullFormat</formatOnVm><fullFormat>com.vmware.vxlan.mcastgroup.joinfail|event.com.vmware.vxlan.mcastgroup.joinfail.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vxlan.port.initfail</description><category>event.com.vmware.vxlan.port.initfail.category</category><formatOnDatacenter>event.com.vmware.vxlan.port.initfail.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vxlan.port.initfail.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vxlan.port.initfail.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vxlan.port.initfail.fullFormat</formatOnVm><fullFormat>com.vmware.vxlan.port.initfail|event.com.vmware.vxlan.port.initfail.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vxlan.vmknic.initfail</description><category>event.com.vmware.vxlan.vmknic.initfail.category</category><formatOnDatacenter>event.com.vmware.vxlan.vmknic.initfail.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vxlan.vmknic.initfail.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vxlan.vmknic.initfail.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vxlan.vmknic.initfail.fullFormat</formatOnVm><fullFormat>com.vmware.vxlan.vmknic.initfail|event.com.vmware.vxlan.vmknic.initfail.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vxlan.vmknic.ipchange</description><category>event.com.vmware.vxlan.vmknic.ipchange.category</category><formatOnDatacenter>event.com.vmware.vxlan.vmknic.ipchange.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vxlan.vmknic.ipchange.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vxlan.vmknic.ipchange.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vxlan.vmknic.ipchange.fullFormat</formatOnVm><fullFormat>com.vmware.vxlan.vmknic.ipchange|event.com.vmware.vxlan.vmknic.ipchange.fullFormat</fullFormat><longDescription></longDescription></eventInfo><eventInfo><key>EventEx</key><description>com.vmware.vxlan.vmknic.noip</description><category>event.com.vmware.vxlan.vmknic.noip.category</category><formatOnDatacenter>event.com.vmware.vxlan.vmknic.noip.fullFormat</formatOnDatacenter><formatOnComputeResource>event.com.vmware.vxlan.vmknic.noip.fullFormat</formatOnComputeResource><formatOnHost>event.com.vmware.vxlan.vmknic.noip.fullFormat</formatOnHost><formatOnVm>event.com.vmware.vxlan.vmknic.noip.fullFormat</formatOnVm><fullFormat>com.vmware.vxlan.vmknic.noip|event.com.vmware.vxlan.vmknic.noip.fullFormat</fullFormat><longDescription></longDescription></eventInfo><enumeratedTypes><key>VirtualMachineFaultToleranceState</key><tags><label>Not\ | |
| \ configured</label><summary>Fault Tolerance has not been configured for this\ | |
| \ virtual machine</summary><key>notConfigured</key></tags><tags><label>Disabled</label><summary>Fault\ | |
| \ Tolerance is disabled</summary><key>disabled</key></tags><tags><label>Enabled</label><summary>Fault\ | |
| \ Tolerance is enabled</summary><key>enabled</key></tags><tags><label>Need\ | |
| \ Secondary VM</label><summary>Fault Tolerant Secondary VM is not running</summary><key>needSecondary</key></tags><tags><label>Starting</label><summary>Fault\ | |
| \ Tolerance is starting</summary><key>starting</key></tags><tags><label>Running</label><summary>Fault\ | |
| \ Tolerance is running</summary><key>running</key></tags></enumeratedTypes></val></propSet></objects></returnval></RetrievePropertiesExResponse>\n\ | |
| </soapenv:Body>\n</soapenv:Envelope>"} | |
| headers: | |
| cache-control: [no-cache] | |
| connection: [Keep-Alive] | |
| content-type: [text/xml; charset=utf-8] | |
| date: ['Thu, 20 Nov 2014 23:40:46 GMT'] | |
| transfer-encoding: [chunked] | |
| status: {code: 200, message: OK} | |
| version: 1 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment