The following table describes important Backup and DR Service event IDs, event messages, and steps to resolve them:
| Event ID | Event message | What to do | 
|---|---|---|
| 5022 | Actifio Connector: Failed in preparing VSS snapshot set | This issue occurs if Windows fails to create a VSS snapshot. To resolve | 
| this issue, do the following: | ||
| * Check UDSAgent.log | ||
| * Check disk space on protected volumes. 300MB may not be enough. | ||
| * Check Windows Event Logs for VSS related errors. | ||
| * vssadmin list writers may show writers in a bad state. | ||
| Usually these errors are accompanied by VSS errors reported in the logs | ||
| such as: VSS_E_VOLUME_NOT_SUPPORTED_BY_PROVIDER | ||
| VSS_E_UNEXPECTED_PROVIDER_ERROR | ||
| First check if all the VSS writers are in a stable state by going to the | ||
| command line and issuing this command. # vssadmin list writers | ||
| Check output to confirm that all the writers are in a stable state. | ||
| Restart VSS service and check if the writers are stable. If not you may | ||
| have to reboot the machine. | ||
| 5024 | Actifio Connector: Failed to create VSS snapshot for backup. Insufficient | This issue occurs if there is insufficient disk space to process a | 
| storage available to create either the shadow copy storage file or other | snapshot. | |
| shadow copy data | ||
| 1. Ensure the drive being backed up is not full. | ||
| 2. Check if all the VSS writers are in a stable state From the Windows | ||
| command line, run: vssadmin list providersvssadmin list writers | ||
| 3. If these services are not running, start them and re-run the job. If | ||
| the writer's State is Not Stable, restart the VSS service. If the | ||
| problem continues after restarting the service, reboot the host. | ||
| Sometimes the message appears when internal VSS errors occur. | ||
| Check the Windows Event Logs for VSS related errors. For errors related | ||
| to VSS, search for related Microsoft patches. Additional VSS | ||
| troubleshooting details can be found on Microsoft TechNet. | ||
| Microsoft recommends at least 320MB on devices specified for saving the | ||
| created VSS snapshot, plus change data that is stored there. | ||
| Actifio recommends the shadow storage space be set to unbounded | ||
| (unlimited) using these commands: vssadmin list shadowstorage | ||
| vssadmin Resize ShadowStorage /On=[drive]: /For=[drive]: / Maxsize=[size] | ||
| To change the storage area size in the Windows UI, refer to | ||
| Configuring Volume Shadow Copy on Windows Server 2008. | ||
| Re-run the backup once the VSS state is stable and shadow storage is set | ||
| to unbounded. | ||
| 5046 | Backup staging LUN is not visible to the Actifio Connector | This issue occurs if the staging LUN is not visible to the UDSAgent on | 
| the application's host and the host is unable to detect the staging LUN | ||
| from the backup/recovery appliance. | ||
| 5049 | Actifio Connector: Failed identifying logical volume on the backup staging | Actifio Connector couldn't see the staging LUN. This can be caused by a | 
| lun | bad connection or by trouble on the LUN. | |
| Verify that FC/iSCSI connectivity is good, then make sure it works by | ||
| mapping the VDisk, partitioning it, formatting it, and copying files to | ||
| it. The steps for partitioning and formatting are OS specific. | ||
| 5078 | Actifio Connector: The staging disk is full | Jobs fail if a file that was modified in the source disk is copied to | 
| the staging disk, but the file is larger than the free space available | ||
| in the staging disk. To fix the issue with full staging disk, increase | ||
| the staging disk. Specify the size of the staging disk in the advanced | ||
| settings for the application. Set the value for staging disk size such | ||
| that it is greater than the sum of size of the source disk and the size | ||
| of the largest file. Note: Changing the staging disk in advanced | ||
| settings provokes a full backup. | ||
| 5087 | Actifio Connector: Failed to write files during a backup (Source File) | Anti-virus programs or third party drivers may have applied file locks | 
| that cannot be overridden. | ||
| Check the UDSAgent.log to see which file couldn't be accessed. Attempt | ||
| to find which process is locking the file using lsof on Unix/ Linux, or | ||
| fltmc on Windows. Exclude the file from the antivirus or capture job | ||
| and re-try the capture. | ||
| The current processes known to Microsoft are listed at: | ||
| Allocated filter altitudes. | ||
| These errors are rarely found on Unix or Linux, but it is possible that | ||
| a process such as database maintenance or patch install / update has | ||
| created an exclusive lock on a file. | ||
| Install the latest Actifio Connector. | ||
| A file system limitation or inconsistency was detected by the host | ||
| operating system. | ||
| Run the Windows Disk defragmenter on the staging disk. | ||
| Low I/O throughput from the hosts disks or transport medium, iSCSI or | ||
| FC. | ||
| Ensure there are no I/O issues in the host's disks or transport medium. | ||
| The transport medium will either be iSCSI or Fibre Channel depending on | ||
| out of band configuration. Consult storage and network administrators as | ||
| needed. | ||
| 5131 - SQL Logs report error 3041 | SQL log backups on instance fail with error 5131 | To resolve this, enable "Don't forcefully unload the user registry at | 
| user logoff", see User Profile Service functionality. | ||
| 5131 - SQL logs show backup/ recovery appliances error 43901 | Snapshot jobs fail with error 5131, SQL logs show backup/ recovery | This is because the ODBC login for the database is failing. Fixing the | 
| "Failed snapshot Job" | ODBC login resolves the issue. | |
| appliances error 43901 | ||
| 5136 | Actifio Connector: The staging volume is not readable | Check /act/logs/UDSAgent.log for details and contact Google support for | 
| the resolution for the issue. | ||
| 5241 | Actifio Connector: Failed to mount/clone applications from mapped image | Invalid username and password being parsed from the control file. | 
| (Source File) | On the source, review the UDSAgent.log to see if the source is | |
| configured with the correct username/password under Advanced Settings in | ||
| the connector properties. | ||
| 5547 | Oracle: Failed to backup archivelog (Source File) | Actifio Connector failed to backup the archive log using RMAN archive | 
| backup commands. The likely causes for this failure are: | ||
| * Connector failed to establish connection to database | ||
| * The archive logs were purged by another application | ||
| * TNS Service name is configured incorrectly, causing backup command | ||
| to be sent to a node where the staging disk isn't mounted | ||
| Search for ORA- or RMAN- errors in the RMAN log. This is the error | ||
| received from Oracle. Use the preferred Oracle resource as these are not | ||
| Backup and DR Service conditions, and hence cannot be resolved within | ||
| Backup and DR Service. | ||
| * Actifio Connector logs: /var/act/log/UDSAgent.log | ||
| * Oracle RMAN logs: /var/act/log/********_rman.log | ||
| 10032 | Snapshot pool exceeded warning level | To reduce consumption of the snapshot pool, do the following: | 
| * Move VMware VMs from a snapshot to a Direct-to-OnVault backup plan. | ||
| Then expire all snapshots to release the space used by the staging | ||
| disks and last snap. This only works for VMware VMs; other | ||
| application types still use some snapshot pool space if protected by | ||
| a Direct-to-OnVault policy. | ||
| * Reduce the number of snaps kept for an application by changing the | ||
| policy template. Applications that have high change rates create | ||
| larger snapshots, so this has the highest benefit for high | ||
| change-rate applications. This does not necessarily lead to a | ||
| different RPO, an OnVault images of each snap can be created before | ||
| they are expired. | ||
| * Delete mounts, clones, and live-clones if they are not needed. | ||
| 10038 | About to exceed VDisk warning limit | To immediately reduce VDisk consumption, do the following: | 
| * Ensure expirations are enabled, both at the global and individual | ||
| application level. | ||
| * Group databases from a single host together into a Consistency Group. | ||
| For example, if a host has 9 databases, create one Consistency Group | ||
| for all 9 databases, then protect that consistency group rather than | ||
| the individual databases. | ||
| * Reduce the number of snapshots kept for an application by changing | ||
| the policy template used by a backup plan. | ||
| * Delete unwanted mounts, clones, and live-clones images. | ||
| * Move VMware VMs from a snapshot to a Direct- OnVault backup plan. | ||
| You need to expire all snapshots to release the VDisks used by the | ||
| staging disks. This only lowers the VDisk count for VMware VMs, and | ||
| still uses VDisks when protected by a direct-to-OnVault policy. | ||
| * Change VMware VMDKs that don't need to be protected to independent | ||
| mode as these cannot be protected by VMware snapshots. | ||
| If this alert repeats daily but the appliance does not reach the maximum | ||
| VDisks, then modify the policies to reduce the number of VDisks used, or | ||
| increase the alert threshold. During a daily snapshot window the VDisk | ||
| count can fluctuate while new VDisks are created for snapshots before | ||
| the old VDisks are removed as a part of snapshot expirations. The daily | ||
| fluctuations varies depending on the number of applications protected. | ||
| 10039 | Network error reaching storage device | A heartbeat ping to monitored storage has failed due to hardware | 
| failure or network issue. Check the network to resolve the issue. | ||
| 10043 | A backup plan violation has been detected | Review the backup plan best practices | 
| and optimize your policies. The are common causes for backup plan | ||
| violations. | ||
| * Job scheduler is not enabled. See to enable the scheduler. | ||
| * The first jobs for new applications can often take a long time: | ||
| Long job times can occur during the first snapshot or dedup job | ||
| for an application. On-ramp settings can be used to prevent | ||
| ingest jobs from locking up slots and locking out ingested | ||
| applications. See [Set priorities for the first new | ||
| applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | ||
| * Applications are inaccessible due to network issues. | ||
| * Policy windows are too small or job run times are too long: While | ||
| you cannot control how long each job takes to run, you can control | ||
| the schedule time for applications that are running. Jobs that run | ||
| for many hours occupy job slots that could be used by other | ||
| applications. Review the [backup plan best | ||
| practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | ||
| and adjust policies accordingly. | ||
| * Replication process sends the data to a remote | ||
| backup/recovery appliance. Ensure that the bandwidth & utilization of | ||
| your replication link is not saturated. | ||
| 10046 | Performance Pool exceeded safe threshold | To reduce consumption of the snapshot pool, do the following: | 
| * Move VMware VMs from a snapshot to a Direct-to-OnVault backup plan. | ||
| Then expire all snapshots to release the space used by the staging | ||
| disks and last snap. This only works for VMware VMs; other | ||
| application types still use some snapshot pool space if protected by | ||
| a Direct-to-OnVault policy. | ||
| * Reduce the number of snaps kept for an application by changing the | ||
| policy template. Applications that have high change rates create | ||
| larger snapshots, so this has the highest benefit for high | ||
| change-rate applications. This does not necessarily lead to a | ||
| different RPO, an OnVault images of each snap can be created before | ||
| they are expired. | ||
| * Delete mounts, clones, and live-clones if they are not needed. | ||
| 10055 | Unable to check remote protection | Each backup/recovery appliance checks the remote appliance hourly for | 
| possible remote protection issues. The appliance communication fails due | ||
| to the following issues: | ||
| * Network error (temporary or permanent). Temporary network error does | ||
| not mean job to fail; jobs are retried, but the hourly check is not | ||
| updated. | ||
| * Certificate error. To fix the certificate error, you need to | ||
| re-exchange the certificate. | ||
| 10070 | Udppm scheduler is off for more than 30 minutes. | The scheduler is off. This may have been set for maintenance. If the | 
| maintenance is complete, you can re-enable the scheduler, see to | ||
| enable the scheduler. | ||
| 10084 | Alert for application (app name) and policy (policyname) job did not run | Review the backup plan best practices | 
| because of unknown reason | and optimize your policies. The are common causes for backup plan | |
| violations. | ||
| * Job scheduler is not enabled. See to enable the scheduler. | ||
| * The first jobs for new applications can often take a long time: | ||
| Long job times can occur during the first snapshot or dedup job | ||
| for an application. On-ramp settings can be used to prevent | ||
| ingest jobs from locking up slots and locking out ingested | ||
| applications. See [Set priorities for the first new | ||
| applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | ||
| * Applications are inaccessible due to network issues. | ||
| * Policy windows are too small or job run times are too long: While | ||
| you cannot control how long each job takes to run, you can control | ||
| the schedule time for applications that are running. Jobs that run | ||
| for many hours occupy job slots that could be used by other | ||
| applications. Review the [backup plan best | ||
| practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | ||
| and adjust policies accordingly. | ||
| * Replication process sends the data to a remote | ||
| backup/recovery appliance. Ensure that the bandwidth & utilization of | ||
| your replication link is not saturated. | ||
| 10085 | Backup Plan violation for application (app name) on host (host name) and | Review the backup plan best practices | 
| policy (policy name). Job did not run because of unknown reason. | and optimize your policies. The are common causes for backup plan | |
| violations. | ||
| * Job scheduler is not enabled. See to enable the scheduler. | ||
| * The first jobs for new applications can often take a long time: | ||
| Long job times can occur during the first snapshot or dedup job | ||
| for an application. On-ramp settings can be used to prevent | ||
| ingest jobs from locking up slots and locking out ingested | ||
| applications. See [Set priorities for the first new | ||
| applications](/backup-disaster-recovery/docs/concepts/appliance-control-panel#set_priorities_for_the_first_new_applications). | ||
| * Applications are inaccessible due to network issues. | ||
| * Policy windows are too small or job run times are too long: While | ||
| you cannot control how long each job takes to run, you can control | ||
| the schedule time for applications that are running. Jobs that run | ||
| for many hours occupy job slots that could be used by other | ||
| applications. Review the [backup plan best | ||
| practices](/backup-disaster-recovery/docs/concepts/backup-plan-policy-best-practices) | ||
| and adjust policies accordingly. | ||
| * Replication process sends the data to a remote | ||
| backup/recovery appliance. Ensure that the bandwidth & utilization of | ||
| your replication link is not saturated. | ||
| 10120 | Psrv started successfully | This is an internal event and can be ignored. | 
| 10220 | NTP Service is not running or not synchronised. | The NTP Service on the backup appliance is not running. The NTP Service | 
| is needed to ensure the backup appliance uses the correct timestamps. A | ||
| Compute Engine appliance should be using metadata.google.internal. | ||
| Follow how to set the NTP server [DNS and NTP | ||
| method](/backup-disaster-recovery/docs/concepts/appliance-network-settings#dns-ntp). | ||
| 10225 | Udp corefiles are found, filename udpengine.(file name) | Internal processes are unexpectedly logging error files. Contact Google | 
| support to get the resolution for this issue. | ||
| 10229 | Exceeded storage, System name: (device name) | This is is an internal event and normally can be ignored. | 
| 10237 | The X job has been running for more than 3 hours. | There are many reasons why a job can run for more than 3 hours. | 
| 11001 | Backup appliance certificate expires in X days. Power on for 24 hours to | The backup/recovery appliance's certificate last update is more than 15 | 
| renew, or contact support. | days old. If the backup/recovery appliance is down, bring it up. | |
| 11004 | System components are down. Please contact support in case backups are | Contact support. | 
| affected. | ||
| 11006 | Not able to sync with host X, regular sync with host required to avoid | The certificate on the host hasn't been updated for more than 7 days. | 
| permanent communication loss between Backup appliance and Host. | Try rebooting and reconnecting to the host. | |
| 20019 | Insufficient CPU / Memory. Minimum number of core required: (cores) Actual | Backup/recovery appliance has been changed and is not the recommended | 
| cores : (cores). Minimum memory size required (GB): (memory) Actual memory : | size. Contact Google support to get the resolution for this issue. | |
| (memory) | ||
| 20025 | Swap usage exceeded | This issue occurs when the swap usage is exceeding the configured | 
| threshold limit that is set for the backup/recoevery appliance. Contact | ||
| Google support to get the resolution for this issue. | ||
| 20030 | tomcat stopped successfully | This is is an internal event and can be ignored. | 
| 20031 | tomcat started successfully | This is is an internal event and can be ignored. | 
| 22001 | OMD started Successfully, sltname: , slpname: . | This is is an internal event and can be ignored. | 
| 42356 | File changes have been detected no deleted files have been detected new | This is is an internal event and can be ignored. | 
| files have been detected. | ||
| 43151 | couldn't add raw device mappings to virtual machine (VM). Error: VM task | Adding a raw device mapping to a VM "stuns" the VM until ESX has had a | 
| failed A general system error occurred: The system returned an error. | chance to add the new resource. To find out why the raw device mapping | |
| couldn't be added, look at the ESX logs for the VM in question | ||
| (vmware.log). | ||
| Refer to the VMware documentation and knowledge base for assistance on | ||
| reviewing the logs for error messages. Also, review the VMware article | ||
| for more information on [collecting VMware | ||
| logs](https://kb.vmware.com/s/article/653). | ||
| 43155 | Error: VM task failed. An error occurred while saving the snapshot: Failed | This is a VMware issue; for additional information, refer to VMware KB | 
| to quiesce the virtual machine. | article - 1015180. | |
| Virtual machine quiesce issues are dependent on the OS type. Additional | ||
| investigation, further VMware KBA searches or contact VMware support to | ||
| resolve this issue. | ||
| 43155 - a | Error: VM task failed. Device scsi3 couldn't be hot-added. | This usually means that the SCSI device you are trying to add to the VM | 
| is already in use by another VM. | ||
| 43155 - b | Error: VM task failed. The virtual disk is either corrupted or not a | This issue occurs if the VM's CTK files are locked, unreadable, or are | 
| supported format. | being committed. To fix this issue, remove and re-create these CTK | |
| files. Refer to the KB article - 2013520 | ||
| for more information. | ||
| 43155 - c | Error: VM task failed. The operation is not allowed in the current state | There are two options for formatting a VMware datastore: NFS and VMFS. | 
| of the datastore." progress ="11" status="running" | With NFS, there are some limitations like not being able to do RDM (Raw | |
| Disk Mapping). This means that you cannot mount from the | ||
| backup/recovery appliance to an NFS datastore. Refer to the following KB | ||
| article - 1001856 for | ||
| additional information. | ||
| 43175 | UDSAgent socket connection got terminated abnormally; while waiting for | The Actifio Connector stops responding between the appliance and a host | 
| the response from agent | with Backup and DR agent is installed. | |
| 1. Restart the UDSAgent Backup and DR agent service on the | ||
| specified host. | ||
| 2. Telnet to tcp port 5106 (UDSAgent communication port) | ||
| ``` | ||
| <P></P> | ||
| Expected output: | ||
| <P></P> | ||
| Trying 10.50.100.67... | ||
| <P></P> | ||
| Connected to dresx2.accu.local. | ||
| <P></P> | ||
| Escape character is '^]'. | ||
| <P></P> | ||
| Connection closed by foreign host. | ||
| ``` | ||
| 3. Verify network connectivity between appliance and host doesn't drop. | ||
| If the problem persists, network analysis will be required. | ||
| 43604 | Failed to verify fingerprint | This occurs when an inconsistency is found between the source and target | 
| data. Contact Google support to get the resolution for this issue. | ||
| 43690 | Host doesn't have any SAN or iSCSI ports defined. | This issue occurs if the backup/recovery appliance is not configured with | 
| iSCSI connection to the target host. | ||
| Ensure that the network ports are open for iSCSI and the target host has | ||
| discovered the backup/recovery appliances. | ||
| 43698 | ESX host is not accessible for NBD mode data movement | The backup/recovery appliance is unable to reach the ESX host over the | 
| network or resolve the ESX hostname using DNS. Contact Google support to | ||
| get the resolution for this issue. | ||
| 43702 | Backup was aborted because there are too many extra files in the home | This is an alert condition generated by Backup and DR Service and is caused | 
| directory of the VM | by leftover delta files in the VM's datastore. Normally, the delta files | |
| are removed after Backup and DR snapshot is consolidated. In | ||
| some instances, these can be left behind by the VMware consolidation, | ||
| and Backup and DR begins failing jobs to prevent exacerbating | ||
| the issue. | ||
| This issue is caused by VMware, refer to the knowledge base article - | ||
| 1002310. | ||
| 43755 | Failed to open VMDK volume; check connectivity to ESX server. | This happens when the ESX server cannot be reached by the controller, | 
| usually because of a physical connection or DNS problem. To fix this | ||
| issue, do the following: | ||
| * Ensure port 902 is open between the backup/recovery appliance and the | ||
| ESX host. | ||
| * Check the current DNS server and ensure it is current and valid. | ||
| * If the vCenter is virtualized, attempt a backup after migrating the | ||
| vCenter to a different ESX host. | ||
| * Ensure SSL required is set to True on the ESX host in the advanced | ||
| settings. | ||
| 43844 | Invalid size vmdk detected for the VM | There are two possible solutions for this situation: | 
| * If consolidation is required for some disks on VM, size is reported | ||
| as zero. To fix this issue, creating and deleting a snapshot of the | ||
| VM. | ||
| * See if the VMDK can be restored from a backup image. | ||
| 43873 | Disk space usage on datastore has grown beyond the critical threshold | This issue occurs when the remaining space on the datastore is less than | 
| the critical threshold. If more storage is not made available soon, then | ||
| jobs start to fail when the remaining space is inadequate to store them. | ||
| This alert is created to help you take action to prevent ESX datastores | ||
| from filling with snapshot data. Increase available space by expanding | ||
| the datastore, migrating some VMs, or deleting old data on the | ||
| datastore. | ||
| Snapshots grow as more change data is added. If a datastore fills up due | ||
| to a growing snapshot, VMs may be taken offline automatically by VMware | ||
| to protect the data. | ||
| 43900 | Retry pending OnVault (log) (jobname for application (appname) on host | Job retries can be caused by many errors. Each 43900 event message | 
| (hostname) Error: (errorID) (Error Description) | includes an error code and an error message. | |
| 43901 | Job failure | Job failures can be caused by many errors. Each 43901 event message | 
| includes an error code and an error message. | ||
| 43903 | Failed expire job | This issue occurs when the image is in use at the time of the | 
| expiration. This can be due to the image is in use by another process or | ||
| operation, such as a mount, clone, restore. The expiration job most | ||
| likely complete successfully on the second attempt. | ||
| Backup and DR does not report the successful completion of | ||
| this second attempt. If you get only one error for an image, it is safe | ||
| to conclude that a second attempt to expire this image was successful. | ||
| If there is a legitimate reason why this image cannot be expired, you | ||
| will get multiple errors related to this image. If you receive more than | ||
| one error, contact Google Support. | ||
| 43905 | Failed mount job | There are many ways a mount job can fail. The error code that | 
| accompanies the event helps to identify the root cause. | ||
| 43908 | Failed restore job | Job failures can be caused by many errors. Each 43908 event message | 
| includes an error code and an error message. | ||
| 43915 | Couldn't connect to backup host. Make sure Backup and DR agent | To initiate backup, the Actifio Connector service must be reachable by | 
| is running on (host) and network port (port) is open | the backup/recovery appliance. This issue occurs, when the required ports | |
| are not open, the incorrect host IP is configured, the | ||
| Backup and DR agent service not running, or the host is out | ||
| of physical resources. To fix this issue, do the following: | ||
| 1. Ensure that the port in use between the host, | ||
| backup/recovery appliance, and Actifio Connector is open. By default, | ||
| the Backup and DR agent uses port 5106 for bi-directional | ||
| communication from the backup/recovery appliance. Make sure your | ||
| firewall permits bi-directional communication through this port. | ||
| 2. Ensure that the correct IP is configured for the host **Manage > | ||
| Appliance > Configure Appliance Networking**. | ||
| 3. Ensure that the Backup and DR agent service is running on | ||
| the target host and restart, if necessary. | ||
| * On Windows, find the UDS Host Agent service in services.msc and | ||
| click Restart. | ||
| *   On Linux, run the command /etc/init.d/udsagent restart | ||
| ``` | ||
| <timestamp> GEN-DEBUG [4400] UDSAgent starting up ... | ||
| <timestamp> GEN-INFO [4400] Locale is initialized to C | ||
| <timestamp> GEN-WARN [4400] VdsServiceObject::initialize | ||
| - LoadService for Vds failed with error 0x80080005 | ||
| <timestamp> GEN-WARN [4400] initialize - Failed to initialize Microsoft Disk | ||
| Management Services: Server execution failed [0x80080005] | ||
| <timestamp> GEN-WARN [4400] Failed initializing VDSMgr, | ||
| err = -1, exiting... | ||
| <timestamp> GEN-INFO [4400] Couldn't connect to namespace: root\mscluster | ||
| <timestamp> GEN-INFO [4400] This host is not part of cluster | ||
| <timestamp> GEN-WARN [4400] Failed initializing connectors,exiting -1 | ||
| ``` | ||
| 4. Retry the backup. | ||
| 43941 | Disk space usage on datastore has grown beyond the critical threshold | This issue occurs when the remaining space on the datastore is less than | 
| the critical threshold. If more storage is not made available soon, then | ||
| jobs start to fail when the remaining space is inadequate to store them. | ||
| This alert is created to help you take action to prevent ESX datastores | ||
| from filling with snapshot data. Increase available space by expanding | ||
| the datastore, migrating some VMs, or deleting old data on the | ||
| datastore. | ||
| Snapshots grow as more change data is added. If a datastore fills up due | ||
| to a growing snapshot, VMs may be taken offline automatically by VMware | ||
| to protect the data. | ||
| 43954 | Failed OnVault job | During a mount job, the backup/recovery appliance is unable to connect to | 
| the OnVault pool. This issue can be occurred due to any of the following | ||
| reasons. | ||
| * No bucket name is specified for the OnVault pool. | ||
| * Invalid credentials-access ID or access key not specified or wrong | ||
| ID for the OnVault pool | ||
| * Invalid bucket in the OnVault pool | ||
| * General authentication issues for the OnVault pool. | ||
| *   DNS server in clusters /etc/resolv.confis either different or the | ||
| forward and reverse DNS zones files are changed. | ||
| 43929 | Snapshot creation of VM failed. Error: VM task failed An error occurred | VM snapshot fails if the ESX server is unable to quiesce the virtual | 
| while saving the snapshot: Failed to quiesce the virtual machine. | machine - either because of too much I/O, or because VMware tools cannot | |
| quiesce the application using VSS in time. Check the event logs on the | ||
| host and check the VM's ESX log (vmware.log). | ||
| Crash-consistent snapshots and connector-based backups show this | ||
| behavior less often. For more information, refer to the VMware knowledge | ||
| base articles - 1018194 and | ||
| 1007696. | ||
| 43933 | Failed to find VM with matching BIOS UUID | This issue occurs if the VM's UUID is modified. To fix this issue, | 
| rediscover the VM and check if it was discovered as a new UUID. You can | ||
| confirm this in the management console by comparing the UUID of the | ||
| newly discovered VM and that of the previously discovered VM. If the | ||
| UUIDs don't match, the VM might have been cloned. | ||
| You can also see this error, if a large number of | ||
| Backup and DR managed VMs are removed from the vCenter. | ||
| 43948 | The number of images not expired awaiting further processing is (x) | "Event ID 43948 is generated when an application begins halting | 
| images ((x) snapshots, (x) onvaults) from (x) unique | expirations as a part of Image Preservation. 'Image Preservation' | |
| applications. (x) snapshots and (x) OnVaults were added in the | preserves snapshot and OnVault images beyond their expiration dates to | |
| last (x) seconds ((x) hours (x) minutes)., sltname: No | ensure that those images are properly processed by the | |
| specific slt, slpname: No specific slp. | backup/recovery appliance. When a new application enters into a preserved | |
| mode, a Warning alert will be generated. The most common cause of this | ||
| is backup plan violations as documented under event ID 10085". | ||
| 43954 | Retry OnVault | An OnVault job needed to be retried. Possible issues could include: The | 
| Service Account being used has the wrong role. The Service Account does | ||
| not have authority to write to the bucket. The Cloud Storage bucket | ||
| no longer exists. | ||
| 43960 | Skipped backing up 6 offline applications for SqlServerWriter application. | Backup of a SQL Server Instance found some databases were offline and | 
| couldn't be backed up. This commonly occurs when the database has been | ||
| deleted on the server side, but is still included on the Backup/DR side. | ||
| The error message contains the names of the offline databases that | ||
| should be investigated. | ||
| 43972 | Metadata upload to bucket failed. | Metadata writes to an OnVault bucket failed. Possible issues could | 
| include: The Service Account being used has the wrong role. The Service | ||
| Account does not have authority to write to the bucket. The | ||
| Cloud Storage bucket no longer exists. | ||
| 43973 | udppm started Successfully | This is is an internal event and can be ignored. | 
| 43999 | Warning: VM is running on a host that is running an outdated version of | Upgrade the VM to a supported version (>=) to ensure the best results. | 
| ESXi , which is not supported by Google. Please upgrade it to a | ||
| supported version (>=) to ensure the best results. | ||
| 44003 | Succeeded Job_xx-xx-xxfor applicationapplication IDon host | This is is a successful status event and can be ignored. | 
| host, sltname:template, slpname:profile. | ||
| 62001 | Streamsnapd daemon started successfully | This is is an internal event and can be ignored. | 
| 90003 | A new update (version X) is available for Backup Appliance | A new update is available. Update your backup/recovery appliances as soon | 
| as possible. | 
What's next
- To view the event logs created on backup/recovery appliances, see Backup and DR event logs
- To configure log-based alerts for Backup and DR Service, based on sample queries, see create a log-based alert.