I am encountering a random issue with the SSHOperator in Apache Airflow. The task fails occasionally with the error:
airflow.exceptions.AirflowException: SSH operator error: exit status = 1
Here is the complete error message:
[2024-11-06, 00:05:06 IST] {taskinstance.py:2890} ERROR - Task failed with exception
Traceback (most recent call last):
File "/home/ubuntu/airflow/venv/lib/python3.9/site-packages/airflow/providers/ssh/operators/ssh.py", line 191, in execute
result = self.run_ssh_client_command(ssh_client, selfmand, context=context)
File "/home/ubuntu/airflow/venv/lib/python3.9/site-packages/airflow/providers/ssh/operators/ssh.py", line 179, in run_ssh_client_command
self.raise_for_status(exit_status, agg_stderr, context=context)
File "/home/ubuntu/airflow/venv/lib/python3.9/site-packages/airflow/providers/ssh/operators/ssh.py", line 173, in raise_for_status
raise AirflowException(f"SSH operator error: exit status = {exit_status}")
airflow.exceptions.AirflowException: SSH operator error: exit status = 1
[2024-11-06, 00:05:06 IST] {standard_task_runner.py:110} ERROR - Failed to execute job 2097570 for task CleanUpDaily_sftp_files (SSH operator error: exit status = 1; 2118240)
I suspect the issue could be related to the load on the EC2 instance since multiple jobs were running during the failure. However, other jobs on the same instance were executing successfully at that time.
Any insights into debugging and resolving this issue would be greatly appreciated!
Verified that the remote server is reachable during failures. Checked server resource utilization, which seems normal. Verified that network connectivity to the server is stable.