Asynchronous parallel SSH client library.
Run SSH commands over many - hundreds/hundreds of thousands - number of servers asynchronously and with minimal system load on the client host.
Native code based clients with extremely high performance, making use of C libraries.
pip install parallel-ssh
An update to pip may be needed to be able to install binary wheels.
pip install -U pip pip install parallel-ssh
See documentation on read the docs for more complete examples.
uname on two hosts in parallel.
from pssh.clients import ParallelSSHClient hosts = ['localhost', 'localhost'] client = ParallelSSHClient(hosts) output = client.run_command('uname') for host_output in output: for line in host_output.stdout: print(line) exit_code = host_output.exit_code
Single host client with similar API can be used if parallel functionality is not needed.
from pssh.clients import SSHClient host = 'localhost' cmd = 'uname' client = SSHClient(host) host_out = client.run_command(cmd) for line in host_out.stdout: print(line) exit_code = host_out.exit_code
Github discussions can be used to discuss, ask questions and share ideas regarding the use of parallel-ssh.
The default client in
parallel-ssh is a native client based on
libssh2 C library - which offers much greater performance and reduced overhead compared to other Python SSH libraries.
See this post for a performance comparison of different Python SSH libraries.
Alternative clients based on
libssh) are also available under
pssh.clients.ssh. See client documentation for a feature comparison of the available clients in the library.
parallel-ssh makes use of clients and an event loop solely based on C libraries providing native code levels of performance and stability with an easy to use Python API.
Because other options are either immature, unstable, lacking in performance or all of the aforementioned.
Certain other self-proclaimed leading Python SSH libraries leave a lot to be desired from a performance and stability point of view, as well as suffering from a lack of maintenance with hundreds of open issues, unresolved pull requests and inherent design flaws.
The SSH libraries
parallel-ssh uses are, on the other hand, long standing mature C libraries in libssh2 and libssh that have been in production use for decades and are part of some of the most widely distributed software available today - Git itself, OpenSSH, Curl and many others.
These low level libraries are far better placed to provide the maturity, stability and performance needed from an SSH client for production use.
parallel-ssh provides easy to use SSH clients that hide the complexity, while offering stability and native code levels of performance and as well as the ability to scale to hundreds or more concurrent hosts.
join function can be used to wait for all commands in output to finish.
join returns, commands have finished and all output can be read without blocking.
Once either standard output is iterated on to completion, or
client.join() is called, exit codes become available in host output.
Iteration ends only when remote command has completed, though it may be interrupted and resumed at any point - see join and output timeouts documentation.
HostOutput.exit_code is a dynamic property and will return
None when exit code is not ready, meaning command has not finished, or unavailable due to error.
Once all output has been gathered exit codes become available even without calling
join as per previous examples.
output = client.run_command('uname') client.join() for host_out in output: for line in host_out.stdout: print(line) print(host_out.exit_code)
Linux 0 Linux 0
Similarly, exit codes are available after
client.join() without reading output.
output = client.run_command('uname') client.join() for host_output in output: print(host_out.exit_code)
There is also a built in host logger that can be enabled to log output from remote hosts for both stdout and stderr. The helper function
pssh.utils.enable_host_logger will enable host logging to stdout.
To log output without having to iterate over output generators, the
consume_output flag must be enabled - for example:
from pssh.utils import enable_host_logger enable_host_logger() client.run_command('uname') client.join(consume_output=True)
SCP is supported - native client only - and provides the best performance for file copying.
Unlike with the SFTP functionality, remote files that already exist are not overwritten and an exception is raised instead.
Note that enabling recursion with SCP requires server SFTP support for creating remote directories.
To copy a local file to remote hosts in parallel with SCP:
from pssh.clients import ParallelSSHClient from gevent import joinall hosts = ['myhost1', 'myhost2'] client = ParallelSSHClient(hosts) cmds = client.scp_send('../test', 'test_dir/test') joinall(cmds, raise_error=True)
See SFTP and SCP documentation for more examples.
SFTP is supported in the native client.
To copy a local file to remote hosts in parallel:
from pssh.clients import ParallelSSHClient from pssh.utils import enable_logger, logger from gevent import joinall enable_logger(logger) hosts = ['myhost1', 'myhost2'] client = ParallelSSHClient(hosts) cmds = client.copy_file('../test', 'test_dir/test') joinall(cmds, raise_error=True)
Copied local file ../test to remote destination myhost1:test_dir/test Copied local file ../test to remote destination myhost2:test_dir/test
There is similar capability to copy remote files to local ones with configurable file names via the copy_remote_file function.
In addition, per-host configurable file name functionality is provided for both SFTP and SCP - see documentation.
Directory recursion is supported in both cases via the
recurse parameter - defaults to off.
See SFTP and SCP documentation for more examples.