diff --git a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/cle/DistributedPyNNCommunicationAdapter.py b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/cle/DistributedPyNNCommunicationAdapter.py index f4c790b6cd2f80647c287ed83b5aac9241fb8e53..48719ecc87cbd24c10af6067b75d75dff73eae82 100644 --- a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/cle/DistributedPyNNCommunicationAdapter.py +++ b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/cle/DistributedPyNNCommunicationAdapter.py @@ -130,10 +130,10 @@ class DistributedPyNNCommunicationAdapter(PyNNNestCommunicationAdapter): Notify remote MPI Brain Processes that they must complete this transfer function connection by duplicating the parameters for this device connection. - :param kind Either 'source' or 'sink'. - :param populations The target population to connect to. - :param device The device to create. - :param params The dictionary of adapter params to use. + :param kind: Either 'source' or 'sink'. + :param populations: The target population to connect to. + :param device: The device to create. + :param params: The dictionary of adapter params to use. """ # timestamp the adapter creation, we have no other way of tracking it between @@ -190,7 +190,7 @@ class DistributedPyNNCommunicationAdapter(PyNNNestCommunicationAdapter): Notify remote MPI Brain Processes that they must delete transfer function connection by duplicating the calls in this process. - :param timestep The creation timestep of the TF to delete. + :param timestep: The creation timestep of the TF to delete. """ # propagate the deletion configuration to all other processes, guaranteed to be diff --git a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DaintLauncher.py b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DaintLauncher.py index e15b26fb3ac7bcbc402f94f58122a871276ef02b..cfdc2be8b61c743e472f2e5ed26291f584e796e2 100644 --- a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DaintLauncher.py +++ b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DaintLauncher.py @@ -58,6 +58,7 @@ class DaintLauncher(object): def add_host(self, hostname, tmpdir, processes=1): """ Ignore add_host in daint. Slurm (srun) manages this internally + :param hostname: ignored :param tmpdir: ignored :param processes: ignored @@ -204,7 +205,7 @@ class DaintLauncher(object): progress bar. Wait for completion of the CLE loading task and mark the MPI process as successfully launched. - :param msg The ros message to parse. + :param msg: The ros message to parse. """ status = json.loads(msg.data) diff --git a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLEProcess.py b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLEProcess.py index 6149f079ce92ac0f600d80a80a019d5474781af7..0da750da715d18180a6f75f98e6895cb5793042a 100644 --- a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLEProcess.py +++ b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLEProcess.py @@ -45,7 +45,7 @@ def launch_cle(argv): # pragma: no cover Launch the distributed CLE process with given assembly class. Process the command line and handles all shutdown events to terminate other MPI processes. - :param assembly_class An invokable CLEGazeboSimulationAssembly class definition to use. + :param assembly_class: An invokable CLEGazeboSimulationAssembly class definition to use. """ # import MPI here, must be done after Nest in subclass adapters diff --git a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLESimulationAssembly.py b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLESimulationAssembly.py index 8bef3d3f341d6a8d61993c8eafabb4865aa50c50..8576239ec801715d7ad8e0b7632edfcb29ac689d 100644 --- a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLESimulationAssembly.py +++ b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/DistributedCLESimulationAssembly.py @@ -46,7 +46,7 @@ class DistributedCLESimulationAssembly(CLEGazeboSimulationAssembly): """ Creates a new simulation assembly to simulate an experiment using the CLE and Gazebo - :param sim_config A sim config object that "similar" to the one used for the CLE process + :param sim_config: A sim config object that "similar" to the one used for the CLE process """ super(DistributedCLESimulationAssembly, self).__init__(sim_config) diff --git a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/MPILauncher.py b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/MPILauncher.py index e804cffddd5a206e38907ca8aa40d4e60bed2e90..66a1f7b5690cccfdfffedfd0f5e0fade8b9adf2e 100644 --- a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/MPILauncher.py +++ b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/MPILauncher.py @@ -66,9 +66,9 @@ class MPILauncher(object): """ Add a target host to the mpi launch configuration. - :param hostname The remote host name or ip. - :param tmpdir A valid temporary directory on the remote host to launch in. - :param processes The number of processes for this host. + :param hostname: The remote host name or ip. + :param tmpdir: A valid temporary directory on the remote host to launch in. + :param processes: The number of processes for this host. """ self._hosts.append('-np {p} -host {h} -wdir {t}' .format(p=processes, h=hostname, t=tmpdir)) @@ -137,7 +137,7 @@ class MPILauncher(object): progress bar. Wait for completion of the CLE loading task and mark the MPI process as successfully launched. - :param msg The ros message to parse. + :param msg: The ros message to parse. """ status = json.loads(msg.data) diff --git a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestBrainProcess.py b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestBrainProcess.py index 5094931f47959fe984d866927d66085eb2396324..2ba2fa225d4db7a577b5184e7825d49a08b3f15f 100644 --- a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestBrainProcess.py +++ b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestBrainProcess.py @@ -58,7 +58,7 @@ class NestBrainProcess(object): Nest will automatically allocate the brain in a round-robin fashion under the hood, we do not need to do anything explicitly. - :param sim_config A sim config object that "similar" to the one used for the CLE process + :param sim_config: A sim config object that "similar" to the one used for the CLE process """ self._sim_config = sim_config @@ -172,7 +172,7 @@ class NestBrainProcess(object): Reflect a transfer function connection made on the CLE side by performing the same connection on this side. - :param params The connectivity/synapse parameters passed by the CLE. + :param params: The connectivity/synapse parameters passed by the CLE. """ # get the population of neurons from params @@ -252,7 +252,7 @@ class NestBrainProcess(object): """ Disconnect the specified device and remove it from tracking. - :param params The device parameters passed by the CLE. + :param params: The device parameters passed by the CLE. """ device = self.devices.pop(params['timestep'], None) @@ -269,7 +269,7 @@ class NestBrainProcess(object): """ Load/reload the brain file and population definitions. - :param params The brain parameters passed by the CLE. + :param params: The brain parameters passed by the CLE. """ # ignore any commands during simulation construction diff --git a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestLauncher.py b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestLauncher.py index 760c78ca006ad1382692be88b130de9a81f46de8..fd62329ce11493ef9e861d4e8b92ef60238069d1 100644 --- a/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestLauncher.py +++ b/hbp_nrp_distributed_nest/hbp_nrp_distributed_nest/launch/NestLauncher.py @@ -55,10 +55,10 @@ class NestLauncher(object): :param exc: the experiment configuration :param bibi: the BIBI configuration. - :param server_host Target Gazebo/brain process host (e.g. local or lugano) - :param reservation Reservation string for cluster backend (None is a valid option) - :param sim_id The id of the simulation/experiment to be launched. - :param timeout The default simulation timeout (time initially allocated). + :param server_host: Target Gazebo/brain process host (e.g. local or lugano) + :param reservation: Reservation string for cluster backend (None is a valid option) + :param sim_id: The id of the simulation/experiment to be launched. + :param timeout: The default simulation timeout (time initially allocated). """ self._sim_config = sim_config