diff --git a/.ycm_extra_conf.py b/.ycm_extra_conf.py index 035b3ed1c5996bd633e7ba77a2273f941375323f..a48a781b9af0e36d56e9b5330aaa93518696fec6 100644 --- a/.ycm_extra_conf.py +++ b/.ycm_extra_conf.py @@ -36,7 +36,7 @@ import ycm_core # CHANGE THIS LIST OF FLAGS. YES, THIS IS THE DROID YOU HAVE BEEN LOOKING FOR. flags = [ '-DNDEBUG', - '-DNMC_HAVE_TBB', + '-DNMC_HAVE_CTHREAD', '-std=c++11', '-x', 'c++', @@ -54,7 +54,7 @@ flags = [ 'modcc', '-I', '/cm/shared/apps/cuda/8.0.44/include', - '-DNMC_HAVE_CUDA' + '-DNMC_HAVE_GPU' ] # Set this to the absolute path to the folder (NOT the file!) containing the diff --git a/miniapp/miniapp.cpp b/miniapp/miniapp.cpp index c9fb6b7c577ceedf2b9afa3f2512866dec156814..474c57f3af6be715c3979e9d0d89e52c231326ba 100644 --- a/miniapp/miniapp.cpp +++ b/miniapp/miniapp.cpp @@ -17,10 +17,10 @@ #include <profiling/profiler.hpp> #include <profiling/meter_manager.hpp> #include <threading/threading.hpp> +#include <util/config.hpp> #include <util/debug.hpp> #include <util/ioutil.hpp> #include <util/nop.hpp> -#include <util/optional.hpp> #include "io.hpp" #include "miniapp_recipes.hpp" @@ -29,12 +29,6 @@ using namespace nest::mc; using global_policy = communication::global_policy; -#ifdef NMC_HAVE_CUDA -using lowered_cell = fvm::fvm_multicell<gpu::backend>; -#else -using lowered_cell = fvm::fvm_multicell<multicore::backend>; -#endif -using model_type = model<lowered_cell>; using sample_trace_type = sample_trace<time_type, double>; using file_export_type = io::exporter_spike_file<global_policy>; void banner(); @@ -101,7 +95,9 @@ int main(int argc, char** argv) { options.file_extension, options.over_write); }; - model_type m(*recipe, util::partition_view(group_divisions)); + model m(*recipe, + util::partition_view(group_divisions), + config::has_cuda? backend_policy::prefer_gpu: backend_policy::use_multicore); if (options.report_compartments) { report_compartment_stats(*recipe); } @@ -202,11 +198,7 @@ void banner() { std::cout << " starting miniapp\n"; std::cout << " - " << threading::description() << " threading support\n"; std::cout << " - communication policy: " << std::to_string(global_policy::kind()) << " (" << global_policy::size() << ")\n"; -#ifdef NMC_HAVE_CUDA - std::cout << " - gpu support: on\n"; -#else - std::cout << " - gpu support: off\n"; -#endif + std::cout << " - gpu support: " << (config::has_cuda? "on": "off") << "\n"; std::cout << "====================\n"; } diff --git a/src/backends.hpp b/src/backends.hpp new file mode 100644 index 0000000000000000000000000000000000000000..0941d5927a19ede939f3b02f93b89ce1276bf305 --- /dev/null +++ b/src/backends.hpp @@ -0,0 +1,22 @@ +#pragma once + +#include <string> +#include <backends/fvm.hpp> + +namespace nest { +namespace mc { + +enum class backend_policy { + use_multicore, // use multicore backend for all computation + prefer_gpu // use gpu back end when supported by cell_group type +}; + +inline std::string to_string(backend_policy p) { + if (p==backend_policy::use_multicore) { + return "use_multicore"; + } + return "prefer_gpu"; +} + +} // namespace mc +} // namespace nest diff --git a/src/backends/fvm.hpp b/src/backends/fvm.hpp index a8124da0815f85dbb674ed5e81c31d5cbf2328b9..adb9838b227a392ba5b1dc6556c937bea303104a 100644 --- a/src/backends/fvm.hpp +++ b/src/backends/fvm.hpp @@ -2,6 +2,43 @@ #include <backends/multicore/fvm.hpp> +namespace nest { +namespace mc { + +// A null back end used as a placeholder for back ends that are not supported +// on the target platform. +struct null_backend: public multicore::backend { + static bool is_supported() { + return false; + } + + static mechanism make_mechanism( + const std::string&, view, view, const std::vector<value_type>&, const std::vector<size_type>&) + { + throw std::runtime_error("attempt to use an unsupported back end"); + } + + static bool has_mechanism(const std::string& name) { + return false; + } + + static std::string name() { + return "null"; + } +}; + +} // namespace mc +} // namespace nest + +// FIXME: This include is where cuda-specific code leaks into the main application. +// e.g.: CUDA kernels, functions marked __host__ __device__, etc. +// Hence why it is guarded with NMC_HAVE_CUDA, and not, NMC_HAVE_GPU, like elsewhere in +// the code. When we implement separate compilation of CUDA, this should be guarded with +// NMC_HAVE_GPU, and the NMC_HAVE_CUDA flag depricated. #ifdef NMC_HAVE_CUDA #include <backends/gpu/fvm.hpp> +#else +namespace nest { namespace mc { namespace gpu { + using backend = null_backend; +}}} // namespace nest::mc::gpu #endif diff --git a/src/backends/gpu/fvm.hpp b/src/backends/gpu/fvm.hpp index 80a8de85f8a1ba59e821fbe0cfd3e46637ea6386..1d932a07b68b39e2ea8a07098358141b61a491d2 100644 --- a/src/backends/gpu/fvm.hpp +++ b/src/backends/gpu/fvm.hpp @@ -17,6 +17,10 @@ namespace mc { namespace gpu { struct backend { + static bool is_supported() { + return true; + } + /// define the real and index types using value_type = double; using size_type = nest::mc::cell_lid_type; diff --git a/src/backends/multicore/fvm.hpp b/src/backends/multicore/fvm.hpp index d04928cf00a64978a1ae076957a8c7e4695b9099..08e919ded66de399494e149291edd2e4f3cfe28a 100644 --- a/src/backends/multicore/fvm.hpp +++ b/src/backends/multicore/fvm.hpp @@ -18,6 +18,10 @@ namespace mc { namespace multicore { struct backend { + static bool is_supported() { + return true; + } + /// define the real and index types using value_type = double; using size_type = nest::mc::cell_lid_type; diff --git a/src/cell_group.hpp b/src/cell_group.hpp index f6e4a5000450fe7409b509aff9a4ab23bd03fe3c..e6153024eff204b3c55bd825bba80ae6ebf826ab 100644 --- a/src/cell_group.hpp +++ b/src/cell_group.hpp @@ -1,272 +1,38 @@ #pragma once -#include <cstdint> -#include <functional> -#include <iterator> +#include <memory> #include <vector> -#include <algorithms.hpp> -#include <cell.hpp> #include <common_types.hpp> #include <event_binner.hpp> #include <event_queue.hpp> +#include <sampler_function.hpp> #include <spike.hpp> -#include <util/debug.hpp> -#include <util/partition.hpp> -#include <util/range.hpp> - -#include <profiling/profiler.hpp> +#include <util/optional.hpp> +#include <util/make_unique.hpp> namespace nest { namespace mc { -template <typename LoweredCell> class cell_group { public: - using lowered_cell_type = LoweredCell; - using value_type = typename lowered_cell_type::value_type; - using size_type = typename lowered_cell_type::value_type; - using source_id_type = cell_member_type; - - using sampler_function = std::function<util::optional<time_type>(time_type, double)>; - - cell_group() = default; - - template <typename Cells> - cell_group(cell_gid_type first_gid, const Cells& cells): - gid_base_{first_gid} - { - // Create lookup structure for probe and target ids. - build_handle_partitions(cells); - std::size_t n_probes = probe_handle_divisions_.back(); - std::size_t n_targets = target_handle_divisions_.back(); - std::size_t n_detectors = algorithms::sum(util::transform_view( - cells, [](const cell& c) { return c.detectors().size(); })); - - // Allocate space to store handles. - target_handles_.resize(n_targets); - probe_handles_.resize(n_probes); - - lowered_.initialize(cells, target_handles_, probe_handles_); - - // Create a list of the global identifiers for the spike sources - auto source_gid = cell_gid_type{gid_base_}; - for (const auto& cell: cells) { - for (cell_lid_type lid=0u; lid<cell.detectors().size(); ++lid) { - spike_sources_.push_back(source_id_type{source_gid, lid}); - } - ++source_gid; - } - EXPECTS(spike_sources_.size()==n_detectors); - } - - void reset() { - spikes_.clear(); - events_.clear(); - reset_samplers(); - binner_.reset(); - lowered_.reset(); - } - - void set_binning_policy(binning_kind policy, time_type bin_interval) { - binner_ = event_binner(policy, bin_interval); - } - - void advance(time_type tfinal, time_type dt) { - EXPECTS(lowered_.state_synchronized()); - - // Bin pending events and enqueue on lowered state. - time_type ev_min_time = lowered_.max_time(); // (but we're synchronized here) - while (auto ev = events_.pop_if_before(tfinal)) { - auto handle = get_target_handle(ev->target); - auto binned_ev_time = binner_.bin(ev->target.gid, ev->time, ev_min_time); - lowered_.add_event(binned_ev_time, handle, ev->weight); - } - - lowered_.setup_integration(tfinal, dt); - - std::vector<sample_event> requeue_sample_events; - while (!lowered_.integration_complete()) { - // Take any pending samples. - // TODO: Placeholder: this will be replaced by a backend polling implementation. - - PE("sampling"); - time_type cell_max_time = lowered_.max_time(); - - requeue_sample_events.clear(); - while (auto m = sample_events_.pop_if_before(cell_max_time)) { - auto& s = samplers_[m->sampler_index]; - EXPECTS((bool)s.sampler); - - time_type cell_time = lowered_.time(s.cell_gid-gid_base_); - if (cell_time<m->time) { - // This cell hasn't reached this sample time yet. - requeue_sample_events.push_back(*m); - } - else { - auto next = s.sampler(cell_time, lowered_.probe(s.handle)); - if (next) { - m->time = std::max(*next, cell_time); - requeue_sample_events.push_back(*m); - } - } - } - for (auto& ev: requeue_sample_events) { - sample_events_.push(std::move(ev)); - } - PL(); - - // Ask lowered_ cell to integrate 'one step', delivering any - // events accordingly. - // TODO: Placeholder: with backend polling for samplers, we will - // request that the lowered cell perform the integration all the - // way to tfinal. - - lowered_.step_integration(); - - if (util::is_debug_mode() && !lowered_.is_physical_solution()) { - std::cerr << "warning: solution out of bounds for cell " - << gid_base_ << " at (max) t " << lowered_.max_time() << " ms\n"; - } - } - - // Copy out spike voltage threshold crossings from the back end, then - // generate spikes with global spike source ids. The threshold crossings - // record the local spike source index, which must be converted to a - // global index for spike communication. - PE("events"); - for (auto c: lowered_.get_spikes()) { - spikes_.push_back({spike_sources_[c.index], time_type(c.time)}); - } - // Now that the spikes have been generated, clear the old crossings - // to get ready to record spikes from the next integration period. - lowered_.clear_spikes(); - PL(); - } - - void enqueue_events(const std::vector<postsynaptic_spike_event>& events) { - for (auto& e: events) { - events_.push(e); - } - } - - const std::vector<spike>& spikes() const { - return spikes_; - } - - void clear_spikes() { - spikes_.clear(); - } - - const std::vector<source_id_type>& spike_sources() const { - return spike_sources_; - } - - void add_sampler(cell_member_type probe_id, sampler_function s, time_type start_time = 0) { - auto handle = get_probe_handle(probe_id); - - using size_type = sample_event::size_type; - auto sampler_index = size_type(samplers_.size()); - samplers_.push_back({handle, probe_id.gid, s}); - sampler_start_times_.push_back(start_time); - sample_events_.push({sampler_index, start_time}); - } - -private: - - // gid of first cell in group. - cell_gid_type gid_base_; - - // The lowered cell state (e.g. FVM) of the cell. - lowered_cell_type lowered_; - - // Spike detectors attached to the cell. - std::vector<source_id_type> spike_sources_; - - // Spikes that are generated. - std::vector<spike> spikes_; - - // Event time binning manager. - event_binner binner_; - - // Pending events to be delivered. - event_queue<postsynaptic_spike_event> events_; - - // Pending samples to be taken. - event_queue<sample_event> sample_events_; - std::vector<time_type> sampler_start_times_; - - // Handles for accessing lowered cell. - using target_handle = typename lowered_cell_type::target_handle; - std::vector<target_handle> target_handles_; - - using probe_handle = typename lowered_cell_type::probe_handle; - std::vector<probe_handle> probe_handles_; - - struct sampler_entry { - typename lowered_cell_type::probe_handle handle; - cell_gid_type cell_gid; - sampler_function sampler; - }; - - // Collection of samplers to be run against probes in this group. - std::vector<sampler_entry> samplers_; - - // Lookup table for probe ids -> local probe handle indices. - std::vector<std::size_t> probe_handle_divisions_; - - // Lookup table for target ids -> local target handle indices. - std::vector<std::size_t> target_handle_divisions_; - - // Build handle index lookup tables. - template <typename Cells> - void build_handle_partitions(const Cells& cells) { - auto probe_counts = - util::transform_view(cells, [](const cell& c) { return c.probes().size(); }); - auto target_counts = - util::transform_view(cells, [](const cell& c) { return c.synapses().size(); }); - - make_partition(probe_handle_divisions_, probe_counts); - make_partition(target_handle_divisions_, target_counts); - } - - // Use handle partition to get index from id. - template <typename Divisions> - std::size_t handle_partition_lookup(const Divisions& divisions, cell_member_type id) const { - // NB: without any assertion checking, this would just be: - // return divisions[id.gid-gid_base_]+id.index; - - EXPECTS(id.gid>=gid_base_); - - auto handle_partition = util::partition_view(divisions); - EXPECTS(id.gid-gid_base_<handle_partition.size()); - - auto ival = handle_partition[id.gid-gid_base_]; - std::size_t i = ival.first + id.index; - EXPECTS(i<ival.second); - - return i; - } - - // Get probe handle from probe id. - probe_handle get_probe_handle(cell_member_type probe_id) const { - return probe_handles_[handle_partition_lookup(probe_handle_divisions_, probe_id)]; - } + virtual ~cell_group() = default; + + virtual void reset() = 0; + virtual void set_binning_policy(binning_kind policy, time_type bin_interval) = 0; + virtual void advance(time_type tfinal, time_type dt) = 0; + virtual void enqueue_events(const std::vector<postsynaptic_spike_event>& events) = 0; + virtual const std::vector<spike>& spikes() const = 0; + virtual void clear_spikes() = 0; + virtual void add_sampler(cell_member_type probe_id, sampler_function s, time_type start_time = 0) = 0; +}; - // Get target handle from target id. - target_handle get_target_handle(cell_member_type target_id) const { - return target_handles_[handle_partition_lookup(target_handle_divisions_, target_id)]; - } +using cell_group_ptr = std::unique_ptr<cell_group>; - void reset_samplers() { - // clear all pending sample events and reset to start at time 0 - sample_events_.clear(); - using size_type = sample_event::size_type; - for(size_type i=0; i<samplers_.size(); ++i) { - sample_events_.push({i, sampler_start_times_[i]}); - } - } -}; +template <typename T, typename... Args> +cell_group_ptr make_cell_group(Args&&... args) { + return cell_group_ptr(new T(std::forward<Args>(args)...)); +} } // namespace mc } // namespace nest diff --git a/src/fvm_multicell.hpp b/src/fvm_multicell.hpp index 78cb84288d5c8fb0ab28a811404dfef686465d69..ba2cb5c3c6a0d74070b81eb823dd3d1544cf4ce4 100644 --- a/src/fvm_multicell.hpp +++ b/src/fvm_multicell.hpp @@ -8,7 +8,6 @@ #include <vector> #include <algorithms.hpp> -#include <backends/fvm.hpp> #include <cell.hpp> #include <compartment.hpp> #include <event_queue.hpp> diff --git a/src/mc_cell_group.hpp b/src/mc_cell_group.hpp new file mode 100644 index 0000000000000000000000000000000000000000..ec3cbc72649f5db0075290ae3b0007d0c9618e5e --- /dev/null +++ b/src/mc_cell_group.hpp @@ -0,0 +1,272 @@ +#pragma once + +#include <cstdint> +#include <functional> +#include <iterator> +#include <vector> + +#include <algorithms.hpp> +#include <cell_group.hpp> +#include <cell.hpp> +#include <common_types.hpp> +#include <event_binner.hpp> +#include <event_queue.hpp> +#include <sampler_function.hpp> +#include <spike.hpp> +#include <util/debug.hpp> +#include <util/partition.hpp> +#include <util/range.hpp> + +#include <profiling/profiler.hpp> + +namespace nest { +namespace mc { + +template <typename LoweredCell> +class mc_cell_group: public cell_group { +public: + using lowered_cell_type = LoweredCell; + using value_type = typename lowered_cell_type::value_type; + using size_type = typename lowered_cell_type::value_type; + using source_id_type = cell_member_type; + + mc_cell_group() = default; + + template <typename Cells> + mc_cell_group(cell_gid_type first_gid, const Cells& cells): + gid_base_{first_gid} + { + // Create lookup structure for probe and target ids. + build_handle_partitions(cells); + std::size_t n_probes = probe_handle_divisions_.back(); + std::size_t n_targets = target_handle_divisions_.back(); + std::size_t n_detectors = algorithms::sum(util::transform_view( + cells, [](const cell& c) { return c.detectors().size(); })); + + // Allocate space to store handles. + target_handles_.resize(n_targets); + probe_handles_.resize(n_probes); + + lowered_.initialize(cells, target_handles_, probe_handles_); + + // Create a list of the global identifiers for the spike sources + auto source_gid = cell_gid_type{gid_base_}; + for (const auto& cell: cells) { + for (cell_lid_type lid=0u; lid<cell.detectors().size(); ++lid) { + spike_sources_.push_back(source_id_type{source_gid, lid}); + } + ++source_gid; + } + EXPECTS(spike_sources_.size()==n_detectors); + } + + void reset() override { + spikes_.clear(); + events_.clear(); + reset_samplers(); + binner_.reset(); + lowered_.reset(); + } + + void set_binning_policy(binning_kind policy, time_type bin_interval) override { + binner_ = event_binner(policy, bin_interval); + } + + void advance(time_type tfinal, time_type dt) override { + EXPECTS(lowered_.state_synchronized()); + + // Bin pending events and enqueue on lowered state. + time_type ev_min_time = lowered_.max_time(); // (but we're synchronized here) + while (auto ev = events_.pop_if_before(tfinal)) { + auto handle = get_target_handle(ev->target); + auto binned_ev_time = binner_.bin(ev->target.gid, ev->time, ev_min_time); + lowered_.add_event(binned_ev_time, handle, ev->weight); + } + + lowered_.setup_integration(tfinal, dt); + + std::vector<sample_event> requeue_sample_events; + while (!lowered_.integration_complete()) { + // Take any pending samples. + // TODO: Placeholder: this will be replaced by a backend polling implementation. + + PE("sampling"); + time_type cell_max_time = lowered_.max_time(); + + requeue_sample_events.clear(); + while (auto m = sample_events_.pop_if_before(cell_max_time)) { + auto& s = samplers_[m->sampler_index]; + EXPECTS((bool)s.sampler); + + time_type cell_time = lowered_.time(s.cell_gid-gid_base_); + if (cell_time<m->time) { + // This cell hasn't reached this sample time yet. + requeue_sample_events.push_back(*m); + } + else { + auto next = s.sampler(cell_time, lowered_.probe(s.handle)); + if (next) { + m->time = std::max(*next, cell_time); + requeue_sample_events.push_back(*m); + } + } + } + for (auto& ev: requeue_sample_events) { + sample_events_.push(std::move(ev)); + } + PL(); + + // Ask lowered_ cell to integrate 'one step', delivering any + // events accordingly. + // TODO: Placeholder: with backend polling for samplers, we will + // request that the lowered cell perform the integration all the + // way to tfinal. + + lowered_.step_integration(); + + if (util::is_debug_mode() && !lowered_.is_physical_solution()) { + std::cerr << "warning: solution out of bounds for cell " + << gid_base_ << " at (max) t " << lowered_.max_time() << " ms\n"; + } + } + + // Copy out spike voltage threshold crossings from the back end, then + // generate spikes with global spike source ids. The threshold crossings + // record the local spike source index, which must be converted to a + // global index for spike communication. + PE("events"); + for (auto c: lowered_.get_spikes()) { + spikes_.push_back({spike_sources_[c.index], time_type(c.time)}); + } + // Now that the spikes have been generated, clear the old crossings + // to get ready to record spikes from the next integration period. + lowered_.clear_spikes(); + PL(); + } + + void enqueue_events(const std::vector<postsynaptic_spike_event>& events) override { + for (auto& e: events) { + events_.push(e); + } + } + + const std::vector<spike>& spikes() const override { + return spikes_; + } + + void clear_spikes() override { + spikes_.clear(); + } + + const std::vector<source_id_type>& spike_sources() const { + return spike_sources_; + } + + void add_sampler(cell_member_type probe_id, sampler_function s, time_type start_time) override { + auto handle = get_probe_handle(probe_id); + + using size_type = sample_event::size_type; + auto sampler_index = size_type(samplers_.size()); + samplers_.push_back({handle, probe_id.gid, s}); + sampler_start_times_.push_back(start_time); + sample_events_.push({sampler_index, start_time}); + } + +private: + + // gid of first cell in group. + cell_gid_type gid_base_; + + // The lowered cell state (e.g. FVM) of the cell. + lowered_cell_type lowered_; + + // Spike detectors attached to the cell. + std::vector<source_id_type> spike_sources_; + + // Spikes that are generated. + std::vector<spike> spikes_; + + // Event time binning manager. + event_binner binner_; + + // Pending events to be delivered. + event_queue<postsynaptic_spike_event> events_; + + // Pending samples to be taken. + event_queue<sample_event> sample_events_; + std::vector<time_type> sampler_start_times_; + + // Handles for accessing lowered cell. + using target_handle = typename lowered_cell_type::target_handle; + std::vector<target_handle> target_handles_; + + using probe_handle = typename lowered_cell_type::probe_handle; + std::vector<probe_handle> probe_handles_; + + struct sampler_entry { + typename lowered_cell_type::probe_handle handle; + cell_gid_type cell_gid; + sampler_function sampler; + }; + + // Collection of samplers to be run against probes in this group. + std::vector<sampler_entry> samplers_; + + // Lookup table for probe ids -> local probe handle indices. + std::vector<std::size_t> probe_handle_divisions_; + + // Lookup table for target ids -> local target handle indices. + std::vector<std::size_t> target_handle_divisions_; + + // Build handle index lookup tables. + template <typename Cells> + void build_handle_partitions(const Cells& cells) { + auto probe_counts = + util::transform_view(cells, [](const cell& c) { return c.probes().size(); }); + auto target_counts = + util::transform_view(cells, [](const cell& c) { return c.synapses().size(); }); + + make_partition(probe_handle_divisions_, probe_counts); + make_partition(target_handle_divisions_, target_counts); + } + + // Use handle partition to get index from id. + template <typename Divisions> + std::size_t handle_partition_lookup(const Divisions& divisions, cell_member_type id) const { + // NB: without any assertion checking, this would just be: + // return divisions[id.gid-gid_base_]+id.index; + + EXPECTS(id.gid>=gid_base_); + + auto handle_partition = util::partition_view(divisions); + EXPECTS(id.gid-gid_base_<handle_partition.size()); + + auto ival = handle_partition[id.gid-gid_base_]; + std::size_t i = ival.first + id.index; + EXPECTS(i<ival.second); + + return i; + } + + // Get probe handle from probe id. + probe_handle get_probe_handle(cell_member_type probe_id) const { + return probe_handles_[handle_partition_lookup(probe_handle_divisions_, probe_id)]; + } + + // Get target handle from target id. + target_handle get_target_handle(cell_member_type target_id) const { + return target_handles_[handle_partition_lookup(target_handle_divisions_, target_id)]; + } + + void reset_samplers() { + // clear all pending sample events and reset to start at time 0 + sample_events_.clear(); + using size_type = sample_event::size_type; + for(size_type i=0; i<samplers_.size(); ++i) { + sample_events_.push({i, sampler_start_times_[i]}); + } + } +}; + +} // namespace mc +} // namespace nest diff --git a/src/memory/allocator.hpp b/src/memory/allocator.hpp index 3755e293f3c951b6510a6c3ed84d77d11769a75b..9add1a67bdd715baa0b0dc95592ea9dfaa02253b 100644 --- a/src/memory/allocator.hpp +++ b/src/memory/allocator.hpp @@ -2,7 +2,7 @@ #include <limits> -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU #include <cuda.h> #include <cuda_runtime.h> #endif @@ -138,7 +138,7 @@ namespace impl { } #endif -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU namespace cuda { template <size_type Alignment> class pinned_policy { @@ -243,7 +243,7 @@ namespace impl { } }; } // namespace cuda -#endif // #ifdef NMC_HAVE_CUDA +#endif // #ifdef NMC_HAVE_GPU } // namespace impl template<typename T, typename Policy > @@ -318,7 +318,7 @@ namespace util { } }; -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU template <size_t Alignment> struct type_printer<impl::cuda::pinned_policy<Alignment>>{ static std::string print() { @@ -357,7 +357,7 @@ template <class T, size_t alignment=(512/8)> using hbw_allocator = allocator<T, impl::knl::hbw_policy<alignment>>; #endif -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU // For pinned and allocation set the default alignment to correspond to // the alignment of 1024 bytes, because pinned memory is allocated at // page boundaries. It is allocated at page boundaries (typically 4k), diff --git a/src/memory/gpu.hpp b/src/memory/gpu.hpp index 8e950d567217459186b7ebf23953104d6366eb3a..c5955eac21d908519e511351fbf4f56f22f09a26 100644 --- a/src/memory/gpu.hpp +++ b/src/memory/gpu.hpp @@ -1,6 +1,6 @@ #pragma once -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU #include <string> #include <cstdint> diff --git a/src/memory/host_coordinator.hpp b/src/memory/host_coordinator.hpp index e9cb069893b2185bbabfd3798919c59306e54dbe..1f45b336f4a8a4319c1f46fae6b5b13fe9fad8cf 100644 --- a/src/memory/host_coordinator.hpp +++ b/src/memory/host_coordinator.hpp @@ -11,7 +11,7 @@ #include "allocator.hpp" #include "util.hpp" -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU #include "gpu.hpp" #endif @@ -23,7 +23,7 @@ namespace memory { template <typename T, class Allocator> class host_coordinator; -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU template <typename T, class Allocator> class device_coordinator; #endif @@ -124,7 +124,7 @@ public: std::copy(from.begin(), from.end(), to.begin()); } -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU // copy memory from device to host template <class Alloc> void copy( diff --git a/src/memory/memory.hpp b/src/memory/memory.hpp index e52619584a97731ac9dc117807799941eee524de..50bbdf14276eb1a390d4e0be427c1be5d5ded515 100644 --- a/src/memory/memory.hpp +++ b/src/memory/memory.hpp @@ -6,7 +6,7 @@ #include "definitions.hpp" #include "host_coordinator.hpp" -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU #include "device_coordinator.hpp" #endif @@ -29,7 +29,7 @@ std::ostream& operator<< (std::ostream& o, host_view<T> const& v) { return o; } -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU // specialization for pinned vectors. Use a host_coordinator, because memory is // in the host memory space, and all of the helpers (copy, set, etc) are the // same with and without page locked memory diff --git a/src/memory/wrappers.hpp b/src/memory/wrappers.hpp index ab463a238efd507a6fbb7b521721afbc33a35fe0..50393017701cb18b0e4b4d7ab9a2026e8264ba5b 100644 --- a/src/memory/wrappers.hpp +++ b/src/memory/wrappers.hpp @@ -5,7 +5,7 @@ #include <memory/memory.hpp> -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU #include <cuda.h> #include <cuda_runtime.h> #endif @@ -96,7 +96,7 @@ namespace util { return is_on_host<typename std::decay<T>::type>::value; } - #ifdef NMC_HAVE_CUDA + #ifdef NMC_HAVE_GPU template <typename T> struct is_on_gpu : std::false_type {}; @@ -132,7 +132,7 @@ auto on_host(const C& c) -> decltype(make_const_view(c)) { return make_const_view(c); } -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU template < typename C, typename = typename std::enable_if<util::is_on_gpu_v<C>()>::type diff --git a/src/model.hpp b/src/model.hpp index 837ddc8d3b71944ce39f1f7046c4dd26883820ce..91d0cd4e150a42ebfa5a62a9777e48651074c5d4 100644 --- a/src/model.hpp +++ b/src/model.hpp @@ -5,29 +5,36 @@ #include <cstdlib> +#include <backends.hpp> +#include <fvm_multicell.hpp> + #include <common_types.hpp> #include <cell.hpp> #include <cell_group.hpp> #include <communication/communicator.hpp> #include <communication/global_policy.hpp> +#include <mc_cell_group.hpp> #include <profiling/profiler.hpp> #include <recipe.hpp> +#include <sampler_function.hpp> #include <thread_private_spike_store.hpp> +#include <threading/threading.hpp> +#include <trace_sampler.hpp> #include <util/nop.hpp> #include <util/partition.hpp> #include <util/range.hpp> -#include "trace_sampler.hpp" - namespace nest { namespace mc { -// FIXME: this is going to go -template <typename Cell> +using gpu_lowered_cell = + mc_cell_group<fvm::fvm_multicell<gpu::backend>>; + +using multicore_lowered_cell = + mc_cell_group<fvm::fvm_multicell<multicore::backend>>; + class model { public: - // FIXME: this is an intermediate step to remove the template parameter from model - using cell_group_type = cell_group<Cell>; // FIXME using communicator_type = communication::communicator<communication::global_policy>; using spike_export_function = std::function<void(const std::vector<spike>&)>; @@ -37,15 +44,19 @@ public: }; template <typename Iter> - model(const recipe& rec, const util::partition_range<Iter>& groups): - cell_group_divisions_(groups.divisions().begin(), groups.divisions().end()) + model( const recipe& rec, + const util::partition_range<Iter>& groups, + backend_policy policy): + cell_group_divisions_(groups.divisions().begin(), groups.divisions().end()), + backend_policy_(policy) { // set up communicator based on partition - communicator_ = communicator_type{gid_partition()}; + communicator_ = communicator_type(gid_partition()); // generate the cell groups in parallel, with one task per cell group - cell_groups_ = std::vector<cell_group_type>{gid_partition().size()}; - threading::parallel_vector<probe_record> probes; + cell_groups_.resize(gid_partition().size()); + // thread safe vector for constructing the list of probes in parallel + threading::parallel_vector<probe_record> probe_tmp; threading::parallel_for::apply(0, cell_groups_.size(), [&](cell_gid_type i) { @@ -61,16 +72,21 @@ public: cell_lid_type j = 0; for (const auto& probe: cells[i].probes()) { cell_member_type probe_id{gid, j++}; - probes.push_back({probe_id, probe}); + probe_tmp.push_back({probe_id, probe}); } } - cell_groups_[i] = cell_group_type(gids.first, cells); + if (backend_policy_==backend_policy::use_multicore) { + cell_groups_[i] = make_cell_group<multicore_lowered_cell>(gids.first, cells); + } + else { + cell_groups_[i] = make_cell_group<gpu_lowered_cell>(gids.first, cells); + } PL(2); }); - // insert probes - probes_.assign(probes.begin(), probes.end()); + // store probes + probes_.assign(probe_tmp.begin(), probe_tmp.end()); // generate the network connections for (cell_gid_type i: util::make_span(gid_partition().bounds())) { @@ -89,14 +105,14 @@ public: } // one cell per group: - model(const recipe& rec): - model(rec, util::partition_view(util::make_span(0, rec.num_cells()+1))) + model(const recipe& rec, backend_policy policy): + model(rec, util::partition_view(util::make_span(0, rec.num_cells()+1)), policy) {} void reset() { t_ = 0.; for (auto& group: cell_groups_) { - group.reset(); + group->reset(); } communicator_.reset(); @@ -132,14 +148,14 @@ public: auto &group = cell_groups_[i]; PE("stepping","events"); - group.enqueue_events(current_events()[i]); + group->enqueue_events(current_events()[i]); PL(); - group.advance(tuntil, dt); + group->advance(tuntil, dt); PE("events"); - current_spikes().insert(group.spikes()); - group.clear_spikes(); + current_spikes().insert(group->spikes()); + group->clear_spikes(); PL(2); }); }; @@ -207,14 +223,13 @@ public: current_spikes().get().push_back({source, tspike}); } - template <typename F> - void attach_sampler(cell_member_type probe_id, F&& f, time_type tfrom = 0) { + void attach_sampler(cell_member_type probe_id, sampler_function f, time_type tfrom = 0) { if (!algorithms::in_interval(probe_id.gid, gid_partition().bounds())) { return; } const auto idx = gid_partition().index(probe_id.gid); - cell_groups_[idx].add_sampler(probe_id, std::forward<F>(f), tfrom); + cell_groups_[idx]->add_sampler(probe_id, f, tfrom); } const std::vector<probe_record>& probes() const { return probes_; } @@ -235,13 +250,13 @@ public: // Set event binning policy on all our groups. void set_binning_policy(binning_kind policy, time_type bin_interval) { for (auto& group: cell_groups_) { - group.set_binning_policy(policy, bin_interval); + group->set_binning_policy(policy, bin_interval); } } // access cell_group directly - cell_group_type& group(int i) { - return cell_groups_[i]; + cell_group& group(int i) { + return *cell_groups_[i]; } // register a callback that will perform a export of the global @@ -258,13 +273,14 @@ public: private: std::vector<cell_gid_type> cell_group_divisions_; + backend_policy backend_policy_; auto gid_partition() const -> decltype(util::partition_view(cell_group_divisions_)) { return util::partition_view(cell_group_divisions_); } time_type t_ = 0.; - std::vector<cell_group_type> cell_groups_; + std::vector<std::unique_ptr<cell_group>> cell_groups_; communicator_type communicator_; std::vector<probe_record> probes_; diff --git a/src/sampler_function.hpp b/src/sampler_function.hpp new file mode 100644 index 0000000000000000000000000000000000000000..6a042b585deb264a0378217b4ff10f0f9d1164df --- /dev/null +++ b/src/sampler_function.hpp @@ -0,0 +1,14 @@ +#pragma once + +#include <functional> + +#include <common_types.hpp> +#include <util/optional.hpp> + +namespace nest { +namespace mc { + +using sampler_function = std::function<util::optional<time_type>(time_type, double)>; + +} // namespace mc +} // namespace nest diff --git a/src/simple_sampler.hpp b/src/simple_sampler.hpp index cb05573468928ca108e585a21fabd9a74d2bac6e..5ffc41d2470dbd55f0b1af732c1f0aa43db11530 100644 --- a/src/simple_sampler.hpp +++ b/src/simple_sampler.hpp @@ -9,6 +9,7 @@ #include <vector> #include <common_types.hpp> +#include <sampler_function.hpp> #include <util/optional.hpp> #include <util/deduce_return.hpp> #include <util/transform.hpp> @@ -43,7 +44,7 @@ class simple_sampler { public: trace_data trace; - simple_sampler(float dt, float t0=0): + simple_sampler(time_type dt, time_type t0=0): t0_(t0), sample_dt_(dt), t_next_sample_(t0) @@ -54,23 +55,22 @@ public: t_next_sample_ = t0_; } - template <typename Time = float, typename Value = double> - std::function<util::optional<Time> (Time, Value)> sampler() { - return [&](Time t, Value v) -> util::optional<Time> { - if (t<(Time)t_next_sample_) { - return (Time)t_next_sample_; + sampler_function sampler() { + return [&](time_type t, double v) -> util::optional<time_type> { + if (t<t_next_sample_) { + return t_next_sample_; } else { - trace.push_back({float(t), double(v)}); + trace.push_back({t, v}); return t_next_sample_+=sample_dt_; } }; } private: - float t0_ = 0; - float sample_dt_ = 0; - float t_next_sample_ = 0; + time_type t0_ = 0; + time_type sample_dt_ = 0; + time_type t_next_sample_ = 0; }; } // namespace mc diff --git a/src/util/config.hpp b/src/util/config.hpp index 4bd3adf3bbd20b7626e5422e27df99b45a9eb3bf..2e113bf485da41c02327ce021d51f569b91a855d 100644 --- a/src/util/config.hpp +++ b/src/util/config.hpp @@ -30,7 +30,7 @@ constexpr bool has_power_measurement = true; constexpr bool has_power_measurement = false; #endif -#ifdef NMC_HAVE_CUDA +#ifdef NMC_HAVE_GPU constexpr bool has_cuda = true; #else constexpr bool has_cuda = false; diff --git a/tests/performance/io/disk_io.cpp b/tests/performance/io/disk_io.cpp index 05b64666c4502264d7d81cb1901e9d1152d2ae47..c671e907475e2de443fceac1d0818958ac0a1cf8 100644 --- a/tests/performance/io/disk_io.cpp +++ b/tests/performance/io/disk_io.cpp @@ -17,8 +17,6 @@ using namespace nest::mc; using global_policy = communication::global_policy; -using lowered_cell = fvm::fvm_multicell<multicore::backend>; -using cell_group_type = cell_group<lowered_cell>; using timer = util::timer_type; int main(int argc, char** argv) { diff --git a/tests/unit/CMakeLists.txt b/tests/unit/CMakeLists.txt index 3ac4c1ff26752c49190ba6279eabf5291769984a..0bae169cdb41837b3f1263743acd37681cfd95a2 100644 --- a/tests/unit/CMakeLists.txt +++ b/tests/unit/CMakeLists.txt @@ -18,7 +18,7 @@ build_modules( set(TEST_CUDA_SOURCES test_atomics.cu - test_cell_group.cu + test_mc_cell_group.cu test_gpu_stack.cu test_matrix.cu test_spikes.cu @@ -31,6 +31,7 @@ set(TEST_CUDA_SOURCES set(TEST_SOURCES # unit tests test_algorithms.cpp + test_backend.cpp test_double_buffer.cpp test_cell.cpp test_compartments.cpp @@ -41,7 +42,7 @@ set(TEST_SOURCES test_event_binner.cpp test_filter.cpp test_fvm_multi.cpp - test_cell_group.cpp + test_mc_cell_group.cpp test_lexcmp.cpp test_mask_stream.cpp test_math.cpp diff --git a/tests/unit/test_backend.cpp b/tests/unit/test_backend.cpp new file mode 100644 index 0000000000000000000000000000000000000000..08679d0d0b6e750e1611cbf0a9cbfc87ca68975f --- /dev/null +++ b/tests/unit/test_backend.cpp @@ -0,0 +1,23 @@ +#include <type_traits> + +#include <backends/fvm.hpp> +#include <memory/memory.hpp> +#include <util/config.hpp> + +#include "../gtest.h" + +TEST(backends, gpu_is_null) { + using backend = nest::mc::gpu::backend; + + static_assert(std::is_same<backend, nest::mc::null_backend>::value || nest::mc::config::has_cuda, + "gpu back should be defined as null when compiling without gpu support."); + + if (not nest::mc::config::has_cuda) { + EXPECT_FALSE(backend::is_supported()); + + EXPECT_FALSE(backend::has_mechanism("hh")); + EXPECT_THROW( + backend::make_mechanism("hh", backend::view(), backend::view(), {}, {}), + std::runtime_error); + } +} diff --git a/tests/unit/test_fvm_multi.cpp b/tests/unit/test_fvm_multi.cpp index 66226d9e9464df8fe014a6ff12c41d8817331bbc..86777590b9c4819dc2ad67a7a13e720aaedae480 100644 --- a/tests/unit/test_fvm_multi.cpp +++ b/tests/unit/test_fvm_multi.cpp @@ -2,8 +2,9 @@ #include "../gtest.h" -#include <common_types.hpp> +#include <backends/multicore/fvm.hpp> #include <cell.hpp> +#include <common_types.hpp> #include <fvm_multicell.hpp> #include <util/rangeutil.hpp> diff --git a/tests/unit/test_cell_group.cpp b/tests/unit/test_mc_cell_group.cpp similarity index 76% rename from tests/unit/test_cell_group.cpp rename to tests/unit/test_mc_cell_group.cpp index 0016b24e2bc74d1edf0989c1f214d3a9c2c9476f..004915370348d038e5f61dec0ac3e5c3ff6dbc1c 100644 --- a/tests/unit/test_cell_group.cpp +++ b/tests/unit/test_mc_cell_group.cpp @@ -1,8 +1,9 @@ #include "../gtest.h" -#include <cell_group.hpp> +#include <backends/multicore/fvm.hpp> #include <common_types.hpp> #include <fvm_multicell.hpp> +#include <mc_cell_group.hpp> #include <util/rangeutil.hpp> #include "common.hpp" @@ -11,19 +12,17 @@ using namespace nest::mc; using fvm_cell = fvm::fvm_multicell<nest::mc::multicore::backend>; -nest::mc::cell make_cell() { - using namespace nest::mc; +cell make_cell() { + auto c = make_cell_ball_and_stick(); - nest::mc::cell cell = make_cell_ball_and_stick(); + c.add_detector({0, 0}, 0); + c.segment(1)->set_compartments(101); - cell.add_detector({0, 0}, 0); - cell.segment(1)->set_compartments(101); - - return cell; + return c; } -TEST(cell_group, test) { - cell_group<fvm_cell> group{0, util::singleton_view(make_cell())}; +TEST(mc_cell_group, test) { + mc_cell_group<fvm_cell> group{0, util::singleton_view(make_cell())}; group.advance(50, 0.01); @@ -32,8 +31,8 @@ TEST(cell_group, test) { EXPECT_EQ(4u, group.spikes().size()); } -TEST(cell_group, sources) { - using cell_group_type = cell_group<fvm_cell>; +TEST(mc_cell_group, sources) { + using cell_group_type = mc_cell_group<fvm_cell>; auto cell = make_cell(); EXPECT_EQ(cell.detectors().size(), 1u); diff --git a/tests/unit/test_cell_group.cu b/tests/unit/test_mc_cell_group.cu similarity index 86% rename from tests/unit/test_cell_group.cu rename to tests/unit/test_mc_cell_group.cu index 16bc124dec87108f97071c7f3e13c193c02402aa..92f7af506db8c5480786e65764c6ca78c29c087d 100644 --- a/tests/unit/test_cell_group.cu +++ b/tests/unit/test_mc_cell_group.cu @@ -1,6 +1,7 @@ #include "../gtest.h" -#include <cell_group.hpp> +#include <backends/gpu/fvm.hpp> +#include <mc_cell_group.hpp> #include <common_types.hpp> #include <fvm_multicell.hpp> #include <util/rangeutil.hpp> @@ -25,7 +26,7 @@ nest::mc::cell make_cell() { TEST(cell_group, test) { - using cell_group_type = cell_group<fvm_cell>; + using cell_group_type = mc_cell_group<fvm_cell>; auto group = cell_group_type{0, util::singleton_view(make_cell())}; group.advance(50, 0.01); diff --git a/tests/unit/test_probe.cpp b/tests/unit/test_probe.cpp index 005eead2b61f2e4eb526edbe96ed2bddd47d611a..8ea355eaf261e2610d8ba736244fe49232b29f1d 100644 --- a/tests/unit/test_probe.cpp +++ b/tests/unit/test_probe.cpp @@ -1,5 +1,6 @@ #include "../gtest.h" +#include <backends/multicore/fvm.hpp> #include <common_types.hpp> #include <cell.hpp> #include <fvm_multicell.hpp> diff --git a/tests/validation/convergence_test.hpp b/tests/validation/convergence_test.hpp index 4d8ea28dffbda02566a7e147ead45eef96acd555..7c9f593634fc80334f4b1cb3afbdaa6673cc82e6 100644 --- a/tests/validation/convergence_test.hpp +++ b/tests/validation/convergence_test.hpp @@ -75,7 +75,7 @@ public: // reset samplers and attach to probe locations for (auto& se: cell_samplers_) { se.sampler.reset(); - m.attach_sampler(se.probe, se.sampler.template sampler<>()); + m.attach_sampler(se.probe, se.sampler.sampler()); } m.run(t_end, dt); diff --git a/tests/validation/validate_ball_and_stick.cpp b/tests/validation/validate_ball_and_stick.cpp index 109155a69a7829b9300665bd5523ec7eaf1c3558..c28530a1afbd0f81c663275bec8819d5ee9a2331 100644 --- a/tests/validation/validate_ball_and_stick.cpp +++ b/tests/validation/validate_ball_and_stick.cpp @@ -3,24 +3,24 @@ #include <fvm_multicell.hpp> -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::multicore::backend>; +const auto backend = nest::mc::backend_policy::use_multicore; TEST(ball_and_stick, neuron_ref) { - validate_ball_and_stick<lowered_cell>(); + validate_ball_and_stick(backend); } TEST(ball_and_taper, neuron_ref) { - validate_ball_and_taper<lowered_cell>(); + validate_ball_and_taper(backend); } TEST(ball_and_3stick, neuron_ref) { - validate_ball_and_3stick<lowered_cell>(); + validate_ball_and_3stick(backend); } TEST(rallpack1, numeric_ref) { - validate_rallpack1<lowered_cell>(); + validate_rallpack1(backend); } TEST(ball_and_squiggle, neuron_ref) { - validate_ball_and_squiggle<lowered_cell>(); + validate_ball_and_squiggle(backend); } diff --git a/tests/validation/validate_ball_and_stick.cu b/tests/validation/validate_ball_and_stick.cu index 52d1bf0dc84925e858c775002e210d3c630fc1d3..3f8e442761e757c2096c2638426f3db08a71d7e7 100644 --- a/tests/validation/validate_ball_and_stick.cu +++ b/tests/validation/validate_ball_and_stick.cu @@ -3,24 +3,24 @@ #include <fvm_multicell.hpp> -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::gpu::backend>; +const auto backend = nest::mc::backend_policy::prefer_gpu; TEST(ball_and_stick, neuron_ref) { - validate_ball_and_stick<lowered_cell>(); + validate_ball_and_stick(backend); } TEST(ball_and_taper, neuron_ref) { - validate_ball_and_taper<lowered_cell>(); + validate_ball_and_taper(backend); } TEST(ball_and_3stick, neuron_ref) { - validate_ball_and_3stick<lowered_cell>(); + validate_ball_and_3stick(backend); } TEST(rallpack1, numeric_ref) { - validate_rallpack1<lowered_cell>(); + validate_rallpack1(backend); } TEST(ball_and_squiggle, neuron_ref) { - validate_ball_and_squiggle<lowered_cell>(); + validate_ball_and_squiggle(backend); } diff --git a/tests/validation/validate_ball_and_stick.hpp b/tests/validation/validate_ball_and_stick.hpp index 0a438f9df90e25ca8ce2ebc68895adae82cff299..da1652e6747e311479b07515faca3c799df8d923 100644 --- a/tests/validation/validate_ball_and_stick.hpp +++ b/tests/validation/validate_ball_and_stick.hpp @@ -15,13 +15,11 @@ #include "trace_analysis.hpp" #include "validation_data.hpp" -template < - typename LoweredCell, - typename SamplerInfoSeq -> +template <typename SamplerInfoSeq> void run_ncomp_convergence_test( const char* model_name, const nest::mc::util::path& ref_data_path, + nest::mc::backend_policy backend, const nest::mc::cell& c, SamplerInfoSeq& samplers, float t_end=100.f) @@ -37,7 +35,7 @@ void run_ncomp_convergence_test( {"dt", dt}, {"sim", "nestmc"}, {"units", "mV"}, - {"backend", LoweredCell::backend::name()} + {"backend_policy", to_string(backend)} }; auto exclude = stimulus_ends(c); @@ -51,7 +49,7 @@ void run_ncomp_convergence_test( seg->set_compartments(ncomp); } } - model<LoweredCell> m(singleton_recipe{c}); + model m(singleton_recipe{c}, backend); runner.run(m, ncomp, t_end, dt, exclude); } @@ -59,8 +57,7 @@ void run_ncomp_convergence_test( runner.assert_all_convergence(); } -template <typename LoweredCell> -void validate_ball_and_stick() { +void validate_ball_and_stick(nest::mc::backend_policy backend) { using namespace nest::mc; cell c = make_cell_ball_and_stick(); @@ -73,15 +70,15 @@ void validate_ball_and_stick() { {"dend.end", {0u, 2u}, simple_sampler(sample_dt)} }; - run_ncomp_convergence_test<LoweredCell>( + run_ncomp_convergence_test( "ball_and_stick", "neuron_ball_and_stick.json", + backend, c, samplers); } -template <typename LoweredCell> -void validate_ball_and_taper() { +void validate_ball_and_taper(nest::mc::backend_policy backend) { using namespace nest::mc; cell c = make_cell_ball_and_taper(); @@ -94,15 +91,15 @@ void validate_ball_and_taper() { {"taper.end", {0u, 2u}, simple_sampler(sample_dt)} }; - run_ncomp_convergence_test<LoweredCell>( + run_ncomp_convergence_test( "ball_and_taper", "neuron_ball_and_taper.json", + backend, c, samplers); } -template <typename LoweredCell> -void validate_ball_and_3stick() { +void validate_ball_and_3stick(nest::mc::backend_policy backend) { using namespace nest::mc; cell c = make_cell_ball_and_3stick(); @@ -119,15 +116,15 @@ void validate_ball_and_3stick() { {"dend3.end", {0u, 6u}, simple_sampler(sample_dt)} }; - run_ncomp_convergence_test<LoweredCell>( + run_ncomp_convergence_test( "ball_and_3stick", "neuron_ball_and_3stick.json", + backend, c, samplers); } -template <typename LoweredCell> -void validate_rallpack1() { +void validate_rallpack1(nest::mc::backend_policy backend) { using namespace nest::mc; cell c = make_cell_simple_cable(); @@ -144,16 +141,16 @@ void validate_rallpack1() { {"cable.x1.0", {0u, 2u}, simple_sampler(sample_dt)}, }; - run_ncomp_convergence_test<LoweredCell>( + run_ncomp_convergence_test( "rallpack1", "numeric_rallpack1.json", + backend, c, samplers, 250.f); } -template <typename LoweredCell> -void validate_ball_and_squiggle() { +void validate_ball_and_squiggle(nest::mc::backend_policy backend) { using namespace nest::mc; cell c = make_cell_ball_and_squiggle(); @@ -177,9 +174,10 @@ void validate_ball_and_squiggle() { samplers); #endif - run_ncomp_convergence_test<LoweredCell>( + run_ncomp_convergence_test( "ball_and_squiggle_integrator", "neuron_ball_and_squiggle.json", + backend, c, samplers); } diff --git a/tests/validation/validate_kinetic.cpp b/tests/validation/validate_kinetic.cpp index 1c2005165dac0295542fdc2335147606efc0538a..b5d97b23ffb87dd3a0a26e1a4dfe1da5f103af37 100644 --- a/tests/validation/validate_kinetic.cpp +++ b/tests/validation/validate_kinetic.cpp @@ -2,12 +2,12 @@ #include "../gtest.h" -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::multicore::backend>; +const auto backend = nest::mc::backend_policy::use_multicore; TEST(kinetic, kin1_numeric_ref) { - validate_kinetic_kin1<lowered_cell>(); + validate_kinetic_kin1(backend); } TEST(kinetic, kinlva_numeric_ref) { - validate_kinetic_kinlva<lowered_cell>(); + validate_kinetic_kinlva(backend); } diff --git a/tests/validation/validate_kinetic.cu b/tests/validation/validate_kinetic.cu index 4c32cd9345582fa0d141fb6e82b099ddf8f7b0ce..a86847f82bd12f0871c0443b3bf40e743f66af7f 100644 --- a/tests/validation/validate_kinetic.cu +++ b/tests/validation/validate_kinetic.cu @@ -2,12 +2,12 @@ #include "../gtest.h" -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::gpu::backend>; +const auto backend = nest::mc::backend_policy::prefer_gpu; TEST(kinetic, kin1_numeric_ref) { - validate_kinetic_kin1<lowered_cell>(); + validate_kinetic_kin1(backend); } TEST(kinetic, kinlva_numeric_ref) { - validate_kinetic_kinlva<lowered_cell>(); + validate_kinetic_kinlva(backend); } diff --git a/tests/validation/validate_kinetic.hpp b/tests/validation/validate_kinetic.hpp index 7ccde7c93e5609fb12364ec308facecab498e2eb..535ed8697abf985893a00108435c51897575c332 100644 --- a/tests/validation/validate_kinetic.hpp +++ b/tests/validation/validate_kinetic.hpp @@ -13,8 +13,13 @@ #include "trace_analysis.hpp" #include "validation_data.hpp" -template <typename LoweredCell> -void run_kinetic_dt(nest::mc::cell& c, float t_end, nlohmann::json meta, const std::string& ref_file) { +void run_kinetic_dt( + nest::mc::backend_policy backend, + nest::mc::cell& c, + float t_end, + nlohmann::json meta, + const std::string& ref_file) +{ using namespace nest::mc; float sample_dt = .025f; @@ -23,11 +28,11 @@ void run_kinetic_dt(nest::mc::cell& c, float t_end, nlohmann::json meta, const s }; meta["sim"] = "nestmc"; - meta["backend"] = LoweredCell::backend::name(); + meta["backend_policy"] = to_string(backend); convergence_test_runner<float> runner("dt", samplers, meta); runner.load_reference_data(ref_file); - model<LoweredCell> model(singleton_recipe{c}); + model model(singleton_recipe{c}, backend); auto exclude = stimulus_ends(c); @@ -49,8 +54,7 @@ end: runner.assert_all_convergence(); } -template <typename LoweredCell> -void validate_kinetic_kin1() { +void validate_kinetic_kin1(nest::mc::backend_policy backend) { using namespace nest::mc; // 20 µm diameter soma with single mechanism, current probe @@ -65,11 +69,10 @@ void validate_kinetic_kin1() { {"units", "nA"} }; - run_kinetic_dt<LoweredCell>(c, 100.f, meta, "numeric_kin1.json"); + run_kinetic_dt(backend, c, 100.f, meta, "numeric_kin1.json"); } -template <typename LoweredCell> -void validate_kinetic_kinlva() { +void validate_kinetic_kinlva(nest::mc::backend_policy backend) { using namespace nest::mc; // 20 µm diameter soma with single mechanism, current probe @@ -85,6 +88,6 @@ void validate_kinetic_kinlva() { {"units", "mV"} }; - run_kinetic_dt<LoweredCell>(c, 300.f, meta, "numeric_kinlva.json"); + run_kinetic_dt(backend, c, 300.f, meta, "numeric_kinlva.json"); } diff --git a/tests/validation/validate_soma.cpp b/tests/validation/validate_soma.cpp index e4e58e7b665c05b0e2afdc99f76e0c4c545e905a..57b094b38a8026cf0ef589308aa3241dea5cbade 100644 --- a/tests/validation/validate_soma.cpp +++ b/tests/validation/validate_soma.cpp @@ -2,8 +2,9 @@ #include "../gtest.h" -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::multicore::backend>; + +const auto backend = nest::mc::backend_policy::use_multicore; TEST(soma, numeric_ref) { - validate_soma<lowered_cell>(); + validate_soma(backend); } diff --git a/tests/validation/validate_soma.cu b/tests/validation/validate_soma.cu index 35355ab93d66715fcf79ce27d064d2cdd119df07..b31b91a42b9a24e8d343ec7a8a1a451177a9f45f 100644 --- a/tests/validation/validate_soma.cu +++ b/tests/validation/validate_soma.cu @@ -2,8 +2,8 @@ #include "../gtest.h" -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::gpu::backend>; +const auto backend = nest::mc::backend_policy::prefer_gpu; TEST(soma, numeric_ref) { - validate_soma<lowered_cell>(); + validate_soma(backend); } diff --git a/tests/validation/validate_soma.hpp b/tests/validation/validate_soma.hpp index eee44e6bd280d63400ff103e441c38ece4c166a3..0c533faa044c9bf84bcfbbd279b77fbddcdc5846 100644 --- a/tests/validation/validate_soma.hpp +++ b/tests/validation/validate_soma.hpp @@ -13,13 +13,12 @@ #include "trace_analysis.hpp" #include "validation_data.hpp" -template <typename LoweredCell> -void validate_soma() { +void validate_soma(nest::mc::backend_policy backend) { using namespace nest::mc; cell c = make_cell_soma_only(); add_common_voltage_probes(c); - model<LoweredCell> model(singleton_recipe{c}); + model model(singleton_recipe{c}, backend); float sample_dt = .025f; sampler_info samplers[] = {{"soma.mid", {0u, 0u}, simple_sampler(sample_dt)}}; @@ -29,7 +28,7 @@ void validate_soma() { {"model", "soma"}, {"sim", "nestmc"}, {"units", "mV"}, - {"backend", LoweredCell::backend::name()} + {"backend_policy", to_string(backend)} }; convergence_test_runner<float> runner("dt", samplers, meta); diff --git a/tests/validation/validate_synapses.cpp b/tests/validation/validate_synapses.cpp index 23413cc91c44a8c890ee9635e6deada92db95e10..d1331b70a10976fe234495bdb9a886b3f9b29602 100644 --- a/tests/validation/validate_synapses.cpp +++ b/tests/validation/validate_synapses.cpp @@ -3,14 +3,14 @@ #include "../gtest.h" #include "validate_synapses.hpp" -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::multicore::backend>; +const auto backend = nest::mc::backend_policy::use_multicore; TEST(simple_synapse, expsyn_neuron_ref) { SCOPED_TRACE("expsyn"); - run_synapse_test<lowered_cell>("expsyn", "neuron_simple_exp_synapse.json"); + run_synapse_test("expsyn", "neuron_simple_exp_synapse.json", backend); } TEST(simple_synapse, exp2syn_neuron_ref) { SCOPED_TRACE("exp2syn"); - run_synapse_test<lowered_cell>("exp2syn", "neuron_simple_exp2_synapse.json"); + run_synapse_test("exp2syn", "neuron_simple_exp2_synapse.json", backend); } diff --git a/tests/validation/validate_synapses.cu b/tests/validation/validate_synapses.cu index 0dedd584268f77011908aa593886a5534965a482..47f0cc67230dd89e09660c66cbaa95f068664385 100644 --- a/tests/validation/validate_synapses.cu +++ b/tests/validation/validate_synapses.cu @@ -3,14 +3,14 @@ #include "../gtest.h" #include "validate_synapses.hpp" -using lowered_cell = nest::mc::fvm::fvm_multicell<nest::mc::gpu::backend>; +const auto backend = nest::mc::backend_policy::prefer_gpu; TEST(simple_synapse, expsyn_neuron_ref) { SCOPED_TRACE("expsyn"); - run_synapse_test<lowered_cell>("expsyn", "neuron_simple_exp_synapse.json"); + run_synapse_test("expsyn", "neuron_simple_exp_synapse.json", backend); } TEST(simple_synapse, exp2syn_neuron_ref) { SCOPED_TRACE("exp2syn"); - run_synapse_test<lowered_cell>("exp2syn", "neuron_simple_exp2_synapse.json"); + run_synapse_test("exp2syn", "neuron_simple_exp2_synapse.json", backend); } diff --git a/tests/validation/validate_synapses.hpp b/tests/validation/validate_synapses.hpp index cee900624b277da18f2b369a73af70b32bea42fd..c3f7f29413526747a982c62733b4e6ae08862969 100644 --- a/tests/validation/validate_synapses.hpp +++ b/tests/validation/validate_synapses.hpp @@ -15,10 +15,10 @@ #include "trace_analysis.hpp" #include "validation_data.hpp" -template <typename LoweredCell> void run_synapse_test( const char* syn_type, const nest::mc::util::path& ref_data_path, + nest::mc::backend_policy backend, float t_end=70.f, float dt=0.001) { @@ -30,7 +30,7 @@ void run_synapse_test( {"model", syn_type}, {"sim", "nestmc"}, {"units", "mV"}, - {"backend", LoweredCell::backend::name()} + {"backend_policy", to_string(backend)} }; cell c = make_cell_ball_and_stick(false); // no stimuli @@ -60,7 +60,7 @@ void run_synapse_test( for (int ncomp = 10; ncomp<max_ncomp; ncomp*=2) { c.cable(1)->set_compartments(ncomp); - model<LoweredCell> m(singleton_recipe{c}); + model m(singleton_recipe{c}, backend); m.group(0).enqueue_events(synthetic_events); runner.run(m, ncomp, t_end, dt, exclude);