Go to the documentation of this file.
43 const std::vector<int>& mech_types,
44 const std::vector<int>& nodecounts);
160 template <
typename F,
typename... Args>
165 #pragma omp parallel for private(i) shared(nrn_threads, job, nrn_nthread, \
166 nrnmpi_myid) schedule(static, 1)
183 extern void nrn_ba(NrnThread*,
int);
195 extern void nonvint(NrnThread* _nt);
196 extern void update(NrnThread*);
199 double x = te - 1e-11;
200 if (x <= nt->_t && x > (nt->
_t - nt->
_dt)) {
NrnFastImem * nrn_fast_imem
void nrncore2nrn_send_init()
void nrn_fixed_single_steps_minimal(int total_sim_steps, double tstop)
--> Coreneuron
size_t * _fornetcon_weight_perm
void dt2thread(double adt)
int _net_send_buffer_size
void nrn_ba(NrnThread *nt, int bat)
for gpu builds with unified memory support
PreSynHelper * presyns_helper
void * setup_tree_matrix_minimal(NrnThread *)
void nrn_solve_minimal(NrnThread *)
void nrn_threads_create(int n)
std::size_t _fornetcon_weight_perm_size
void direct_mode_initialize()
All state from NEURON necessary to continue a run.
std::vector< int > _pnt_offset
void nrn_mk_table_check()
THIS FILE IS AUTO GENERATED DONT MODIFY IT.
TrajectoryRequests * trajec_requests
void nrn_fixed_step_minimal()
NrnThreadBAList * tbl[BEFORE_AFTER_SIZE]
Memb_list * _ecell_memb_list
void update(NrnThread *_nt)
void nrn_multithread_job(F &&job, Args &&... args)
constexpr int at_time(NrnThread *nt, double te)
std::size_t _fornetcon_perm_indices_size
NrnThreadMembList * create_tml(NrnThread &nt, int mech_id, Memb_func &memb_func, int &shadow_rhs_cnt, const std::vector< int > &mech_types, const std::vector< int > &nodecounts)
void nrncore2nrn_send_values(NrnThread *nth)
size_t * _fornetcon_perm_indices
void nrn_fixed_step_group_minimal(int total_sim_steps)
#define BEFORE_AFTER_SIZE
bool use_solve_interleave
void nonvint(NrnThread *_nt)
void * nrn_fixed_step_lastpart(NrnThread *nth)
std::unique_ptr< SummationReportMapping > summation_report_handler_
void nrn_thread_table_check()
void nrn_finitialize(int setv, double v)