@@ -33,7 +33,7 @@ static void nrn_rhs(NrnThread* _nt) {
3333 int * parent_index = _nt->_v_parent_index ;
3434
3535 nrn_pragma_acc (parallel loop present (vec_rhs [0 :i3], vec_d [0 :i3]) if (_nt->compute_gpu )
36- async (_nt->streams [_nt->stream_id ])
36+ async (_nt->streams [_nt->stream_id ]))
3737 nrn_pragma_omp (target teams distribute parallel for if (_nt->compute_gpu ))
3838 for (int i = i1; i < i3; ++i) {
3939 vec_rhs[i] = 0 .;
@@ -45,7 +45,7 @@ static void nrn_rhs(NrnThread* _nt) {
4545 double * fast_imem_rhs = _nt->nrn_fast_imem ->nrn_sav_rhs ;
4646 nrn_pragma_acc (
4747 parallel loop present (fast_imem_d [i1:i3], fast_imem_rhs [i1:i3]) if (_nt->compute_gpu )
48- async (_nt->streams [_nt->stream_id ])
48+ async (_nt->streams [_nt->stream_id ]))
4949 nrn_pragma_omp (target teams distribute parallel for if (_nt->compute_gpu ))
5050 for (int i = i1; i < i3; ++i) {
5151 fast_imem_d[i] = 0 .;
@@ -75,7 +75,7 @@ static void nrn_rhs(NrnThread* _nt) {
7575 */
7676 double * p = _nt->nrn_fast_imem ->nrn_sav_rhs ;
7777 nrn_pragma_acc (parallel loop present (p, vec_rhs) if (_nt->compute_gpu )
78- async (_nt->streams [_nt->stream_id ])
78+ async (_nt->streams [_nt->stream_id ]))
7979 nrn_pragma_omp (target teams distribute parallel for if (_nt->compute_gpu ))
8080 for (int i = i1; i < i3; ++i) {
8181 p[i] -= vec_rhs[i];
@@ -92,7 +92,7 @@ static void nrn_rhs(NrnThread* _nt) {
9292 vec_b [0 :i3],
9393 vec_v [0 :i3],
9494 parent_index [0 :i3]) if (_nt->compute_gpu )
95- async (_nt->streams [_nt->stream_id ])
95+ async (_nt->streams [_nt->stream_id ]))
9696 nrn_pragma_omp (target teams distribute parallel for if (_nt->compute_gpu ))
9797 for (int i = i2; i < i3; ++i) {
9898 double dv = vec_v[parent_index[i]] - vec_v[i];
@@ -152,7 +152,7 @@ static void nrn_lhs(NrnThread* _nt) {
152152 so here we transform so it only has membrane current contribution
153153 */
154154 double * p = _nt->nrn_fast_imem ->nrn_sav_d ;
155- nrn_pragma_acc (parallel loop present (p, vec_d) if (_nt->compute_gpu ) async (_nt->streams [_nt->stream_id ])
155+ nrn_pragma_acc (parallel loop present (p, vec_d) if (_nt->compute_gpu ) async (_nt->streams [_nt->stream_id ]))
156156 nrn_pragma_omp (target teams distribute parallel for if (_nt->compute_gpu ))
157157 for (int i = i1; i < i3; ++i) {
158158 p[i] += vec_d[i];
@@ -162,7 +162,7 @@ static void nrn_lhs(NrnThread* _nt) {
162162 /* now add the axial currents */
163163 nrn_pragma_acc (parallel loop present (
164164 vec_d [0 :i3], vec_a [0 :i3], vec_b [0 :i3], parent_index [0 :i3]) if (_nt->compute_gpu )
165- async (_nt->streams [_nt->stream_id ])
165+ async (_nt->streams [_nt->stream_id ]))
166166 nrn_pragma_omp (target teams distribute parallel for if (_nt->compute_gpu ))
167167 for (int i = i2; i < i3; ++i) {
168168 nrn_pragma_acc (atomic update)
0 commit comments