Loading [MathJax]/extensions/tex2jax.js
Neko 0.9.99
A portable framework for high-order spectral element flow simulations
All Classes Namespaces Files Functions Variables Typedefs Enumerator Macros Pages
fusedcg_aux.cu
Go to the documentation of this file.
1/*
2 Copyright (c) 2021-2025, The Neko Authors
3 All rights reserved.
4
5 Redistribution and use in source and binary forms, with or without
6 modification, are permitted provided that the following conditions
7 are met:
8
9 * Redistributions of source code must retain the above copyright
10 notice, this list of conditions and the following disclaimer.
11
12 * Redistributions in binary form must reproduce the above
13 copyright notice, this list of conditions and the following
14 disclaimer in the documentation and/or other materials provided
15 with the distribution.
16
17 * Neither the name of the authors nor the names of its
18 contributors may be used to endorse or promote products derived
19 from this software without specific prior written permission.
20
21 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
24 FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
25 COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
26 INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
27 BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
28 LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
29 CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30 LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
31 ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
32 POSSIBILITY OF SUCH DAMAGE.
33*/
34
35#include "fusedcg_kernel.h"
37#include <device/cuda/check.h>
38
39#ifdef HAVE_NVSHMEM
40#include <nvshmem.h>
41#include <nvshmemx.h>
42#endif
43
50
51extern "C" {
52
55
56#ifdef HAVE_NCCL
59#endif
60
61 void cuda_fusedcg_update_p(void *p, void *z, void *po, real *beta, int *n) {
62
63 const dim3 nthrds(1024, 1, 1);
64 const dim3 nblcks(((*n)+1024 - 1)/ 1024, 1, 1);
65 const cudaStream_t stream = (cudaStream_t) glb_cmd_queue;
66
68 <<<nblcks, nthrds, 0, stream>>>((real *) p, (real *) z,
69 (real *) po, *beta, *n);
71
72 }
73
74 void cuda_fusedcg_update_x(void *x, void *p, void *alpha, int *p_cur, int *n) {
75
76 const dim3 nthrds(1024, 1, 1);
77 const dim3 nblcks(((*n)+1024 - 1)/ 1024, 1, 1);
78 const cudaStream_t stream = (cudaStream_t) glb_cmd_queue;
79
81 <<<nblcks, nthrds, 0, stream>>>((real *) x, (const real **) p,
82 (const real *) alpha, *p_cur, *n);
84 }
85
86
87 real cuda_fusedcg_part2(void *a, void *b, void *c,
88 void *alpha_d , real *alpha, int *p_cur, int * n) {
89
90 const dim3 nthrds(1024, 1, 1);
91 const dim3 nblcks(((*n)+1024 - 1)/ 1024, 1, 1);
92 const int nb = ((*n) + 1024 - 1)/ 1024;
93 const cudaStream_t stream = (cudaStream_t) glb_cmd_queue;
94
95 if (fusedcg_buf != NULL && fusedcg_buf_len < nb) {
97#ifdef HAVE_NVSHMEM
99#else
101#endif
103 }
104
105 if (fusedcg_buf == NULL){
107#ifdef HAVE_NVSHMEM
108 fusedcg_buf_d = (real *) nvshmem_malloc(nb*sizeof(real));
109#else
111#endif
113 }
114
115 /* Store alpha(p_cur) in pinned memory */
116 fusedcg_buf[1] = (*alpha);
117
118 /* Update alpha_d(p_cur) = alpha(p_cur) */
119 real *alpha_d_p_cur = ((real *) alpha_d) + ((*p_cur - 1));
121 cudaMemcpyHostToDevice, stream));
122
123
125 <<<nblcks, nthrds, 0, stream>>>((real *) a, (real *) b,
126 (real *) c, *alpha,
127 (real *) fusedcg_buf_d, * n);
129
130 reduce_kernel<real><<<1, 1024, 0, stream>>>((real *) fusedcg_buf_d, nb);
132
133#ifdef HAVE_NCCL
135 DEVICE_NCCL_SUM, stream);
137 cudaMemcpyDeviceToHost, stream));
138 cudaStreamSynchronize(stream);
139#elif HAVE_NVSHMEM
140 if (sizeof(real) == sizeof(float)) {
142 (float *) fusedcg_buf_d,
143 (float *) fusedcg_buf_d, 1, stream);
144 }
145 else if (sizeof(real) == sizeof(double)) {
147 (double *) fusedcg_buf_d,
148 (double *) fusedcg_buf_d, 1, stream);
149
150 }
152 sizeof(real), cudaMemcpyDeviceToHost, stream));
153 cudaStreamSynchronize(stream);
154#elif HAVE_DEVICE_MPI
155 cudaStreamSynchronize(stream);
157 sizeof(real), DEVICE_MPI_SUM);
158#else
160 cudaMemcpyDeviceToHost, stream));
161 cudaStreamSynchronize(stream);
162#endif
163
164 return fusedcg_buf[0];
165 }
166}
__global__ void const T *__restrict__ x
#define CUDA_CHECK(err)
Definition check.h:6
__global__ void dirichlet_apply_scalar_kernel(const int *__restrict__ msk, T *__restrict__ x, const T g, const int m)
double real
#define DEVICE_MPI_SUM
void device_mpi_allreduce(void *buf_d, void *buf, int count, int nbytes, int op)
#define DEVICE_NCCL_SUM
void device_nccl_allreduce(void *sbuf_d, void *rbuf_d, int count, int nbytes, int op, void *stream)
real cuda_fusedcg_part2(void *a, void *b, void *c, void *alpha_d, real *alpha, int *p_cur, int *n)
int fusedcg_buf_len
real * fusedcg_buf
void * fusedcg_buf_d
void cuda_fusedcg_update_x(void *x, void *p, void *alpha, int *p_cur, int *n)
void cuda_fusedcg_update_p(void *p, void *z, void *po, real *beta, int *n)