1010#include "qat_bl.h"
1111#include "qat_crypto.h"
1212
13- void qat_bl_free_bufl (struct qat_crypto_instance * inst ,
14- struct qat_crypto_request * qat_req )
13+ void qat_bl_free_bufl (struct adf_accel_dev * accel_dev ,
14+ struct qat_crypto_request_buffs * buf )
1515{
16- struct device * dev = & GET_DEV (inst -> accel_dev );
17- struct qat_alg_buf_list * bl = qat_req -> buf . bl ;
18- struct qat_alg_buf_list * blout = qat_req -> buf . blout ;
19- dma_addr_t blp = qat_req -> buf . blp ;
20- dma_addr_t blpout = qat_req -> buf . bloutp ;
21- size_t sz = qat_req -> buf . sz ;
22- size_t sz_out = qat_req -> buf . sz_out ;
16+ struct device * dev = & GET_DEV (accel_dev );
17+ struct qat_alg_buf_list * bl = buf -> bl ;
18+ struct qat_alg_buf_list * blout = buf -> blout ;
19+ dma_addr_t blp = buf -> blp ;
20+ dma_addr_t blpout = buf -> bloutp ;
21+ size_t sz = buf -> sz ;
22+ size_t sz_out = buf -> sz_out ;
2323 int bl_dma_dir ;
2424 int i ;
2525
@@ -31,7 +31,7 @@ void qat_bl_free_bufl(struct qat_crypto_instance *inst,
3131
3232 dma_unmap_single (dev , blp , sz , DMA_TO_DEVICE );
3333
34- if (!qat_req -> buf . sgl_src_valid )
34+ if (!buf -> sgl_src_valid )
3535 kfree (bl );
3636
3737 if (blp != blpout ) {
@@ -45,18 +45,18 @@ void qat_bl_free_bufl(struct qat_crypto_instance *inst,
4545 }
4646 dma_unmap_single (dev , blpout , sz_out , DMA_TO_DEVICE );
4747
48- if (!qat_req -> buf . sgl_dst_valid )
48+ if (!buf -> sgl_dst_valid )
4949 kfree (blout );
5050 }
5151}
5252
53- int qat_bl_sgl_to_bufl (struct qat_crypto_instance * inst ,
53+ int qat_bl_sgl_to_bufl (struct adf_accel_dev * accel_dev ,
5454 struct scatterlist * sgl ,
5555 struct scatterlist * sglout ,
56- struct qat_crypto_request * qat_req ,
56+ struct qat_crypto_request_buffs * buf ,
5757 gfp_t flags )
5858{
59- struct device * dev = & GET_DEV (inst -> accel_dev );
59+ struct device * dev = & GET_DEV (accel_dev );
6060 int i , sg_nctr = 0 ;
6161 int n = sg_nents (sgl );
6262 struct qat_alg_buf_list * bufl ;
@@ -65,23 +65,23 @@ int qat_bl_sgl_to_bufl(struct qat_crypto_instance *inst,
6565 dma_addr_t bloutp = DMA_MAPPING_ERROR ;
6666 struct scatterlist * sg ;
6767 size_t sz_out , sz = struct_size (bufl , bufers , n );
68- int node = dev_to_node (& GET_DEV (inst -> accel_dev ));
68+ int node = dev_to_node (& GET_DEV (accel_dev ));
6969 int bufl_dma_dir ;
7070
7171 if (unlikely (!n ))
7272 return - EINVAL ;
7373
74- qat_req -> buf . sgl_src_valid = false;
75- qat_req -> buf . sgl_dst_valid = false;
74+ buf -> sgl_src_valid = false;
75+ buf -> sgl_dst_valid = false;
7676
7777 if (n > QAT_MAX_BUFF_DESC ) {
7878 bufl = kzalloc_node (sz , flags , node );
7979 if (unlikely (!bufl ))
8080 return - ENOMEM ;
8181 } else {
82- bufl = & qat_req -> buf . sgl_src .sgl_hdr ;
82+ bufl = & buf -> sgl_src .sgl_hdr ;
8383 memset (bufl , 0 , sizeof (struct qat_alg_buf_list ));
84- qat_req -> buf . sgl_src_valid = true;
84+ buf -> sgl_src_valid = true;
8585 }
8686
8787 bufl_dma_dir = sgl != sglout ? DMA_TO_DEVICE : DMA_BIDIRECTIONAL ;
@@ -107,9 +107,9 @@ int qat_bl_sgl_to_bufl(struct qat_crypto_instance *inst,
107107 blp = dma_map_single (dev , bufl , sz , DMA_TO_DEVICE );
108108 if (unlikely (dma_mapping_error (dev , blp )))
109109 goto err_in ;
110- qat_req -> buf . bl = bufl ;
111- qat_req -> buf . blp = blp ;
112- qat_req -> buf . sz = sz ;
110+ buf -> bl = bufl ;
111+ buf -> blp = blp ;
112+ buf -> sz = sz ;
113113 /* Handle out of place operation */
114114 if (sgl != sglout ) {
115115 struct qat_alg_buf * bufers ;
@@ -123,9 +123,9 @@ int qat_bl_sgl_to_bufl(struct qat_crypto_instance *inst,
123123 if (unlikely (!buflout ))
124124 goto err_in ;
125125 } else {
126- buflout = & qat_req -> buf . sgl_dst .sgl_hdr ;
126+ buflout = & buf -> sgl_dst .sgl_hdr ;
127127 memset (buflout , 0 , sizeof (struct qat_alg_buf_list ));
128- qat_req -> buf . sgl_dst_valid = true;
128+ buf -> sgl_dst_valid = true;
129129 }
130130
131131 bufers = buflout -> bufers ;
@@ -151,13 +151,13 @@ int qat_bl_sgl_to_bufl(struct qat_crypto_instance *inst,
151151 bloutp = dma_map_single (dev , buflout , sz_out , DMA_TO_DEVICE );
152152 if (unlikely (dma_mapping_error (dev , bloutp )))
153153 goto err_out ;
154- qat_req -> buf . blout = buflout ;
155- qat_req -> buf . bloutp = bloutp ;
156- qat_req -> buf . sz_out = sz_out ;
154+ buf -> blout = buflout ;
155+ buf -> bloutp = bloutp ;
156+ buf -> sz_out = sz_out ;
157157 } else {
158158 /* Otherwise set the src and dst to the same address */
159- qat_req -> buf . bloutp = qat_req -> buf . blp ;
160- qat_req -> buf . sz_out = 0 ;
159+ buf -> bloutp = buf -> blp ;
160+ buf -> sz_out = 0 ;
161161 }
162162 return 0 ;
163163
@@ -172,7 +172,7 @@ int qat_bl_sgl_to_bufl(struct qat_crypto_instance *inst,
172172 buflout -> bufers [i ].len ,
173173 DMA_FROM_DEVICE );
174174
175- if (!qat_req -> buf . sgl_dst_valid )
175+ if (!buf -> sgl_dst_valid )
176176 kfree (buflout );
177177
178178err_in :
@@ -186,7 +186,7 @@ int qat_bl_sgl_to_bufl(struct qat_crypto_instance *inst,
186186 bufl -> bufers [i ].len ,
187187 bufl_dma_dir );
188188
189- if (!qat_req -> buf . sgl_src_valid )
189+ if (!buf -> sgl_src_valid )
190190 kfree (bufl );
191191
192192 dev_err (dev , "Failed to map buf for dma\n" );
0 commit comments