/linux-master/fs/btrfs/ |
H A D | zlib.c | 23 /* workspace buffer size for s390 zlib hardware support */ 26 struct workspace { struct 39 struct workspace *workspace = list_entry(ws, struct workspace, list); local 41 workspace->level = level; 48 struct workspace *workspace = list_entry(ws, struct workspace, list); local 50 kvfree(workspace 57 struct workspace *workspace; local 98 struct workspace *workspace = list_entry(ws, struct workspace, list); local 273 struct workspace *workspace = list_entry(ws, struct workspace, list); local 360 struct workspace *workspace = list_entry(ws, struct workspace, list); local [all...] |
H A D | zstd.c | 43 struct workspace { struct 62 * workspace. 64 * Getting a workspace is done by using the bitmap to identify the levels that 66 * workspaces because of the monotonic memory guarantee. A workspace's 68 * level. Putting a workspace involves adding it back to the appropriate places 90 static inline struct workspace *list_to_workspace(struct list_head *list) 92 return container_of(list, struct workspace, list); 103 * This scans the lru_list and attempts to reclaim any workspace that hasn't 121 struct workspace *victim = container_of(pos, struct workspace, 203 struct workspace *workspace; local 236 struct workspace *workspace; local 314 struct workspace *workspace = list_to_workspace(ws); local 344 struct workspace *workspace = list_entry(ws, struct workspace, list); local 353 struct workspace *workspace; local 381 struct workspace *workspace = list_entry(ws, struct workspace, list); local 547 struct workspace *workspace = list_entry(ws, struct workspace, list); local 625 struct workspace *workspace = list_entry(ws, struct workspace, list); local [all...] |
H A D | lzo.c | 64 struct workspace { struct 75 struct workspace *workspace = list_entry(ws, struct workspace, list); local 77 kvfree(workspace->buf); 78 kvfree(workspace->cbuf); 79 kvfree(workspace->mem); 80 kfree(workspace); 85 struct workspace *workspace; local 216 struct workspace *workspace = list_entry(ws, struct workspace, list); local 332 struct workspace *workspace = list_entry(ws, struct workspace, list); local 431 struct workspace *workspace = list_entry(ws, struct workspace, list); local [all...] |
H A D | compression.c | 688 struct heuristic_ws *workspace; local 690 workspace = list_entry(ws, struct heuristic_ws, list); 692 kvfree(workspace->sample); 693 kfree(workspace->bucket); 694 kfree(workspace->bucket_b); 695 kfree(workspace); 772 struct list_head *workspace; local 781 * Preallocate one workspace for each compression type so we can 784 workspace = alloc_workspace(type, 0); 785 if (IS_ERR(workspace)) { 818 struct list_head *workspace; local 1005 struct list_head *workspace; local 1018 struct list_head *workspace; local 1040 struct list_head *workspace; local [all...] |
/linux-master/lib/zstd/ |
H A D | zstd_decompress_module.c | 47 zstd_dctx *zstd_init_dctx(void *workspace, size_t workspace_size) argument 49 if (workspace == NULL) 51 return ZSTD_initStaticDCtx(workspace, workspace_size); 68 zstd_dstream *zstd_init_dstream(size_t max_window_size, void *workspace, argument 71 if (workspace == NULL) 74 return ZSTD_initStaticDStream(workspace, workspace_size);
|
H A D | zstd_compress_module.c | 88 zstd_cctx *zstd_init_cctx(void *workspace, size_t workspace_size) argument 90 if (workspace == NULL) 92 return ZSTD_initStaticCCtx(workspace, workspace_size); 111 unsigned long long pledged_src_size, void *workspace, size_t workspace_size) 115 if (workspace == NULL) 118 cstream = ZSTD_initStaticCStream(workspace, workspace_size); 110 zstd_init_cstream(const zstd_parameters *parameters, unsigned long long pledged_src_size, void *workspace, size_t workspace_size) argument
|
/linux-master/lib/zlib_inflate/ |
H A D | infutil.h | 37 #define WS(strm) ((struct inflate_workspace *)(strm->workspace))
|
H A D | infutil.c | 20 strm->workspace = kmalloc(zlib_inflate_workspacesize(), GFP_KERNEL); 21 if (strm->workspace == NULL) 44 kfree(strm->workspace);
|
/linux-master/include/linux/ |
H A D | zstd.h | 151 * Return: A lower bound on the size of the workspace that is passed to 158 * @workspace: The workspace to emplace the context into. It must outlive 160 * @workspace_size: The size of workspace. Use zstd_cctx_workspace_bound() to 161 * determine how large the workspace must be. 165 zstd_cctx *zstd_init_cctx(void *workspace, size_t workspace_size); 190 * Return: A lower bound on the size of the workspace that is passed to 197 * @workspace: The workspace to emplace the context into. It must outlive 199 * @workspace_size: The size of workspace [all...] |
H A D | zlib.h | 96 void *workspace; /* memory allocated for this stream */ member in struct:z_stream_s 185 stream workspace with the specified parameters. A pointer to this 186 number of bytes should be returned in stream->workspace before 318 stream workspace. A pointer to this number of bytes should be 319 returned in stream->workspace before calling zlib_inflateInit(). 326 next_in, avail_in, and workspace must be initialized before by
|
/linux-master/fs/cramfs/ |
H A D | uncompress.c | 62 stream.workspace = vmalloc(zlib_inflate_workspacesize()); 63 if (!stream.workspace) { 78 vfree(stream.workspace);
|
/linux-master/fs/jffs2/ |
H A D | compr_zlib.c | 45 def_strm.workspace = vmalloc(zlib_deflate_workspacesize(MAX_WBITS, 47 if (!def_strm.workspace) 50 jffs2_dbg(1, "Allocated %d bytes for deflate workspace\n", 52 inf_strm.workspace = vmalloc(zlib_inflate_workspacesize()); 53 if (!inf_strm.workspace) { 54 vfree(def_strm.workspace); 57 jffs2_dbg(1, "Allocated %d bytes for inflate workspace\n", 64 vfree(def_strm.workspace); 65 vfree(inf_strm.workspace);
|
/linux-master/fs/squashfs/ |
H A D | zstd_wrapper.c | 23 struct workspace { struct 31 struct workspace *wksp = kmalloc(sizeof(*wksp), GFP_KERNEL); 45 ERROR("Failed to allocate zstd workspace\n"); 53 struct workspace *wksp = strm; 65 struct workspace *wksp = strm;
|
H A D | zlib_wrapper.c | 29 stream->workspace = vmalloc(zlib_inflate_workspacesize()); 30 if (stream->workspace == NULL) 36 ERROR("Failed to allocate zlib workspace\n"); 47 vfree(stream->workspace);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cudbg_zlib.c | 45 compress_stream.workspace = pdbg_init->workspace;
|
H A D | cudbg_if.h | 84 void *workspace; /* Workspace for zlib */ member in struct:cudbg_init
|
/linux-master/crypto/ |
H A D | deflate.c | 47 stream->workspace = vzalloc(zlib_deflate_workspacesize( 49 if (!stream->workspace) { 63 vfree(stream->workspace); 72 stream->workspace = vzalloc(zlib_inflate_workspacesize()); 73 if (!stream->workspace) { 85 vfree(stream->workspace); 92 vfree(ctx->comp_stream.workspace); 98 vfree(ctx->decomp_stream.workspace);
|
/linux-master/fs/bcachefs/ |
H A D | compress.c | 145 static inline void zlib_set_workspace(z_stream *strm, void *workspace) argument 148 strm->workspace = workspace; 158 void *workspace; local 179 workspace = mempool_alloc(&c->decompress_workspace, GFP_NOFS); 181 zlib_set_workspace(&strm, workspace); 185 mempool_free(workspace, &c->decompress_workspace); 198 workspace = mempool_alloc(&c->decompress_workspace, GFP_NOFS); 199 ctx = zstd_init_dctx(workspace, zstd_dctx_workspace_bound()); 205 mempool_free(workspace, 295 attempt_compress(struct bch_fs *c, void *workspace, void *dst, size_t dst_len, void *src, size_t src_len, struct bch_compression_opt compression) argument 391 void *workspace; local [all...] |
/linux-master/fs/pstore/ |
H A D | platform.c | 181 .workspace = compress_workspace, 238 pr_err("Failed to allocate zlib deflate workspace\n"); 587 char *unzipped, *workspace; local 600 if (!zstream->workspace) { 613 workspace = kvzalloc(max_uncompressed_size + record->ecc_notice_size, 615 if (!workspace) 620 zstream->next_out = workspace; 626 kvfree(workspace); 633 memcpy(workspace + unzipped_len, record->buf + record->size, 637 unzipped = kvmemdup(workspace, unzipped_le [all...] |
/linux-master/arch/arm/kernel/ |
H A D | head-inflate-data.c | 44 strm->workspace = &state;
|
/linux-master/lib/zstd/compress/ |
H A D | zstd_cwksp.h | 24 /* Since the workspace is effectively its own little malloc implementation / 26 * each internal element of the workspace, so ASAN will catch overruns that 27 * reach outside an object but that stay inside the workspace. 49 * Used to describe whether the workspace is statically allocated (and will not 62 * is called the workspace. 65 * from this workspace for each internal datastructure: 85 * - We would like to reuse the objects in the workspace for multiple 89 * - We would like to be able to efficiently reuse the workspace across 98 * [ ... workspace ... ] 101 * The various objects that live in the workspace ar 143 void* workspace; member in struct:__anon211 509 void* workspace = ZSTD_customMalloc(size, customMem); local [all...] |
H A D | zstd_compress_superblock.c | 305 void* workspace, size_t wkspSize, 308 unsigned* const countWksp = (unsigned*)workspace; 315 size_t const largest = HIST_count_wksp (countWksp, &maxSymbolValue, (const BYTE*)literals, litSize, workspace, wkspSize); 330 void* workspace, size_t wkspSize) 332 unsigned* const countWksp = (unsigned*)workspace; 339 HIST_countFast_wksp(countWksp, &max, codeTable, nbSeq, workspace, wkspSize); /* can't fail */ 366 void* workspace, size_t wkspSize, 375 workspace, wkspSize); 379 workspace, wkspSize); 383 workspace, wkspSiz 302 ZSTD_estimateSubBlockSize_literal(const BYTE* literals, size_t litSize, const ZSTD_hufCTables_t* huf, const ZSTD_hufCTablesMetadata_t* hufMetadata, void* workspace, size_t wkspSize, int writeEntropy) argument 325 ZSTD_estimateSubBlockSize_symbolType(symbolEncodingType_e type, const BYTE* codeTable, unsigned maxCode, size_t nbSeq, const FSE_CTable* fseCTable, const U8* additionalBits, short const* defaultNorm, U32 defaultNormLog, U32 defaultMax, void* workspace, size_t wkspSize) argument 360 ZSTD_estimateSubBlockSize_sequences(const BYTE* ofCodeTable, const BYTE* llCodeTable, const BYTE* mlCodeTable, size_t nbSeq, const ZSTD_fseCTables_t* fseTables, const ZSTD_fseCTablesMetadata_t* fseMetadata, void* workspace, size_t wkspSize, int writeEntropy) argument 388 ZSTD_estimateSubBlockSize(const BYTE* literals, size_t litSize, const BYTE* ofCodeTable, const BYTE* llCodeTable, const BYTE* mlCodeTable, size_t nbSeq, const ZSTD_entropyCTables_t* entropy, const ZSTD_entropyCTablesMetadata_t* entropyMetadata, void* workspace, size_t wkspSize, int writeLitEntropy, int writeSeqEntropy) argument 425 ZSTD_compressSubBlock_multi(const seqStore_t* seqStorePtr, const ZSTD_compressedBlockState_t* prevCBlock, ZSTD_compressedBlockState_t* nextCBlock, const ZSTD_entropyCTablesMetadata_t* entropyMetadata, const ZSTD_CCtx_params* cctxParams, void* dst, size_t dstCapacity, const void* src, size_t srcSize, const int bmi2, U32 lastBlock, void* workspace, size_t wkspSize) argument [all...] |
/linux-master/lib/ |
H A D | decompress_inflate.c | 82 strm->workspace = malloc(flush ? zlib_inflate_workspacesize() : 84 /* Always allocate the full workspace for DFLTCC */ 89 if (strm->workspace == NULL) { 90 error("Out of memory while allocating workspace"); 186 free(strm->workspace);
|
/linux-master/arch/xtensa/boot/lib/ |
H A D | zmem.c | 63 s.workspace = zalloc(zlib_inflate_workspacesize());
|
/linux-master/drivers/net/ppp/ |
H A D | ppp_deflate.c | 66 vfree(state->strm.workspace); 107 state->strm.workspace = vmalloc(zlib_deflate_workspacesize(-w_size, 8)); 108 if (state->strm.workspace == NULL) 282 vfree(state->strm.workspace); 322 state->strm.workspace = vmalloc(zlib_inflate_workspacesize()); 323 if (state->strm.workspace == NULL)
|