allocation.inline.hpp revision 4820:a837fa3d3f86
1/*
2 * Copyright (c) 1997, 2013, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25#ifndef SHARE_VM_MEMORY_ALLOCATION_INLINE_HPP
26#define SHARE_VM_MEMORY_ALLOCATION_INLINE_HPP
27
28#include "runtime/atomic.inline.hpp"
29#include "runtime/os.hpp"
30
31// Explicit C-heap memory management
32
33void trace_heap_malloc(size_t size, const char* name, void *p);
34void trace_heap_free(void *p);
35
36#ifndef PRODUCT
37// Increments unsigned long value for statistics (not atomic on MP).
38inline void inc_stat_counter(volatile julong* dest, julong add_value) {
39#if defined(SPARC) || defined(X86)
40  // Sparc and X86 have atomic jlong (8 bytes) instructions
41  julong value = Atomic::load((volatile jlong*)dest);
42  value += add_value;
43  Atomic::store((jlong)value, (volatile jlong*)dest);
44#else
45  // possible word-tearing during load/store
46  *dest += add_value;
47#endif
48}
49#endif
50
51// allocate using malloc; will fail if no memory available
52inline char* AllocateHeap(size_t size, MEMFLAGS flags, address pc = 0,
53    AllocFailType alloc_failmode = AllocFailStrategy::EXIT_OOM) {
54  if (pc == 0) {
55    pc = CURRENT_PC;
56  }
57  char* p = (char*) os::malloc(size, flags, pc);
58  #ifdef ASSERT
59  if (PrintMallocFree) trace_heap_malloc(size, "AllocateHeap", p);
60  #endif
61  if (p == NULL && alloc_failmode == AllocFailStrategy::EXIT_OOM) {
62    vm_exit_out_of_memory(size, OOM_MALLOC_ERROR, "AllocateHeap");
63  }
64  return p;
65}
66
67inline char* ReallocateHeap(char *old, size_t size, MEMFLAGS flags,
68    AllocFailType alloc_failmode = AllocFailStrategy::EXIT_OOM) {
69  char* p = (char*) os::realloc(old, size, flags, CURRENT_PC);
70  #ifdef ASSERT
71  if (PrintMallocFree) trace_heap_malloc(size, "ReallocateHeap", p);
72  #endif
73  if (p == NULL && alloc_failmode == AllocFailStrategy::EXIT_OOM) {
74    vm_exit_out_of_memory(size, OOM_MALLOC_ERROR, "ReallocateHeap");
75  }
76  return p;
77}
78
79inline void FreeHeap(void* p, MEMFLAGS memflags = mtInternal) {
80  #ifdef ASSERT
81  if (PrintMallocFree) trace_heap_free(p);
82  #endif
83  os::free(p, memflags);
84}
85
86
87template <MEMFLAGS F> void* CHeapObj<F>::operator new(size_t size,
88      address caller_pc){
89    void* p = (void*)AllocateHeap(size, F, (caller_pc != 0 ? caller_pc : CALLER_PC));
90#ifdef ASSERT
91    if (PrintMallocFree) trace_heap_malloc(size, "CHeapObj-new", p);
92#endif
93    return p;
94  }
95
96template <MEMFLAGS F> void* CHeapObj<F>::operator new (size_t size,
97  const std::nothrow_t&  nothrow_constant, address caller_pc) {
98  void* p = (void*)AllocateHeap(size, F, (caller_pc != 0 ? caller_pc : CALLER_PC),
99      AllocFailStrategy::RETURN_NULL);
100#ifdef ASSERT
101    if (PrintMallocFree) trace_heap_malloc(size, "CHeapObj-new", p);
102#endif
103    return p;
104}
105
106template <MEMFLAGS F> void* CHeapObj<F>::operator new [](size_t size,
107      address caller_pc){
108    return CHeapObj<F>::operator new(size, caller_pc);
109}
110
111template <MEMFLAGS F> void* CHeapObj<F>::operator new [](size_t size,
112  const std::nothrow_t&  nothrow_constant, address caller_pc) {
113    return CHeapObj<F>::operator new(size, nothrow_constant, caller_pc);
114}
115
116template <MEMFLAGS F> void CHeapObj<F>::operator delete(void* p){
117    FreeHeap(p, F);
118}
119
120template <MEMFLAGS F> void CHeapObj<F>::operator delete [](void* p){
121    FreeHeap(p, F);
122}
123
124template <class E, MEMFLAGS F>
125E* ArrayAllocator<E, F>::allocate(size_t length) {
126  assert(_addr == NULL, "Already in use");
127
128  _size = sizeof(E) * length;
129  _use_malloc = _size < ArrayAllocatorMallocLimit;
130
131  if (_use_malloc) {
132    _addr = AllocateHeap(_size, F);
133    if (_addr == NULL && _size >=  (size_t)os::vm_allocation_granularity()) {
134      // malloc failed let's try with mmap instead
135      _use_malloc = false;
136    } else {
137      return (E*)_addr;
138    }
139  }
140
141  int alignment = os::vm_allocation_granularity();
142  _size = align_size_up(_size, alignment);
143
144  _addr = os::reserve_memory(_size, NULL, alignment, F);
145  if (_addr == NULL) {
146    vm_exit_out_of_memory(_size, OOM_MMAP_ERROR, "Allocator (reserve)");
147  }
148
149  os::commit_memory_or_exit(_addr, _size, !ExecMem, "Allocator (commit)");
150
151  return (E*)_addr;
152}
153
154template<class E, MEMFLAGS F>
155void ArrayAllocator<E, F>::free() {
156  if (_addr != NULL) {
157    if (_use_malloc) {
158      FreeHeap(_addr, F);
159    } else {
160      os::release_memory(_addr, _size);
161    }
162    _addr = NULL;
163  }
164}
165
166#endif // SHARE_VM_MEMORY_ALLOCATION_INLINE_HPP
167