Deleted Added
sdiff udiff text old ( 0:a61af66fc99e ) new ( 113:ba764ed4b6f2 )
full compact
1/*
2 * Copyright 1997-2006 Sun Microsystems, Inc. All Rights Reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *

--- 11 unchanged lines hidden (view full) ---

20 * CA 95054 USA or visit www.sun.com if you need additional information or
21 * have any questions.
22 *
23 */
24
25# include "incls/_precompiled.incl"
26# include "incls/_instanceRefKlass.cpp.incl"
27
28template <class T>
29static void specialized_oop_follow_contents(instanceRefKlass* ref, oop obj) {
30 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
31 oop referent = oopDesc::load_decode_heap_oop(referent_addr);
32 debug_only(
33 if(TraceReferenceGC && PrintGCDetails) {
34 gclog_or_tty->print_cr("instanceRefKlass::oop_follow_contents " INTPTR_FORMAT, obj);
35 }
36 )
37 if (referent != NULL) {
38 if (!referent->is_gc_marked() &&
39 MarkSweep::ref_processor()->
40 discover_reference(obj, ref->reference_type())) {
41 // reference already enqueued, referent will be traversed later
42 ref->instanceKlass::oop_follow_contents(obj);
43 debug_only(
44 if(TraceReferenceGC && PrintGCDetails) {
45 gclog_or_tty->print_cr(" Non NULL enqueued " INTPTR_FORMAT, obj);
46 }
47 )
48 return;
49 } else {
50 // treat referent as normal oop
51 debug_only(
52 if(TraceReferenceGC && PrintGCDetails) {
53 gclog_or_tty->print_cr(" Non NULL normal " INTPTR_FORMAT, obj);
54 }
55 )
56 MarkSweep::mark_and_push(referent_addr);
57 }
58 }
59 // treat next as normal oop. next is a link in the pending list.
60 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
61 debug_only(
62 if(TraceReferenceGC && PrintGCDetails) {
63 gclog_or_tty->print_cr(" Process next as normal " INTPTR_FORMAT, next_addr);
64 }
65 )
66 MarkSweep::mark_and_push(next_addr);
67 ref->instanceKlass::oop_follow_contents(obj);
68}
69
70void instanceRefKlass::oop_follow_contents(oop obj) {
71 if (UseCompressedOops) {
72 specialized_oop_follow_contents<narrowOop>(this, obj);
73 } else {
74 specialized_oop_follow_contents<oop>(this, obj);
75 }
76}
77
78#ifndef SERIALGC
79template <class T>
80static void specialized_oop_follow_contents(instanceRefKlass* ref,
81 ParCompactionManager* cm,
82 oop obj) {
83 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
84 oop referent = oopDesc::load_decode_heap_oop(referent_addr);
85 debug_only(
86 if(TraceReferenceGC && PrintGCDetails) {
87 gclog_or_tty->print_cr("instanceRefKlass::oop_follow_contents " INTPTR_FORMAT, obj);
88 }
89 )
90 if (referent != NULL) {
91 if (PSParallelCompact::mark_bitmap()->is_unmarked(referent) &&
92 PSParallelCompact::ref_processor()->
93 discover_reference(obj, ref->reference_type())) {
94 // reference already enqueued, referent will be traversed later
95 ref->instanceKlass::oop_follow_contents(cm, obj);
96 debug_only(
97 if(TraceReferenceGC && PrintGCDetails) {
98 gclog_or_tty->print_cr(" Non NULL enqueued " INTPTR_FORMAT, obj);
99 }
100 )
101 return;
102 } else {
103 // treat referent as normal oop
104 debug_only(
105 if(TraceReferenceGC && PrintGCDetails) {
106 gclog_or_tty->print_cr(" Non NULL normal " INTPTR_FORMAT, obj);
107 }
108 )
109 PSParallelCompact::mark_and_push(cm, referent_addr);
110 }
111 }
112 // treat next as normal oop. next is a link in the pending list.
113 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
114 debug_only(
115 if(TraceReferenceGC && PrintGCDetails) {
116 gclog_or_tty->print_cr(" Process next as normal " INTPTR_FORMAT, next_addr);
117 }
118 )
119 PSParallelCompact::mark_and_push(cm, next_addr);
120 ref->instanceKlass::oop_follow_contents(cm, obj);
121}
122
123void instanceRefKlass::oop_follow_contents(ParCompactionManager* cm,
124 oop obj) {
125 if (UseCompressedOops) {
126 specialized_oop_follow_contents<narrowOop>(this, cm, obj);
127 } else {
128 specialized_oop_follow_contents<oop>(this, cm, obj);
129 }
130}
131#endif // SERIALGC
132
133#ifdef ASSERT
134template <class T> void trace_reference_gc(const char *s, oop obj,
135 T* referent_addr,
136 T* next_addr,
137 T* discovered_addr) {
138 if(TraceReferenceGC && PrintGCDetails) {
139 gclog_or_tty->print_cr("%s obj " INTPTR_FORMAT, s, (address)obj);
140 gclog_or_tty->print_cr(" referent_addr/* " INTPTR_FORMAT " / "
141 INTPTR_FORMAT, referent_addr,
142 referent_addr ?
143 (address)oopDesc::load_decode_heap_oop(referent_addr) : NULL);
144 gclog_or_tty->print_cr(" next_addr/* " INTPTR_FORMAT " / "
145 INTPTR_FORMAT, next_addr,
146 next_addr ? (address)oopDesc::load_decode_heap_oop(next_addr) : NULL);
147 gclog_or_tty->print_cr(" discovered_addr/* " INTPTR_FORMAT " / "
148 INTPTR_FORMAT, discovered_addr,
149 discovered_addr ?
150 (address)oopDesc::load_decode_heap_oop(discovered_addr) : NULL);
151 }
152}
153#endif
154
155template <class T> void specialized_oop_adjust_pointers(instanceRefKlass *ref, oop obj) {
156 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
157 MarkSweep::adjust_pointer(referent_addr);
158 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
159 MarkSweep::adjust_pointer(next_addr);
160 T* discovered_addr = (T*)java_lang_ref_Reference::discovered_addr(obj);
161 MarkSweep::adjust_pointer(discovered_addr);
162 debug_only(trace_reference_gc("instanceRefKlass::oop_adjust_pointers", obj,
163 referent_addr, next_addr, discovered_addr);)
164}
165
166int instanceRefKlass::oop_adjust_pointers(oop obj) {
167 int size = size_helper();
168 instanceKlass::oop_adjust_pointers(obj);
169
170 if (UseCompressedOops) {
171 specialized_oop_adjust_pointers<narrowOop>(this, obj);
172 } else {
173 specialized_oop_adjust_pointers<oop>(this, obj);
174 }
175 return size;
176}
177
178#define InstanceRefKlass_SPECIALIZED_OOP_ITERATE(T, nv_suffix, contains) \
179 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj); \
180 oop referent = oopDesc::load_decode_heap_oop(referent_addr); \
181 if (referent != NULL && contains(referent_addr)) { \
182 ReferenceProcessor* rp = closure->_ref_processor; \
183 if (!referent->is_gc_marked() && (rp != NULL) && \
184 rp->discover_reference(obj, reference_type())) { \
185 return size; \
186 } else { \
187 /* treat referent as normal oop */ \
188 SpecializationStats::record_do_oop_call##nv_suffix(SpecializationStats::irk);\
189 closure->do_oop##nv_suffix(referent_addr); \
190 } \
191 } \
192 /* treat next as normal oop */ \
193 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj); \
194 if (contains(next_addr)) { \
195 SpecializationStats::record_do_oop_call##nv_suffix(SpecializationStats::irk); \
196 closure->do_oop##nv_suffix(next_addr); \
197 } \
198 return size; \
199
200
201template <class T> bool contains(T *t) { return true; }
202
203// Macro to define instanceRefKlass::oop_oop_iterate for virtual/nonvirtual for
204// all closures. Macros calling macros above for each oop size.
205
206#define InstanceRefKlass_OOP_OOP_ITERATE_DEFN(OopClosureType, nv_suffix) \
207 \
208int instanceRefKlass:: \
209oop_oop_iterate##nv_suffix(oop obj, OopClosureType* closure) { \
210 /* Get size before changing pointers */ \
211 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk);\
212 \
213 int size = instanceKlass::oop_oop_iterate##nv_suffix(obj, closure); \
214 \
215 if (UseCompressedOops) { \
216 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, contains); \
217 } else { \
218 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, contains); \
219 } \
220}
221
222#define InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m(OopClosureType, nv_suffix) \
223 \
224int instanceRefKlass:: \
225oop_oop_iterate##nv_suffix##_m(oop obj, \
226 OopClosureType* closure, \
227 MemRegion mr) { \
228 SpecializationStats::record_iterate_call##nv_suffix(SpecializationStats::irk);\
229 \
230 int size = instanceKlass::oop_oop_iterate##nv_suffix##_m(obj, closure, mr); \
231 if (UseCompressedOops) { \
232 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(narrowOop, nv_suffix, mr.contains); \
233 } else { \
234 InstanceRefKlass_SPECIALIZED_OOP_ITERATE(oop, nv_suffix, mr.contains); \
235 } \
236}
237
238ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceRefKlass_OOP_OOP_ITERATE_DEFN)
239ALL_OOP_OOP_ITERATE_CLOSURES_3(InstanceRefKlass_OOP_OOP_ITERATE_DEFN)
240ALL_OOP_OOP_ITERATE_CLOSURES_1(InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m)
241ALL_OOP_OOP_ITERATE_CLOSURES_3(InstanceRefKlass_OOP_OOP_ITERATE_DEFN_m)
242
243#ifndef SERIALGC
244template <class T>
245void specialized_oop_copy_contents(instanceRefKlass *ref,
246 PSPromotionManager* pm, oop obj) {
247 assert(!pm->depth_first(), "invariant");
248 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
249 if (PSScavenge::should_scavenge(referent_addr)) {
250 ReferenceProcessor* rp = PSScavenge::reference_processor();
251 if (rp->discover_reference(obj, ref->reference_type())) {
252 // reference already enqueued, referent and next will be traversed later
253 ref->instanceKlass::oop_copy_contents(pm, obj);
254 return;
255 } else {
256 // treat referent as normal oop
257 pm->claim_or_forward_breadth(referent_addr);
258 }
259 }
260 // treat next as normal oop
261 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
262 if (PSScavenge::should_scavenge(next_addr)) {
263 pm->claim_or_forward_breadth(next_addr);
264 }
265 ref->instanceKlass::oop_copy_contents(pm, obj);
266}
267
268void instanceRefKlass::oop_copy_contents(PSPromotionManager* pm, oop obj) {
269 if (UseCompressedOops) {
270 specialized_oop_copy_contents<narrowOop>(this, pm, obj);
271 } else {
272 specialized_oop_copy_contents<oop>(this, pm, obj);
273 }
274}
275
276template <class T>
277void specialized_oop_push_contents(instanceRefKlass *ref,
278 PSPromotionManager* pm, oop obj) {
279 assert(pm->depth_first(), "invariant");
280 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
281 if (PSScavenge::should_scavenge(referent_addr)) {
282 ReferenceProcessor* rp = PSScavenge::reference_processor();
283 if (rp->discover_reference(obj, ref->reference_type())) {
284 // reference already enqueued, referent and next will be traversed later
285 ref->instanceKlass::oop_push_contents(pm, obj);
286 return;
287 } else {
288 // treat referent as normal oop
289 pm->claim_or_forward_depth(referent_addr);
290 }
291 }
292 // treat next as normal oop
293 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
294 if (PSScavenge::should_scavenge(next_addr)) {
295 pm->claim_or_forward_depth(next_addr);
296 }
297 ref->instanceKlass::oop_push_contents(pm, obj);
298}
299
300void instanceRefKlass::oop_push_contents(PSPromotionManager* pm, oop obj) {
301 if (UseCompressedOops) {
302 specialized_oop_push_contents<narrowOop>(this, pm, obj);
303 } else {
304 specialized_oop_push_contents<oop>(this, pm, obj);
305 }
306}
307
308template <class T>
309void specialized_oop_update_pointers(instanceRefKlass *ref,
310 ParCompactionManager* cm, oop obj) {
311 T* referent_addr = (T*)java_lang_ref_Reference::referent_addr(obj);
312 PSParallelCompact::adjust_pointer(referent_addr);
313 T* next_addr = (T*)java_lang_ref_Reference::next_addr(obj);
314 PSParallelCompact::adjust_pointer(next_addr);
315 T* discovered_addr = (T*)java_lang_ref_Reference::discovered_addr(obj);
316 PSParallelCompact::adjust_pointer(discovered_addr);
317 debug_only(trace_reference_gc("instanceRefKlass::oop_update_ptrs", obj,
318 referent_addr, next_addr, discovered_addr);)
319}
320
321int instanceRefKlass::oop_update_pointers(ParCompactionManager* cm, oop obj) {
322 instanceKlass::oop_update_pointers(cm, obj);
323 if (UseCompressedOops) {
324 specialized_oop_update_pointers<narrowOop>(this, cm, obj);
325 } else {
326 specialized_oop_update_pointers<oop>(this, cm, obj);
327 }
328 return size_helper();
329}
330
331
332template <class T> void
333specialized_oop_update_pointers(ParCompactionManager* cm, oop obj,
334 HeapWord* beg_addr, HeapWord* end_addr) {
335 T* p;
336 T* referent_addr = p = (T*)java_lang_ref_Reference::referent_addr(obj);
337 PSParallelCompact::adjust_pointer(p, beg_addr, end_addr);
338 T* next_addr = p = (T*)java_lang_ref_Reference::next_addr(obj);
339 PSParallelCompact::adjust_pointer(p, beg_addr, end_addr);
340 T* discovered_addr = p = (T*)java_lang_ref_Reference::discovered_addr(obj);
341 PSParallelCompact::adjust_pointer(p, beg_addr, end_addr);
342 debug_only(trace_reference_gc("instanceRefKlass::oop_update_ptrs", obj,
343 referent_addr, next_addr, discovered_addr);)
344}
345
346int
347instanceRefKlass::oop_update_pointers(ParCompactionManager* cm, oop obj,
348 HeapWord* beg_addr, HeapWord* end_addr) {
349 instanceKlass::oop_update_pointers(cm, obj, beg_addr, end_addr);
350 if (UseCompressedOops) {
351 specialized_oop_update_pointers<narrowOop>(cm, obj, beg_addr, end_addr);
352 } else {
353 specialized_oop_update_pointers<oop>(cm, obj, beg_addr, end_addr);
354 }
355 return size_helper();
356}
357#endif // SERIALGC
358
359void instanceRefKlass::update_nonstatic_oop_maps(klassOop k) {
360 // Clear the nonstatic oop-map entries corresponding to referent
361 // and nextPending field. They are treated specially by the
362 // garbage collector.

--- 9 unchanged lines hidden (view full) ---

372 assert(ik->nonstatic_oop_map_size() == 1, "just checking");
373
374 OopMapBlock* map = ik->start_of_nonstatic_oop_maps();
375
376 // Check that the current map is (2,4) - currently points at field with
377 // offset 2 (words) and has 4 map entries.
378 debug_only(int offset = java_lang_ref_Reference::referent_offset);
379 debug_only(int length = ((java_lang_ref_Reference::discovered_offset -
380 java_lang_ref_Reference::referent_offset)/heapOopSize) + 1);
381
382 if (UseSharedSpaces) {
383 assert(map->offset() == java_lang_ref_Reference::queue_offset &&
384 map->length() == 1, "just checking");
385 } else {
386 assert(map->offset() == offset && map->length() == length,
387 "just checking");
388

--- 13 unchanged lines hidden (view full) ---

402
403 // We should make this general to all heaps
404 GenCollectedHeap* gch = NULL;
405 if (Universe::heap()->kind() == CollectedHeap::GenCollectedHeap)
406 gch = GenCollectedHeap::heap();
407
408 if (referent != NULL) {
409 guarantee(referent->is_oop(), "referent field heap failed");
410 if (gch != NULL && !gch->is_in_youngest(obj)) {
411 // We do a specific remembered set check here since the referent
412 // field is not part of the oop mask and therefore skipped by the
413 // regular verify code.
414 if (UseCompressedOops) {
415 narrowOop* referent_addr = (narrowOop*)java_lang_ref_Reference::referent_addr(obj);
416 obj->verify_old_oop(referent_addr, true);
417 } else {
418 oop* referent_addr = (oop*)java_lang_ref_Reference::referent_addr(obj);
419 obj->verify_old_oop(referent_addr, true);
420 }
421 }
422 }
423 // Verify next field
424 oop next = java_lang_ref_Reference::next(obj);
425 if (next != NULL) {
426 guarantee(next->is_oop(), "next field verify fa iled");
427 guarantee(next->is_instanceRef(), "next field verify failed");
428 if (gch != NULL && !gch->is_in_youngest(obj)) {
429 // We do a specific remembered set check here since the next field is
430 // not part of the oop mask and therefore skipped by the regular
431 // verify code.
432 if (UseCompressedOops) {
433 narrowOop* next_addr = (narrowOop*)java_lang_ref_Reference::next_addr(obj);
434 obj->verify_old_oop(next_addr, true);
435 } else {
436 oop* next_addr = (oop*)java_lang_ref_Reference::next_addr(obj);
437 obj->verify_old_oop(next_addr, true);
438 }
439 }
440 }
441}
442
443void instanceRefKlass::acquire_pending_list_lock(BasicLock *pending_list_basic_lock) {
444 // we may enter this with pending exception set
445 PRESERVE_EXCEPTION_MARK; // exceptions are never thrown, needed for TRAPS argument
446 Handle h_lock(THREAD, java_lang_ref_Reference::pending_list_lock());

--- 23 unchanged lines hidden ---