1 // Copyright 2012-2013 The Rust Project Developers. See the COPYRIGHT
2 // file at the top-level directory of this distribution and at
3 // http://rust-lang.org/COPYRIGHT.
5 // Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or
6 // http://www.apache.org/licenses/LICENSE-2.0> or the MIT license
7 // <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your
8 // option. This file may not be copied, modified, or distributed
9 // except according to those terms.
13 // Code relating to drop glue.
19 use llvm
::{ValueRef, get_param}
;
20 use metadata
::csearch
;
21 use middle
::lang_items
::ExchangeFreeFnLangItem
;
23 use middle
::subst
::{Subst, Substs}
;
24 use middle
::ty
::{self, Ty}
;
26 use trans
::adt
::GetDtorType
; // for tcx.dtor_type()
31 use trans
::cleanup
::CleanupMethods
;
33 use trans
::debuginfo
::DebugLoc
;
38 use trans
::machine
::*;
39 use trans
::monomorphize
;
40 use trans
::type_of
::{type_of, type_of_dtor, sizing_type_of, align_of}
;
41 use trans
::type_
::Type
;
43 use arena
::TypedArena
;
47 pub fn trans_exchange_free_dyn
<'blk
, 'tcx
>(cx
: Block
<'blk
, 'tcx
>,
52 -> Block
<'blk
, 'tcx
> {
53 let _icx
= push_ctxt("trans_exchange_free");
55 callee
::trans_lang_call(cx
,
56 langcall(cx
, None
, "", ExchangeFreeFnLangItem
),
57 &[PointerCast(cx
, v
, Type
::i8p(ccx
)), size
, align
],
62 pub fn trans_exchange_free
<'blk
, 'tcx
>(cx
: Block
<'blk
, 'tcx
>,
67 -> Block
<'blk
, 'tcx
> {
68 trans_exchange_free_dyn(cx
,
70 C_uint(cx
.ccx(), size
),
71 C_uint(cx
.ccx(), align
),
75 pub fn trans_exchange_free_ty
<'blk
, 'tcx
>(bcx
: Block
<'blk
, 'tcx
>,
79 -> Block
<'blk
, 'tcx
> {
80 assert
!(type_is_sized(bcx
.ccx().tcx(), content_ty
));
81 let sizing_type
= sizing_type_of(bcx
.ccx(), content_ty
);
82 let content_size
= llsize_of_alloc(bcx
.ccx(), sizing_type
);
84 // `Box<ZeroSizeType>` does not allocate.
85 if content_size
!= 0 {
86 let content_align
= align_of(bcx
.ccx(), content_ty
);
87 trans_exchange_free(bcx
, ptr
, content_size
, content_align
, debug_loc
)
93 pub fn get_drop_glue_type
<'a
, 'tcx
>(ccx
: &CrateContext
<'a
, 'tcx
>,
94 t
: Ty
<'tcx
>) -> Ty
<'tcx
> {
96 // Even if there is no dtor for t, there might be one deeper down and we
97 // might need to pass in the vtable ptr.
98 if !type_is_sized(tcx
, t
) {
102 // FIXME (#22815): note that type_needs_drop conservatively
103 // approximates in some cases and may say a type expression
104 // requires drop glue when it actually does not.
106 // (In this case it is not clear whether any harm is done, i.e.
107 // erroneously returning `t` in some cases where we could have
108 // returned `tcx.types.i8` does not appear unsound. The impact on
109 // code quality is unknown at this time.)
111 if !type_needs_drop(tcx
, t
) {
115 ty
::TyBox(typ
) if !type_needs_drop(tcx
, typ
)
116 && type_is_sized(tcx
, typ
) => {
117 let llty
= sizing_type_of(ccx
, typ
);
118 // `Box<ZeroSizeType>` does not allocate.
119 if llsize_of_alloc(ccx
, llty
) == 0 {
129 pub fn drop_ty
<'blk
, 'tcx
>(bcx
: Block
<'blk
, 'tcx
>,
132 debug_loc
: DebugLoc
) -> Block
<'blk
, 'tcx
> {
133 drop_ty_core(bcx
, v
, t
, debug_loc
, false, None
)
136 pub fn drop_ty_core
<'blk
, 'tcx
>(bcx
: Block
<'blk
, 'tcx
>,
141 drop_hint
: Option
<cleanup
::DropHintValue
>)
142 -> Block
<'blk
, 'tcx
> {
143 // NB: v is an *alias* of type t here, not a direct value.
144 debug
!("drop_ty_core(t={:?}, skip_dtor={} drop_hint={:?})", t
, skip_dtor
, drop_hint
);
145 let _icx
= push_ctxt("drop_ty");
147 if bcx
.fcx
.type_needs_drop(t
) {
149 let g
= if skip_dtor
{
150 DropGlueKind
::TyContents(t
)
154 let glue
= get_drop_glue_core(ccx
, g
);
155 let glue_type
= get_drop_glue_type(ccx
, t
);
156 let ptr
= if glue_type
!= t
{
157 PointerCast(bcx
, v
, type_of(ccx
, glue_type
).ptr_to())
164 let hint_val
= load_ty(bcx
, drop_hint
.value(), bcx
.tcx().types
.u8);
166 C_integral(Type
::i8(bcx
.ccx()), adt
::DTOR_MOVED_HINT
as u64, false);
168 ICmp(bcx
, llvm
::IntNE
, hint_val
, moved_val
, DebugLoc
::None
);
169 bcx
= with_cond(bcx
, may_need_drop
, |cx
| {
170 Call(cx
, glue
, &[ptr
], None
, debug_loc
);
175 // No drop-hint ==> call standard drop glue
176 Call(bcx
, glue
, &[ptr
], None
, debug_loc
);
183 pub fn drop_ty_immediate
<'blk
, 'tcx
>(bcx
: Block
<'blk
, 'tcx
>,
188 -> Block
<'blk
, 'tcx
> {
189 let _icx
= push_ctxt("drop_ty_immediate");
190 let vp
= alloca(bcx
, type_of(bcx
.ccx(), t
), "");
191 store_ty(bcx
, v
, vp
, t
);
192 drop_ty_core(bcx
, vp
, t
, debug_loc
, skip_dtor
, None
)
195 pub fn get_drop_glue
<'a
, 'tcx
>(ccx
: &CrateContext
<'a
, 'tcx
>, t
: Ty
<'tcx
>) -> ValueRef
{
196 get_drop_glue_core(ccx
, DropGlueKind
::Ty(t
))
199 #[derive(Copy, Clone, PartialEq, Eq, Hash, Debug)]
200 pub enum DropGlueKind
<'tcx
> {
201 /// The normal path; runs the dtor, and then recurs on the contents
203 /// Skips the dtor, if any, for ty; drops the contents directly.
204 /// Note that the dtor is only skipped at the most *shallow*
205 /// level, namely, an `impl Drop for Ty` itself. So, for example,
206 /// if Ty is Newtype(S) then only the Drop impl for for Newtype
207 /// itself will be skipped, while the Drop impl for S, if any,
209 TyContents(Ty
<'tcx
>),
212 impl<'tcx
> DropGlueKind
<'tcx
> {
213 fn ty(&self) -> Ty
<'tcx
> {
214 match *self { DropGlueKind::Ty(t) | DropGlueKind::TyContents(t) => t }
217 fn map_ty
<F
>(&self, mut f
: F
) -> DropGlueKind
<'tcx
> where F
: FnMut(Ty
<'tcx
>) -> Ty
<'tcx
>
220 DropGlueKind
::Ty(t
) => DropGlueKind
::Ty(f(t
)),
221 DropGlueKind
::TyContents(t
) => DropGlueKind
::TyContents(f(t
)),
226 fn get_drop_glue_core
<'a
, 'tcx
>(ccx
: &CrateContext
<'a
, 'tcx
>,
227 g
: DropGlueKind
<'tcx
>) -> ValueRef
{
228 debug
!("make drop glue for {:?}", g
);
229 let g
= g
.map_ty(|t
| get_drop_glue_type(ccx
, t
));
230 debug
!("drop glue type {:?}", g
);
231 match ccx
.drop_glues().borrow().get(&g
) {
232 Some(&glue
) => return glue
,
237 let llty
= if type_is_sized(ccx
.tcx(), t
) {
238 type_of(ccx
, t
).ptr_to()
240 type_of(ccx
, ccx
.tcx().mk_box(t
)).ptr_to()
243 let llfnty
= Type
::glue_fn(ccx
, llty
);
245 // To avoid infinite recursion, don't `make_drop_glue` until after we've
246 // added the entry to the `drop_glues` cache.
247 if let Some(old_sym
) = ccx
.available_drop_glues().borrow().get(&g
) {
248 let llfn
= declare
::declare_cfn(ccx
, &old_sym
, llfnty
, ccx
.tcx().mk_nil());
249 ccx
.drop_glues().borrow_mut().insert(g
, llfn
);
253 let fn_nm
= mangle_internal_name_by_type_and_seq(ccx
, t
, "drop");
254 let llfn
= declare
::define_cfn(ccx
, &fn_nm
, llfnty
, ccx
.tcx().mk_nil()).unwrap_or_else(||{
255 ccx
.sess().bug(&format
!("symbol `{}` already defined", fn_nm
));
257 ccx
.available_drop_glues().borrow_mut().insert(g
, fn_nm
);
259 let _s
= StatRecorder
::new(ccx
, format
!("drop {:?}", t
));
261 let empty_substs
= ccx
.tcx().mk_substs(Substs
::trans_empty());
262 let (arena
, fcx
): (TypedArena
<_
>, FunctionContext
);
263 arena
= TypedArena
::new();
264 fcx
= new_fn_ctxt(ccx
, llfn
, ast
::DUMMY_NODE_ID
, false,
265 ty
::FnConverging(ccx
.tcx().mk_nil()),
266 empty_substs
, None
, &arena
);
268 let bcx
= init_function(&fcx
, false, ty
::FnConverging(ccx
.tcx().mk_nil()));
270 update_linkage(ccx
, llfn
, None
, OriginalTranslation
);
272 ccx
.stats().n_glues_created
.set(ccx
.stats().n_glues_created
.get() + 1);
273 // All glue functions take values passed *by alias*; this is a
274 // requirement since in many contexts glue is invoked indirectly and
275 // the caller has no idea if it's dealing with something that can be
278 // llfn is expected be declared to take a parameter of the appropriate
279 // type, so we don't need to explicitly cast the function parameter.
281 let llrawptr0
= get_param(llfn
, fcx
.arg_offset() as c_uint
);
282 let bcx
= make_drop_glue(bcx
, llrawptr0
, g
);
283 finish_fn(&fcx
, bcx
, ty
::FnConverging(ccx
.tcx().mk_nil()), DebugLoc
::None
);
288 fn trans_struct_drop_flag
<'blk
, 'tcx
>(mut bcx
: Block
<'blk
, 'tcx
>,
290 struct_data
: ValueRef
,
291 dtor_did
: ast
::DefId
,
292 class_did
: ast
::DefId
,
293 substs
: &subst
::Substs
<'tcx
>)
294 -> Block
<'blk
, 'tcx
> {
295 assert
!(type_is_sized(bcx
.tcx(), t
), "Precondition: caller must ensure t is sized");
297 let repr
= adt
::represent_type(bcx
.ccx(), t
);
298 let drop_flag
= unpack_datum
!(bcx
, adt
::trans_drop_flag_ptr(bcx
, &*repr
, struct_data
));
299 let loaded
= load_ty(bcx
, drop_flag
.val
, bcx
.tcx().dtor_type());
300 let drop_flag_llty
= type_of(bcx
.fcx
.ccx
, bcx
.tcx().dtor_type());
301 let init_val
= C_integral(drop_flag_llty
, adt
::DTOR_NEEDED
as u64, false);
303 let bcx
= if !bcx
.ccx().check_drop_flag_for_sanity() {
306 let drop_flag_llty
= type_of(bcx
.fcx
.ccx
, bcx
.tcx().dtor_type());
307 let done_val
= C_integral(drop_flag_llty
, adt
::DTOR_DONE
as u64, false);
308 let not_init
= ICmp(bcx
, llvm
::IntNE
, loaded
, init_val
, DebugLoc
::None
);
309 let not_done
= ICmp(bcx
, llvm
::IntNE
, loaded
, done_val
, DebugLoc
::None
);
310 let drop_flag_neither_initialized_nor_cleared
=
311 And(bcx
, not_init
, not_done
, DebugLoc
::None
);
312 with_cond(bcx
, drop_flag_neither_initialized_nor_cleared
, |cx
| {
313 let llfn
= cx
.ccx().get_intrinsic(&("llvm.debugtrap"));
314 Call(cx
, llfn
, &[], None
, DebugLoc
::None
);
319 let drop_flag_dtor_needed
= ICmp(bcx
, llvm
::IntEQ
, loaded
, init_val
, DebugLoc
::None
);
320 with_cond(bcx
, drop_flag_dtor_needed
, |cx
| {
321 trans_struct_drop(cx
, t
, struct_data
, dtor_did
, class_did
, substs
)
325 pub fn get_res_dtor
<'a
, 'tcx
>(ccx
: &CrateContext
<'a
, 'tcx
>,
328 parent_id
: ast
::DefId
,
329 substs
: &Substs
<'tcx
>)
331 let _icx
= push_ctxt("trans_res_dtor");
332 let did
= inline
::maybe_instantiate_inline(ccx
, did
);
334 if !substs
.types
.is_empty() {
335 assert_eq
!(did
.krate
, ast
::LOCAL_CRATE
);
337 // Since we're in trans we don't care for any region parameters
338 let substs
= ccx
.tcx().mk_substs(Substs
::erased(substs
.types
.clone()));
340 let (val
, _
, _
) = monomorphize
::monomorphic_fn(ccx
, did
, substs
, None
);
343 } else if did
.krate
== ast
::LOCAL_CRATE
{
344 get_item_val(ccx
, did
.node
)
347 let name
= csearch
::get_symbol(&ccx
.sess().cstore
, did
);
348 let class_ty
= tcx
.lookup_item_type(parent_id
).ty
.subst(tcx
, substs
);
349 let llty
= type_of_dtor(ccx
, class_ty
);
350 let dtor_ty
= ccx
.tcx().mk_ctor_fn(did
,
351 &[get_drop_glue_type(ccx
, t
)],
353 foreign
::get_extern_fn(ccx
, &mut *ccx
.externs().borrow_mut(), &name
[..], llvm
::CCallConv
,
358 fn trans_struct_drop
<'blk
, 'tcx
>(bcx
: Block
<'blk
, 'tcx
>,
361 dtor_did
: ast
::DefId
,
362 class_did
: ast
::DefId
,
363 substs
: &subst
::Substs
<'tcx
>)
366 debug
!("trans_struct_drop t: {}", t
);
368 // Find and call the actual destructor
369 let dtor_addr
= get_res_dtor(bcx
.ccx(), dtor_did
, t
, class_did
, substs
);
371 // Class dtors have no explicit args, so the params should
372 // just consist of the environment (self).
373 let params
= unsafe {
374 let ty
= Type
::from_ref(llvm
::LLVMTypeOf(dtor_addr
));
375 ty
.element_type().func_params()
377 assert_eq
!(params
.len(), if type_is_sized(bcx
.tcx(), t
) { 1 }
else { 2 }
);
379 // Be sure to put the contents into a scope so we can use an invoke
380 // instruction to call the user destructor but still call the field
381 // destructors if the user destructor panics.
383 // FIXME (#14875) panic-in-drop semantics might be unsupported; we
384 // might well consider changing below to more direct code.
385 let contents_scope
= bcx
.fcx
.push_custom_cleanup_scope();
387 // Issue #23611: schedule cleanup of contents, re-inspecting the
388 // discriminant (if any) in case of variant swap in drop code.
389 bcx
.fcx
.schedule_drop_adt_contents(cleanup
::CustomScope(contents_scope
), v0
, t
);
391 let glue_type
= get_drop_glue_type(bcx
.ccx(), t
);
392 let dtor_ty
= bcx
.tcx().mk_ctor_fn(class_did
, &[glue_type
], bcx
.tcx().mk_nil());
393 let (_
, bcx
) = if type_is_sized(bcx
.tcx(), t
) {
394 invoke(bcx
, dtor_addr
, &[v0
], dtor_ty
, DebugLoc
::None
)
396 let args
= [Load(bcx
, expr
::get_dataptr(bcx
, v0
)), Load(bcx
, expr
::get_len(bcx
, v0
))];
397 invoke(bcx
, dtor_addr
, &args
, dtor_ty
, DebugLoc
::None
)
400 bcx
.fcx
.pop_and_trans_custom_cleanup_scope(bcx
, contents_scope
)
403 pub fn size_and_align_of_dst
<'blk
, 'tcx
>(bcx
: Block
<'blk
, 'tcx
>, t
: Ty
<'tcx
>, info
: ValueRef
)
404 -> (ValueRef
, ValueRef
) {
405 debug
!("calculate size of DST: {}; with lost info: {}",
406 t
, bcx
.val_to_string(info
));
407 if type_is_sized(bcx
.tcx(), t
) {
408 let sizing_type
= sizing_type_of(bcx
.ccx(), t
);
409 let size
= C_uint(bcx
.ccx(), llsize_of_alloc(bcx
.ccx(), sizing_type
));
410 let align
= C_uint(bcx
.ccx(), align_of(bcx
.ccx(), t
));
411 return (size
, align
);
414 ty
::TyStruct(id
, substs
) => {
416 // First get the size of all statically known fields.
417 // Don't use type_of::sizing_type_of because that expects t to be sized.
418 assert
!(!t
.is_simd(bcx
.tcx()));
419 let repr
= adt
::represent_type(ccx
, t
);
420 let sizing_type
= adt
::sizing_type_of(ccx
, &*repr
, true);
421 let sized_size
= C_uint(ccx
, llsize_of_alloc(ccx
, sizing_type
));
422 let sized_align
= C_uint(ccx
, llalign_of_min(ccx
, sizing_type
));
424 // Recurse to get the size of the dynamically sized field (must be
426 let fields
= bcx
.tcx().struct_fields(id
, substs
);
427 let last_field
= fields
[fields
.len()-1];
428 let field_ty
= last_field
.mt
.ty
;
429 let (unsized_size
, unsized_align
) = size_and_align_of_dst(bcx
, field_ty
, info
);
431 // Return the sum of sizes and max of aligns.
432 let size
= Add(bcx
, sized_size
, unsized_size
, DebugLoc
::None
);
433 let align
= Select(bcx
,
444 // info points to the vtable and the second entry in the vtable is the
445 // dynamic size of the object.
446 let info
= PointerCast(bcx
, info
, Type
::int(bcx
.ccx()).ptr_to());
447 let size_ptr
= GEPi(bcx
, info
, &[1]);
448 let align_ptr
= GEPi(bcx
, info
, &[2]);
449 (Load(bcx
, size_ptr
), Load(bcx
, align_ptr
))
451 ty
::TySlice(_
) | ty
::TyStr
=> {
452 let unit_ty
= t
.sequence_element_type(bcx
.tcx());
453 // The info in this case is the length of the str, so the size is that
454 // times the unit size.
455 let llunit_ty
= sizing_type_of(bcx
.ccx(), unit_ty
);
456 let unit_align
= llalign_of_min(bcx
.ccx(), llunit_ty
);
457 let unit_size
= llsize_of_alloc(bcx
.ccx(), llunit_ty
);
458 (Mul(bcx
, info
, C_uint(bcx
.ccx(), unit_size
), DebugLoc
::None
),
459 C_uint(bcx
.ccx(), unit_align
))
461 _
=> bcx
.sess().bug(&format
!("Unexpected unsized type, found {}", t
))
465 fn make_drop_glue
<'blk
, 'tcx
>(bcx
: Block
<'blk
, 'tcx
>, v0
: ValueRef
, g
: DropGlueKind
<'tcx
>)
466 -> Block
<'blk
, 'tcx
> {
468 let skip_dtor
= match g { DropGlueKind::Ty(_) => false, DropGlueKind::TyContents(_) => true }
;
469 // NB: v0 is an *alias* of type t here, not a direct value.
470 let _icx
= push_ctxt("make_drop_glue");
472 // Only drop the value when it ... well, we used to check for
473 // non-null, (and maybe we need to continue doing so), but we now
474 // must definitely check for special bit-patterns corresponding to
475 // the special dtor markings.
477 let inttype
= Type
::int(bcx
.ccx());
478 let dropped_pattern
= C_integral(inttype
, adt
::dtor_done_usize(bcx
.fcx
.ccx
) as u64, false);
481 ty
::TyBox(content_ty
) => {
482 // Support for TyBox is built-in and its drop glue is
483 // special. It may move to library and have Drop impl. As
484 // a safe-guard, assert TyBox not used with TyContents.
486 if !type_is_sized(bcx
.tcx(), content_ty
) {
487 let llval
= GEPi(bcx
, v0
, &[0, abi
::FAT_PTR_ADDR
]);
488 let llbox
= Load(bcx
, llval
);
489 let llbox_as_usize
= PtrToInt(bcx
, llbox
, Type
::int(bcx
.ccx()));
490 let drop_flag_not_dropped_already
=
491 ICmp(bcx
, llvm
::IntNE
, llbox_as_usize
, dropped_pattern
, DebugLoc
::None
);
492 with_cond(bcx
, drop_flag_not_dropped_already
, |bcx
| {
493 let bcx
= drop_ty(bcx
, v0
, content_ty
, DebugLoc
::None
);
494 let info
= GEPi(bcx
, v0
, &[0, abi
::FAT_PTR_EXTRA
]);
495 let info
= Load(bcx
, info
);
496 let (llsize
, llalign
) = size_and_align_of_dst(bcx
, content_ty
, info
);
498 // `Box<ZeroSizeType>` does not allocate.
499 let needs_free
= ICmp(bcx
,
502 C_uint(bcx
.ccx(), 0u64),
504 with_cond(bcx
, needs_free
, |bcx
| {
505 trans_exchange_free_dyn(bcx
, llbox
, llsize
, llalign
, DebugLoc
::None
)
510 let llbox
= Load(bcx
, llval
);
511 let llbox_as_usize
= PtrToInt(bcx
, llbox
, inttype
);
512 let drop_flag_not_dropped_already
=
513 ICmp(bcx
, llvm
::IntNE
, llbox_as_usize
, dropped_pattern
, DebugLoc
::None
);
514 with_cond(bcx
, drop_flag_not_dropped_already
, |bcx
| {
515 let bcx
= drop_ty(bcx
, llbox
, content_ty
, DebugLoc
::None
);
516 trans_exchange_free_ty(bcx
, llbox
, content_ty
, DebugLoc
::None
)
520 ty
::TyStruct(did
, substs
) | ty
::TyEnum(did
, substs
) => {
522 match (tcx
.ty_dtor(did
), skip_dtor
) {
523 (ty
::TraitDtor(dtor
, true), false) => {
524 // FIXME(16758) Since the struct is unsized, it is hard to
525 // find the drop flag (which is at the end of the struct).
526 // Lets just ignore the flag and pretend everything will be
528 if type_is_sized(bcx
.tcx(), t
) {
529 trans_struct_drop_flag(bcx
, t
, v0
, dtor
, did
, substs
)
531 // Give the user a heads up that we are doing something
532 // stupid and dangerous.
533 bcx
.sess().warn(&format
!("Ignoring drop flag in destructor for {}\
534 because the struct is unsized. See issue\
536 trans_struct_drop(bcx
, t
, v0
, dtor
, did
, substs
)
539 (ty
::TraitDtor(dtor
, false), false) => {
540 trans_struct_drop(bcx
, t
, v0
, dtor
, did
, substs
)
542 (ty
::NoDtor
, _
) | (_
, true) => {
543 // No dtor? Just the default case
544 iter_structural_ty(bcx
, v0
, t
, |bb
, vv
, tt
| drop_ty(bb
, vv
, tt
, DebugLoc
::None
))
549 // No support in vtable for distinguishing destroying with
550 // versus without calling Drop::drop. Assert caller is
551 // okay with always calling the Drop impl, if any.
553 let data_ptr
= GEPi(bcx
, v0
, &[0, abi
::FAT_PTR_ADDR
]);
554 let vtable_ptr
= Load(bcx
, GEPi(bcx
, v0
, &[0, abi
::FAT_PTR_EXTRA
]));
555 let dtor
= Load(bcx
, vtable_ptr
);
558 &[PointerCast(bcx
, Load(bcx
, data_ptr
), Type
::i8p(bcx
.ccx()))],
564 if bcx
.fcx
.type_needs_drop(t
) {
565 iter_structural_ty(bcx
,
568 |bb
, vv
, tt
| drop_ty(bb
, vv
, tt
, DebugLoc
::None
))