@@ -1014,7 +1014,7 @@ static void jl_compilation_sig(
1014
1014
}
1015
1015
}
1016
1016
else if (jl_is_kind (elt )) {
1017
- // not triggered for isdispatchtuple (tt), this attempts to handle
1017
+ // not triggered for isindivisibletype (tt), this attempts to handle
1018
1018
// some cases of adapting a random signature into a compilation signature
1019
1019
// if we get a kind, where we don't expect to accept one, widen it to something more expected (Type{T})
1020
1020
if (!(jl_subtype (elt , type_i ) && !jl_subtype ((jl_value_t * )jl_type_type , type_i ))) {
@@ -1044,11 +1044,11 @@ static void jl_compilation_sig(
1044
1044
}
1045
1045
1046
1046
if (jl_types_equal (elt , (jl_value_t * )jl_type_type )) { // elt == Type{T} where T
1047
- // not triggered for isdispatchtuple (tt), this attempts to handle
1047
+ // not triggered for isindivisibletype (tt), this attempts to handle
1048
1048
// some cases of adapting a random signature into a compilation signature
1049
1049
}
1050
1050
else if (!jl_is_datatype (elt ) && jl_subtype (elt , (jl_value_t * )jl_type_type )) { // elt <: Type{T}
1051
- // not triggered for isdispatchtuple (tt), this attempts to handle
1051
+ // not triggered for isindivisibletype (tt), this attempts to handle
1052
1052
// some cases of adapting a random signature into a compilation signature
1053
1053
if (!* newparams ) * newparams = jl_svec_copy (tt -> parameters );
1054
1054
jl_svecset (* newparams , i , jl_type_type );
@@ -1197,7 +1197,7 @@ JL_DLLEXPORT int jl_isa_compileable_sig(
1197
1197
if (definition -> generator ) {
1198
1198
// staged functions aren't optimized
1199
1199
// so assume the caller was intelligent about calling us
1200
- return (definition -> isva ? np >= nargs - 1 : np == nargs ) && type -> isdispatchtuple ;
1200
+ return (definition -> isva ? np >= nargs - 1 : np == nargs ) && type -> isindivisibletype ;
1201
1201
}
1202
1202
1203
1203
// for varargs methods, only specialize up to max_args (>= nargs + 1).
@@ -1636,9 +1636,9 @@ static jl_method_instance_t *jl_mt_assoc_by_type(jl_methtable_t *mt JL_PROPAGATE
1636
1636
jl_method_match_t * matc = NULL ;
1637
1637
JL_GC_PUSH2 (& tt , & matc );
1638
1638
JL_LOCK (& mt -> writelock );
1639
- assert (tt -> isdispatchtuple || tt -> hasfreetypevars );
1639
+ assert (tt -> isindivisibletype || tt -> hasfreetypevars );
1640
1640
jl_method_instance_t * mi = NULL ;
1641
- if (tt -> isdispatchtuple ) {
1641
+ if (tt -> isindivisibletype ) {
1642
1642
jl_genericmemory_t * leafcache = jl_atomic_load_relaxed (& mt -> leafcache );
1643
1643
jl_typemap_entry_t * entry = lookup_leafcache (leafcache , (jl_value_t * )tt , world );
1644
1644
if (entry )
@@ -3132,7 +3132,7 @@ JL_DLLEXPORT jl_value_t *jl_normalize_to_compilable_sig(jl_methtable_t *mt, jl_t
3132
3132
jl_methtable_t * kwmt = mt == jl_kwcall_mt ? jl_kwmethod_table_for (m -> sig ) : mt ;
3133
3133
intptr_t max_varargs = get_max_varargs (m , kwmt , mt , NULL );
3134
3134
jl_compilation_sig (ti , env , m , max_varargs , & newparams );
3135
- int is_compileable = ((jl_datatype_t * )ti )-> isdispatchtuple ;
3135
+ int is_compileable = ((jl_datatype_t * )ti )-> isindivisibletype ;
3136
3136
if (newparams ) {
3137
3137
tt = (jl_datatype_t * )jl_apply_tuple_type (newparams , 1 );
3138
3138
if (!is_compileable ) {
@@ -3184,7 +3184,7 @@ JL_DLLEXPORT jl_method_instance_t *jl_method_match_to_mi(jl_method_match_t *matc
3184
3184
assert (mt != NULL );
3185
3185
if ((jl_value_t * )mt != jl_nothing ) {
3186
3186
// get the specialization, possibly also caching it
3187
- if (mt_cache && ((jl_datatype_t * )ti )-> isdispatchtuple ) {
3187
+ if (mt_cache && ((jl_datatype_t * )ti )-> isindivisibletype ) {
3188
3188
// Since we also use this presence in the cache
3189
3189
// to trigger compilation when producing `.ji` files,
3190
3190
// inject it there now if we think it will be
@@ -3334,7 +3334,7 @@ JL_DLLEXPORT void jl_compile_method_instance(jl_method_instance_t *mi, jl_tuplet
3334
3334
// In addition to full compilation of the compilation-signature, if `types` is more specific (e.g. due to nospecialize),
3335
3335
// also run inference now on the original `types`, since that may help us guide inference to find
3336
3336
// additional useful methods that should be compiled
3337
- //ALT: if (jl_is_datatype(types) && ((jl_datatype_t*)types)->isdispatchtuple && !jl_egal(mi->specTypes, types))
3337
+ //ALT: if (jl_is_datatype(types) && ((jl_datatype_t*)types)->isindivisibletype && !jl_egal(mi->specTypes, types))
3338
3338
//ALT: if (jl_subtype(types, mi->specTypes))
3339
3339
if (types && !jl_subtype (mi -> specTypes , (jl_value_t * )types )) {
3340
3340
jl_svec_t * tpenv2 = jl_emptysvec ;
@@ -3882,7 +3882,7 @@ static int ml_matches_visitor(jl_typemap_entry_t *ml, struct typemap_intersectio
3882
3882
return 1 ;
3883
3883
}
3884
3884
jl_method_t * meth = ml -> func .method ;
3885
- if (closure -> lim >= 0 && jl_is_dispatch_tupletype (meth -> sig )) {
3885
+ if (closure -> lim >= 0 && jl_is_indivisible_type (meth -> sig )) {
3886
3886
int replaced = 0 ;
3887
3887
// check if this is replaced, in which case we need to avoid double-counting it against the limit
3888
3888
// (although it will figure out later which one to keep and return)
@@ -4233,7 +4233,7 @@ static jl_value_t *ml_matches(jl_methtable_t *mt,
4233
4233
4234
4234
if (mt ) {
4235
4235
// check the leaf cache if this type can be in there
4236
- if (((jl_datatype_t * )unw )-> isdispatchtuple ) {
4236
+ if (((jl_datatype_t * )unw )-> isindivisibletype ) {
4237
4237
jl_genericmemory_t * leafcache = jl_atomic_load_relaxed (& mt -> leafcache );
4238
4238
jl_typemap_entry_t * entry = lookup_leafcache (leafcache , (jl_value_t * )type , world );
4239
4239
if (entry ) {
@@ -4266,12 +4266,12 @@ static jl_value_t *ml_matches(jl_methtable_t *mt,
4266
4266
}
4267
4267
}
4268
4268
// then check the full cache if it seems profitable
4269
- if (((jl_datatype_t * )unw )-> isdispatchtuple ) {
4269
+ if (((jl_datatype_t * )unw )-> isindivisibletype ) {
4270
4270
jl_typemap_entry_t * entry = jl_typemap_assoc_by_type (jl_atomic_load_relaxed (& mt -> cache ), & search , jl_cachearg_offset (mt ), /*subtype*/ 1 );
4271
- if (entry && (((jl_datatype_t * )unw )-> isdispatchtuple || entry -> guardsigs == jl_emptysvec )) {
4271
+ if (entry && (((jl_datatype_t * )unw )-> isindivisibletype || entry -> guardsigs == jl_emptysvec )) {
4272
4272
jl_method_instance_t * mi = entry -> func .linfo ;
4273
4273
jl_method_t * meth = mi -> def .method ;
4274
- if (!jl_is_unionall (meth -> sig ) && ((jl_datatype_t * )unw )-> isdispatchtuple ) {
4274
+ if (!jl_is_unionall (meth -> sig ) && ((jl_datatype_t * )unw )-> isindivisibletype ) {
4275
4275
env .match .env = jl_emptysvec ;
4276
4276
env .match .ti = unw ;
4277
4277
}
@@ -4560,7 +4560,7 @@ static jl_value_t *ml_matches(jl_methtable_t *mt,
4560
4560
if (env .match .max_valid > max_world )
4561
4561
env .match .max_valid = max_world ;
4562
4562
}
4563
- if (mt && cache_result && ((jl_datatype_t * )unw )-> isdispatchtuple ) { // cache_result parameter keeps this from being recursive
4563
+ if (mt && cache_result && ((jl_datatype_t * )unw )-> isindivisibletype ) { // cache_result parameter keeps this from being recursive
4564
4564
if (len == 1 && !has_ambiguity ) {
4565
4565
env .matc = (jl_method_match_t * )jl_array_ptr_ref (env .t , 0 );
4566
4566
jl_method_t * meth = env .matc -> method ;
0 commit comments