mirror of
https://github.com/python/cpython.git
synced 2024-12-01 11:15:56 +01:00
2761 lines
96 KiB
C
2761 lines
96 KiB
C
#include "Python.h"
|
|
|
|
#include "opcode.h"
|
|
|
|
#include "pycore_code.h"
|
|
#include "pycore_descrobject.h" // _PyMethodWrapper_Type
|
|
#include "pycore_dict.h" // DICT_KEYS_UNICODE
|
|
#include "pycore_function.h" // _PyFunction_GetVersionForCurrentState()
|
|
#include "pycore_long.h" // _PyLong_IsNonNegativeCompact()
|
|
#include "pycore_moduleobject.h"
|
|
#include "pycore_object.h"
|
|
#include "pycore_opcode_metadata.h" // _PyOpcode_Caches
|
|
#include "pycore_uop_metadata.h" // _PyOpcode_uop_name
|
|
#include "pycore_uop_ids.h" // MAX_UOP_ID
|
|
#include "pycore_opcode_utils.h" // RESUME_AT_FUNC_START
|
|
#include "pycore_pylifecycle.h" // _PyOS_URandomNonblock()
|
|
#include "pycore_runtime.h" // _Py_ID()
|
|
|
|
#include <stdlib.h> // rand()
|
|
|
|
extern const char *_PyUOpName(int index);
|
|
|
|
/* For guidance on adding or extending families of instructions see
|
|
* ./adaptive.md
|
|
*/
|
|
|
|
#ifdef Py_STATS
|
|
GCStats _py_gc_stats[NUM_GENERATIONS] = { 0 };
|
|
static PyStats _Py_stats_struct = { .gc_stats = _py_gc_stats };
|
|
PyStats *_Py_stats = NULL;
|
|
|
|
#if PYSTATS_MAX_UOP_ID < MAX_UOP_ID
|
|
#error "Not enough space allocated for pystats. Increase PYSTATS_MAX_UOP_ID to at least MAX_UOP_ID"
|
|
#endif
|
|
|
|
#define ADD_STAT_TO_DICT(res, field) \
|
|
do { \
|
|
PyObject *val = PyLong_FromUnsignedLongLong(stats->field); \
|
|
if (val == NULL) { \
|
|
Py_DECREF(res); \
|
|
return NULL; \
|
|
} \
|
|
if (PyDict_SetItemString(res, #field, val) == -1) { \
|
|
Py_DECREF(res); \
|
|
Py_DECREF(val); \
|
|
return NULL; \
|
|
} \
|
|
Py_DECREF(val); \
|
|
} while(0);
|
|
|
|
static PyObject*
|
|
stats_to_dict(SpecializationStats *stats)
|
|
{
|
|
PyObject *res = PyDict_New();
|
|
if (res == NULL) {
|
|
return NULL;
|
|
}
|
|
ADD_STAT_TO_DICT(res, success);
|
|
ADD_STAT_TO_DICT(res, failure);
|
|
ADD_STAT_TO_DICT(res, hit);
|
|
ADD_STAT_TO_DICT(res, deferred);
|
|
ADD_STAT_TO_DICT(res, miss);
|
|
ADD_STAT_TO_DICT(res, deopt);
|
|
PyObject *failure_kinds = PyTuple_New(SPECIALIZATION_FAILURE_KINDS);
|
|
if (failure_kinds == NULL) {
|
|
Py_DECREF(res);
|
|
return NULL;
|
|
}
|
|
for (int i = 0; i < SPECIALIZATION_FAILURE_KINDS; i++) {
|
|
PyObject *stat = PyLong_FromUnsignedLongLong(stats->failure_kinds[i]);
|
|
if (stat == NULL) {
|
|
Py_DECREF(res);
|
|
Py_DECREF(failure_kinds);
|
|
return NULL;
|
|
}
|
|
PyTuple_SET_ITEM(failure_kinds, i, stat);
|
|
}
|
|
if (PyDict_SetItemString(res, "failure_kinds", failure_kinds)) {
|
|
Py_DECREF(res);
|
|
Py_DECREF(failure_kinds);
|
|
return NULL;
|
|
}
|
|
Py_DECREF(failure_kinds);
|
|
return res;
|
|
}
|
|
#undef ADD_STAT_TO_DICT
|
|
|
|
static int
|
|
add_stat_dict(
|
|
PyObject *res,
|
|
int opcode,
|
|
const char *name) {
|
|
|
|
SpecializationStats *stats = &_Py_stats_struct.opcode_stats[opcode].specialization;
|
|
PyObject *d = stats_to_dict(stats);
|
|
if (d == NULL) {
|
|
return -1;
|
|
}
|
|
int err = PyDict_SetItemString(res, name, d);
|
|
Py_DECREF(d);
|
|
return err;
|
|
}
|
|
|
|
PyObject*
|
|
_Py_GetSpecializationStats(void) {
|
|
PyObject *stats = PyDict_New();
|
|
if (stats == NULL) {
|
|
return NULL;
|
|
}
|
|
int err = 0;
|
|
err += add_stat_dict(stats, CONTAINS_OP, "contains_op");
|
|
err += add_stat_dict(stats, LOAD_SUPER_ATTR, "load_super_attr");
|
|
err += add_stat_dict(stats, LOAD_ATTR, "load_attr");
|
|
err += add_stat_dict(stats, LOAD_GLOBAL, "load_global");
|
|
err += add_stat_dict(stats, BINARY_SUBSCR, "binary_subscr");
|
|
err += add_stat_dict(stats, STORE_SUBSCR, "store_subscr");
|
|
err += add_stat_dict(stats, STORE_ATTR, "store_attr");
|
|
err += add_stat_dict(stats, CALL, "call");
|
|
err += add_stat_dict(stats, CALL_KW, "call_kw");
|
|
err += add_stat_dict(stats, BINARY_OP, "binary_op");
|
|
err += add_stat_dict(stats, COMPARE_OP, "compare_op");
|
|
err += add_stat_dict(stats, UNPACK_SEQUENCE, "unpack_sequence");
|
|
err += add_stat_dict(stats, FOR_ITER, "for_iter");
|
|
err += add_stat_dict(stats, TO_BOOL, "to_bool");
|
|
err += add_stat_dict(stats, SEND, "send");
|
|
if (err < 0) {
|
|
Py_DECREF(stats);
|
|
return NULL;
|
|
}
|
|
return stats;
|
|
}
|
|
|
|
|
|
#define PRINT_STAT(i, field) \
|
|
if (stats[i].field) { \
|
|
fprintf(out, " opcode[%s]." #field " : %" PRIu64 "\n", _PyOpcode_OpName[i], stats[i].field); \
|
|
}
|
|
|
|
static void
|
|
print_spec_stats(FILE *out, OpcodeStats *stats)
|
|
{
|
|
/* Mark some opcodes as specializable for stats,
|
|
* even though we don't specialize them yet. */
|
|
fprintf(out, "opcode[BINARY_SLICE].specializable : 1\n");
|
|
fprintf(out, "opcode[STORE_SLICE].specializable : 1\n");
|
|
for (int i = 0; i < 256; i++) {
|
|
if (_PyOpcode_Caches[i]) {
|
|
/* Ignore jumps as they cannot be specialized */
|
|
switch (i) {
|
|
case POP_JUMP_IF_FALSE:
|
|
case POP_JUMP_IF_TRUE:
|
|
case POP_JUMP_IF_NONE:
|
|
case POP_JUMP_IF_NOT_NONE:
|
|
case JUMP_BACKWARD:
|
|
break;
|
|
default:
|
|
fprintf(out, "opcode[%s].specializable : 1\n", _PyOpcode_OpName[i]);
|
|
}
|
|
}
|
|
PRINT_STAT(i, specialization.success);
|
|
PRINT_STAT(i, specialization.failure);
|
|
PRINT_STAT(i, specialization.hit);
|
|
PRINT_STAT(i, specialization.deferred);
|
|
PRINT_STAT(i, specialization.miss);
|
|
PRINT_STAT(i, specialization.deopt);
|
|
PRINT_STAT(i, execution_count);
|
|
for (int j = 0; j < SPECIALIZATION_FAILURE_KINDS; j++) {
|
|
uint64_t val = stats[i].specialization.failure_kinds[j];
|
|
if (val) {
|
|
fprintf(out, " opcode[%s].specialization.failure_kinds[%d] : %"
|
|
PRIu64 "\n", _PyOpcode_OpName[i], j, val);
|
|
}
|
|
}
|
|
for (int j = 0; j < 256; j++) {
|
|
if (stats[i].pair_count[j]) {
|
|
fprintf(out, "opcode[%s].pair_count[%s] : %" PRIu64 "\n",
|
|
_PyOpcode_OpName[i], _PyOpcode_OpName[j], stats[i].pair_count[j]);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
#undef PRINT_STAT
|
|
|
|
|
|
static void
|
|
print_call_stats(FILE *out, CallStats *stats)
|
|
{
|
|
fprintf(out, "Calls to PyEval_EvalDefault: %" PRIu64 "\n", stats->pyeval_calls);
|
|
fprintf(out, "Calls to Python functions inlined: %" PRIu64 "\n", stats->inlined_py_calls);
|
|
fprintf(out, "Frames pushed: %" PRIu64 "\n", stats->frames_pushed);
|
|
fprintf(out, "Frame objects created: %" PRIu64 "\n", stats->frame_objects_created);
|
|
for (int i = 0; i < EVAL_CALL_KINDS; i++) {
|
|
fprintf(out, "Calls via PyEval_EvalFrame[%d] : %" PRIu64 "\n", i, stats->eval_calls[i]);
|
|
}
|
|
}
|
|
|
|
static void
|
|
print_object_stats(FILE *out, ObjectStats *stats)
|
|
{
|
|
fprintf(out, "Object allocations from freelist: %" PRIu64 "\n", stats->from_freelist);
|
|
fprintf(out, "Object frees to freelist: %" PRIu64 "\n", stats->to_freelist);
|
|
fprintf(out, "Object allocations: %" PRIu64 "\n", stats->allocations);
|
|
fprintf(out, "Object allocations to 512 bytes: %" PRIu64 "\n", stats->allocations512);
|
|
fprintf(out, "Object allocations to 4 kbytes: %" PRIu64 "\n", stats->allocations4k);
|
|
fprintf(out, "Object allocations over 4 kbytes: %" PRIu64 "\n", stats->allocations_big);
|
|
fprintf(out, "Object frees: %" PRIu64 "\n", stats->frees);
|
|
fprintf(out, "Object inline values: %" PRIu64 "\n", stats->inline_values);
|
|
fprintf(out, "Object interpreter increfs: %" PRIu64 "\n", stats->interpreter_increfs);
|
|
fprintf(out, "Object interpreter decrefs: %" PRIu64 "\n", stats->interpreter_decrefs);
|
|
fprintf(out, "Object increfs: %" PRIu64 "\n", stats->increfs);
|
|
fprintf(out, "Object decrefs: %" PRIu64 "\n", stats->decrefs);
|
|
fprintf(out, "Object materialize dict (on request): %" PRIu64 "\n", stats->dict_materialized_on_request);
|
|
fprintf(out, "Object materialize dict (new key): %" PRIu64 "\n", stats->dict_materialized_new_key);
|
|
fprintf(out, "Object materialize dict (too big): %" PRIu64 "\n", stats->dict_materialized_too_big);
|
|
fprintf(out, "Object materialize dict (str subclass): %" PRIu64 "\n", stats->dict_materialized_str_subclass);
|
|
fprintf(out, "Object method cache hits: %" PRIu64 "\n", stats->type_cache_hits);
|
|
fprintf(out, "Object method cache misses: %" PRIu64 "\n", stats->type_cache_misses);
|
|
fprintf(out, "Object method cache collisions: %" PRIu64 "\n", stats->type_cache_collisions);
|
|
fprintf(out, "Object method cache dunder hits: %" PRIu64 "\n", stats->type_cache_dunder_hits);
|
|
fprintf(out, "Object method cache dunder misses: %" PRIu64 "\n", stats->type_cache_dunder_misses);
|
|
}
|
|
|
|
static void
|
|
print_gc_stats(FILE *out, GCStats *stats)
|
|
{
|
|
for (int i = 0; i < NUM_GENERATIONS; i++) {
|
|
fprintf(out, "GC[%d] collections: %" PRIu64 "\n", i, stats[i].collections);
|
|
fprintf(out, "GC[%d] object visits: %" PRIu64 "\n", i, stats[i].object_visits);
|
|
fprintf(out, "GC[%d] objects collected: %" PRIu64 "\n", i, stats[i].objects_collected);
|
|
}
|
|
}
|
|
|
|
#ifdef _Py_TIER2
|
|
static void
|
|
print_histogram(FILE *out, const char *name, uint64_t hist[_Py_UOP_HIST_SIZE])
|
|
{
|
|
for (int i = 0; i < _Py_UOP_HIST_SIZE; i++) {
|
|
fprintf(out, "%s[%" PRIu64"]: %" PRIu64 "\n", name, (uint64_t)1 << i, hist[i]);
|
|
}
|
|
}
|
|
|
|
static void
|
|
print_optimization_stats(FILE *out, OptimizationStats *stats)
|
|
{
|
|
fprintf(out, "Optimization attempts: %" PRIu64 "\n", stats->attempts);
|
|
fprintf(out, "Optimization traces created: %" PRIu64 "\n", stats->traces_created);
|
|
fprintf(out, "Optimization traces executed: %" PRIu64 "\n", stats->traces_executed);
|
|
fprintf(out, "Optimization uops executed: %" PRIu64 "\n", stats->uops_executed);
|
|
fprintf(out, "Optimization trace stack overflow: %" PRIu64 "\n", stats->trace_stack_overflow);
|
|
fprintf(out, "Optimization trace stack underflow: %" PRIu64 "\n", stats->trace_stack_underflow);
|
|
fprintf(out, "Optimization trace too long: %" PRIu64 "\n", stats->trace_too_long);
|
|
fprintf(out, "Optimization trace too short: %" PRIu64 "\n", stats->trace_too_short);
|
|
fprintf(out, "Optimization inner loop: %" PRIu64 "\n", stats->inner_loop);
|
|
fprintf(out, "Optimization recursive call: %" PRIu64 "\n", stats->recursive_call);
|
|
fprintf(out, "Optimization low confidence: %" PRIu64 "\n", stats->low_confidence);
|
|
fprintf(out, "Executors invalidated: %" PRIu64 "\n", stats->executors_invalidated);
|
|
|
|
print_histogram(out, "Trace length", stats->trace_length_hist);
|
|
print_histogram(out, "Trace run length", stats->trace_run_length_hist);
|
|
print_histogram(out, "Optimized trace length", stats->optimized_trace_length_hist);
|
|
|
|
fprintf(out, "Optimization optimizer attempts: %" PRIu64 "\n", stats->optimizer_attempts);
|
|
fprintf(out, "Optimization optimizer successes: %" PRIu64 "\n", stats->optimizer_successes);
|
|
fprintf(out, "Optimization optimizer failure no memory: %" PRIu64 "\n",
|
|
stats->optimizer_failure_reason_no_memory);
|
|
fprintf(out, "Optimizer remove globals builtins changed: %" PRIu64 "\n", stats->remove_globals_builtins_changed);
|
|
fprintf(out, "Optimizer remove globals incorrect keys: %" PRIu64 "\n", stats->remove_globals_incorrect_keys);
|
|
for (int i = 0; i <= MAX_UOP_ID; i++) {
|
|
if (stats->opcode[i].execution_count) {
|
|
fprintf(out, "uops[%s].execution_count : %" PRIu64 "\n", _PyUOpName(i), stats->opcode[i].execution_count);
|
|
}
|
|
if (stats->opcode[i].miss) {
|
|
fprintf(out, "uops[%s].specialization.miss : %" PRIu64 "\n", _PyUOpName(i), stats->opcode[i].miss);
|
|
}
|
|
}
|
|
for (int i = 0; i < 256; i++) {
|
|
if (stats->unsupported_opcode[i]) {
|
|
fprintf(
|
|
out,
|
|
"unsupported_opcode[%s].count : %" PRIu64 "\n",
|
|
_PyOpcode_OpName[i],
|
|
stats->unsupported_opcode[i]
|
|
);
|
|
}
|
|
}
|
|
|
|
for (int i = 1; i <= MAX_UOP_ID; i++){
|
|
for (int j = 1; j <= MAX_UOP_ID; j++) {
|
|
if (stats->opcode[i].pair_count[j]) {
|
|
fprintf(out, "uop[%s].pair_count[%s] : %" PRIu64 "\n",
|
|
_PyOpcode_uop_name[i], _PyOpcode_uop_name[j], stats->opcode[i].pair_count[j]);
|
|
}
|
|
}
|
|
}
|
|
for (int i = 0; i < MAX_UOP_ID; i++) {
|
|
if (stats->error_in_opcode[i]) {
|
|
fprintf(
|
|
out,
|
|
"error_in_opcode[%s].count : %" PRIu64 "\n",
|
|
_PyUOpName(i),
|
|
stats->error_in_opcode[i]
|
|
);
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static void
|
|
print_rare_event_stats(FILE *out, RareEventStats *stats)
|
|
{
|
|
fprintf(out, "Rare event (set_class): %" PRIu64 "\n", stats->set_class);
|
|
fprintf(out, "Rare event (set_bases): %" PRIu64 "\n", stats->set_bases);
|
|
fprintf(out, "Rare event (set_eval_frame_func): %" PRIu64 "\n", stats->set_eval_frame_func);
|
|
fprintf(out, "Rare event (builtin_dict): %" PRIu64 "\n", stats->builtin_dict);
|
|
fprintf(out, "Rare event (func_modification): %" PRIu64 "\n", stats->func_modification);
|
|
fprintf(out, "Rare event (watched_dict_modification): %" PRIu64 "\n", stats->watched_dict_modification);
|
|
fprintf(out, "Rare event (watched_globals_modification): %" PRIu64 "\n", stats->watched_globals_modification);
|
|
}
|
|
|
|
static void
|
|
print_stats(FILE *out, PyStats *stats)
|
|
{
|
|
print_spec_stats(out, stats->opcode_stats);
|
|
print_call_stats(out, &stats->call_stats);
|
|
print_object_stats(out, &stats->object_stats);
|
|
print_gc_stats(out, stats->gc_stats);
|
|
#ifdef _Py_TIER2
|
|
print_optimization_stats(out, &stats->optimization_stats);
|
|
#endif
|
|
print_rare_event_stats(out, &stats->rare_event_stats);
|
|
}
|
|
|
|
void
|
|
_Py_StatsOn(void)
|
|
{
|
|
_Py_stats = &_Py_stats_struct;
|
|
}
|
|
|
|
void
|
|
_Py_StatsOff(void)
|
|
{
|
|
_Py_stats = NULL;
|
|
}
|
|
|
|
void
|
|
_Py_StatsClear(void)
|
|
{
|
|
memset(&_py_gc_stats, 0, sizeof(_py_gc_stats));
|
|
memset(&_Py_stats_struct, 0, sizeof(_Py_stats_struct));
|
|
_Py_stats_struct.gc_stats = _py_gc_stats;
|
|
}
|
|
|
|
static int
|
|
mem_is_zero(unsigned char *ptr, size_t size)
|
|
{
|
|
for (size_t i=0; i < size; i++) {
|
|
if (*ptr != 0) {
|
|
return 0;
|
|
}
|
|
ptr++;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
int
|
|
_Py_PrintSpecializationStats(int to_file)
|
|
{
|
|
PyStats *stats = &_Py_stats_struct;
|
|
#define MEM_IS_ZERO(DATA) mem_is_zero((unsigned char*)DATA, sizeof(*(DATA)))
|
|
int is_zero = (
|
|
MEM_IS_ZERO(stats->gc_stats) // is a pointer
|
|
&& MEM_IS_ZERO(&stats->opcode_stats)
|
|
&& MEM_IS_ZERO(&stats->call_stats)
|
|
&& MEM_IS_ZERO(&stats->object_stats)
|
|
);
|
|
#undef MEM_IS_ZERO
|
|
if (is_zero) {
|
|
// gh-108753: -X pystats command line was used, but then _stats_off()
|
|
// and _stats_clear() have been called: in this case, avoid printing
|
|
// useless "all zeros" statistics.
|
|
return 0;
|
|
}
|
|
|
|
FILE *out = stderr;
|
|
if (to_file) {
|
|
/* Write to a file instead of stderr. */
|
|
# ifdef MS_WINDOWS
|
|
const char *dirname = "c:\\temp\\py_stats\\";
|
|
# else
|
|
const char *dirname = "/tmp/py_stats/";
|
|
# endif
|
|
/* Use random 160 bit number as file name,
|
|
* to avoid both accidental collisions and
|
|
* symlink attacks. */
|
|
unsigned char rand[20];
|
|
char hex_name[41];
|
|
_PyOS_URandomNonblock(rand, 20);
|
|
for (int i = 0; i < 20; i++) {
|
|
hex_name[2*i] = Py_hexdigits[rand[i]&15];
|
|
hex_name[2*i+1] = Py_hexdigits[(rand[i]>>4)&15];
|
|
}
|
|
hex_name[40] = '\0';
|
|
char buf[64];
|
|
assert(strlen(dirname) + 40 + strlen(".txt") < 64);
|
|
sprintf(buf, "%s%s.txt", dirname, hex_name);
|
|
FILE *fout = fopen(buf, "w");
|
|
if (fout) {
|
|
out = fout;
|
|
}
|
|
}
|
|
else {
|
|
fprintf(out, "Specialization stats:\n");
|
|
}
|
|
print_stats(out, stats);
|
|
if (out != stderr) {
|
|
fclose(out);
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
#define SPECIALIZATION_FAIL(opcode, kind) \
|
|
do { \
|
|
if (_Py_stats) { \
|
|
_Py_stats->opcode_stats[opcode].specialization.failure_kinds[kind]++; \
|
|
} \
|
|
} while (0)
|
|
|
|
#endif // Py_STATS
|
|
|
|
|
|
#ifndef SPECIALIZATION_FAIL
|
|
# define SPECIALIZATION_FAIL(opcode, kind) ((void)0)
|
|
#endif
|
|
|
|
// Initialize warmup counters and insert superinstructions. This cannot fail.
|
|
void
|
|
_PyCode_Quicken(PyCodeObject *code)
|
|
{
|
|
#if ENABLE_SPECIALIZATION
|
|
int opcode = 0;
|
|
_Py_CODEUNIT *instructions = _PyCode_CODE(code);
|
|
/* The last code unit cannot have a cache, so we don't need to check it */
|
|
for (int i = 0; i < Py_SIZE(code)-1; i++) {
|
|
opcode = instructions[i].op.code;
|
|
int caches = _PyOpcode_Caches[opcode];
|
|
if (caches) {
|
|
// The initial value depends on the opcode
|
|
switch (opcode) {
|
|
case JUMP_BACKWARD:
|
|
instructions[i + 1].counter = initial_jump_backoff_counter();
|
|
break;
|
|
case POP_JUMP_IF_FALSE:
|
|
case POP_JUMP_IF_TRUE:
|
|
case POP_JUMP_IF_NONE:
|
|
case POP_JUMP_IF_NOT_NONE:
|
|
instructions[i + 1].cache = 0x5555; // Alternating 0, 1 bits
|
|
break;
|
|
default:
|
|
instructions[i + 1].counter = adaptive_counter_warmup();
|
|
break;
|
|
}
|
|
i += caches;
|
|
}
|
|
}
|
|
#endif /* ENABLE_SPECIALIZATION */
|
|
}
|
|
|
|
#define SIMPLE_FUNCTION 0
|
|
|
|
/* Common */
|
|
|
|
#define SPEC_FAIL_OTHER 0
|
|
#define SPEC_FAIL_NO_DICT 1
|
|
#define SPEC_FAIL_OVERRIDDEN 2
|
|
#define SPEC_FAIL_OUT_OF_VERSIONS 3
|
|
#define SPEC_FAIL_OUT_OF_RANGE 4
|
|
#define SPEC_FAIL_EXPECTED_ERROR 5
|
|
#define SPEC_FAIL_WRONG_NUMBER_ARGUMENTS 6
|
|
#define SPEC_FAIL_CODE_COMPLEX_PARAMETERS 7
|
|
#define SPEC_FAIL_CODE_NOT_OPTIMIZED 8
|
|
|
|
|
|
#define SPEC_FAIL_LOAD_GLOBAL_NON_DICT 17
|
|
#define SPEC_FAIL_LOAD_GLOBAL_NON_STRING_OR_SPLIT 18
|
|
|
|
/* Super */
|
|
|
|
#define SPEC_FAIL_SUPER_BAD_CLASS 9
|
|
#define SPEC_FAIL_SUPER_SHADOWED 10
|
|
|
|
/* Attributes */
|
|
|
|
#define SPEC_FAIL_ATTR_OVERRIDING_DESCRIPTOR 9
|
|
#define SPEC_FAIL_ATTR_NON_OVERRIDING_DESCRIPTOR 10
|
|
#define SPEC_FAIL_ATTR_NOT_DESCRIPTOR 11
|
|
#define SPEC_FAIL_ATTR_METHOD 12
|
|
#define SPEC_FAIL_ATTR_MUTABLE_CLASS 13
|
|
#define SPEC_FAIL_ATTR_PROPERTY 14
|
|
#define SPEC_FAIL_ATTR_NON_OBJECT_SLOT 15
|
|
#define SPEC_FAIL_ATTR_READ_ONLY 16
|
|
#define SPEC_FAIL_ATTR_AUDITED_SLOT 17
|
|
#define SPEC_FAIL_ATTR_NOT_MANAGED_DICT 18
|
|
#define SPEC_FAIL_ATTR_NON_STRING 19
|
|
#define SPEC_FAIL_ATTR_MODULE_ATTR_NOT_FOUND 20
|
|
#define SPEC_FAIL_ATTR_SHADOWED 21
|
|
#define SPEC_FAIL_ATTR_BUILTIN_CLASS_METHOD 22
|
|
#define SPEC_FAIL_ATTR_CLASS_METHOD_OBJ 23
|
|
#define SPEC_FAIL_ATTR_OBJECT_SLOT 24
|
|
|
|
#define SPEC_FAIL_ATTR_INSTANCE_ATTRIBUTE 26
|
|
#define SPEC_FAIL_ATTR_METACLASS_ATTRIBUTE 27
|
|
#define SPEC_FAIL_ATTR_PROPERTY_NOT_PY_FUNCTION 28
|
|
#define SPEC_FAIL_ATTR_NOT_IN_KEYS 29
|
|
#define SPEC_FAIL_ATTR_NOT_IN_DICT 30
|
|
#define SPEC_FAIL_ATTR_CLASS_ATTR_SIMPLE 31
|
|
#define SPEC_FAIL_ATTR_CLASS_ATTR_DESCRIPTOR 32
|
|
#define SPEC_FAIL_ATTR_BUILTIN_CLASS_METHOD_OBJ 33
|
|
#define SPEC_FAIL_ATTR_METACLASS_OVERRIDDEN 34
|
|
#define SPEC_FAIL_ATTR_SPLIT_DICT 35
|
|
|
|
/* Binary subscr and store subscr */
|
|
|
|
#define SPEC_FAIL_SUBSCR_ARRAY_INT 9
|
|
#define SPEC_FAIL_SUBSCR_ARRAY_SLICE 10
|
|
#define SPEC_FAIL_SUBSCR_LIST_SLICE 11
|
|
#define SPEC_FAIL_SUBSCR_TUPLE_SLICE 12
|
|
#define SPEC_FAIL_SUBSCR_STRING_SLICE 14
|
|
#define SPEC_FAIL_SUBSCR_BUFFER_INT 15
|
|
#define SPEC_FAIL_SUBSCR_BUFFER_SLICE 16
|
|
#define SPEC_FAIL_SUBSCR_SEQUENCE_INT 17
|
|
|
|
/* Store subscr */
|
|
#define SPEC_FAIL_SUBSCR_BYTEARRAY_INT 18
|
|
#define SPEC_FAIL_SUBSCR_BYTEARRAY_SLICE 19
|
|
#define SPEC_FAIL_SUBSCR_PY_SIMPLE 20
|
|
#define SPEC_FAIL_SUBSCR_PY_OTHER 21
|
|
#define SPEC_FAIL_SUBSCR_DICT_SUBCLASS_NO_OVERRIDE 22
|
|
#define SPEC_FAIL_SUBSCR_NOT_HEAP_TYPE 23
|
|
|
|
/* Binary op */
|
|
|
|
#define SPEC_FAIL_BINARY_OP_ADD_DIFFERENT_TYPES 9
|
|
#define SPEC_FAIL_BINARY_OP_ADD_OTHER 10
|
|
#define SPEC_FAIL_BINARY_OP_AND_DIFFERENT_TYPES 11
|
|
#define SPEC_FAIL_BINARY_OP_AND_INT 12
|
|
#define SPEC_FAIL_BINARY_OP_AND_OTHER 13
|
|
#define SPEC_FAIL_BINARY_OP_FLOOR_DIVIDE 14
|
|
#define SPEC_FAIL_BINARY_OP_LSHIFT 15
|
|
#define SPEC_FAIL_BINARY_OP_MATRIX_MULTIPLY 16
|
|
#define SPEC_FAIL_BINARY_OP_MULTIPLY_DIFFERENT_TYPES 17
|
|
#define SPEC_FAIL_BINARY_OP_MULTIPLY_OTHER 18
|
|
#define SPEC_FAIL_BINARY_OP_OR 19
|
|
#define SPEC_FAIL_BINARY_OP_POWER 20
|
|
#define SPEC_FAIL_BINARY_OP_REMAINDER 21
|
|
#define SPEC_FAIL_BINARY_OP_RSHIFT 22
|
|
#define SPEC_FAIL_BINARY_OP_SUBTRACT_DIFFERENT_TYPES 23
|
|
#define SPEC_FAIL_BINARY_OP_SUBTRACT_OTHER 24
|
|
#define SPEC_FAIL_BINARY_OP_TRUE_DIVIDE_DIFFERENT_TYPES 25
|
|
#define SPEC_FAIL_BINARY_OP_TRUE_DIVIDE_FLOAT 26
|
|
#define SPEC_FAIL_BINARY_OP_TRUE_DIVIDE_OTHER 27
|
|
#define SPEC_FAIL_BINARY_OP_XOR 28
|
|
|
|
/* Calls */
|
|
|
|
#define SPEC_FAIL_CALL_INSTANCE_METHOD 11
|
|
#define SPEC_FAIL_CALL_CMETHOD 12
|
|
#define SPEC_FAIL_CALL_CFUNC_VARARGS 13
|
|
#define SPEC_FAIL_CALL_CFUNC_VARARGS_KEYWORDS 14
|
|
#define SPEC_FAIL_CALL_CFUNC_NOARGS 15
|
|
#define SPEC_FAIL_CALL_CFUNC_METHOD_FASTCALL_KEYWORDS 16
|
|
#define SPEC_FAIL_CALL_METH_DESCR_VARARGS 17
|
|
#define SPEC_FAIL_CALL_METH_DESCR_VARARGS_KEYWORDS 18
|
|
#define SPEC_FAIL_CALL_METH_DESCR_METHOD_FASTCALL_KEYWORDS 19
|
|
#define SPEC_FAIL_CALL_BAD_CALL_FLAGS 20
|
|
#define SPEC_FAIL_CALL_INIT_NOT_PYTHON 21
|
|
#define SPEC_FAIL_CALL_PEP_523 22
|
|
#define SPEC_FAIL_CALL_BOUND_METHOD 23
|
|
#define SPEC_FAIL_CALL_CLASS_MUTABLE 26
|
|
#define SPEC_FAIL_CALL_METHOD_WRAPPER 28
|
|
#define SPEC_FAIL_CALL_OPERATOR_WRAPPER 29
|
|
#define SPEC_FAIL_CALL_INIT_NOT_SIMPLE 30
|
|
#define SPEC_FAIL_CALL_METACLASS 31
|
|
#define SPEC_FAIL_CALL_INIT_NOT_INLINE_VALUES 32
|
|
|
|
/* COMPARE_OP */
|
|
#define SPEC_FAIL_COMPARE_OP_DIFFERENT_TYPES 12
|
|
#define SPEC_FAIL_COMPARE_OP_STRING 13
|
|
#define SPEC_FAIL_COMPARE_OP_BIG_INT 14
|
|
#define SPEC_FAIL_COMPARE_OP_BYTES 15
|
|
#define SPEC_FAIL_COMPARE_OP_TUPLE 16
|
|
#define SPEC_FAIL_COMPARE_OP_LIST 17
|
|
#define SPEC_FAIL_COMPARE_OP_SET 18
|
|
#define SPEC_FAIL_COMPARE_OP_BOOL 19
|
|
#define SPEC_FAIL_COMPARE_OP_BASEOBJECT 20
|
|
#define SPEC_FAIL_COMPARE_OP_FLOAT_LONG 21
|
|
#define SPEC_FAIL_COMPARE_OP_LONG_FLOAT 22
|
|
|
|
/* FOR_ITER and SEND */
|
|
#define SPEC_FAIL_ITER_GENERATOR 10
|
|
#define SPEC_FAIL_ITER_COROUTINE 11
|
|
#define SPEC_FAIL_ITER_ASYNC_GENERATOR 12
|
|
#define SPEC_FAIL_ITER_LIST 13
|
|
#define SPEC_FAIL_ITER_TUPLE 14
|
|
#define SPEC_FAIL_ITER_SET 15
|
|
#define SPEC_FAIL_ITER_STRING 16
|
|
#define SPEC_FAIL_ITER_BYTES 17
|
|
#define SPEC_FAIL_ITER_RANGE 18
|
|
#define SPEC_FAIL_ITER_ITERTOOLS 19
|
|
#define SPEC_FAIL_ITER_DICT_KEYS 20
|
|
#define SPEC_FAIL_ITER_DICT_ITEMS 21
|
|
#define SPEC_FAIL_ITER_DICT_VALUES 22
|
|
#define SPEC_FAIL_ITER_ENUMERATE 23
|
|
#define SPEC_FAIL_ITER_MAP 24
|
|
#define SPEC_FAIL_ITER_ZIP 25
|
|
#define SPEC_FAIL_ITER_SEQ_ITER 26
|
|
#define SPEC_FAIL_ITER_REVERSED_LIST 27
|
|
#define SPEC_FAIL_ITER_CALLABLE 28
|
|
#define SPEC_FAIL_ITER_ASCII_STRING 29
|
|
#define SPEC_FAIL_ITER_ASYNC_GENERATOR_SEND 30
|
|
|
|
// UNPACK_SEQUENCE
|
|
|
|
#define SPEC_FAIL_UNPACK_SEQUENCE_ITERATOR 9
|
|
#define SPEC_FAIL_UNPACK_SEQUENCE_SEQUENCE 10
|
|
|
|
// TO_BOOL
|
|
#define SPEC_FAIL_TO_BOOL_BYTEARRAY 9
|
|
#define SPEC_FAIL_TO_BOOL_BYTES 10
|
|
#define SPEC_FAIL_TO_BOOL_DICT 11
|
|
#define SPEC_FAIL_TO_BOOL_FLOAT 12
|
|
#define SPEC_FAIL_TO_BOOL_MAPPING 13
|
|
#define SPEC_FAIL_TO_BOOL_MEMORY_VIEW 14
|
|
#define SPEC_FAIL_TO_BOOL_NUMBER 15
|
|
#define SPEC_FAIL_TO_BOOL_SEQUENCE 16
|
|
#define SPEC_FAIL_TO_BOOL_SET 17
|
|
#define SPEC_FAIL_TO_BOOL_TUPLE 18
|
|
|
|
// CONTAINS_OP
|
|
#define SPEC_FAIL_CONTAINS_OP_STR 9
|
|
#define SPEC_FAIL_CONTAINS_OP_TUPLE 10
|
|
#define SPEC_FAIL_CONTAINS_OP_LIST 11
|
|
#define SPEC_FAIL_CONTAINS_OP_USER_CLASS 12
|
|
|
|
static int function_kind(PyCodeObject *code);
|
|
static bool function_check_args(PyObject *o, int expected_argcount, int opcode);
|
|
static uint32_t function_get_version(PyObject *o, int opcode);
|
|
static uint32_t type_get_version(PyTypeObject *t, int opcode);
|
|
|
|
static int
|
|
specialize_module_load_attr(
|
|
PyObject *owner, _Py_CODEUNIT *instr, PyObject *name
|
|
) {
|
|
_PyAttrCache *cache = (_PyAttrCache *)(instr + 1);
|
|
PyModuleObject *m = (PyModuleObject *)owner;
|
|
assert((Py_TYPE(owner)->tp_flags & Py_TPFLAGS_MANAGED_DICT) == 0);
|
|
PyDictObject *dict = (PyDictObject *)m->md_dict;
|
|
if (dict == NULL) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_NO_DICT);
|
|
return -1;
|
|
}
|
|
if (dict->ma_keys->dk_kind != DICT_KEYS_UNICODE) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_NON_STRING);
|
|
return -1;
|
|
}
|
|
Py_ssize_t index = _PyDict_LookupIndex(dict, &_Py_ID(__getattr__));
|
|
assert(index != DKIX_ERROR);
|
|
if (index != DKIX_EMPTY) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_MODULE_ATTR_NOT_FOUND);
|
|
return -1;
|
|
}
|
|
index = _PyDict_LookupIndex(dict, name);
|
|
assert (index != DKIX_ERROR);
|
|
if (index != (uint16_t)index) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR,
|
|
index == DKIX_EMPTY ?
|
|
SPEC_FAIL_ATTR_MODULE_ATTR_NOT_FOUND :
|
|
SPEC_FAIL_OUT_OF_RANGE);
|
|
return -1;
|
|
}
|
|
uint32_t keys_version = _PyDictKeys_GetVersionForCurrentState(
|
|
_PyInterpreterState_GET(), dict->ma_keys);
|
|
if (keys_version == 0) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
return -1;
|
|
}
|
|
write_u32(cache->version, keys_version);
|
|
cache->index = (uint16_t)index;
|
|
instr->op.code = LOAD_ATTR_MODULE;
|
|
return 0;
|
|
}
|
|
|
|
|
|
|
|
/* Attribute specialization */
|
|
|
|
void
|
|
_Py_Specialize_LoadSuperAttr(_PyStackRef global_super_st, _PyStackRef cls_st, _Py_CODEUNIT *instr, int load_method) {
|
|
PyObject *global_super = PyStackRef_AsPyObjectBorrow(global_super_st);
|
|
PyObject *cls = PyStackRef_AsPyObjectBorrow(cls_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[LOAD_SUPER_ATTR] == INLINE_CACHE_ENTRIES_LOAD_SUPER_ATTR);
|
|
_PySuperAttrCache *cache = (_PySuperAttrCache *)(instr + 1);
|
|
if (global_super != (PyObject *)&PySuper_Type) {
|
|
SPECIALIZATION_FAIL(LOAD_SUPER_ATTR, SPEC_FAIL_SUPER_SHADOWED);
|
|
goto fail;
|
|
}
|
|
if (!PyType_Check(cls)) {
|
|
SPECIALIZATION_FAIL(LOAD_SUPER_ATTR, SPEC_FAIL_SUPER_BAD_CLASS);
|
|
goto fail;
|
|
}
|
|
instr->op.code = load_method ? LOAD_SUPER_ATTR_METHOD : LOAD_SUPER_ATTR_ATTR;
|
|
goto success;
|
|
|
|
fail:
|
|
STAT_INC(LOAD_SUPER_ATTR, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = LOAD_SUPER_ATTR;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(LOAD_SUPER_ATTR, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
typedef enum {
|
|
OVERRIDING, /* Is an overriding descriptor, and will remain so. */
|
|
METHOD, /* Attribute has Py_TPFLAGS_METHOD_DESCRIPTOR set */
|
|
PROPERTY, /* Is a property */
|
|
OBJECT_SLOT, /* Is an object slot descriptor */
|
|
OTHER_SLOT, /* Is a slot descriptor of another type */
|
|
NON_OVERRIDING, /* Is another non-overriding descriptor, and is an instance of an immutable class*/
|
|
BUILTIN_CLASSMETHOD, /* Builtin methods with METH_CLASS */
|
|
PYTHON_CLASSMETHOD, /* Python classmethod(func) object */
|
|
NON_DESCRIPTOR, /* Is not a descriptor, and is an instance of an immutable class */
|
|
MUTABLE, /* Instance of a mutable class; might, or might not, be a descriptor */
|
|
ABSENT, /* Attribute is not present on the class */
|
|
DUNDER_CLASS, /* __class__ attribute */
|
|
GETSET_OVERRIDDEN, /* __getattribute__ or __setattr__ has been overridden */
|
|
GETATTRIBUTE_IS_PYTHON_FUNCTION /* Descriptor requires calling a Python __getattribute__ */
|
|
} DescriptorClassification;
|
|
|
|
|
|
static DescriptorClassification
|
|
classify_descriptor(PyObject *descriptor, bool has_getattr)
|
|
{
|
|
if (descriptor == NULL) {
|
|
return ABSENT;
|
|
}
|
|
PyTypeObject *desc_cls = Py_TYPE(descriptor);
|
|
if (!(desc_cls->tp_flags & Py_TPFLAGS_IMMUTABLETYPE)) {
|
|
return MUTABLE;
|
|
}
|
|
if (desc_cls->tp_descr_set) {
|
|
if (desc_cls == &PyMemberDescr_Type) {
|
|
PyMemberDescrObject *member = (PyMemberDescrObject *)descriptor;
|
|
struct PyMemberDef *dmem = member->d_member;
|
|
if (dmem->type == Py_T_OBJECT_EX || dmem->type == _Py_T_OBJECT) {
|
|
return OBJECT_SLOT;
|
|
}
|
|
return OTHER_SLOT;
|
|
}
|
|
if (desc_cls == &PyProperty_Type) {
|
|
/* We can't detect at runtime whether an attribute exists
|
|
with property. So that means we may have to call
|
|
__getattr__. */
|
|
return has_getattr ? GETSET_OVERRIDDEN : PROPERTY;
|
|
}
|
|
return OVERRIDING;
|
|
}
|
|
if (desc_cls->tp_descr_get) {
|
|
if (desc_cls->tp_flags & Py_TPFLAGS_METHOD_DESCRIPTOR) {
|
|
return METHOD;
|
|
}
|
|
if (Py_IS_TYPE(descriptor, &PyClassMethodDescr_Type)) {
|
|
return BUILTIN_CLASSMETHOD;
|
|
}
|
|
if (Py_IS_TYPE(descriptor, &PyClassMethod_Type)) {
|
|
return PYTHON_CLASSMETHOD;
|
|
}
|
|
return NON_OVERRIDING;
|
|
}
|
|
return NON_DESCRIPTOR;
|
|
}
|
|
|
|
static DescriptorClassification
|
|
analyze_descriptor(PyTypeObject *type, PyObject *name, PyObject **descr, int store)
|
|
{
|
|
bool has_getattr = false;
|
|
if (store) {
|
|
if (type->tp_setattro != PyObject_GenericSetAttr) {
|
|
*descr = NULL;
|
|
return GETSET_OVERRIDDEN;
|
|
}
|
|
}
|
|
else {
|
|
getattrofunc getattro_slot = type->tp_getattro;
|
|
if (getattro_slot == PyObject_GenericGetAttr) {
|
|
/* Normal attribute lookup; */
|
|
has_getattr = false;
|
|
}
|
|
else if (getattro_slot == _Py_slot_tp_getattr_hook ||
|
|
getattro_slot == _Py_slot_tp_getattro) {
|
|
/* One or both of __getattribute__ or __getattr__ may have been
|
|
overridden See typeobject.c for why these functions are special. */
|
|
PyObject *getattribute = _PyType_Lookup(type,
|
|
&_Py_ID(__getattribute__));
|
|
PyInterpreterState *interp = _PyInterpreterState_GET();
|
|
bool has_custom_getattribute = getattribute != NULL &&
|
|
getattribute != interp->callable_cache.object__getattribute__;
|
|
has_getattr = _PyType_Lookup(type, &_Py_ID(__getattr__)) != NULL;
|
|
if (has_custom_getattribute) {
|
|
if (getattro_slot == _Py_slot_tp_getattro &&
|
|
!has_getattr &&
|
|
Py_IS_TYPE(getattribute, &PyFunction_Type)) {
|
|
*descr = getattribute;
|
|
return GETATTRIBUTE_IS_PYTHON_FUNCTION;
|
|
}
|
|
/* Potentially both __getattr__ and __getattribute__ are set.
|
|
Too complicated */
|
|
*descr = NULL;
|
|
return GETSET_OVERRIDDEN;
|
|
}
|
|
/* Potentially has __getattr__ but no custom __getattribute__.
|
|
Fall through to usual descriptor analysis.
|
|
Usual attribute lookup should only be allowed at runtime
|
|
if we can guarantee that there is no way an exception can be
|
|
raised. This means some specializations, e.g. specializing
|
|
for property() isn't safe.
|
|
*/
|
|
}
|
|
else {
|
|
*descr = NULL;
|
|
return GETSET_OVERRIDDEN;
|
|
}
|
|
}
|
|
PyObject *descriptor = _PyType_Lookup(type, name);
|
|
*descr = descriptor;
|
|
if (PyUnicode_CompareWithASCIIString(name, "__class__") == 0) {
|
|
if (descriptor == _PyType_Lookup(&PyBaseObject_Type, name)) {
|
|
return DUNDER_CLASS;
|
|
}
|
|
}
|
|
return classify_descriptor(descriptor, has_getattr);
|
|
}
|
|
|
|
static int
|
|
specialize_dict_access(
|
|
PyObject *owner, _Py_CODEUNIT *instr, PyTypeObject *type,
|
|
DescriptorClassification kind, PyObject *name,
|
|
int base_op, int values_op, int hint_op)
|
|
{
|
|
assert(kind == NON_OVERRIDING || kind == NON_DESCRIPTOR || kind == ABSENT ||
|
|
kind == BUILTIN_CLASSMETHOD || kind == PYTHON_CLASSMETHOD ||
|
|
kind == METHOD);
|
|
// No descriptor, or non overriding.
|
|
if ((type->tp_flags & Py_TPFLAGS_MANAGED_DICT) == 0) {
|
|
SPECIALIZATION_FAIL(base_op, SPEC_FAIL_ATTR_NOT_MANAGED_DICT);
|
|
return 0;
|
|
}
|
|
_PyAttrCache *cache = (_PyAttrCache *)(instr + 1);
|
|
if (type->tp_flags & Py_TPFLAGS_INLINE_VALUES && _PyObject_InlineValues(owner)->valid) {
|
|
PyDictKeysObject *keys = ((PyHeapTypeObject *)type)->ht_cached_keys;
|
|
assert(PyUnicode_CheckExact(name));
|
|
Py_ssize_t index = _PyDictKeys_StringLookup(keys, name);
|
|
assert (index != DKIX_ERROR);
|
|
if (index == DKIX_EMPTY) {
|
|
SPECIALIZATION_FAIL(base_op, SPEC_FAIL_ATTR_NOT_IN_KEYS);
|
|
return 0;
|
|
}
|
|
assert(index >= 0);
|
|
char *value_addr = (char *)&_PyObject_InlineValues(owner)->values[index];
|
|
Py_ssize_t offset = value_addr - (char *)owner;
|
|
if (offset != (uint16_t)offset) {
|
|
SPECIALIZATION_FAIL(base_op, SPEC_FAIL_OUT_OF_RANGE);
|
|
return 0;
|
|
}
|
|
write_u32(cache->version, type->tp_version_tag);
|
|
cache->index = (uint16_t)offset;
|
|
instr->op.code = values_op;
|
|
}
|
|
else {
|
|
PyDictObject *dict = _PyObject_GetManagedDict(owner);
|
|
if (dict == NULL || !PyDict_CheckExact(dict)) {
|
|
SPECIALIZATION_FAIL(base_op, SPEC_FAIL_NO_DICT);
|
|
return 0;
|
|
}
|
|
// We found an instance with a __dict__.
|
|
if (dict->ma_values) {
|
|
SPECIALIZATION_FAIL(base_op, SPEC_FAIL_ATTR_SPLIT_DICT);
|
|
return 0;
|
|
}
|
|
Py_ssize_t index =
|
|
_PyDict_LookupIndex(dict, name);
|
|
if (index != (uint16_t)index) {
|
|
SPECIALIZATION_FAIL(base_op,
|
|
index == DKIX_EMPTY ?
|
|
SPEC_FAIL_ATTR_NOT_IN_DICT :
|
|
SPEC_FAIL_OUT_OF_RANGE);
|
|
return 0;
|
|
}
|
|
cache->index = (uint16_t)index;
|
|
write_u32(cache->version, type->tp_version_tag);
|
|
instr->op.code = hint_op;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static int specialize_attr_loadclassattr(PyObject* owner, _Py_CODEUNIT* instr, PyObject* name,
|
|
PyObject* descr, DescriptorClassification kind, bool is_method);
|
|
static int specialize_class_load_attr(PyObject* owner, _Py_CODEUNIT* instr, PyObject* name);
|
|
|
|
/* Returns true if instances of obj's class are
|
|
* likely to have `name` in their __dict__.
|
|
* For objects with inline values, we check in the shared keys.
|
|
* For other objects, we check their actual dictionary.
|
|
*/
|
|
static bool
|
|
instance_has_key(PyObject *obj, PyObject* name)
|
|
{
|
|
PyTypeObject *cls = Py_TYPE(obj);
|
|
if ((cls->tp_flags & Py_TPFLAGS_MANAGED_DICT) == 0) {
|
|
return false;
|
|
}
|
|
if (cls->tp_flags & Py_TPFLAGS_INLINE_VALUES) {
|
|
PyDictKeysObject *keys = ((PyHeapTypeObject *)cls)->ht_cached_keys;
|
|
Py_ssize_t index = _PyDictKeys_StringLookup(keys, name);
|
|
return index >= 0;
|
|
}
|
|
PyDictObject *dict = _PyObject_GetManagedDict(obj);
|
|
if (dict == NULL || !PyDict_CheckExact(dict)) {
|
|
return false;
|
|
}
|
|
if (dict->ma_values) {
|
|
return false;
|
|
}
|
|
Py_ssize_t index = _PyDict_LookupIndex(dict, name);
|
|
if (index < 0) {
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static int
|
|
specialize_instance_load_attr(PyObject* owner, _Py_CODEUNIT* instr, PyObject* name)
|
|
{
|
|
_PyAttrCache *cache = (_PyAttrCache *)(instr + 1);
|
|
PyTypeObject *type = Py_TYPE(owner);
|
|
bool shadow = instance_has_key(owner, name);
|
|
PyObject *descr = NULL;
|
|
DescriptorClassification kind = analyze_descriptor(type, name, &descr, 0);
|
|
assert(descr != NULL || kind == ABSENT || kind == GETSET_OVERRIDDEN);
|
|
if (type_get_version(type, LOAD_ATTR) == 0) {
|
|
return -1;
|
|
}
|
|
switch(kind) {
|
|
case OVERRIDING:
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_OVERRIDING_DESCRIPTOR);
|
|
return -1;
|
|
case METHOD:
|
|
{
|
|
if (shadow) {
|
|
goto try_instance;
|
|
}
|
|
int oparg = instr->op.arg;
|
|
if (oparg & 1) {
|
|
if (specialize_attr_loadclassattr(owner, instr, name, descr, kind, true)) {
|
|
return 0;
|
|
}
|
|
else {
|
|
return -1;
|
|
}
|
|
}
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_METHOD);
|
|
return -1;
|
|
}
|
|
case PROPERTY:
|
|
{
|
|
_PyLoadMethodCache *lm_cache = (_PyLoadMethodCache *)(instr + 1);
|
|
assert(Py_TYPE(descr) == &PyProperty_Type);
|
|
PyObject *fget = ((_PyPropertyObject *)descr)->prop_get;
|
|
if (fget == NULL) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_EXPECTED_ERROR);
|
|
return -1;
|
|
}
|
|
if (!Py_IS_TYPE(fget, &PyFunction_Type)) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_PROPERTY_NOT_PY_FUNCTION);
|
|
return -1;
|
|
}
|
|
if (!function_check_args(fget, 1, LOAD_ATTR)) {
|
|
return -1;
|
|
}
|
|
if (instr->op.arg & 1) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_METHOD);
|
|
return -1;
|
|
}
|
|
if (_PyInterpreterState_GET()->eval_frame) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OTHER);
|
|
return -1;
|
|
}
|
|
assert(type->tp_version_tag != 0);
|
|
write_u32(lm_cache->type_version, type->tp_version_tag);
|
|
/* borrowed */
|
|
write_obj(lm_cache->descr, fget);
|
|
instr->op.code = LOAD_ATTR_PROPERTY;
|
|
return 0;
|
|
}
|
|
case OBJECT_SLOT:
|
|
{
|
|
PyMemberDescrObject *member = (PyMemberDescrObject *)descr;
|
|
struct PyMemberDef *dmem = member->d_member;
|
|
Py_ssize_t offset = dmem->offset;
|
|
if (!PyObject_TypeCheck(owner, member->d_common.d_type)) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_EXPECTED_ERROR);
|
|
return -1;
|
|
}
|
|
if (dmem->flags & Py_AUDIT_READ) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_AUDITED_SLOT);
|
|
return -1;
|
|
}
|
|
if (offset != (uint16_t)offset) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OUT_OF_RANGE);
|
|
return -1;
|
|
}
|
|
assert(dmem->type == Py_T_OBJECT_EX || dmem->type == _Py_T_OBJECT);
|
|
assert(offset > 0);
|
|
cache->index = (uint16_t)offset;
|
|
write_u32(cache->version, type->tp_version_tag);
|
|
instr->op.code = LOAD_ATTR_SLOT;
|
|
return 0;
|
|
}
|
|
case DUNDER_CLASS:
|
|
{
|
|
Py_ssize_t offset = offsetof(PyObject, ob_type);
|
|
assert(offset == (uint16_t)offset);
|
|
cache->index = (uint16_t)offset;
|
|
write_u32(cache->version, type->tp_version_tag);
|
|
instr->op.code = LOAD_ATTR_SLOT;
|
|
return 0;
|
|
}
|
|
case OTHER_SLOT:
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_NON_OBJECT_SLOT);
|
|
return -1;
|
|
case MUTABLE:
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_MUTABLE_CLASS);
|
|
return -1;
|
|
case GETSET_OVERRIDDEN:
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OVERRIDDEN);
|
|
return -1;
|
|
case GETATTRIBUTE_IS_PYTHON_FUNCTION:
|
|
{
|
|
assert(type->tp_getattro == _Py_slot_tp_getattro);
|
|
assert(Py_IS_TYPE(descr, &PyFunction_Type));
|
|
_PyLoadMethodCache *lm_cache = (_PyLoadMethodCache *)(instr + 1);
|
|
if (!function_check_args(descr, 2, LOAD_ATTR)) {
|
|
return -1;
|
|
}
|
|
if (instr->op.arg & 1) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_METHOD);
|
|
return -1;
|
|
}
|
|
uint32_t version = function_get_version(descr, LOAD_ATTR);
|
|
if (version == 0) {
|
|
return -1;
|
|
}
|
|
if (_PyInterpreterState_GET()->eval_frame) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OTHER);
|
|
return -1;
|
|
}
|
|
write_u32(lm_cache->keys_version, version);
|
|
/* borrowed */
|
|
write_obj(lm_cache->descr, descr);
|
|
write_u32(lm_cache->type_version, type->tp_version_tag);
|
|
instr->op.code = LOAD_ATTR_GETATTRIBUTE_OVERRIDDEN;
|
|
return 0;
|
|
}
|
|
case BUILTIN_CLASSMETHOD:
|
|
case PYTHON_CLASSMETHOD:
|
|
case NON_OVERRIDING:
|
|
if (shadow) {
|
|
goto try_instance;
|
|
}
|
|
return -1;
|
|
case NON_DESCRIPTOR:
|
|
if (shadow) {
|
|
goto try_instance;
|
|
}
|
|
if ((instr->op.arg & 1) == 0) {
|
|
if (specialize_attr_loadclassattr(owner, instr, name, descr, kind, false)) {
|
|
return 0;
|
|
}
|
|
}
|
|
return -1;
|
|
case ABSENT:
|
|
if (shadow) {
|
|
goto try_instance;
|
|
}
|
|
return 0;
|
|
}
|
|
Py_UNREACHABLE();
|
|
try_instance:
|
|
if (specialize_dict_access(owner, instr, type, kind, name, LOAD_ATTR,
|
|
LOAD_ATTR_INSTANCE_VALUE, LOAD_ATTR_WITH_HINT))
|
|
{
|
|
return 0;
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_LoadAttr(_PyStackRef owner_st, _Py_CODEUNIT *instr, PyObject *name)
|
|
{
|
|
_PyAttrCache *cache = (_PyAttrCache *)(instr + 1);
|
|
PyObject *owner = PyStackRef_AsPyObjectBorrow(owner_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[LOAD_ATTR] == INLINE_CACHE_ENTRIES_LOAD_ATTR);
|
|
PyTypeObject *type = Py_TYPE(owner);
|
|
bool fail;
|
|
if (!_PyType_IsReady(type)) {
|
|
// We *might* not really need this check, but we inherited it from
|
|
// PyObject_GenericGetAttr and friends... and this way we still do the
|
|
// right thing if someone forgets to call PyType_Ready(type):
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OTHER);
|
|
fail = true;
|
|
}
|
|
else if (PyModule_CheckExact(owner)) {
|
|
fail = specialize_module_load_attr(owner, instr, name);
|
|
}
|
|
else if (PyType_Check(owner)) {
|
|
fail = specialize_class_load_attr(owner, instr, name);
|
|
}
|
|
else {
|
|
fail = specialize_instance_load_attr(owner, instr, name);
|
|
}
|
|
|
|
if (fail) {
|
|
STAT_INC(LOAD_ATTR, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = LOAD_ATTR;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
}
|
|
else {
|
|
STAT_INC(LOAD_ATTR, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_StoreAttr(_PyStackRef owner_st, _Py_CODEUNIT *instr, PyObject *name)
|
|
{
|
|
PyObject *owner = PyStackRef_AsPyObjectBorrow(owner_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[STORE_ATTR] == INLINE_CACHE_ENTRIES_STORE_ATTR);
|
|
_PyAttrCache *cache = (_PyAttrCache *)(instr + 1);
|
|
PyTypeObject *type = Py_TYPE(owner);
|
|
if (!_PyType_IsReady(type)) {
|
|
// We *might* not really need this check, but we inherited it from
|
|
// PyObject_GenericSetAttr and friends... and this way we still do the
|
|
// right thing if someone forgets to call PyType_Ready(type):
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_OTHER);
|
|
goto fail;
|
|
}
|
|
if (PyModule_CheckExact(owner)) {
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_OVERRIDDEN);
|
|
goto fail;
|
|
}
|
|
PyObject *descr;
|
|
DescriptorClassification kind = analyze_descriptor(type, name, &descr, 1);
|
|
if (type_get_version(type, STORE_ATTR) == 0) {
|
|
goto fail;
|
|
}
|
|
switch(kind) {
|
|
case OVERRIDING:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_OVERRIDING_DESCRIPTOR);
|
|
goto fail;
|
|
case METHOD:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_METHOD);
|
|
goto fail;
|
|
case PROPERTY:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_PROPERTY);
|
|
goto fail;
|
|
case OBJECT_SLOT:
|
|
{
|
|
PyMemberDescrObject *member = (PyMemberDescrObject *)descr;
|
|
struct PyMemberDef *dmem = member->d_member;
|
|
Py_ssize_t offset = dmem->offset;
|
|
if (!PyObject_TypeCheck(owner, member->d_common.d_type)) {
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_EXPECTED_ERROR);
|
|
goto fail;
|
|
}
|
|
if (dmem->flags & Py_READONLY) {
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_READ_ONLY);
|
|
goto fail;
|
|
}
|
|
if (offset != (uint16_t)offset) {
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
assert(dmem->type == Py_T_OBJECT_EX || dmem->type == _Py_T_OBJECT);
|
|
assert(offset > 0);
|
|
cache->index = (uint16_t)offset;
|
|
write_u32(cache->version, type->tp_version_tag);
|
|
instr->op.code = STORE_ATTR_SLOT;
|
|
goto success;
|
|
}
|
|
case DUNDER_CLASS:
|
|
case OTHER_SLOT:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_NON_OBJECT_SLOT);
|
|
goto fail;
|
|
case MUTABLE:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_MUTABLE_CLASS);
|
|
goto fail;
|
|
case GETATTRIBUTE_IS_PYTHON_FUNCTION:
|
|
case GETSET_OVERRIDDEN:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_OVERRIDDEN);
|
|
goto fail;
|
|
case BUILTIN_CLASSMETHOD:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_BUILTIN_CLASS_METHOD_OBJ);
|
|
goto fail;
|
|
case PYTHON_CLASSMETHOD:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_CLASS_METHOD_OBJ);
|
|
goto fail;
|
|
case NON_OVERRIDING:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_CLASS_ATTR_DESCRIPTOR);
|
|
goto fail;
|
|
case NON_DESCRIPTOR:
|
|
SPECIALIZATION_FAIL(STORE_ATTR, SPEC_FAIL_ATTR_CLASS_ATTR_SIMPLE);
|
|
goto fail;
|
|
case ABSENT:
|
|
if (specialize_dict_access(owner, instr, type, kind, name, STORE_ATTR,
|
|
STORE_ATTR_INSTANCE_VALUE, STORE_ATTR_WITH_HINT))
|
|
{
|
|
goto success;
|
|
}
|
|
}
|
|
fail:
|
|
STAT_INC(STORE_ATTR, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = STORE_ATTR;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(STORE_ATTR, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
|
|
#ifdef Py_STATS
|
|
static int
|
|
load_attr_fail_kind(DescriptorClassification kind)
|
|
{
|
|
switch (kind) {
|
|
case OVERRIDING:
|
|
return SPEC_FAIL_ATTR_OVERRIDING_DESCRIPTOR;
|
|
case METHOD:
|
|
return SPEC_FAIL_ATTR_METHOD;
|
|
case PROPERTY:
|
|
return SPEC_FAIL_ATTR_PROPERTY;
|
|
case OBJECT_SLOT:
|
|
return SPEC_FAIL_ATTR_OBJECT_SLOT;
|
|
case OTHER_SLOT:
|
|
return SPEC_FAIL_ATTR_NON_OBJECT_SLOT;
|
|
case DUNDER_CLASS:
|
|
return SPEC_FAIL_OTHER;
|
|
case MUTABLE:
|
|
return SPEC_FAIL_ATTR_MUTABLE_CLASS;
|
|
case GETSET_OVERRIDDEN:
|
|
case GETATTRIBUTE_IS_PYTHON_FUNCTION:
|
|
return SPEC_FAIL_OVERRIDDEN;
|
|
case BUILTIN_CLASSMETHOD:
|
|
return SPEC_FAIL_ATTR_BUILTIN_CLASS_METHOD;
|
|
case PYTHON_CLASSMETHOD:
|
|
return SPEC_FAIL_ATTR_CLASS_METHOD_OBJ;
|
|
case NON_OVERRIDING:
|
|
return SPEC_FAIL_ATTR_NON_OVERRIDING_DESCRIPTOR;
|
|
case NON_DESCRIPTOR:
|
|
return SPEC_FAIL_ATTR_NOT_DESCRIPTOR;
|
|
case ABSENT:
|
|
return SPEC_FAIL_ATTR_INSTANCE_ATTRIBUTE;
|
|
}
|
|
Py_UNREACHABLE();
|
|
}
|
|
#endif // Py_STATS
|
|
|
|
static int
|
|
specialize_class_load_attr(PyObject *owner, _Py_CODEUNIT *instr,
|
|
PyObject *name)
|
|
{
|
|
assert(PyType_Check(owner));
|
|
PyTypeObject *cls = (PyTypeObject *)owner;
|
|
_PyLoadMethodCache *cache = (_PyLoadMethodCache *)(instr + 1);
|
|
if (Py_TYPE(cls)->tp_getattro != _Py_type_getattro) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_METACLASS_OVERRIDDEN);
|
|
return -1;
|
|
}
|
|
PyObject *metadescriptor = _PyType_Lookup(Py_TYPE(cls), name);
|
|
DescriptorClassification metakind = classify_descriptor(metadescriptor, false);
|
|
switch (metakind) {
|
|
case METHOD:
|
|
case NON_DESCRIPTOR:
|
|
case NON_OVERRIDING:
|
|
case BUILTIN_CLASSMETHOD:
|
|
case PYTHON_CLASSMETHOD:
|
|
case ABSENT:
|
|
break;
|
|
default:
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_METACLASS_ATTRIBUTE);
|
|
return -1;
|
|
}
|
|
PyObject *descr = NULL;
|
|
DescriptorClassification kind = 0;
|
|
kind = analyze_descriptor(cls, name, &descr, 0);
|
|
if (type_get_version(cls, LOAD_ATTR) == 0) {
|
|
return -1;
|
|
}
|
|
bool metaclass_check = false;
|
|
if ((Py_TYPE(cls)->tp_flags & Py_TPFLAGS_IMMUTABLETYPE) == 0) {
|
|
metaclass_check = true;
|
|
if (type_get_version(Py_TYPE(cls), LOAD_ATTR) == 0) {
|
|
return -1;
|
|
}
|
|
}
|
|
switch (kind) {
|
|
case METHOD:
|
|
case NON_DESCRIPTOR:
|
|
write_u32(cache->type_version, cls->tp_version_tag);
|
|
write_obj(cache->descr, descr);
|
|
if (metaclass_check) {
|
|
write_u32(cache->keys_version, Py_TYPE(cls)->tp_version_tag);
|
|
instr->op.code = LOAD_ATTR_CLASS_WITH_METACLASS_CHECK;
|
|
}
|
|
else {
|
|
instr->op.code = LOAD_ATTR_CLASS;
|
|
}
|
|
return 0;
|
|
#ifdef Py_STATS
|
|
case ABSENT:
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_EXPECTED_ERROR);
|
|
return -1;
|
|
#endif
|
|
default:
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, load_attr_fail_kind(kind));
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
// Please collect stats carefully before and after modifying. A subtle change
|
|
// can cause a significant drop in cache hits. A possible test is
|
|
// python.exe -m test_typing test_re test_dis test_zlib.
|
|
static int
|
|
specialize_attr_loadclassattr(PyObject *owner, _Py_CODEUNIT *instr, PyObject *name,
|
|
PyObject *descr, DescriptorClassification kind, bool is_method)
|
|
{
|
|
_PyLoadMethodCache *cache = (_PyLoadMethodCache *)(instr + 1);
|
|
PyTypeObject *owner_cls = Py_TYPE(owner);
|
|
|
|
assert(descr != NULL);
|
|
assert((is_method && kind == METHOD) || (!is_method && kind == NON_DESCRIPTOR));
|
|
if (owner_cls->tp_flags & Py_TPFLAGS_INLINE_VALUES) {
|
|
PyDictKeysObject *keys = ((PyHeapTypeObject *)owner_cls)->ht_cached_keys;
|
|
assert(_PyDictKeys_StringLookup(keys, name) < 0);
|
|
uint32_t keys_version = _PyDictKeys_GetVersionForCurrentState(
|
|
_PyInterpreterState_GET(), keys);
|
|
if (keys_version == 0) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
return 0;
|
|
}
|
|
write_u32(cache->keys_version, keys_version);
|
|
instr->op.code = is_method ? LOAD_ATTR_METHOD_WITH_VALUES : LOAD_ATTR_NONDESCRIPTOR_WITH_VALUES;
|
|
}
|
|
else {
|
|
Py_ssize_t dictoffset;
|
|
if (owner_cls->tp_flags & Py_TPFLAGS_MANAGED_DICT) {
|
|
dictoffset = MANAGED_DICT_OFFSET;
|
|
}
|
|
else {
|
|
dictoffset = owner_cls->tp_dictoffset;
|
|
if (dictoffset < 0 || dictoffset > INT16_MAX + MANAGED_DICT_OFFSET) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_OUT_OF_RANGE);
|
|
return 0;
|
|
}
|
|
}
|
|
if (dictoffset == 0) {
|
|
instr->op.code = is_method ? LOAD_ATTR_METHOD_NO_DICT : LOAD_ATTR_NONDESCRIPTOR_NO_DICT;
|
|
}
|
|
else if (is_method) {
|
|
PyObject *dict = *(PyObject **) ((char *)owner + dictoffset);
|
|
if (dict) {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_NOT_MANAGED_DICT);
|
|
return 0;
|
|
}
|
|
/* Cache entries must be unsigned values, so we offset the
|
|
* dictoffset by MANAGED_DICT_OFFSET.
|
|
* We do the reverse offset in LOAD_ATTR_METHOD_LAZY_DICT */
|
|
dictoffset -= MANAGED_DICT_OFFSET;
|
|
assert(((uint16_t)dictoffset) == dictoffset);
|
|
cache->dict_offset = (uint16_t)dictoffset;
|
|
instr->op.code = LOAD_ATTR_METHOD_LAZY_DICT;
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(LOAD_ATTR, SPEC_FAIL_ATTR_CLASS_ATTR_SIMPLE);
|
|
return 0;
|
|
}
|
|
}
|
|
/* `descr` is borrowed. This is safe for methods (even inherited ones from
|
|
* super classes!) as long as tp_version_tag is validated for two main reasons:
|
|
*
|
|
* 1. The class will always hold a reference to the method so it will
|
|
* usually not be GC-ed. Should it be deleted in Python, e.g.
|
|
* `del obj.meth`, tp_version_tag will be invalidated, because of reason 2.
|
|
*
|
|
* 2. The pre-existing type method cache (MCACHE) uses the same principles
|
|
* of caching a borrowed descriptor. The MCACHE infrastructure does all the
|
|
* heavy lifting for us. E.g. it invalidates tp_version_tag on any MRO
|
|
* modification, on any type object change along said MRO, etc. (see
|
|
* PyType_Modified usages in typeobject.c). The MCACHE has been
|
|
* working since Python 2.6 and it's battle-tested.
|
|
*/
|
|
write_u32(cache->type_version, owner_cls->tp_version_tag);
|
|
write_obj(cache->descr, descr);
|
|
return 1;
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_LoadGlobal(
|
|
PyObject *globals, PyObject *builtins,
|
|
_Py_CODEUNIT *instr, PyObject *name)
|
|
{
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[LOAD_GLOBAL] == INLINE_CACHE_ENTRIES_LOAD_GLOBAL);
|
|
/* Use inline cache */
|
|
_PyLoadGlobalCache *cache = (_PyLoadGlobalCache *)(instr + 1);
|
|
assert(PyUnicode_CheckExact(name));
|
|
if (!PyDict_CheckExact(globals)) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_LOAD_GLOBAL_NON_DICT);
|
|
goto fail;
|
|
}
|
|
PyDictKeysObject * globals_keys = ((PyDictObject *)globals)->ma_keys;
|
|
if (!DK_IS_UNICODE(globals_keys)) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_LOAD_GLOBAL_NON_STRING_OR_SPLIT);
|
|
goto fail;
|
|
}
|
|
Py_ssize_t index = _PyDictKeys_StringLookup(globals_keys, name);
|
|
if (index == DKIX_ERROR) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_EXPECTED_ERROR);
|
|
goto fail;
|
|
}
|
|
PyInterpreterState *interp = _PyInterpreterState_GET();
|
|
if (index != DKIX_EMPTY) {
|
|
if (index != (uint16_t)index) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
uint32_t keys_version = _PyDictKeys_GetVersionForCurrentState(
|
|
interp, globals_keys);
|
|
if (keys_version == 0) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
goto fail;
|
|
}
|
|
if (keys_version != (uint16_t)keys_version) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
cache->index = (uint16_t)index;
|
|
cache->module_keys_version = (uint16_t)keys_version;
|
|
instr->op.code = LOAD_GLOBAL_MODULE;
|
|
goto success;
|
|
}
|
|
if (!PyDict_CheckExact(builtins)) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_LOAD_GLOBAL_NON_DICT);
|
|
goto fail;
|
|
}
|
|
PyDictKeysObject * builtin_keys = ((PyDictObject *)builtins)->ma_keys;
|
|
if (!DK_IS_UNICODE(builtin_keys)) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_LOAD_GLOBAL_NON_STRING_OR_SPLIT);
|
|
goto fail;
|
|
}
|
|
index = _PyDictKeys_StringLookup(builtin_keys, name);
|
|
if (index == DKIX_ERROR) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_EXPECTED_ERROR);
|
|
goto fail;
|
|
}
|
|
if (index != (uint16_t)index) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
uint32_t globals_version = _PyDictKeys_GetVersionForCurrentState(
|
|
interp, globals_keys);
|
|
if (globals_version == 0) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
goto fail;
|
|
}
|
|
if (globals_version != (uint16_t)globals_version) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
uint32_t builtins_version = _PyDictKeys_GetVersionForCurrentState(
|
|
interp, builtin_keys);
|
|
if (builtins_version == 0) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
goto fail;
|
|
}
|
|
if (builtins_version > UINT16_MAX) {
|
|
SPECIALIZATION_FAIL(LOAD_GLOBAL, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
cache->index = (uint16_t)index;
|
|
cache->module_keys_version = (uint16_t)globals_version;
|
|
cache->builtin_keys_version = (uint16_t)builtins_version;
|
|
instr->op.code = LOAD_GLOBAL_BUILTIN;
|
|
goto success;
|
|
fail:
|
|
STAT_INC(LOAD_GLOBAL, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = LOAD_GLOBAL;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(LOAD_GLOBAL, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
#ifdef Py_STATS
|
|
static int
|
|
binary_subscr_fail_kind(PyTypeObject *container_type, PyObject *sub)
|
|
{
|
|
if (strcmp(container_type->tp_name, "array.array") == 0) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
return SPEC_FAIL_SUBSCR_ARRAY_INT;
|
|
}
|
|
if (PySlice_Check(sub)) {
|
|
return SPEC_FAIL_SUBSCR_ARRAY_SLICE;
|
|
}
|
|
return SPEC_FAIL_OTHER;
|
|
}
|
|
else if (container_type->tp_as_buffer) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
return SPEC_FAIL_SUBSCR_BUFFER_INT;
|
|
}
|
|
if (PySlice_Check(sub)) {
|
|
return SPEC_FAIL_SUBSCR_BUFFER_SLICE;
|
|
}
|
|
return SPEC_FAIL_OTHER;
|
|
}
|
|
else if (container_type->tp_as_sequence) {
|
|
if (PyLong_CheckExact(sub) && container_type->tp_as_sequence->sq_item) {
|
|
return SPEC_FAIL_SUBSCR_SEQUENCE_INT;
|
|
}
|
|
}
|
|
return SPEC_FAIL_OTHER;
|
|
}
|
|
#endif // Py_STATS
|
|
|
|
static int
|
|
function_kind(PyCodeObject *code) {
|
|
int flags = code->co_flags;
|
|
if ((flags & (CO_VARKEYWORDS | CO_VARARGS)) || code->co_kwonlyargcount) {
|
|
return SPEC_FAIL_CODE_COMPLEX_PARAMETERS;
|
|
}
|
|
if ((flags & CO_OPTIMIZED) == 0) {
|
|
return SPEC_FAIL_CODE_NOT_OPTIMIZED;
|
|
}
|
|
return SIMPLE_FUNCTION;
|
|
}
|
|
|
|
/* Returning false indicates a failure. */
|
|
static bool
|
|
function_check_args(PyObject *o, int expected_argcount, int opcode)
|
|
{
|
|
assert(Py_IS_TYPE(o, &PyFunction_Type));
|
|
PyFunctionObject *func = (PyFunctionObject *)o;
|
|
PyCodeObject *fcode = (PyCodeObject *)func->func_code;
|
|
int kind = function_kind(fcode);
|
|
if (kind != SIMPLE_FUNCTION) {
|
|
SPECIALIZATION_FAIL(opcode, kind);
|
|
return false;
|
|
}
|
|
if (fcode->co_argcount != expected_argcount) {
|
|
SPECIALIZATION_FAIL(opcode, SPEC_FAIL_WRONG_NUMBER_ARGUMENTS);
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/* Returning 0 indicates a failure. */
|
|
static uint32_t
|
|
function_get_version(PyObject *o, int opcode)
|
|
{
|
|
assert(Py_IS_TYPE(o, &PyFunction_Type));
|
|
PyFunctionObject *func = (PyFunctionObject *)o;
|
|
uint32_t version = _PyFunction_GetVersionForCurrentState(func);
|
|
if (version == 0) {
|
|
SPECIALIZATION_FAIL(opcode, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
return 0;
|
|
}
|
|
return version;
|
|
}
|
|
|
|
/* Returning 0 indicates a failure. */
|
|
static uint32_t
|
|
type_get_version(PyTypeObject *t, int opcode)
|
|
{
|
|
uint32_t version = t->tp_version_tag;
|
|
if (version == 0) {
|
|
SPECIALIZATION_FAIL(opcode, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
return 0;
|
|
}
|
|
return version;
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_BinarySubscr(
|
|
_PyStackRef container_st, _PyStackRef sub_st, _Py_CODEUNIT *instr)
|
|
{
|
|
PyObject *container = PyStackRef_AsPyObjectBorrow(container_st);
|
|
PyObject *sub = PyStackRef_AsPyObjectBorrow(sub_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[BINARY_SUBSCR] ==
|
|
INLINE_CACHE_ENTRIES_BINARY_SUBSCR);
|
|
_PyBinarySubscrCache *cache = (_PyBinarySubscrCache *)(instr + 1);
|
|
PyTypeObject *container_type = Py_TYPE(container);
|
|
if (container_type == &PyList_Type) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
if (_PyLong_IsNonNegativeCompact((PyLongObject *)sub)) {
|
|
instr->op.code = BINARY_SUBSCR_LIST_INT;
|
|
goto success;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR,
|
|
PySlice_Check(sub) ? SPEC_FAIL_SUBSCR_LIST_SLICE : SPEC_FAIL_OTHER);
|
|
goto fail;
|
|
}
|
|
if (container_type == &PyTuple_Type) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
if (_PyLong_IsNonNegativeCompact((PyLongObject *)sub)) {
|
|
instr->op.code = BINARY_SUBSCR_TUPLE_INT;
|
|
goto success;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR,
|
|
PySlice_Check(sub) ? SPEC_FAIL_SUBSCR_TUPLE_SLICE : SPEC_FAIL_OTHER);
|
|
goto fail;
|
|
}
|
|
if (container_type == &PyUnicode_Type) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
if (_PyLong_IsNonNegativeCompact((PyLongObject *)sub)) {
|
|
instr->op.code = BINARY_SUBSCR_STR_INT;
|
|
goto success;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR,
|
|
PySlice_Check(sub) ? SPEC_FAIL_SUBSCR_STRING_SLICE : SPEC_FAIL_OTHER);
|
|
goto fail;
|
|
}
|
|
if (container_type == &PyDict_Type) {
|
|
instr->op.code = BINARY_SUBSCR_DICT;
|
|
goto success;
|
|
}
|
|
PyTypeObject *cls = Py_TYPE(container);
|
|
PyObject *descriptor = _PyType_Lookup(cls, &_Py_ID(__getitem__));
|
|
if (descriptor && Py_TYPE(descriptor) == &PyFunction_Type) {
|
|
if (!(container_type->tp_flags & Py_TPFLAGS_HEAPTYPE)) {
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, SPEC_FAIL_SUBSCR_NOT_HEAP_TYPE);
|
|
goto fail;
|
|
}
|
|
PyFunctionObject *func = (PyFunctionObject *)descriptor;
|
|
PyCodeObject *fcode = (PyCodeObject *)func->func_code;
|
|
int kind = function_kind(fcode);
|
|
if (kind != SIMPLE_FUNCTION) {
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, kind);
|
|
goto fail;
|
|
}
|
|
if (fcode->co_argcount != 2) {
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, SPEC_FAIL_WRONG_NUMBER_ARGUMENTS);
|
|
goto fail;
|
|
}
|
|
uint32_t version = _PyFunction_GetVersionForCurrentState(func);
|
|
if (version == 0) {
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
goto fail;
|
|
}
|
|
if (_PyInterpreterState_GET()->eval_frame) {
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR, SPEC_FAIL_OTHER);
|
|
goto fail;
|
|
}
|
|
PyHeapTypeObject *ht = (PyHeapTypeObject *)container_type;
|
|
// This pointer is invalidated by PyType_Modified (see the comment on
|
|
// struct _specialization_cache):
|
|
ht->_spec_cache.getitem = descriptor;
|
|
ht->_spec_cache.getitem_version = version;
|
|
instr->op.code = BINARY_SUBSCR_GETITEM;
|
|
goto success;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_SUBSCR,
|
|
binary_subscr_fail_kind(container_type, sub));
|
|
fail:
|
|
STAT_INC(BINARY_SUBSCR, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = BINARY_SUBSCR;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(BINARY_SUBSCR, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_StoreSubscr(_PyStackRef container_st, _PyStackRef sub_st, _Py_CODEUNIT *instr)
|
|
{
|
|
PyObject *container = PyStackRef_AsPyObjectBorrow(container_st);
|
|
PyObject *sub = PyStackRef_AsPyObjectBorrow(sub_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
_PyStoreSubscrCache *cache = (_PyStoreSubscrCache *)(instr + 1);
|
|
PyTypeObject *container_type = Py_TYPE(container);
|
|
if (container_type == &PyList_Type) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
if (_PyLong_IsNonNegativeCompact((PyLongObject *)sub)
|
|
&& ((PyLongObject *)sub)->long_value.ob_digit[0] < (size_t)PyList_GET_SIZE(container))
|
|
{
|
|
instr->op.code = STORE_SUBSCR_LIST_INT;
|
|
goto success;
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_OUT_OF_RANGE);
|
|
goto fail;
|
|
}
|
|
}
|
|
else if (PySlice_Check(sub)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_LIST_SLICE);
|
|
goto fail;
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_OTHER);
|
|
goto fail;
|
|
}
|
|
}
|
|
if (container_type == &PyDict_Type) {
|
|
instr->op.code = STORE_SUBSCR_DICT;
|
|
goto success;
|
|
}
|
|
#ifdef Py_STATS
|
|
PyMappingMethods *as_mapping = container_type->tp_as_mapping;
|
|
if (as_mapping && (as_mapping->mp_ass_subscript
|
|
== PyDict_Type.tp_as_mapping->mp_ass_subscript)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_DICT_SUBCLASS_NO_OVERRIDE);
|
|
goto fail;
|
|
}
|
|
if (PyObject_CheckBuffer(container)) {
|
|
if (PyLong_CheckExact(sub) && (!_PyLong_IsNonNegativeCompact((PyLongObject *)sub))) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_OUT_OF_RANGE);
|
|
}
|
|
else if (strcmp(container_type->tp_name, "array.array") == 0) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_ARRAY_INT);
|
|
}
|
|
else if (PySlice_Check(sub)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_ARRAY_SLICE);
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_OTHER);
|
|
}
|
|
}
|
|
else if (PyByteArray_CheckExact(container)) {
|
|
if (PyLong_CheckExact(sub)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_BYTEARRAY_INT);
|
|
}
|
|
else if (PySlice_Check(sub)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_BYTEARRAY_SLICE);
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_OTHER);
|
|
}
|
|
}
|
|
else {
|
|
if (PyLong_CheckExact(sub)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_BUFFER_INT);
|
|
}
|
|
else if (PySlice_Check(sub)) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_BUFFER_SLICE);
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_OTHER);
|
|
}
|
|
}
|
|
goto fail;
|
|
}
|
|
PyObject *descriptor = _PyType_Lookup(container_type, &_Py_ID(__setitem__));
|
|
if (descriptor && Py_TYPE(descriptor) == &PyFunction_Type) {
|
|
PyFunctionObject *func = (PyFunctionObject *)descriptor;
|
|
PyCodeObject *code = (PyCodeObject *)func->func_code;
|
|
int kind = function_kind(code);
|
|
if (kind == SIMPLE_FUNCTION) {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_PY_SIMPLE);
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_SUBSCR_PY_OTHER);
|
|
}
|
|
goto fail;
|
|
}
|
|
#endif // Py_STATS
|
|
SPECIALIZATION_FAIL(STORE_SUBSCR, SPEC_FAIL_OTHER);
|
|
fail:
|
|
STAT_INC(STORE_SUBSCR, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = STORE_SUBSCR;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(STORE_SUBSCR, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
/* Returns a borrowed reference.
|
|
* The reference is only valid if guarded by a type version check.
|
|
*/
|
|
static PyFunctionObject *
|
|
get_init_for_simple_managed_python_class(PyTypeObject *tp)
|
|
{
|
|
assert(tp->tp_new == PyBaseObject_Type.tp_new);
|
|
if (tp->tp_alloc != PyType_GenericAlloc) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_OVERRIDDEN);
|
|
return NULL;
|
|
}
|
|
if ((tp->tp_flags & Py_TPFLAGS_INLINE_VALUES) == 0) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CALL_INIT_NOT_INLINE_VALUES);
|
|
return NULL;
|
|
}
|
|
if (!(tp->tp_flags & Py_TPFLAGS_HEAPTYPE)) {
|
|
/* Is this possible? */
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_EXPECTED_ERROR);
|
|
return NULL;
|
|
}
|
|
PyObject *init = _PyType_Lookup(tp, &_Py_ID(__init__));
|
|
if (init == NULL || !PyFunction_Check(init)) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CALL_INIT_NOT_PYTHON);
|
|
return NULL;
|
|
}
|
|
int kind = function_kind((PyCodeObject *)PyFunction_GET_CODE(init));
|
|
if (kind != SIMPLE_FUNCTION) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CALL_INIT_NOT_SIMPLE);
|
|
return NULL;
|
|
}
|
|
((PyHeapTypeObject *)tp)->_spec_cache.init = init;
|
|
return (PyFunctionObject *)init;
|
|
}
|
|
|
|
static int
|
|
specialize_class_call(PyObject *callable, _Py_CODEUNIT *instr, int nargs)
|
|
{
|
|
assert(PyType_Check(callable));
|
|
PyTypeObject *tp = _PyType_CAST(callable);
|
|
if (tp->tp_flags & Py_TPFLAGS_IMMUTABLETYPE) {
|
|
int oparg = instr->op.arg;
|
|
if (nargs == 1 && oparg == 1) {
|
|
if (tp == &PyUnicode_Type) {
|
|
instr->op.code = CALL_STR_1;
|
|
return 0;
|
|
}
|
|
else if (tp == &PyType_Type) {
|
|
instr->op.code = CALL_TYPE_1;
|
|
return 0;
|
|
}
|
|
else if (tp == &PyTuple_Type) {
|
|
instr->op.code = CALL_TUPLE_1;
|
|
return 0;
|
|
}
|
|
}
|
|
if (tp->tp_vectorcall != NULL) {
|
|
instr->op.code = CALL_BUILTIN_CLASS;
|
|
return 0;
|
|
}
|
|
goto generic;
|
|
}
|
|
if (Py_TYPE(tp) != &PyType_Type) {
|
|
goto generic;
|
|
}
|
|
if (tp->tp_new == PyBaseObject_Type.tp_new) {
|
|
PyFunctionObject *init = get_init_for_simple_managed_python_class(tp);
|
|
if (type_get_version(tp, CALL) == 0) {
|
|
return -1;
|
|
}
|
|
if (init != NULL) {
|
|
_PyCallCache *cache = (_PyCallCache *)(instr + 1);
|
|
write_u32(cache->func_version, tp->tp_version_tag);
|
|
_Py_SET_OPCODE(*instr, CALL_ALLOC_AND_ENTER_INIT);
|
|
return 0;
|
|
}
|
|
}
|
|
generic:
|
|
instr->op.code = CALL_NON_PY_GENERAL;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
specialize_method_descriptor(PyMethodDescrObject *descr, _Py_CODEUNIT *instr,
|
|
int nargs)
|
|
{
|
|
switch (descr->d_method->ml_flags &
|
|
(METH_VARARGS | METH_FASTCALL | METH_NOARGS | METH_O |
|
|
METH_KEYWORDS | METH_METHOD)) {
|
|
case METH_NOARGS: {
|
|
if (nargs != 1) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_WRONG_NUMBER_ARGUMENTS);
|
|
return -1;
|
|
}
|
|
instr->op.code = CALL_METHOD_DESCRIPTOR_NOARGS;
|
|
return 0;
|
|
}
|
|
case METH_O: {
|
|
if (nargs != 2) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_WRONG_NUMBER_ARGUMENTS);
|
|
return -1;
|
|
}
|
|
PyInterpreterState *interp = _PyInterpreterState_GET();
|
|
PyObject *list_append = interp->callable_cache.list_append;
|
|
_Py_CODEUNIT next = instr[INLINE_CACHE_ENTRIES_CALL + 1];
|
|
bool pop = (next.op.code == POP_TOP);
|
|
int oparg = instr->op.arg;
|
|
if ((PyObject *)descr == list_append && oparg == 1 && pop) {
|
|
instr->op.code = CALL_LIST_APPEND;
|
|
return 0;
|
|
}
|
|
instr->op.code = CALL_METHOD_DESCRIPTOR_O;
|
|
return 0;
|
|
}
|
|
case METH_FASTCALL: {
|
|
instr->op.code = CALL_METHOD_DESCRIPTOR_FAST;
|
|
return 0;
|
|
}
|
|
case METH_FASTCALL | METH_KEYWORDS: {
|
|
instr->op.code = CALL_METHOD_DESCRIPTOR_FAST_WITH_KEYWORDS;
|
|
return 0;
|
|
}
|
|
}
|
|
instr->op.code = CALL_NON_PY_GENERAL;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
specialize_py_call(PyFunctionObject *func, _Py_CODEUNIT *instr, int nargs,
|
|
bool bound_method)
|
|
{
|
|
_PyCallCache *cache = (_PyCallCache *)(instr + 1);
|
|
PyCodeObject *code = (PyCodeObject *)func->func_code;
|
|
int kind = function_kind(code);
|
|
/* Don't specialize if PEP 523 is active */
|
|
if (_PyInterpreterState_GET()->eval_frame) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CALL_PEP_523);
|
|
return -1;
|
|
}
|
|
int argcount = -1;
|
|
if (kind == SPEC_FAIL_CODE_NOT_OPTIMIZED) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CODE_NOT_OPTIMIZED);
|
|
return -1;
|
|
}
|
|
if (kind == SIMPLE_FUNCTION) {
|
|
argcount = code->co_argcount;
|
|
}
|
|
int version = _PyFunction_GetVersionForCurrentState(func);
|
|
if (version == 0) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
return -1;
|
|
}
|
|
write_u32(cache->func_version, version);
|
|
if (argcount == nargs + bound_method) {
|
|
instr->op.code = bound_method ? CALL_BOUND_METHOD_EXACT_ARGS : CALL_PY_EXACT_ARGS;
|
|
}
|
|
else {
|
|
instr->op.code = bound_method ? CALL_BOUND_METHOD_GENERAL : CALL_PY_GENERAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
specialize_py_call_kw(PyFunctionObject *func, _Py_CODEUNIT *instr, int nargs,
|
|
bool bound_method)
|
|
{
|
|
_PyCallCache *cache = (_PyCallCache *)(instr + 1);
|
|
PyCodeObject *code = (PyCodeObject *)func->func_code;
|
|
int kind = function_kind(code);
|
|
/* Don't specialize if PEP 523 is active */
|
|
if (_PyInterpreterState_GET()->eval_frame) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CALL_PEP_523);
|
|
return -1;
|
|
}
|
|
if (kind == SPEC_FAIL_CODE_NOT_OPTIMIZED) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CODE_NOT_OPTIMIZED);
|
|
return -1;
|
|
}
|
|
int version = _PyFunction_GetVersionForCurrentState(func);
|
|
if (version == 0) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
return -1;
|
|
}
|
|
write_u32(cache->func_version, version);
|
|
instr->op.code = bound_method ? CALL_KW_BOUND_METHOD : CALL_KW_PY;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
specialize_c_call(PyObject *callable, _Py_CODEUNIT *instr, int nargs)
|
|
{
|
|
if (PyCFunction_GET_FUNCTION(callable) == NULL) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_OTHER);
|
|
return 1;
|
|
}
|
|
switch (PyCFunction_GET_FLAGS(callable) &
|
|
(METH_VARARGS | METH_FASTCALL | METH_NOARGS | METH_O |
|
|
METH_KEYWORDS | METH_METHOD)) {
|
|
case METH_O: {
|
|
if (nargs != 1) {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_WRONG_NUMBER_ARGUMENTS);
|
|
return 1;
|
|
}
|
|
/* len(o) */
|
|
PyInterpreterState *interp = _PyInterpreterState_GET();
|
|
if (callable == interp->callable_cache.len) {
|
|
instr->op.code = CALL_LEN;
|
|
return 0;
|
|
}
|
|
instr->op.code = CALL_BUILTIN_O;
|
|
return 0;
|
|
}
|
|
case METH_FASTCALL: {
|
|
if (nargs == 2) {
|
|
/* isinstance(o1, o2) */
|
|
PyInterpreterState *interp = _PyInterpreterState_GET();
|
|
if (callable == interp->callable_cache.isinstance) {
|
|
instr->op.code = CALL_ISINSTANCE;
|
|
return 0;
|
|
}
|
|
}
|
|
instr->op.code = CALL_BUILTIN_FAST;
|
|
return 0;
|
|
}
|
|
case METH_FASTCALL | METH_KEYWORDS: {
|
|
instr->op.code = CALL_BUILTIN_FAST_WITH_KEYWORDS;
|
|
return 0;
|
|
}
|
|
default:
|
|
instr->op.code = CALL_NON_PY_GENERAL;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_Call(_PyStackRef callable_st, _Py_CODEUNIT *instr, int nargs)
|
|
{
|
|
PyObject *callable = PyStackRef_AsPyObjectBorrow(callable_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[CALL] == INLINE_CACHE_ENTRIES_CALL);
|
|
assert(_Py_OPCODE(*instr) != INSTRUMENTED_CALL);
|
|
_PyCallCache *cache = (_PyCallCache *)(instr + 1);
|
|
int fail;
|
|
if (PyCFunction_CheckExact(callable)) {
|
|
fail = specialize_c_call(callable, instr, nargs);
|
|
}
|
|
else if (PyFunction_Check(callable)) {
|
|
fail = specialize_py_call((PyFunctionObject *)callable, instr, nargs, false);
|
|
}
|
|
else if (PyType_Check(callable)) {
|
|
fail = specialize_class_call(callable, instr, nargs);
|
|
}
|
|
else if (Py_IS_TYPE(callable, &PyMethodDescr_Type)) {
|
|
fail = specialize_method_descriptor((PyMethodDescrObject *)callable, instr, nargs);
|
|
}
|
|
else if (PyMethod_Check(callable)) {
|
|
PyObject *func = ((PyMethodObject *)callable)->im_func;
|
|
if (PyFunction_Check(func)) {
|
|
fail = specialize_py_call((PyFunctionObject *)func, instr, nargs, true);
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(CALL, SPEC_FAIL_CALL_BOUND_METHOD);
|
|
fail = -1;
|
|
}
|
|
}
|
|
else {
|
|
instr->op.code = CALL_NON_PY_GENERAL;
|
|
fail = 0;
|
|
}
|
|
if (fail) {
|
|
STAT_INC(CALL, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = CALL;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
}
|
|
else {
|
|
STAT_INC(CALL, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_CallKw(_PyStackRef callable_st, _Py_CODEUNIT *instr, int nargs)
|
|
{
|
|
PyObject *callable = PyStackRef_AsPyObjectBorrow(callable_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[CALL_KW] == INLINE_CACHE_ENTRIES_CALL_KW);
|
|
assert(_Py_OPCODE(*instr) != INSTRUMENTED_CALL_KW);
|
|
_PyCallCache *cache = (_PyCallCache *)(instr + 1);
|
|
int fail;
|
|
if (PyFunction_Check(callable)) {
|
|
fail = specialize_py_call_kw((PyFunctionObject *)callable, instr, nargs, false);
|
|
}
|
|
else if (PyMethod_Check(callable)) {
|
|
PyObject *func = ((PyMethodObject *)callable)->im_func;
|
|
if (PyFunction_Check(func)) {
|
|
fail = specialize_py_call_kw((PyFunctionObject *)func, instr, nargs, true);
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(CALL_KW, SPEC_FAIL_CALL_BOUND_METHOD);
|
|
fail = -1;
|
|
}
|
|
}
|
|
else {
|
|
instr->op.code = CALL_KW_NON_PY;
|
|
fail = 0;
|
|
}
|
|
if (fail) {
|
|
STAT_INC(CALL, failure);
|
|
assert(!PyErr_Occurred());
|
|
instr->op.code = CALL_KW;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
}
|
|
else {
|
|
STAT_INC(CALL, success);
|
|
assert(!PyErr_Occurred());
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
}
|
|
|
|
#ifdef Py_STATS
|
|
static int
|
|
binary_op_fail_kind(int oparg, PyObject *lhs, PyObject *rhs)
|
|
{
|
|
switch (oparg) {
|
|
case NB_ADD:
|
|
case NB_INPLACE_ADD:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
return SPEC_FAIL_BINARY_OP_ADD_DIFFERENT_TYPES;
|
|
}
|
|
return SPEC_FAIL_BINARY_OP_ADD_OTHER;
|
|
case NB_AND:
|
|
case NB_INPLACE_AND:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
return SPEC_FAIL_BINARY_OP_AND_DIFFERENT_TYPES;
|
|
}
|
|
if (PyLong_CheckExact(lhs)) {
|
|
return SPEC_FAIL_BINARY_OP_AND_INT;
|
|
}
|
|
return SPEC_FAIL_BINARY_OP_AND_OTHER;
|
|
case NB_FLOOR_DIVIDE:
|
|
case NB_INPLACE_FLOOR_DIVIDE:
|
|
return SPEC_FAIL_BINARY_OP_FLOOR_DIVIDE;
|
|
case NB_LSHIFT:
|
|
case NB_INPLACE_LSHIFT:
|
|
return SPEC_FAIL_BINARY_OP_LSHIFT;
|
|
case NB_MATRIX_MULTIPLY:
|
|
case NB_INPLACE_MATRIX_MULTIPLY:
|
|
return SPEC_FAIL_BINARY_OP_MATRIX_MULTIPLY;
|
|
case NB_MULTIPLY:
|
|
case NB_INPLACE_MULTIPLY:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
return SPEC_FAIL_BINARY_OP_MULTIPLY_DIFFERENT_TYPES;
|
|
}
|
|
return SPEC_FAIL_BINARY_OP_MULTIPLY_OTHER;
|
|
case NB_OR:
|
|
case NB_INPLACE_OR:
|
|
return SPEC_FAIL_BINARY_OP_OR;
|
|
case NB_POWER:
|
|
case NB_INPLACE_POWER:
|
|
return SPEC_FAIL_BINARY_OP_POWER;
|
|
case NB_REMAINDER:
|
|
case NB_INPLACE_REMAINDER:
|
|
return SPEC_FAIL_BINARY_OP_REMAINDER;
|
|
case NB_RSHIFT:
|
|
case NB_INPLACE_RSHIFT:
|
|
return SPEC_FAIL_BINARY_OP_RSHIFT;
|
|
case NB_SUBTRACT:
|
|
case NB_INPLACE_SUBTRACT:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
return SPEC_FAIL_BINARY_OP_SUBTRACT_DIFFERENT_TYPES;
|
|
}
|
|
return SPEC_FAIL_BINARY_OP_SUBTRACT_OTHER;
|
|
case NB_TRUE_DIVIDE:
|
|
case NB_INPLACE_TRUE_DIVIDE:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
return SPEC_FAIL_BINARY_OP_TRUE_DIVIDE_DIFFERENT_TYPES;
|
|
}
|
|
if (PyFloat_CheckExact(lhs)) {
|
|
return SPEC_FAIL_BINARY_OP_TRUE_DIVIDE_FLOAT;
|
|
}
|
|
return SPEC_FAIL_BINARY_OP_TRUE_DIVIDE_OTHER;
|
|
case NB_XOR:
|
|
case NB_INPLACE_XOR:
|
|
return SPEC_FAIL_BINARY_OP_XOR;
|
|
}
|
|
Py_UNREACHABLE();
|
|
}
|
|
#endif // Py_STATS
|
|
|
|
void
|
|
_Py_Specialize_BinaryOp(_PyStackRef lhs_st, _PyStackRef rhs_st, _Py_CODEUNIT *instr,
|
|
int oparg, _PyStackRef *locals)
|
|
{
|
|
PyObject *lhs = PyStackRef_AsPyObjectBorrow(lhs_st);
|
|
PyObject *rhs = PyStackRef_AsPyObjectBorrow(rhs_st);
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[BINARY_OP] == INLINE_CACHE_ENTRIES_BINARY_OP);
|
|
_PyBinaryOpCache *cache = (_PyBinaryOpCache *)(instr + 1);
|
|
switch (oparg) {
|
|
case NB_ADD:
|
|
case NB_INPLACE_ADD:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
break;
|
|
}
|
|
if (PyUnicode_CheckExact(lhs)) {
|
|
_Py_CODEUNIT next = instr[INLINE_CACHE_ENTRIES_BINARY_OP + 1];
|
|
bool to_store = (next.op.code == STORE_FAST);
|
|
if (to_store && PyStackRef_AsPyObjectBorrow(locals[next.op.arg]) == lhs) {
|
|
instr->op.code = BINARY_OP_INPLACE_ADD_UNICODE;
|
|
goto success;
|
|
}
|
|
instr->op.code = BINARY_OP_ADD_UNICODE;
|
|
goto success;
|
|
}
|
|
if (PyLong_CheckExact(lhs)) {
|
|
instr->op.code = BINARY_OP_ADD_INT;
|
|
goto success;
|
|
}
|
|
if (PyFloat_CheckExact(lhs)) {
|
|
instr->op.code = BINARY_OP_ADD_FLOAT;
|
|
goto success;
|
|
}
|
|
break;
|
|
case NB_MULTIPLY:
|
|
case NB_INPLACE_MULTIPLY:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
break;
|
|
}
|
|
if (PyLong_CheckExact(lhs)) {
|
|
instr->op.code = BINARY_OP_MULTIPLY_INT;
|
|
goto success;
|
|
}
|
|
if (PyFloat_CheckExact(lhs)) {
|
|
instr->op.code = BINARY_OP_MULTIPLY_FLOAT;
|
|
goto success;
|
|
}
|
|
break;
|
|
case NB_SUBTRACT:
|
|
case NB_INPLACE_SUBTRACT:
|
|
if (!Py_IS_TYPE(lhs, Py_TYPE(rhs))) {
|
|
break;
|
|
}
|
|
if (PyLong_CheckExact(lhs)) {
|
|
instr->op.code = BINARY_OP_SUBTRACT_INT;
|
|
goto success;
|
|
}
|
|
if (PyFloat_CheckExact(lhs)) {
|
|
instr->op.code = BINARY_OP_SUBTRACT_FLOAT;
|
|
goto success;
|
|
}
|
|
break;
|
|
}
|
|
SPECIALIZATION_FAIL(BINARY_OP, binary_op_fail_kind(oparg, lhs, rhs));
|
|
STAT_INC(BINARY_OP, failure);
|
|
instr->op.code = BINARY_OP;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(BINARY_OP, success);
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
|
|
#ifdef Py_STATS
|
|
static int
|
|
compare_op_fail_kind(PyObject *lhs, PyObject *rhs)
|
|
{
|
|
if (Py_TYPE(lhs) != Py_TYPE(rhs)) {
|
|
if (PyFloat_CheckExact(lhs) && PyLong_CheckExact(rhs)) {
|
|
return SPEC_FAIL_COMPARE_OP_FLOAT_LONG;
|
|
}
|
|
if (PyLong_CheckExact(lhs) && PyFloat_CheckExact(rhs)) {
|
|
return SPEC_FAIL_COMPARE_OP_LONG_FLOAT;
|
|
}
|
|
return SPEC_FAIL_COMPARE_OP_DIFFERENT_TYPES;
|
|
}
|
|
if (PyBytes_CheckExact(lhs)) {
|
|
return SPEC_FAIL_COMPARE_OP_BYTES;
|
|
}
|
|
if (PyTuple_CheckExact(lhs)) {
|
|
return SPEC_FAIL_COMPARE_OP_TUPLE;
|
|
}
|
|
if (PyList_CheckExact(lhs)) {
|
|
return SPEC_FAIL_COMPARE_OP_LIST;
|
|
}
|
|
if (PySet_CheckExact(lhs) || PyFrozenSet_CheckExact(lhs)) {
|
|
return SPEC_FAIL_COMPARE_OP_SET;
|
|
}
|
|
if (PyBool_Check(lhs)) {
|
|
return SPEC_FAIL_COMPARE_OP_BOOL;
|
|
}
|
|
if (Py_TYPE(lhs)->tp_richcompare == PyBaseObject_Type.tp_richcompare) {
|
|
return SPEC_FAIL_COMPARE_OP_BASEOBJECT;
|
|
}
|
|
return SPEC_FAIL_OTHER;
|
|
}
|
|
#endif // Py_STATS
|
|
|
|
void
|
|
_Py_Specialize_CompareOp(_PyStackRef lhs_st, _PyStackRef rhs_st, _Py_CODEUNIT *instr,
|
|
int oparg)
|
|
{
|
|
PyObject *lhs = PyStackRef_AsPyObjectBorrow(lhs_st);
|
|
PyObject *rhs = PyStackRef_AsPyObjectBorrow(rhs_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[COMPARE_OP] == INLINE_CACHE_ENTRIES_COMPARE_OP);
|
|
// All of these specializations compute boolean values, so they're all valid
|
|
// regardless of the fifth-lowest oparg bit.
|
|
_PyCompareOpCache *cache = (_PyCompareOpCache *)(instr + 1);
|
|
if (Py_TYPE(lhs) != Py_TYPE(rhs)) {
|
|
SPECIALIZATION_FAIL(COMPARE_OP, compare_op_fail_kind(lhs, rhs));
|
|
goto failure;
|
|
}
|
|
if (PyFloat_CheckExact(lhs)) {
|
|
instr->op.code = COMPARE_OP_FLOAT;
|
|
goto success;
|
|
}
|
|
if (PyLong_CheckExact(lhs)) {
|
|
if (_PyLong_IsCompact((PyLongObject *)lhs) && _PyLong_IsCompact((PyLongObject *)rhs)) {
|
|
instr->op.code = COMPARE_OP_INT;
|
|
goto success;
|
|
}
|
|
else {
|
|
SPECIALIZATION_FAIL(COMPARE_OP, SPEC_FAIL_COMPARE_OP_BIG_INT);
|
|
goto failure;
|
|
}
|
|
}
|
|
if (PyUnicode_CheckExact(lhs)) {
|
|
int cmp = oparg >> 5;
|
|
if (cmp != Py_EQ && cmp != Py_NE) {
|
|
SPECIALIZATION_FAIL(COMPARE_OP, SPEC_FAIL_COMPARE_OP_STRING);
|
|
goto failure;
|
|
}
|
|
else {
|
|
instr->op.code = COMPARE_OP_STR;
|
|
goto success;
|
|
}
|
|
}
|
|
SPECIALIZATION_FAIL(COMPARE_OP, compare_op_fail_kind(lhs, rhs));
|
|
failure:
|
|
STAT_INC(COMPARE_OP, failure);
|
|
instr->op.code = COMPARE_OP;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(COMPARE_OP, success);
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
#ifdef Py_STATS
|
|
static int
|
|
unpack_sequence_fail_kind(PyObject *seq)
|
|
{
|
|
if (PySequence_Check(seq)) {
|
|
return SPEC_FAIL_UNPACK_SEQUENCE_SEQUENCE;
|
|
}
|
|
if (PyIter_Check(seq)) {
|
|
return SPEC_FAIL_UNPACK_SEQUENCE_ITERATOR;
|
|
}
|
|
return SPEC_FAIL_OTHER;
|
|
}
|
|
#endif // Py_STATS
|
|
|
|
void
|
|
_Py_Specialize_UnpackSequence(_PyStackRef seq_st, _Py_CODEUNIT *instr, int oparg)
|
|
{
|
|
PyObject *seq = PyStackRef_AsPyObjectBorrow(seq_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[UNPACK_SEQUENCE] ==
|
|
INLINE_CACHE_ENTRIES_UNPACK_SEQUENCE);
|
|
_PyUnpackSequenceCache *cache = (_PyUnpackSequenceCache *)(instr + 1);
|
|
if (PyTuple_CheckExact(seq)) {
|
|
if (PyTuple_GET_SIZE(seq) != oparg) {
|
|
SPECIALIZATION_FAIL(UNPACK_SEQUENCE, SPEC_FAIL_EXPECTED_ERROR);
|
|
goto failure;
|
|
}
|
|
if (PyTuple_GET_SIZE(seq) == 2) {
|
|
instr->op.code = UNPACK_SEQUENCE_TWO_TUPLE;
|
|
goto success;
|
|
}
|
|
instr->op.code = UNPACK_SEQUENCE_TUPLE;
|
|
goto success;
|
|
}
|
|
if (PyList_CheckExact(seq)) {
|
|
if (PyList_GET_SIZE(seq) != oparg) {
|
|
SPECIALIZATION_FAIL(UNPACK_SEQUENCE, SPEC_FAIL_EXPECTED_ERROR);
|
|
goto failure;
|
|
}
|
|
instr->op.code = UNPACK_SEQUENCE_LIST;
|
|
goto success;
|
|
}
|
|
SPECIALIZATION_FAIL(UNPACK_SEQUENCE, unpack_sequence_fail_kind(seq));
|
|
failure:
|
|
STAT_INC(UNPACK_SEQUENCE, failure);
|
|
instr->op.code = UNPACK_SEQUENCE;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(UNPACK_SEQUENCE, success);
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
#ifdef Py_STATS
|
|
int
|
|
_PySpecialization_ClassifyIterator(PyObject *iter)
|
|
{
|
|
if (PyGen_CheckExact(iter)) {
|
|
return SPEC_FAIL_ITER_GENERATOR;
|
|
}
|
|
if (PyCoro_CheckExact(iter)) {
|
|
return SPEC_FAIL_ITER_COROUTINE;
|
|
}
|
|
if (PyAsyncGen_CheckExact(iter)) {
|
|
return SPEC_FAIL_ITER_ASYNC_GENERATOR;
|
|
}
|
|
if (PyAsyncGenASend_CheckExact(iter)) {
|
|
return SPEC_FAIL_ITER_ASYNC_GENERATOR_SEND;
|
|
}
|
|
PyTypeObject *t = Py_TYPE(iter);
|
|
if (t == &PyListIter_Type) {
|
|
return SPEC_FAIL_ITER_LIST;
|
|
}
|
|
if (t == &PyTupleIter_Type) {
|
|
return SPEC_FAIL_ITER_TUPLE;
|
|
}
|
|
if (t == &PyDictIterKey_Type) {
|
|
return SPEC_FAIL_ITER_DICT_KEYS;
|
|
}
|
|
if (t == &PyDictIterValue_Type) {
|
|
return SPEC_FAIL_ITER_DICT_VALUES;
|
|
}
|
|
if (t == &PyDictIterItem_Type) {
|
|
return SPEC_FAIL_ITER_DICT_ITEMS;
|
|
}
|
|
if (t == &PySetIter_Type) {
|
|
return SPEC_FAIL_ITER_SET;
|
|
}
|
|
if (t == &PyUnicodeIter_Type) {
|
|
return SPEC_FAIL_ITER_STRING;
|
|
}
|
|
if (t == &PyBytesIter_Type) {
|
|
return SPEC_FAIL_ITER_BYTES;
|
|
}
|
|
if (t == &PyRangeIter_Type) {
|
|
return SPEC_FAIL_ITER_RANGE;
|
|
}
|
|
if (t == &PyEnum_Type) {
|
|
return SPEC_FAIL_ITER_ENUMERATE;
|
|
}
|
|
if (t == &PyMap_Type) {
|
|
return SPEC_FAIL_ITER_MAP;
|
|
}
|
|
if (t == &PyZip_Type) {
|
|
return SPEC_FAIL_ITER_ZIP;
|
|
}
|
|
if (t == &PySeqIter_Type) {
|
|
return SPEC_FAIL_ITER_SEQ_ITER;
|
|
}
|
|
if (t == &PyListRevIter_Type) {
|
|
return SPEC_FAIL_ITER_REVERSED_LIST;
|
|
}
|
|
if (t == &_PyUnicodeASCIIIter_Type) {
|
|
return SPEC_FAIL_ITER_ASCII_STRING;
|
|
}
|
|
const char *name = t->tp_name;
|
|
if (strncmp(name, "itertools", 9) == 0) {
|
|
return SPEC_FAIL_ITER_ITERTOOLS;
|
|
}
|
|
if (strncmp(name, "callable_iterator", 17) == 0) {
|
|
return SPEC_FAIL_ITER_CALLABLE;
|
|
}
|
|
return SPEC_FAIL_OTHER;
|
|
}
|
|
#endif // Py_STATS
|
|
|
|
void
|
|
_Py_Specialize_ForIter(_PyStackRef iter, _Py_CODEUNIT *instr, int oparg)
|
|
{
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[FOR_ITER] == INLINE_CACHE_ENTRIES_FOR_ITER);
|
|
_PyForIterCache *cache = (_PyForIterCache *)(instr + 1);
|
|
PyObject *iter_o = PyStackRef_AsPyObjectBorrow(iter);
|
|
PyTypeObject *tp = Py_TYPE(iter_o);
|
|
if (tp == &PyListIter_Type) {
|
|
instr->op.code = FOR_ITER_LIST;
|
|
goto success;
|
|
}
|
|
else if (tp == &PyTupleIter_Type) {
|
|
instr->op.code = FOR_ITER_TUPLE;
|
|
goto success;
|
|
}
|
|
else if (tp == &PyRangeIter_Type) {
|
|
instr->op.code = FOR_ITER_RANGE;
|
|
goto success;
|
|
}
|
|
else if (tp == &PyGen_Type && oparg <= SHRT_MAX) {
|
|
assert(instr[oparg + INLINE_CACHE_ENTRIES_FOR_ITER + 1].op.code == END_FOR ||
|
|
instr[oparg + INLINE_CACHE_ENTRIES_FOR_ITER + 1].op.code == INSTRUMENTED_END_FOR
|
|
);
|
|
if (_PyInterpreterState_GET()->eval_frame) {
|
|
SPECIALIZATION_FAIL(FOR_ITER, SPEC_FAIL_OTHER);
|
|
goto failure;
|
|
}
|
|
instr->op.code = FOR_ITER_GEN;
|
|
goto success;
|
|
}
|
|
SPECIALIZATION_FAIL(FOR_ITER,
|
|
_PySpecialization_ClassifyIterator(iter_o));
|
|
failure:
|
|
STAT_INC(FOR_ITER, failure);
|
|
instr->op.code = FOR_ITER;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(FOR_ITER, success);
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_Send(_PyStackRef receiver_st, _Py_CODEUNIT *instr)
|
|
{
|
|
PyObject *receiver = PyStackRef_AsPyObjectBorrow(receiver_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[SEND] == INLINE_CACHE_ENTRIES_SEND);
|
|
_PySendCache *cache = (_PySendCache *)(instr + 1);
|
|
PyTypeObject *tp = Py_TYPE(receiver);
|
|
if (tp == &PyGen_Type || tp == &PyCoro_Type) {
|
|
if (_PyInterpreterState_GET()->eval_frame) {
|
|
SPECIALIZATION_FAIL(SEND, SPEC_FAIL_OTHER);
|
|
goto failure;
|
|
}
|
|
instr->op.code = SEND_GEN;
|
|
goto success;
|
|
}
|
|
SPECIALIZATION_FAIL(SEND,
|
|
_PySpecialization_ClassifyIterator(receiver));
|
|
failure:
|
|
STAT_INC(SEND, failure);
|
|
instr->op.code = SEND;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(SEND, success);
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
void
|
|
_Py_Specialize_ToBool(_PyStackRef value_o, _Py_CODEUNIT *instr)
|
|
{
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[TO_BOOL] == INLINE_CACHE_ENTRIES_TO_BOOL);
|
|
_PyToBoolCache *cache = (_PyToBoolCache *)(instr + 1);
|
|
PyObject *value = PyStackRef_AsPyObjectBorrow(value_o);
|
|
if (PyBool_Check(value)) {
|
|
instr->op.code = TO_BOOL_BOOL;
|
|
goto success;
|
|
}
|
|
if (PyLong_CheckExact(value)) {
|
|
instr->op.code = TO_BOOL_INT;
|
|
goto success;
|
|
}
|
|
if (PyList_CheckExact(value)) {
|
|
instr->op.code = TO_BOOL_LIST;
|
|
goto success;
|
|
}
|
|
if (Py_IsNone(value)) {
|
|
instr->op.code = TO_BOOL_NONE;
|
|
goto success;
|
|
}
|
|
if (PyUnicode_CheckExact(value)) {
|
|
instr->op.code = TO_BOOL_STR;
|
|
goto success;
|
|
}
|
|
if (PyType_HasFeature(Py_TYPE(value), Py_TPFLAGS_HEAPTYPE)) {
|
|
PyNumberMethods *nb = Py_TYPE(value)->tp_as_number;
|
|
if (nb && nb->nb_bool) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_NUMBER);
|
|
goto failure;
|
|
}
|
|
PyMappingMethods *mp = Py_TYPE(value)->tp_as_mapping;
|
|
if (mp && mp->mp_length) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_MAPPING);
|
|
goto failure;
|
|
}
|
|
PySequenceMethods *sq = Py_TYPE(value)->tp_as_sequence;
|
|
if (sq && sq->sq_length) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_SEQUENCE);
|
|
goto failure;
|
|
}
|
|
if (!PyUnstable_Type_AssignVersionTag(Py_TYPE(value))) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_OUT_OF_VERSIONS);
|
|
goto failure;
|
|
}
|
|
uint32_t version = type_get_version(Py_TYPE(value), TO_BOOL);
|
|
if (version == 0) {
|
|
goto failure;
|
|
}
|
|
instr->op.code = TO_BOOL_ALWAYS_TRUE;
|
|
write_u32(cache->version, version);
|
|
assert(version);
|
|
goto success;
|
|
}
|
|
#ifdef Py_STATS
|
|
if (PyByteArray_CheckExact(value)) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_BYTEARRAY);
|
|
goto failure;
|
|
}
|
|
if (PyBytes_CheckExact(value)) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_BYTES);
|
|
goto failure;
|
|
}
|
|
if (PyDict_CheckExact(value)) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_DICT);
|
|
goto failure;
|
|
}
|
|
if (PyFloat_CheckExact(value)) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_FLOAT);
|
|
goto failure;
|
|
}
|
|
if (PyMemoryView_Check(value)) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_MEMORY_VIEW);
|
|
goto failure;
|
|
}
|
|
if (PyAnySet_CheckExact(value)) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_SET);
|
|
goto failure;
|
|
}
|
|
if (PyTuple_CheckExact(value)) {
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_TO_BOOL_TUPLE);
|
|
goto failure;
|
|
}
|
|
SPECIALIZATION_FAIL(TO_BOOL, SPEC_FAIL_OTHER);
|
|
#endif // Py_STATS
|
|
failure:
|
|
STAT_INC(TO_BOOL, failure);
|
|
instr->op.code = TO_BOOL;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(TO_BOOL, success);
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
#ifdef Py_STATS
|
|
static int containsop_fail_kind(PyObject *value) {
|
|
if (PyUnicode_CheckExact(value)) {
|
|
return SPEC_FAIL_CONTAINS_OP_STR;
|
|
}
|
|
if (PyList_CheckExact(value)) {
|
|
return SPEC_FAIL_CONTAINS_OP_LIST;
|
|
}
|
|
if (PyTuple_CheckExact(value)) {
|
|
return SPEC_FAIL_CONTAINS_OP_TUPLE;
|
|
}
|
|
if (PyType_Check(value)) {
|
|
return SPEC_FAIL_CONTAINS_OP_USER_CLASS;
|
|
}
|
|
return SPEC_FAIL_OTHER;
|
|
}
|
|
#endif // Py_STATS
|
|
|
|
void
|
|
_Py_Specialize_ContainsOp(_PyStackRef value_st, _Py_CODEUNIT *instr)
|
|
{
|
|
PyObject *value = PyStackRef_AsPyObjectBorrow(value_st);
|
|
|
|
assert(ENABLE_SPECIALIZATION);
|
|
assert(_PyOpcode_Caches[CONTAINS_OP] == INLINE_CACHE_ENTRIES_COMPARE_OP);
|
|
_PyContainsOpCache *cache = (_PyContainsOpCache *)(instr + 1);
|
|
if (PyDict_CheckExact(value)) {
|
|
instr->op.code = CONTAINS_OP_DICT;
|
|
goto success;
|
|
}
|
|
if (PySet_CheckExact(value) || PyFrozenSet_CheckExact(value)) {
|
|
instr->op.code = CONTAINS_OP_SET;
|
|
goto success;
|
|
}
|
|
|
|
SPECIALIZATION_FAIL(CONTAINS_OP, containsop_fail_kind(value));
|
|
STAT_INC(CONTAINS_OP, failure);
|
|
instr->op.code = CONTAINS_OP;
|
|
cache->counter = adaptive_counter_backoff(cache->counter);
|
|
return;
|
|
success:
|
|
STAT_INC(CONTAINS_OP, success);
|
|
cache->counter = adaptive_counter_cooldown();
|
|
}
|
|
|
|
/* Code init cleanup.
|
|
* CALL_ALLOC_AND_ENTER_INIT will set up
|
|
* the frame to execute the EXIT_INIT_CHECK
|
|
* instruction.
|
|
* Ends with a RESUME so that it is not traced.
|
|
* This is used as a plain code object, not a function,
|
|
* so must not access globals or builtins.
|
|
*/
|
|
|
|
#define NO_LOC_4 (128 | (PY_CODE_LOCATION_INFO_NONE << 3) | 3)
|
|
|
|
static const PyBytesObject no_location = {
|
|
PyVarObject_HEAD_INIT(&PyBytes_Type, 1)
|
|
.ob_sval = { NO_LOC_4 }
|
|
};
|
|
|
|
const struct _PyCode8 _Py_InitCleanup = {
|
|
_PyVarObject_HEAD_INIT(&PyCode_Type, 3),
|
|
.co_consts = (PyObject *)&_Py_SINGLETON(tuple_empty),
|
|
.co_names = (PyObject *)&_Py_SINGLETON(tuple_empty),
|
|
.co_exceptiontable = (PyObject *)&_Py_SINGLETON(bytes_empty),
|
|
.co_flags = CO_OPTIMIZED | CO_NO_MONITORING_EVENTS,
|
|
.co_localsplusnames = (PyObject *)&_Py_SINGLETON(tuple_empty),
|
|
.co_localspluskinds = (PyObject *)&_Py_SINGLETON(bytes_empty),
|
|
.co_filename = &_Py_ID(__init__),
|
|
.co_name = &_Py_ID(__init__),
|
|
.co_qualname = &_Py_ID(__init__),
|
|
.co_linetable = (PyObject *)&no_location,
|
|
._co_firsttraceable = 4,
|
|
.co_stacksize = 2,
|
|
.co_framesize = 2 + FRAME_SPECIALS_SIZE,
|
|
.co_code_adaptive = {
|
|
EXIT_INIT_CHECK, 0,
|
|
RETURN_VALUE, 0,
|
|
RESUME, RESUME_AT_FUNC_START,
|
|
}
|
|
};
|