mirror of
https://github.com/mongodb/mongo.git
synced 2024-11-24 00:17:37 +01:00
6982155e42
GitOrigin-RevId: c571ee716edb731ab2c0dbb7ab83d4ba56dd82fa
6955 lines
246 KiB
Python
6955 lines
246 KiB
Python
# -*- mode: python; -*-
|
|
|
|
import atexit
|
|
import copy
|
|
import errno
|
|
import functools
|
|
import json
|
|
import os
|
|
import pathlib
|
|
import platform
|
|
import re
|
|
import shlex
|
|
import shutil
|
|
import subprocess
|
|
import sys
|
|
import textwrap
|
|
import uuid
|
|
from datetime import datetime
|
|
from glob import glob
|
|
|
|
import SCons
|
|
import SCons.Script
|
|
from mongo_tooling_metrics.lib.top_level_metrics import SConsToolingMetrics
|
|
from pkg_resources import parse_version
|
|
|
|
from site_scons.mongo import build_profiles
|
|
|
|
# This must be first, even before EnsureSConsVersion, if
|
|
# we are to avoid bulk loading all tools in the DefaultEnvironment.
|
|
DefaultEnvironment(tools=[])
|
|
|
|
# These come from site_scons/mongo. Import these things
|
|
# after calling DefaultEnvironment, for the sake of paranoia.
|
|
import mongo
|
|
import mongo.generators as mongo_generators
|
|
import mongo.install_actions as install_actions
|
|
import mongo.platform as mongo_platform
|
|
import mongo.toolchain as mongo_toolchain
|
|
|
|
EnsurePythonVersion(3, 10)
|
|
EnsureSConsVersion(3, 1, 1)
|
|
|
|
utc_starttime = datetime.utcnow()
|
|
|
|
|
|
# Monkey patch SCons.FS.File.release_target_info to be a no-op.
|
|
# See https://github.com/SCons/scons/issues/3454
|
|
def release_target_info_noop(self):
|
|
pass
|
|
|
|
|
|
SCons.Node.FS.File.release_target_info = release_target_info_noop
|
|
|
|
import psutil
|
|
|
|
from buildscripts import moduleconfig, utils
|
|
|
|
scons_invocation = "{} {}".format(sys.executable, " ".join(sys.argv))
|
|
print("scons: running with args {}".format(scons_invocation))
|
|
|
|
atexit.register(mongo.print_build_failures)
|
|
|
|
# An extra instance of the SCons parser is used to manually validate options
|
|
# flags. We use it detect some common misspellings/unknown options and
|
|
# communicate with the user more effectively than just allowing Configure to
|
|
# fail.
|
|
# This is to work around issue #4187
|
|
# (https://github.com/SCons/scons/issues/4187). Upon a future upgrade to SCons
|
|
# that incorporates #4187, we should replace this solution with that.
|
|
_parser = SCons.Script.SConsOptions.Parser("")
|
|
|
|
|
|
def add_option(name, **kwargs):
|
|
_parser.add_option("--" + name, **{"default": None, **kwargs})
|
|
|
|
if "dest" not in kwargs:
|
|
kwargs["dest"] = name
|
|
|
|
if "metavar" not in kwargs and kwargs.get("type", None) == "choice":
|
|
kwargs["metavar"] = "[" + "|".join(kwargs["choices"]) + "]"
|
|
|
|
AddOption("--" + name, **kwargs)
|
|
|
|
|
|
def get_option(name):
|
|
return GetOption(name)
|
|
|
|
|
|
def has_option(name):
|
|
optval = GetOption(name)
|
|
# Options with nargs=0 are true when their value is the empty tuple. Otherwise,
|
|
# if the value is falsish (empty string, None, etc.), coerce to False.
|
|
return True if optval == () else bool(optval)
|
|
|
|
|
|
def use_system_version_of_library(name):
|
|
# Disabled during Bazel migration
|
|
return False
|
|
|
|
|
|
def using_system_version_of_cxx_libraries():
|
|
# Disabled during Bazel migration
|
|
return False
|
|
|
|
|
|
def make_variant_dir_generator():
|
|
memoized_variant_dir = [False]
|
|
|
|
def generate_variant_dir(target, source, env, for_signature):
|
|
if not memoized_variant_dir[0]:
|
|
memoized_variant_dir[0] = env.subst("$BUILD_ROOT/$VARIANT_DIR")
|
|
return memoized_variant_dir[0]
|
|
|
|
return generate_variant_dir
|
|
|
|
|
|
# Always randomize the build order to shake out missing edges, and to help the cache:
|
|
# http://scons.org/doc/production/HTML/scons-user/ch24s06.html
|
|
SetOption("random", 1)
|
|
|
|
# Options TODOs:
|
|
#
|
|
# - We should either alphabetize the entire list of options, or split them into logical groups
|
|
# with clear boundaries, and then alphabetize the groups. There is no way in SCons though to
|
|
# inform it of options groups.
|
|
#
|
|
# - Many of these options are currently only either present or absent. This is not good for
|
|
# scripting the build invocation because it means you need to interpolate in the presence of
|
|
# the whole option. It is better to make all options take an optional on/off or true/false
|
|
# using the nargs='const' mechanism.
|
|
#
|
|
|
|
add_option(
|
|
"build-profile",
|
|
choices=[type for type in build_profiles.BuildProfileType],
|
|
default=build_profiles.BuildProfileType.DEFAULT,
|
|
type="choice",
|
|
help="""Short hand for common build configurations. These profiles are well supported by the build
|
|
and are kept up to date. The 'default' profile should be used unless you have the required
|
|
prerequisites in place to use the other profiles, i.e. having the mongodbtoolchain installed
|
|
and being connected to an icecream cluster. For mongodb developers, it is recommended to use
|
|
the 'san' (sanitizer) profile to identify bugs as soon as possible. Check out
|
|
site_scons/mongo/build_profiles.py to see each profile.""",
|
|
)
|
|
|
|
add_option(
|
|
"evergreen-tmp-dir",
|
|
help="Configures the path to the evergreen configured tmp directory.",
|
|
default=None,
|
|
)
|
|
# Preload to perform early fetch fo repositories
|
|
tool = Tool("integrate_bazel")
|
|
tool.exists(DefaultEnvironment())
|
|
mongo_toolchain_execroot = DefaultEnvironment().PrefetchToolchain()
|
|
|
|
build_profile = build_profiles.get_build_profile(get_option("build-profile"))
|
|
|
|
add_option(
|
|
"ninja",
|
|
choices=["enabled", "disabled"],
|
|
default=build_profile.ninja,
|
|
nargs="?",
|
|
const="enabled",
|
|
type="choice",
|
|
help="Enable the build.ninja generator tool stable or canary version",
|
|
)
|
|
|
|
add_option(
|
|
"force-jobs",
|
|
help="Allow more jobs than available cpu's when icecream is not enabled.",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"build-tools",
|
|
choices=["stable", "next"],
|
|
default="stable",
|
|
type="choice",
|
|
help="Enable experimental build tools",
|
|
)
|
|
|
|
add_option(
|
|
"legacy-tarball",
|
|
choices=["true", "false"],
|
|
default="false",
|
|
const="true",
|
|
nargs="?",
|
|
type="choice",
|
|
help="Build a tarball matching the old MongoDB dist targets",
|
|
)
|
|
|
|
add_option(
|
|
"lint-scope",
|
|
choices=["all", "changed"],
|
|
default="all",
|
|
type="choice",
|
|
help="Lint files in the current git diff instead of all files",
|
|
)
|
|
|
|
add_option(
|
|
"bazel-includes-info",
|
|
action="append",
|
|
help="write included headers in bazel label format to put files ([library].bazel_includes)",
|
|
default=[],
|
|
)
|
|
|
|
add_option(
|
|
"install-mode",
|
|
choices=["hygienic"],
|
|
default="hygienic",
|
|
help="select type of installation",
|
|
nargs=1,
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"install-action",
|
|
choices=([*install_actions.available_actions] + ["default"]),
|
|
default="hardlink",
|
|
help="select mechanism to use to install files (advanced option to reduce disk IO and utilization)",
|
|
nargs=1,
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"build-dir",
|
|
default="#build",
|
|
help="build output directory",
|
|
)
|
|
|
|
add_option(
|
|
"release",
|
|
choices=["on", "off"],
|
|
const="on",
|
|
default=build_profile.release,
|
|
help="release build",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"remote-exec-release",
|
|
choices=["on", "off"],
|
|
const="off",
|
|
default=build_profile.remote_exec_release,
|
|
help="Turn on bazel remote execution for release",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"lto",
|
|
help="enable full link time optimizations (experimental, except with MSVC)",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"thin-lto",
|
|
help="enable thin link time optimizations (experimental)",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"endian",
|
|
choices=["big", "little", "auto"],
|
|
default="auto",
|
|
help="endianness of target platform",
|
|
nargs=1,
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"disable-minimum-compiler-version-enforcement",
|
|
help="allow use of unsupported older compilers (NEVER for production builds)",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"ssl",
|
|
help="Enable or Disable SSL",
|
|
choices=["on", "off"],
|
|
default="on",
|
|
const="on",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"wiredtiger",
|
|
choices=["on", "off"],
|
|
const="on",
|
|
default="on",
|
|
help="Enable wiredtiger",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"ocsp-stapling",
|
|
choices=["on", "off"],
|
|
default="on",
|
|
help="Enable OCSP Stapling on servers",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
js_engine_choices = ["mozjs", "none"]
|
|
add_option(
|
|
"js-engine",
|
|
choices=js_engine_choices,
|
|
default=js_engine_choices[0],
|
|
help="JavaScript scripting engine implementation",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"server-js",
|
|
choices=["on", "off"],
|
|
default="on",
|
|
help="Build mongod without JavaScript support",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"libc++",
|
|
help="use libc++ (experimental, requires clang)",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"use-glibcxx-debug",
|
|
help="Enable the glibc++ debug implementations of the C++ standard libary",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"noshell",
|
|
help="don't build shell",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"dbg",
|
|
choices=["on", "off"],
|
|
const="on",
|
|
default=build_profile.dbg,
|
|
help="Enable runtime debugging checks",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
|
|
add_option(
|
|
"debug-symbols",
|
|
choices=["on", "off"],
|
|
const="on",
|
|
default=build_profile.debug_symbols,
|
|
help="Enable producing debug symbols",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"disable-ref-track",
|
|
help="Disables runtime tracking of REF state changes for pages within wiredtiger. "
|
|
"Tracking the REF state changes is useful for debugging but there is a small performance cost.",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"separate-debug",
|
|
choices=["on", "off"],
|
|
const="on",
|
|
default="off",
|
|
help="Produce separate debug files",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"spider-monkey-dbg",
|
|
choices=["on", "off"],
|
|
const="on",
|
|
default="off",
|
|
help="Enable SpiderMonkey debug mode",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"opt",
|
|
choices=["on", "debug", "size", "off", "auto"],
|
|
const="on",
|
|
default=build_profile.opt,
|
|
help="Enable compile-time optimization",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
experimental_optimizations = [
|
|
"O3",
|
|
"builtin-memcmp",
|
|
"fnsi",
|
|
"nofp",
|
|
"nordyn",
|
|
"sandybridge",
|
|
"tbaa",
|
|
"treevec",
|
|
"vishidden",
|
|
]
|
|
experimental_optimization_choices = ["*"]
|
|
experimental_optimization_choices.extend("+" + opt for opt in experimental_optimizations)
|
|
experimental_optimization_choices.extend("-" + opt for opt in experimental_optimizations)
|
|
|
|
add_option(
|
|
"experimental-optimization",
|
|
action="append",
|
|
choices=experimental_optimization_choices,
|
|
const=experimental_optimization_choices[0],
|
|
default=["+sandybridge"],
|
|
help="Enable experimental optimizations",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"debug-compress",
|
|
action="append",
|
|
choices=["off", "as", "ld"],
|
|
default=["auto"],
|
|
help="Compress debug sections",
|
|
)
|
|
|
|
add_option(
|
|
"coverity-build",
|
|
action="store_true",
|
|
default=False,
|
|
help=(
|
|
"Enable coverity build mode, which only means the bazel build will not run. "
|
|
"The bazel build is expected to be run in a prior separate coverity enabled bazel build."
|
|
),
|
|
)
|
|
|
|
add_option(
|
|
"sanitize",
|
|
help="enable selected sanitizers",
|
|
metavar="san1,san2,...sanN",
|
|
default=build_profile.sanitize,
|
|
)
|
|
|
|
add_option(
|
|
"sanitize-coverage",
|
|
help="enable selected coverage sanitizers",
|
|
metavar="cov1,cov2,...covN",
|
|
)
|
|
|
|
add_option(
|
|
"shared-libsan",
|
|
choices=["on", "off"],
|
|
default="off",
|
|
nargs="?",
|
|
const="on",
|
|
help="dynamically link to sanitizer runtime(s)",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"allocator",
|
|
choices=["auto", "system", "tcmalloc-google", "tcmalloc-gperf"],
|
|
default=build_profile.allocator,
|
|
help='allocator to use (use "auto" for best choice for current platform)',
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"gdbserver",
|
|
help="build in gdb server support",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"lldb-server",
|
|
help="build in lldb server support",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"wait-for-debugger",
|
|
help="Wait for debugger attach on process startup",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"gcov",
|
|
help="compile with flags for gcov",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"pgo-profile",
|
|
help="compile with pgo profiling",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"pgo",
|
|
help="compile with pgo. Assumes profile file default.profdata at root of repository",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"enable-http-client",
|
|
choices=["auto", "on", "off"],
|
|
default="auto",
|
|
help="Enable support for HTTP client requests (required WinHTTP or cURL)",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"consolidated-test-bins",
|
|
choices=["on", "off"],
|
|
default="off",
|
|
help="Test binaries should build consolidated versions of themselves as defined by CONSOLIDATED_TARGET",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"use-sasl-client",
|
|
help="Support SASL authentication in the client library",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"use-tracing-profiler",
|
|
choices=["on", "off"],
|
|
default="off",
|
|
help="Enable tracing profiler statistic collection",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"build-fast-and-loose",
|
|
choices=["on", "off", "auto"],
|
|
const="on",
|
|
default="auto",
|
|
help="looser dependency checking",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"disable-warnings-as-errors",
|
|
action="append",
|
|
choices=["configure", "source"],
|
|
const="source",
|
|
default=build_profile.disable_warnings_as_errors,
|
|
help="Don't add a warnings-as-errors flag to compiler command lines in selected contexts; defaults to 'source' if no argument is provided",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"detect-odr-violations",
|
|
help="Have the linker try to detect ODR violations, if supported",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"variables-help",
|
|
help="Print the help text for SCons variables",
|
|
nargs=0,
|
|
)
|
|
|
|
add_option(
|
|
"osx-version-min",
|
|
help="minimum OS X version to support",
|
|
)
|
|
|
|
# https://docs.microsoft.com/en-us/cpp/porting/modifying-winver-and-win32-winnt?view=vs-2017
|
|
# https://docs.microsoft.com/en-us/windows-server/get-started/windows-server-release-info
|
|
win_version_min_choices = {
|
|
"win10": ("0A00", "0000"),
|
|
"ws2016": ("0A00", "1607"),
|
|
"ws2019": ("0A00", "1809"),
|
|
}
|
|
|
|
add_option(
|
|
"win-version-min",
|
|
choices=list(win_version_min_choices.keys()),
|
|
default=None,
|
|
help="minimum Windows version to support",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"cache",
|
|
choices=["all", "nolinked"],
|
|
const="all",
|
|
help="Use an object cache rather than a per-build variant directory (experimental)",
|
|
nargs="?",
|
|
)
|
|
|
|
add_option(
|
|
"cache-dir",
|
|
default="$BUILD_ROOT/scons/cache",
|
|
help="Specify the directory to use for caching objects if --cache is in use",
|
|
)
|
|
|
|
add_option(
|
|
"cache-signature-mode",
|
|
choices=["none", "validate"],
|
|
default="none",
|
|
help="Extra check to validate integrity of cache files after pulling from cache",
|
|
)
|
|
|
|
add_option(
|
|
"cxx-std",
|
|
choices=["20"],
|
|
default="20",
|
|
help="Select the C++ language standard to build with",
|
|
)
|
|
|
|
|
|
def find_mongo_custom_variables():
|
|
files = []
|
|
paths = [path for path in sys.path if "site_scons" in path]
|
|
for path in paths:
|
|
probe = os.path.join(path, "mongo_custom_variables.py")
|
|
if os.path.isfile(probe):
|
|
files.append(probe)
|
|
return files
|
|
|
|
|
|
add_option(
|
|
"variables-files",
|
|
default=build_profile.variables_files,
|
|
action="append",
|
|
help="Specify variables files to load.",
|
|
)
|
|
|
|
add_option(
|
|
"streams-release-build",
|
|
default=False,
|
|
action="store_true",
|
|
help="If set, will include the enterprise streams module in a release build.",
|
|
)
|
|
|
|
link_model_choices = ["auto", "object", "static", "dynamic", "dynamic-strict", "dynamic-sdk"]
|
|
add_option(
|
|
"link-model",
|
|
choices=link_model_choices,
|
|
default=build_profile.link_model,
|
|
help="Select the linking model for the project",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"linker",
|
|
choices=["auto", "gold", "lld", "bfd"],
|
|
default="auto",
|
|
help="Specify the type of linker to use.",
|
|
type="choice",
|
|
)
|
|
|
|
variable_parse_mode_choices = ["auto", "posix", "other"]
|
|
add_option(
|
|
"variable-parse-mode",
|
|
choices=variable_parse_mode_choices,
|
|
default=variable_parse_mode_choices[0],
|
|
help="Select which parsing mode is used to interpret command line variables",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"modules",
|
|
help="Comma-separated list of modules to build. Empty means none. Default is all.",
|
|
)
|
|
|
|
add_option(
|
|
"runtime-hardening",
|
|
choices=["on", "off"],
|
|
default="on",
|
|
help="Enable runtime hardening features (e.g. stack smash protection)",
|
|
type="choice",
|
|
)
|
|
|
|
experimental_runtime_hardenings = [
|
|
"cfex",
|
|
"controlflow",
|
|
"stackclash",
|
|
]
|
|
experimental_runtime_hardening_choices = ["*"]
|
|
experimental_runtime_hardening_choices.extend("+" + opt for opt in experimental_runtime_hardenings)
|
|
experimental_runtime_hardening_choices.extend("-" + opt for opt in experimental_runtime_hardenings)
|
|
|
|
add_option(
|
|
"experimental-runtime-hardening",
|
|
action="append",
|
|
choices=experimental_runtime_hardening_choices,
|
|
const=experimental_runtime_hardening_choices[0],
|
|
default=[],
|
|
help="Enable experimental runtime hardenings",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"use-hardware-crc32",
|
|
choices=["on", "off"],
|
|
default="on",
|
|
help="Enable CRC32 hardware acceleration",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"xray",
|
|
choices=["on", "off"],
|
|
default="off",
|
|
help="Build with LLVM XRay support",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"xray-instruction-threshold",
|
|
help="XRay instrumentation instruction threshold",
|
|
default=1,
|
|
nargs="?",
|
|
type=int,
|
|
)
|
|
|
|
add_option(
|
|
"git-decider",
|
|
choices=["on", "off"],
|
|
const="on",
|
|
default="off",
|
|
help="Use git metadata for out-of-date detection for source files",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"toolchain-root",
|
|
default=mongo_toolchain_execroot if mongo_toolchain_execroot else "",
|
|
help="Name a toolchain root for use with toolchain selection Variables files in etc/scons",
|
|
)
|
|
|
|
if mongo_toolchain_execroot:
|
|
bin_dir = os.path.join(mongo_toolchain_execroot, "external/mongo_toolchain/v4/bin")
|
|
gcc_path = os.path.dirname(
|
|
os.path.realpath(os.path.join(bin_dir, os.readlink(os.path.join(bin_dir, "g++"))))
|
|
)
|
|
clang_path = os.path.dirname(
|
|
os.path.realpath(os.path.join(bin_dir, os.readlink(os.path.join(bin_dir, "clang++"))))
|
|
)
|
|
else:
|
|
gcc_path = ""
|
|
clang_path = ""
|
|
|
|
add_option(
|
|
"bazel-toolchain-clang",
|
|
default=clang_path,
|
|
help="used in Variables files to help find the real bazel toolchain location.",
|
|
)
|
|
|
|
add_option(
|
|
"bazel-toolchain-gcc",
|
|
default=gcc_path,
|
|
help="used in Variables files to help find the real bazel toolchain location.",
|
|
)
|
|
|
|
|
|
add_option(
|
|
"msvc-debugging-format",
|
|
choices=["codeview", "pdb"],
|
|
default="codeview",
|
|
help="Debugging format in debug builds using msvc. Codeview (/Z7) or Program database (/Zi). Default is codeview.",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"use-libunwind",
|
|
choices=["on", "off", "auto"],
|
|
const="on",
|
|
default=build_profile.libunwind,
|
|
help="Enable libunwind for backtraces",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"jlink",
|
|
help="Limit link concurrency. Takes either an integer to limit to or a"
|
|
" float between 0 and 1.0 whereby jobs will be multiplied to get the final"
|
|
" jlink value."
|
|
"\n\nExample: --jlink=0.75 --jobs 8 will result in a jlink value of 6",
|
|
const=0.5,
|
|
default=build_profile.jlink,
|
|
nargs="?",
|
|
type=float,
|
|
)
|
|
|
|
add_option(
|
|
"enable-usdt-probes",
|
|
choices=["on", "off", "auto"],
|
|
default="auto",
|
|
help="Enable USDT probes. Default is auto, which is enabled only on Linux with SystemTap headers",
|
|
type="choice",
|
|
nargs="?",
|
|
const="on",
|
|
)
|
|
|
|
add_option(
|
|
"libdeps-debug",
|
|
choices=["on", "off"],
|
|
const="off",
|
|
help="Print way too much debugging information on how libdeps is handling dependencies.",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"libdeps-linting",
|
|
choices=["on", "off", "print"],
|
|
const="on",
|
|
default="on",
|
|
help="Enable linting of libdeps. Default is on, optionally 'print' will not stop the build.",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"build-metrics",
|
|
metavar="FILE",
|
|
const="build-metrics.json",
|
|
default="",
|
|
help="Enable tracking of build performance and output data as json."
|
|
' Use "-" to output json to stdout, or supply a path to the desired'
|
|
" file to output to. If no argument is supplied, the default log"
|
|
' file will be "build-metrics.json".',
|
|
nargs="?",
|
|
type=str,
|
|
)
|
|
|
|
add_option(
|
|
"visibility-support",
|
|
choices=["auto", "on", "off"],
|
|
const="auto",
|
|
default="auto",
|
|
help="Enable visibility annotations",
|
|
nargs="?",
|
|
type="choice",
|
|
)
|
|
|
|
add_option(
|
|
"force-macos-dynamic-link",
|
|
default=False,
|
|
action="store_true",
|
|
help="Bypass link-model=dynamic check for macos versions <12.",
|
|
)
|
|
|
|
add_option(
|
|
"bazel-dynamic-execution",
|
|
default=False,
|
|
action="store_true",
|
|
help="use bazel dynamic execution experimental feature",
|
|
)
|
|
|
|
|
|
# --build-mongot is a compile flag used by the evergreen build variants that run end-to-end search
|
|
# suites, as it downloads the necessary mongot binary.
|
|
add_option(
|
|
"build-mongot",
|
|
choices=["latest", "release"],
|
|
default=None,
|
|
type="choice",
|
|
help="Installs the appropriate mongot for your architecture",
|
|
)
|
|
|
|
add_option(
|
|
"patch-build-mongot-url",
|
|
default=None,
|
|
help="Installs mongot binary from upstream patch on mongot-master for your architecture",
|
|
)
|
|
|
|
try:
|
|
with open("version.json", "r") as version_fp:
|
|
version_data = json.load(version_fp)
|
|
|
|
if "version" not in version_data:
|
|
print("version.json does not contain a version string")
|
|
Exit(1)
|
|
if "githash" not in version_data:
|
|
version_data["githash"] = utils.get_git_version()
|
|
|
|
except IOError as e:
|
|
# If the file error wasn't because the file is missing, error out
|
|
if e.errno != errno.ENOENT:
|
|
print(("Error opening version.json: {0}".format(e.strerror)))
|
|
Exit(1)
|
|
|
|
version_data = {
|
|
"version": utils.get_git_describe()[1:],
|
|
"githash": utils.get_git_version(),
|
|
}
|
|
|
|
except ValueError as e:
|
|
print(("Error decoding version.json: {0}".format(e)))
|
|
Exit(1)
|
|
|
|
|
|
def to_boolean(s):
|
|
if isinstance(s, bool):
|
|
return s
|
|
elif s.lower() in ("1", "on", "true", "yes"):
|
|
return True
|
|
elif s.lower() in ("0", "off", "false", "no"):
|
|
return False
|
|
raise ValueError(f"Invalid value {s}, must be a boolean-like string")
|
|
|
|
|
|
# Setup the command-line variables
|
|
def variable_shlex_converter(val):
|
|
# If the argument is something other than a string, propagate
|
|
# it literally.
|
|
if not isinstance(val, str):
|
|
return val
|
|
parse_mode = get_option("variable-parse-mode")
|
|
if parse_mode == "auto":
|
|
parse_mode = "other" if mongo_platform.is_running_os("windows") else "posix"
|
|
return shlex.split(val, posix=(parse_mode == "posix"))
|
|
|
|
|
|
# Setup the command-line variables
|
|
def where_is_converter(val):
|
|
path = WhereIs(val)
|
|
if path:
|
|
return os.path.abspath(path)
|
|
return val
|
|
|
|
|
|
def variable_arch_converter(val):
|
|
arches = {
|
|
"x86_64": "x86_64",
|
|
"amd64": "x86_64",
|
|
"emt64": "x86_64",
|
|
"x86": "i386",
|
|
}
|
|
val = val.lower()
|
|
|
|
if val in arches:
|
|
return arches[val]
|
|
|
|
# Uname returns a bunch of possible x86's on Linux.
|
|
# Check whether the value is an i[3456]86 processor.
|
|
if re.match(r"^i[3-6]86$", val):
|
|
return "i386"
|
|
|
|
# Return whatever val is passed in - hopefully it's legit
|
|
return val
|
|
|
|
|
|
def bool_var_converter(val, var):
|
|
try:
|
|
return to_boolean(val)
|
|
except ValueError as exc:
|
|
if val.lower() != "auto":
|
|
raise ValueError(
|
|
f'Invalid {var} value {s}, must be a boolean-like string or "auto"'
|
|
) from exc
|
|
return "auto"
|
|
|
|
|
|
# The Scons 'default' tool enables a lot of tools that we don't actually need to enable.
|
|
# On platforms like Solaris, it actually does the wrong thing by enabling the sunstudio
|
|
# toolchain first. As such it is simpler and more efficient to manually load the precise
|
|
# set of tools we need for each platform.
|
|
# If we aren't on a platform where we know the minimal set of tools, we fall back to loading
|
|
# the 'default' tool.
|
|
def decide_platform_tools():
|
|
if mongo_platform.is_running_os("windows"):
|
|
# we only support MS toolchain on windows
|
|
return ["msvc", "mslink", "mslib", "masm", "vcredist"]
|
|
elif mongo_platform.is_running_os("linux", "solaris"):
|
|
return ["gcc", "g++", "gnulink", "ar", "gas"]
|
|
elif mongo_platform.is_running_os("darwin"):
|
|
return ["gcc", "g++", "applelink", "ar", "libtool", "as", "xcode"]
|
|
else:
|
|
return ["default"]
|
|
|
|
|
|
def variable_tools_converter(val):
|
|
tool_list = shlex.split(val)
|
|
# This list is intentionally not sorted; the order of tool loading
|
|
# matters as some of the tools have dependencies on other tools.
|
|
return tool_list + [
|
|
"distsrc",
|
|
"gziptool",
|
|
"mongo_consolidated_targets",
|
|
"mongo_test_execution",
|
|
"mongo_test_list",
|
|
"mongo_benchmark",
|
|
"mongo_integrationtest",
|
|
"mongo_unittest",
|
|
"mongo_libfuzzer",
|
|
"mongo_pretty_printer_tests",
|
|
"textfile",
|
|
"mongo_workload_simulator",
|
|
]
|
|
|
|
|
|
def variable_distsrc_converter(val):
|
|
if not val.endswith("/"):
|
|
return val + "/"
|
|
return val
|
|
|
|
|
|
def fatal_error(env, msg, *args):
|
|
print(msg.format(*args))
|
|
Exit(1)
|
|
|
|
|
|
# Apply the default variables files, and walk the provided
|
|
# arguments. Interpret any falsy argument (like the empty string) as
|
|
# resetting any prior state. This makes the argument
|
|
# --variables-files= destructive of any prior variables files
|
|
# arguments, including the default.
|
|
variables_files_args = get_option("variables-files")
|
|
variables_files = find_mongo_custom_variables()
|
|
for variables_file in variables_files_args:
|
|
if variables_file:
|
|
variables_files.append(variables_file)
|
|
else:
|
|
variables_files = []
|
|
for vf in variables_files:
|
|
if not os.path.isfile(vf):
|
|
fatal_error(None, f"Specified variables file '{vf}' does not exist")
|
|
print(f"Using variable customization file {vf}")
|
|
|
|
env_vars = Variables(
|
|
files=variables_files,
|
|
args=ARGUMENTS,
|
|
)
|
|
|
|
sconsflags = os.environ.get("SCONSFLAGS", None)
|
|
if sconsflags:
|
|
print(("Using SCONSFLAGS environment variable arguments: %s" % sconsflags))
|
|
|
|
env_vars.Add(
|
|
"ABIDW",
|
|
help="Configures the path to the 'abidw' (a libabigail) utility",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"AR",
|
|
help="Sets path for the archiver",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ARFLAGS",
|
|
help="Sets flags for the archiver",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CCACHE",
|
|
help="Tells SCons where the ccache binary is",
|
|
default=build_profile.CCACHE,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CACHE_SIZE",
|
|
help="Maximum size of the SCons cache (in gigabytes)",
|
|
default=32,
|
|
converter=lambda x: int(x),
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CACHE_PRUNE_TARGET",
|
|
help="Maximum percent in-use in SCons cache after pruning",
|
|
default=66,
|
|
converter=lambda x: int(x),
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CC",
|
|
help="Selects the C compiler to use",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CCFLAGS",
|
|
help="Sets flags for the C and C++ compiler",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"NON_CONF_CCFLAGS",
|
|
help="Sets flags for the C and C++ compiler that are not used in configure checks",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"TOOLCHAIN_CCFLAGS",
|
|
help="Sets flags for the C and C++ compiler specific to the toolchain",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"NON_CONF_LINKFLAGS",
|
|
help="Sets flags for the C and C++ linker that are not used in configure checks",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"TOOLCHAIN_LINKFLAGS",
|
|
help="Sets flags for the C and C++ linker specific to the toolchain",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ASFLAGS",
|
|
help="Sets assembler specific flags",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CFLAGS",
|
|
help="Sets flags for the C compiler",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CPPDEFINES",
|
|
help="Sets pre-processor definitions for C and C++",
|
|
converter=variable_shlex_converter,
|
|
default=[],
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CPPPATH",
|
|
help="Adds paths to the preprocessor search path",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CXX",
|
|
help="Selects the C++ compiler to use",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"CXXFLAGS",
|
|
help="Sets flags for the C++ compiler",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"UNITTESTS_COMPILE_CONCURRENCY",
|
|
help="Sets the ratio of total jobs for max concurrency when compiling unittests source files. Should be float between 0 and 1.",
|
|
default="1",
|
|
converter=lambda val: float(val) if val != "" else "",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"DESTDIR",
|
|
help="Where builds will install files",
|
|
default="$BUILD_ROOT/install",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"BAZEL_FLAGS",
|
|
help="Flags specific to bazel to pass through to the underlying bazel build command.",
|
|
default="",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"BAZEL_INTEGRATION_DEBUG",
|
|
help="Enable SCons/Bazel integration debug output",
|
|
converter=functools.partial(bool_var_converter, var="BAZEL_INTEGRATION_DEBUG"),
|
|
default="0",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"DSYMUTIL",
|
|
help="Path to the dsymutil utility",
|
|
)
|
|
|
|
|
|
def validate_dwarf_version(key, val, env):
|
|
if val == "4" or val == "5" or val == "":
|
|
return
|
|
|
|
print(f"Invalid DWARF_VERSION '{val}'. Only valid versions are 4 or 5.")
|
|
Exit(1)
|
|
|
|
|
|
env_vars.Add(
|
|
"DWARF_VERSION",
|
|
help="Sets the DWARF version (non-Windows). Incompatible with SPLIT_DWARF=1.",
|
|
validator=validate_dwarf_version,
|
|
converter=lambda val: int(val) if val != "" else "",
|
|
default="",
|
|
)
|
|
|
|
|
|
def validate_dwarf_width(key, val, env):
|
|
if val == "32" or val == "64" or val == "":
|
|
return
|
|
|
|
print(f"Invalid DWARF_WIDTH '{val}'. Only valid versions are 32 or 64.")
|
|
Exit(1)
|
|
|
|
|
|
env_vars.Add(
|
|
"DWARF_WIDTH",
|
|
help="Sets the DWARF addressing mode to either 32-bit or 64-bit (non-Windows)",
|
|
validator=validate_dwarf_width,
|
|
converter=lambda val: int(val) if val != "" else "",
|
|
default="",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"READELF",
|
|
help="Path to readelf",
|
|
default="readelf",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"GITDIFFFLAGS",
|
|
help="Sets flags for git diff",
|
|
default="",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"REVISION",
|
|
help="Base git revision",
|
|
default="",
|
|
)
|
|
|
|
# Note: This probably is only really meaningful when configured via a variables file. It will
|
|
# also override whatever the SCons platform defaults would be.
|
|
env_vars.Add(
|
|
"ENV",
|
|
help="Sets the environment for subprocesses",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"FRAMEWORKPATH",
|
|
help="Adds paths to the linker search path for darwin frameworks",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"FRAMEWORKS",
|
|
help="Adds extra darwin frameworks to link against",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"HOST_ARCH",
|
|
help="Sets the native architecture of the compiler",
|
|
converter=variable_arch_converter,
|
|
default=None,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ICECC",
|
|
help="Tells SCons where icecream icecc tool is",
|
|
default=build_profile.ICECC,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ICERUN",
|
|
help="Tells SCons where icecream icerun tool is",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ICECC_CREATE_ENV",
|
|
help="Tells SCons where icecc-create-env tool is",
|
|
default="icecc-create-env",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ICECC_DEBUG",
|
|
help="Tell ICECC to create debug logs (auto, on/off true/false 1/0)",
|
|
default=False,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ICECC_SCHEDULER",
|
|
help="Tells ICECC where the scheduler daemon is running",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ICECC_VERSION",
|
|
help="Tells ICECC where the compiler package is",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ICECC_VERSION_ARCH",
|
|
help="Tells ICECC the target architecture for the compiler package, if non-native",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"LIBPATH",
|
|
help="Adds paths to the linker search path",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"LIBS",
|
|
help="Adds extra libraries to link against",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"LINKFLAGS",
|
|
help="Sets flags for the linker",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"LLVM_SYMBOLIZER",
|
|
help="Name of or path to the LLVM symbolizer",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MAXLINELENGTH",
|
|
help="Maximum line length before using temp files",
|
|
# This is very small, but appears to be the least upper bound
|
|
# across our platforms.
|
|
#
|
|
# See https://support.microsoft.com/en-us/help/830473/command-prompt-cmd.-exe-command-line-string-limitation
|
|
default=4095,
|
|
)
|
|
|
|
# Note: This is only really meaningful when configured via a variables file. See the
|
|
# default_buildinfo_environment_data() function for examples of how to use this.
|
|
env_vars.Add(
|
|
"MONGO_BUILDINFO_ENVIRONMENT_DATA",
|
|
help="Sets the info returned from the buildInfo command and --version command-line flag",
|
|
default=mongo_generators.default_buildinfo_environment_data(),
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MONGO_DIST_SRC_PREFIX",
|
|
help="Sets the prefix for files in the source distribution archive",
|
|
converter=variable_distsrc_converter,
|
|
default="mongodb-src-r${MONGO_VERSION}",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MONGO_DISTARCH",
|
|
help="Adds a string representing the target processor architecture to the dist archive",
|
|
default="$TARGET_ARCH",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MONGO_DISTMOD",
|
|
help="Adds a string that will be embedded in the dist archive naming",
|
|
default="",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MONGO_DISTNAME",
|
|
help="Sets the version string to be used in dist archive naming",
|
|
default="$MONGO_VERSION",
|
|
)
|
|
|
|
|
|
def validate_mongo_version(key, val, env):
|
|
valid_version_re = re.compile(r"^(\d+)\.(\d+)\.(\d+)-?((?:(rc)(\d+))?.*)?$", re.MULTILINE)
|
|
invalid_version_re = re.compile(r"^0\.0\.0(?:-.*)?", re.MULTILINE)
|
|
if not valid_version_re.match(val) or invalid_version_re.match(val):
|
|
print(
|
|
(
|
|
"Invalid MONGO_VERSION '{}', or could not derive from version.json or git metadata. Please add a conforming MONGO_VERSION=x.y.z[-extra] as an argument to SCons".format(
|
|
val
|
|
)
|
|
)
|
|
)
|
|
Exit(1)
|
|
|
|
|
|
env_vars.Add(
|
|
"MONGO_VERSION",
|
|
help="Sets the version string for MongoDB",
|
|
default=version_data["version"],
|
|
validator=validate_mongo_version,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MONGO_GIT_HASH",
|
|
help="Sets the githash to store in the MongoDB version information",
|
|
default=version_data["githash"],
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MSVC_USE_SCRIPT",
|
|
help="Sets the script used to setup Visual Studio.",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"MSVC_VERSION",
|
|
help="Sets the version of Visual C++ to use (e.g. 14.2 for VS2019, 14.3 for VS2022)",
|
|
default="14.3",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"LINKFLAGS_COMPILER_EXEC_PREFIX",
|
|
help="Specify the search path to be injected into the LINKFLAGS",
|
|
default="",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"COMPILER_EXEC_PREFIX_OPT",
|
|
help="Specify the option sign for compiler exec search paths.",
|
|
default="-B",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"NINJA_BUILDDIR",
|
|
help="Location for shared Ninja state",
|
|
default="$BUILD_ROOT/ninja",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"NINJA_PREFIX",
|
|
default=build_profile.NINJA_PREFIX,
|
|
help="""A prefix to add to the beginning of generated ninja
|
|
files. Useful for when compiling multiple build ninja files for
|
|
different configurations, for instance:
|
|
|
|
scons --sanitize=asan --ninja NINJA_PREFIX=asan asan.ninja
|
|
scons --sanitize=tsan --ninja NINJA_PREFIX=tsan tsan.ninja
|
|
|
|
Will generate the files (respectively):
|
|
|
|
asan.ninja
|
|
tsan.ninja
|
|
|
|
Defaults to build. Best used with the --ninja flag so you don't have to
|
|
reiterate the prefix in the target name and variable.
|
|
""",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"NINJA_SUFFIX",
|
|
help="""A suffix to add to the end of generated build.ninja
|
|
files. Useful for when compiling multiple build ninja files for
|
|
different configurations, for instance:
|
|
|
|
scons --sanitize=asan --ninja NINJA_SUFFIX=asan build.ninja
|
|
scons --sanitize=tsan --ninja NINJA_SUFFIX=tsan build.ninja
|
|
|
|
Will generate the files (respectively):
|
|
|
|
build.ninja.asan
|
|
build.ninja.tsan
|
|
""",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"__NINJA_NO",
|
|
help="Disables the Ninja tool unconditionally. Not intended for human use.",
|
|
default=0,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"COMPILATIONDB_IGNORE_WRAPPERS",
|
|
help="Comma separated list of variables which reference wrapper binaries that should be excluded when generating compile_commands.json",
|
|
default="$ICECC,$ICERUN,$ICECREAM_RUN_ICECC,$CCACHE",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"OBJCOPY",
|
|
help="Sets the path to objcopy",
|
|
default=WhereIs("objcopy"),
|
|
)
|
|
|
|
env_vars.Add(
|
|
"PKGDIR",
|
|
help="Directory in which to build packages and archives",
|
|
default="$BUILD_DIR/pkgs",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"PREFIX",
|
|
help="Final installation location of files. Will be made into a sub dir of $DESTDIR",
|
|
default=".",
|
|
)
|
|
|
|
# Exposed to be able to cross compile Android/*nix from Windows without ending up with the .exe suffix.
|
|
env_vars.Add(
|
|
"PROGSUFFIX",
|
|
help="Sets the suffix for built executable files",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"RPATH",
|
|
help="Set the RPATH for dynamic libraries and executables",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"SHCCFLAGS",
|
|
help="Sets flags for the C and C++ compiler when building shared libraries",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"SHCFLAGS",
|
|
help="Sets flags for the C compiler when building shared libraries",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"SHCXXFLAGS",
|
|
help="Sets flags for the C++ compiler when building shared libraries",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"SHELL",
|
|
help="Picks the shell to use when spawning commands",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"SHLINKFLAGS",
|
|
help="Sets flags for the linker when building shared libraries",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"SHLINKFLAGS_EXTRA",
|
|
help="Adds additional flags for shared links without overwriting tool configured SHLINKFLAGS values",
|
|
converter=variable_shlex_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"STRIP",
|
|
help="Path to the strip utility (non-darwin platforms probably use OBJCOPY for this)",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"PROTOC",
|
|
default="$$PROTOC_VAR_GEN",
|
|
help="Path to protobuf compiler.",
|
|
converter=where_is_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"PROTOC_GRPC_PLUGIN",
|
|
default="$$PROTOC_GRPC_PLUGIN_GEN",
|
|
help="Path to protobuf compiler grpc plugin.",
|
|
converter=where_is_converter,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"SPLIT_DWARF",
|
|
help="Set the boolean (auto, on/off true/false 1/0) to enable gsplit-dwarf (non-Windows). Incompatible with DWARF_VERSION=5",
|
|
converter=functools.partial(bool_var_converter, var="SPLIT_DWARF"),
|
|
default="auto",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ENABLE_GRPC_BUILD",
|
|
help="Set the boolean (auto, on/off true/false 1/0) to enable building grpc and protobuf compiler.",
|
|
converter=functools.partial(bool_var_converter, var="ENABLE_GRPC_BUILD"),
|
|
default="0",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"GDB",
|
|
help="Configures the path to the 'gdb' debugger binary.",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"GDB_INDEX",
|
|
help="Set the boolean (auto, on/off true/false 1/0) to enable creation of a gdb_index in binaries.",
|
|
converter=functools.partial(bool_var_converter, var="GDB_INDEX"),
|
|
default="auto",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"GDB_PPTEST_PYONLY",
|
|
help="""Set the boolean (on/off true/false 1/0) to enable SCons to only emit the .py files
|
|
needed for testing GDB pretty printers.
|
|
|
|
Useful for when the executable files for testing GDB pretty printers are more rapidly built
|
|
using Ninja and the install-dist-test target (separately from SCons).""",
|
|
converter=functools.partial(bool_var_converter, var="GDB_PPTEST_PYONLY"),
|
|
default="False",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"ENABLE_BUILD_RETRY",
|
|
help="Set the boolean (auto, on/off true/false 1/0) to enable retrying a compile or link commands failures.",
|
|
converter=functools.partial(bool_var_converter, var="ENABLE_BUILD_RETRY"),
|
|
default="False",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"TAPI",
|
|
help="Configures the path to the 'tapi' (an Xcode) utility",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"TARGET_ARCH",
|
|
help="Sets the architecture to build for",
|
|
converter=variable_arch_converter,
|
|
default=None,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"TARGET_OS",
|
|
help="Sets the target OS to build for",
|
|
default=mongo_platform.get_running_os_name(),
|
|
)
|
|
|
|
env_vars.Add(
|
|
"TOOLS",
|
|
help="Sets the list of SCons tools to add to the environment",
|
|
converter=variable_tools_converter,
|
|
default=decide_platform_tools(),
|
|
)
|
|
|
|
env_vars.Add(
|
|
"VARIANT_DIR",
|
|
help="Sets the name (or generator function) for the variant directory",
|
|
default=build_profile.VARIANT_DIR,
|
|
)
|
|
|
|
env_vars.Add(
|
|
"VERBOSE",
|
|
help="Controls build verbosity (auto, on/off true/false 1/0)",
|
|
default="auto",
|
|
)
|
|
|
|
env_vars.Add(
|
|
"WINDOWS_OPENSSL_BIN",
|
|
help="Sets the path to the openssl binaries for packaging",
|
|
default="c:/openssl/bin",
|
|
)
|
|
|
|
env_vars.Add(
|
|
PathVariable(
|
|
"LOCAL_TMPDIR",
|
|
help="Set the TMPDIR when running tests.",
|
|
default="$BUILD_ROOT/tmp_test_data",
|
|
validator=PathVariable.PathAccept,
|
|
),
|
|
)
|
|
|
|
env_vars.AddVariables(
|
|
("BUILD_METRICS_EVG_TASK_ID", "Evergreen task ID to add to build metrics data."),
|
|
("BUILD_METRICS_EVG_BUILD_VARIANT", "Evergreen build variant to add to build metrics data."),
|
|
)
|
|
for tool in ["build_metrics", "split_dwarf"]:
|
|
try:
|
|
Tool(tool).options(env_vars)
|
|
except ImportError as exc:
|
|
print(f"WARNING: The {tool} tool might not work as intended due to a failed import:\n{exc}")
|
|
pass
|
|
|
|
# -- Validate user provided options --
|
|
|
|
# A dummy environment that should *only* have the variables we have set. In practice it has
|
|
# some other things because SCons isn't quite perfect about keeping variable initialization
|
|
# scoped to Tools, but it should be good enough to do validation on any Variable values that
|
|
# came from the command line or from loaded files.
|
|
variables_only_env = Environment(
|
|
# Disable platform specific variable injection
|
|
platform=(lambda x: ()),
|
|
# But do *not* load any tools, since those might actually set variables. Note that this
|
|
# causes the value of our TOOLS variable to have no effect.
|
|
tools=[],
|
|
# Use the Variables specified above.
|
|
variables=env_vars,
|
|
)
|
|
|
|
# don't run configure if user calls --help
|
|
if GetOption("help"):
|
|
try:
|
|
Help("\nThe following variables may also be set like scons VARIABLE=value\n", append=True)
|
|
Help(env_vars.GenerateHelpText(variables_only_env, sort=True), append=True)
|
|
Help(
|
|
"\nThe 'list-targets' target can be built to list useful comprehensive build targets\n",
|
|
append=True,
|
|
)
|
|
except TypeError:
|
|
# The append=true kwarg is only supported in scons>=2.4. Without it, calls to Help() clobber
|
|
# the automatically generated options help, which we don't want. Users on older scons
|
|
# versions will need to use --variables-help to learn about which variables we support.
|
|
pass
|
|
|
|
Return()
|
|
|
|
if ("CC" in variables_only_env) != ("CXX" in variables_only_env):
|
|
print("Cannot customize C compiler without customizing C++ compiler, and vice versa")
|
|
Exit(1)
|
|
|
|
# --- environment setup ---
|
|
|
|
# If the user isn't using the # to indicate top-of-tree or $ to expand a variable, forbid
|
|
# relative paths. Relative paths don't really work as expected, because they end up relative to
|
|
# the top level SConstruct, not the invoker's CWD. We could in theory fix this with
|
|
# GetLaunchDir, but that seems a step too far.
|
|
buildDir = get_option("build-dir").rstrip("/")
|
|
if buildDir[0] not in ["$", "#"]:
|
|
if not os.path.isabs(buildDir):
|
|
print("Do not use relative paths with --build-dir")
|
|
Exit(1)
|
|
|
|
cacheDir = get_option("cache-dir").rstrip("/")
|
|
if cacheDir[0] not in ["$", "#"]:
|
|
if not os.path.isabs(cacheDir):
|
|
print("Do not use relative paths with --cache-dir")
|
|
Exit(1)
|
|
|
|
sconsDataDir = Dir(buildDir).Dir("scons")
|
|
SConsignFile(str(sconsDataDir.File("sconsign.py3")))
|
|
|
|
|
|
def printLocalInfo():
|
|
import sys
|
|
|
|
import SCons
|
|
|
|
print(("scons version: " + SCons.__version__))
|
|
print(("python version: " + " ".join([repr(i) for i in sys.version_info])))
|
|
|
|
|
|
printLocalInfo()
|
|
|
|
boostLibs = ["filesystem", "program_options", "system", "iostreams", "thread", "log"]
|
|
|
|
onlyServer = len(COMMAND_LINE_TARGETS) == 0 or (
|
|
len(COMMAND_LINE_TARGETS) == 1 and str(COMMAND_LINE_TARGETS[0]) in ["mongod", "mongos", "test"]
|
|
)
|
|
|
|
noshell = has_option("noshell")
|
|
|
|
jsEngine = get_option("js-engine")
|
|
|
|
serverJs = get_option("server-js") == "on"
|
|
|
|
if not serverJs and not jsEngine:
|
|
print("Warning: --server-js=off is not needed with --js-engine=none")
|
|
|
|
# We defer building the env until we have determined whether we want certain values. Some values
|
|
# in the env actually have semantics for 'None' that differ from being absent, so it is better
|
|
# to build it up via a dict, and then construct the Environment in one shot with kwargs.
|
|
#
|
|
# Yes, BUILD_ROOT vs BUILD_DIR is confusing. Ideally, BUILD_DIR would actually be called
|
|
# VARIANT_DIR, and at some point we should probably do that renaming. Until we do though, we
|
|
# also need an Environment variable for the argument to --build-dir, which is the parent of all
|
|
# variant dirs. For now, we call that BUILD_ROOT. If and when we s/BUILD_DIR/VARIANT_DIR/g,
|
|
# then also s/BUILD_ROOT/BUILD_DIR/g.
|
|
envDict = dict(
|
|
BUILD_ROOT=buildDir,
|
|
BUILD_DIR=make_variant_dir_generator(),
|
|
DIST_ARCHIVE_SUFFIX=".tgz",
|
|
MODULE_BANNERS=[],
|
|
MODULE_INJECTORS=dict(),
|
|
PYTHON="$( {} $)".format(sys.executable),
|
|
SERVER_ARCHIVE="${SERVER_DIST_BASENAME}${DIST_ARCHIVE_SUFFIX}",
|
|
UNITTEST_ALIAS="install-unittests",
|
|
# TODO: Move unittests.txt to $BUILD_DIR, but that requires
|
|
# changes to MCI.
|
|
UNITTEST_LIST="$BUILD_ROOT/unittests.txt",
|
|
PRETTY_PRINTER_TEST_ALIAS="install-pretty-printer-tests",
|
|
PRETTY_PRINTER_TEST_LIST="$BUILD_ROOT/pretty_printer_tests.txt",
|
|
LIBFUZZER_TEST_ALIAS="install-fuzzertests",
|
|
LIBFUZZER_TEST_LIST="$BUILD_ROOT/libfuzzer_tests.txt",
|
|
INTEGRATION_TEST_ALIAS="install-integration-tests",
|
|
INTEGRATION_TEST_LIST="$BUILD_ROOT/integration_tests.txt",
|
|
BENCHMARK_ALIAS="install-benchmarks",
|
|
BENCHMARK_LIST="$BUILD_ROOT/benchmarks.txt",
|
|
CONFIGUREDIR="$BUILD_ROOT/scons/$VARIANT_DIR/sconf_temp",
|
|
CONFIGURELOG="$BUILD_ROOT/scons/config.log",
|
|
LIBDEPS_TAG_EXPANSIONS=[],
|
|
)
|
|
|
|
# By default, we will get the normal SCons tool search. But if the
|
|
# user has opted into the next gen tools, add our experimental tool
|
|
# directory into the default toolpath, ahead of whatever is already in
|
|
# there so it overrides it.
|
|
if get_option("build-tools") == "next":
|
|
SCons.Tool.DefaultToolpath.insert(0, os.path.abspath("site_scons/site_tools/next"))
|
|
|
|
env = Environment(variables=env_vars, **envDict)
|
|
del envDict
|
|
env.AddMethod(lambda env, name, **kwargs: add_option(name, **kwargs), "AddOption")
|
|
|
|
env.Prepend(CCFLAGS="$TOOLCHAIN_CCFLAGS")
|
|
env.Prepend(LINKFLAGS="$TOOLCHAIN_LINKFLAGS")
|
|
|
|
if ARGUMENTS.get("CC") and ARGUMENTS.get("CXX"):
|
|
os.environ["CC"] = env.get("CC")
|
|
os.environ["CXX"] = env.get("CXX")
|
|
os.environ["USE_NATIVE_TOOLCHAIN"] = "1"
|
|
|
|
# Early load to setup env functions
|
|
tool = Tool("integrate_bazel")
|
|
tool.exists(env)
|
|
|
|
# The placement of this is intentional. Here we setup an atexit method to store tooling metrics.
|
|
# We should only register this function after env, env_vars and the parser have been properly initialized.
|
|
SConsToolingMetrics.register_metrics(
|
|
utc_starttime=datetime.utcnow(),
|
|
artifact_dir=env.Dir("$BUILD_DIR").get_abspath(),
|
|
env_vars=env_vars,
|
|
env=env,
|
|
parser=_parser,
|
|
)
|
|
|
|
if get_option("build-metrics"):
|
|
env["BUILD_METRICS_ARTIFACTS_DIR"] = "$BUILD_ROOT/$VARIANT_DIR"
|
|
env.Tool("build_metrics")
|
|
env.AddBuildMetricsMetaData("evg_id", env.get("BUILD_METRICS_EVG_TASK_ID", "UNKNOWN"))
|
|
env.AddBuildMetricsMetaData("variant", env.get("BUILD_METRICS_EVG_BUILD_VARIANT", "UNKNOWN"))
|
|
|
|
env.Execute(SCons.Defaults.Mkdir(env.Dir("$LOCAL_TMPDIR")))
|
|
|
|
if get_option("cache-signature-mode") == "validate":
|
|
validate_cache_dir = Tool("validate_cache_dir")
|
|
if validate_cache_dir.exists(env):
|
|
validate_cache_dir(env)
|
|
else:
|
|
env.FatalError("Failed to enable validate_cache_dir tool.")
|
|
|
|
# Only print the spinner if stdout is a tty
|
|
if sys.stdout.isatty():
|
|
Progress(["-\r", "\\\r", "|\r", "/\r"], interval=50)
|
|
|
|
|
|
# We are going to start running conf tests soon, so setup
|
|
# --disable-warnings-as-errors as soon as possible.
|
|
def create_werror_generator(flagname):
|
|
werror_conftests = "configure" not in get_option("disable-warnings-as-errors")
|
|
werror_source = "source" not in get_option("disable-warnings-as-errors")
|
|
|
|
def generator(target, source, env, for_signature):
|
|
if werror_conftests and "conftest" in str(target[0]):
|
|
return flagname
|
|
|
|
if werror_source:
|
|
return flagname
|
|
|
|
return str()
|
|
|
|
return generator
|
|
|
|
|
|
env.Append(
|
|
CCFLAGS=["$CCFLAGS_GENERATE_WERROR"],
|
|
CCFLAGS_GENERATE_WERROR=create_werror_generator("$CCFLAGS_WERROR"),
|
|
CXXFLAGS=["$CXXFLAGS_GENERATE_WERROR"],
|
|
CXXFLAGS_GENERATE_WERROR=create_werror_generator("$CXXFLAGS_WERROR"),
|
|
LINKFLAGS=["$LINKFLAGS_GENERATE_WERROR"],
|
|
LINKFLAGS_GENERATE_WERROR=create_werror_generator("$LINKFLAGS_WERROR"),
|
|
)
|
|
|
|
|
|
def non_conf_ccflags_gen(target, source, env, for_signature):
|
|
if "conftest" in str(target[0]):
|
|
return ""
|
|
return "$NON_CONF_CCFLAGS"
|
|
|
|
|
|
def non_conf_linkflags_gen(target, source, env, for_signature):
|
|
if "conftest" in str(target[0]):
|
|
return ""
|
|
return "$NON_CONF_LINKFLAGS"
|
|
|
|
|
|
env["_NON_CONF_CCFLAGS_GEN"] = non_conf_ccflags_gen
|
|
env["_NON_CONF_LINKFLAGS_GEN"] = non_conf_linkflags_gen
|
|
|
|
env.Append(CCFLAGS="$_NON_CONF_CCFLAGS_GEN")
|
|
env.Append(LINKFLAGS="$_NON_CONF_LINKFLAGS_GEN")
|
|
|
|
for var in ["CC", "CXX"]:
|
|
if var not in env:
|
|
continue
|
|
path = env[var]
|
|
print("{} is {}".format(var, path))
|
|
if not os.path.isabs(path):
|
|
which = shutil.which(path)
|
|
if which is None:
|
|
print("{} was not found in $PATH".format(path))
|
|
else:
|
|
print("{} found in $PATH at {}".format(path, which))
|
|
path = which
|
|
|
|
realpath = os.path.realpath(path)
|
|
if realpath != path:
|
|
print("{} resolves to {}".format(path, realpath))
|
|
|
|
env.AddMethod(mongo_platform.env_os_is_wrapper, "TargetOSIs")
|
|
env.AddMethod(mongo_platform.env_get_os_name_wrapper, "GetTargetOSName")
|
|
|
|
|
|
def conf_error(env, msg, *args):
|
|
print(msg.format(*args))
|
|
print("See {0} for details".format(env.File("$CONFIGURELOG").abspath))
|
|
Exit(1)
|
|
|
|
|
|
env.AddMethod(fatal_error, "FatalError")
|
|
env.AddMethod(conf_error, "ConfError")
|
|
|
|
# Normalize the VERBOSE Option, and make its value available as a
|
|
# function.
|
|
if env["VERBOSE"] == "auto":
|
|
env["VERBOSE"] = not sys.stdout.isatty() and env.get("__NINJA_NO") != "1"
|
|
else:
|
|
try:
|
|
env["VERBOSE"] = to_boolean(env["VERBOSE"])
|
|
except ValueError as e:
|
|
env.FatalError(f"Error setting VERBOSE variable: {e}")
|
|
env.AddMethod(lambda env: env["VERBOSE"], "Verbose")
|
|
|
|
env.Append(
|
|
LINKFLAGS=['${_concat(COMPILER_EXEC_PREFIX_OPT, LINKFLAGS_COMPILER_EXEC_PREFIX, "", __env__)}']
|
|
)
|
|
|
|
# Normalize the ICECC_DEBUG option
|
|
try:
|
|
env["ICECC_DEBUG"] = to_boolean(env["ICECC_DEBUG"])
|
|
except ValueError as e:
|
|
env.FatalError(f"Error setting ICECC_DEBUG variable: {e}")
|
|
|
|
if has_option("variables-help"):
|
|
print(env_vars.GenerateHelpText(env))
|
|
Exit(0)
|
|
|
|
unknown_vars = env_vars.UnknownVariables()
|
|
if unknown_vars:
|
|
env.FatalError("Unknown variables specified: {0}", ", ".join(list(unknown_vars.keys())))
|
|
|
|
install_actions.setup(env, get_option("install-action"))
|
|
|
|
|
|
if env.TargetOSIs("windows") and os.path.exists(
|
|
env.File("#/src/mongo/db/modules/enterprise/SConscript").abspath
|
|
):
|
|
# the sasl zip can be rebuilt by following the instructions at:
|
|
# https://github.com/mongodb-forks/cyrus-sasl/blob/mongo-sasl-2-1-28/README.md
|
|
import mongo.download_windows_sasl
|
|
|
|
mongo.download_windows_sasl.download_sasl(env)
|
|
|
|
detectEnv = env.Clone()
|
|
|
|
# Identify the toolchain in use. We currently support the following:
|
|
# These macros came from
|
|
# http://nadeausoftware.com/articles/2012/10/c_c_tip_how_detect_compiler_name_and_version_using_compiler_predefined_macros
|
|
toolchain_macros = {
|
|
"GCC": "defined(__GNUC__) && !defined(__clang__)",
|
|
"clang": "defined(__clang__)",
|
|
"MSVC": "defined(_MSC_VER)",
|
|
}
|
|
|
|
|
|
def CheckForToolchain(context, toolchain, lang_name, compiler_var, source_suffix):
|
|
test_body = textwrap.dedent(
|
|
"""
|
|
#if {0}
|
|
/* we are using toolchain {0} */
|
|
#else
|
|
#error
|
|
#endif
|
|
""".format(toolchain_macros[toolchain])
|
|
)
|
|
|
|
print_tuple = (lang_name, context.env[compiler_var], toolchain)
|
|
context.Message('Checking if %s compiler "%s" is %s... ' % print_tuple)
|
|
|
|
# Strip indentation from the test body to ensure that the newline at the end of the
|
|
# endif is the last character in the file (rather than a line of spaces with no
|
|
# newline), and that all of the preprocessor directives start at column zero. Both of
|
|
# these issues can trip up older toolchains.
|
|
result = context.TryCompile(test_body, source_suffix)
|
|
context.Result(result)
|
|
return result
|
|
|
|
|
|
endian = get_option("endian")
|
|
|
|
if endian == "auto":
|
|
endian = sys.byteorder
|
|
|
|
processor_macros = {
|
|
"aarch64": {"endian": "little", "check": "(defined(__arm64__) || defined(__aarch64__))"},
|
|
"emscripten": {"endian": "little", "check": "(defined(__EMSCRIPTEN__))"},
|
|
"ppc64le": {"endian": "little", "check": "(defined(__powerpc64__))"},
|
|
"riscv64": {"endian": "little", "check": "(defined(__riscv)) && (__riscv_xlen == 64)"},
|
|
"s390x": {"endian": "big", "check": "(defined(__s390x__))"},
|
|
"x86_64": {"endian": "little", "check": "(defined(__x86_64) || defined(_M_AMD64))"},
|
|
}
|
|
|
|
|
|
def CheckForProcessor(context, which_arch):
|
|
def run_compile_check(arch):
|
|
if not endian == processor_macros[arch]["endian"]:
|
|
return False
|
|
|
|
test_body = """
|
|
#if {0}
|
|
/* Detected {1} */
|
|
#else
|
|
#error not {1}
|
|
#endif
|
|
""".format(processor_macros[arch]["check"], arch)
|
|
|
|
return context.TryCompile(textwrap.dedent(test_body), ".c")
|
|
|
|
if which_arch:
|
|
ret = run_compile_check(which_arch)
|
|
context.Message("Checking if target processor is %s " % which_arch)
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
for k in list(processor_macros.keys()):
|
|
ret = run_compile_check(k)
|
|
if ret:
|
|
context.Result("Detected a %s processor" % k)
|
|
return k
|
|
|
|
context.Result("Could not detect processor model/architecture")
|
|
return False
|
|
|
|
|
|
# Taken from http://nadeausoftware.com/articles/2012/01/c_c_tip_how_use_compiler_predefined_macros_detect_operating_system
|
|
os_macros = {
|
|
"windows": "defined(_WIN32)",
|
|
"solaris": "defined(__sun)",
|
|
"freebsd": "defined(__FreeBSD__)",
|
|
"openbsd": "defined(__OpenBSD__)",
|
|
"iOS": "defined(__APPLE__) && TARGET_OS_IOS && !TARGET_OS_SIMULATOR",
|
|
"iOS-sim": "defined(__APPLE__) && TARGET_OS_IOS && TARGET_OS_SIMULATOR",
|
|
"tvOS": "defined(__APPLE__) && TARGET_OS_TV && !TARGET_OS_SIMULATOR",
|
|
"tvOS-sim": "defined(__APPLE__) && TARGET_OS_TV && TARGET_OS_SIMULATOR",
|
|
"watchOS": "defined(__APPLE__) && TARGET_OS_WATCH && !TARGET_OS_SIMULATOR",
|
|
"watchOS-sim": "defined(__APPLE__) && TARGET_OS_WATCH && TARGET_OS_SIMULATOR",
|
|
# NOTE: Once we have XCode 8 required, we can rely on the value of TARGET_OS_OSX. In case
|
|
# we are on an older XCode, use TARGET_OS_MAC and TARGET_OS_IPHONE. We don't need to correct
|
|
# the above declarations since we will never target them with anything other than XCode 8.
|
|
"macOS": "defined(__APPLE__) && (TARGET_OS_OSX || (TARGET_OS_MAC && !TARGET_OS_IPHONE))",
|
|
"linux": "defined(__linux__)",
|
|
"android": "defined(__ANDROID__)",
|
|
"emscripten": "defined(__EMSCRIPTEN__)",
|
|
}
|
|
|
|
|
|
def CheckForOS(context, which_os):
|
|
test_body = """
|
|
#if defined(__APPLE__)
|
|
#include <TargetConditionals.h>
|
|
#endif
|
|
#if {0}
|
|
/* detected {1} */
|
|
#else
|
|
#error
|
|
#endif
|
|
""".format(os_macros[which_os], which_os)
|
|
context.Message("Checking if target OS {0} is supported by the toolchain... ".format(which_os))
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".c")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
|
|
def CheckForCXXLink(context):
|
|
test_body = """
|
|
#include <iostream>
|
|
#include <cstdlib>
|
|
|
|
int main() {
|
|
std::cout << "Hello, World" << std::endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
"""
|
|
context.Message("Checking that the C++ compiler can link a C++ program... ")
|
|
ret = context.TryLink(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
|
|
detectSystem = Configure(
|
|
detectEnv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckForToolchain": CheckForToolchain,
|
|
"CheckForProcessor": CheckForProcessor,
|
|
"CheckForOS": CheckForOS,
|
|
},
|
|
)
|
|
|
|
toolchain_search_sequence = ["GCC", "clang"]
|
|
if mongo_platform.is_running_os("windows"):
|
|
toolchain_search_sequence = ["MSVC", "clang", "GCC"]
|
|
|
|
detected_toolchain = None
|
|
for candidate_toolchain in toolchain_search_sequence:
|
|
if detectSystem.CheckForToolchain(candidate_toolchain, "C++", "CXX", ".cpp"):
|
|
detected_toolchain = candidate_toolchain
|
|
break
|
|
|
|
if not detected_toolchain:
|
|
env.ConfError("Couldn't identify the C++ compiler")
|
|
|
|
if not detectSystem.CheckForToolchain(detected_toolchain, "C", "CC", ".c"):
|
|
env.ConfError("C compiler does not match identified C++ compiler")
|
|
|
|
|
|
# Now that we've detected the toolchain, we add methods to the env
|
|
# to get the canonical name of the toolchain and to test whether
|
|
# scons is using a particular toolchain.
|
|
def get_toolchain_name(self):
|
|
return detected_toolchain.lower()
|
|
|
|
|
|
def is_toolchain(self, *args):
|
|
actual_toolchain = self.ToolchainName()
|
|
for v in args:
|
|
if v.lower() == actual_toolchain:
|
|
return True
|
|
return False
|
|
|
|
|
|
env.AddMethod(get_toolchain_name, "ToolchainName")
|
|
env.AddMethod(is_toolchain, "ToolchainIs")
|
|
|
|
releaseBuild = get_option("release") == "on"
|
|
debugBuild = get_option("dbg") == "on"
|
|
debug_symbols = get_option("debug-symbols") == "on"
|
|
optBuild = mongo_generators.get_opt_options(env)
|
|
|
|
if env.get("ENABLE_BUILD_RETRY"):
|
|
if get_option("ninja") != "disabled":
|
|
print("ENABLE_BUILD_RETRY not compatible with ninja, disabling ENABLE_BUILD_RETRY.")
|
|
else:
|
|
env["BUILD_RETRY_ATTEMPTS"] = 10
|
|
env["BUILD_RETRY_MAX_DELAY_SECONDS"] = 120
|
|
env.Tool("build_auto_retry")
|
|
|
|
if env.ToolchainIs("clang"):
|
|
# LLVM utilizes the stack extensively without optimization enabled, which
|
|
# causes the built product to easily blow through our 1M stack size whenever
|
|
# either gcov or sanitizers are enabled. Ref: SERVER-65684
|
|
if has_option("gcov") and optBuild not in ("on", "debug"):
|
|
env.FatalError(
|
|
"Error: A clang --gcov build must have either --opt=debug or --opt=on to "
|
|
+ "prevent crashes due to excessive stack usage"
|
|
)
|
|
|
|
if has_option("sanitize") and get_option("sanitize") != "" and optBuild not in ("on", "debug"):
|
|
env.FatalError(
|
|
"Error: A clang --sanitize build must have either --opt=debug or --opt=on "
|
|
+ "to prevent crashes due to excessive stack usage"
|
|
)
|
|
|
|
if releaseBuild and (debugBuild or optBuild != "on"):
|
|
env.FatalError(
|
|
"Error: A --release build may not have debugging, and must have full optimization"
|
|
)
|
|
|
|
if env["TARGET_ARCH"]:
|
|
if not detectSystem.CheckForProcessor(env["TARGET_ARCH"]):
|
|
env.ConfError("Could not detect processor specified in TARGET_ARCH variable")
|
|
else:
|
|
detected_processor = detectSystem.CheckForProcessor(None)
|
|
if not detected_processor:
|
|
env.ConfError("Failed to detect a supported target architecture")
|
|
env["TARGET_ARCH"] = detected_processor
|
|
|
|
if env["TARGET_OS"] not in os_macros:
|
|
print("No special config for [{0}] which probably means it won't work".format(env["TARGET_OS"]))
|
|
elif not detectSystem.CheckForOS(env["TARGET_OS"]):
|
|
env.ConfError("TARGET_OS ({0}) is not supported by compiler", env["TARGET_OS"])
|
|
|
|
detectSystem.Finish()
|
|
|
|
if env.TargetOSIs("posix"):
|
|
if env.ToolchainIs("gcc", "clang"):
|
|
env.Append(
|
|
CCFLAGS_WERROR=["-Werror"],
|
|
CXXFLAGS_WERROR=["-Werror=unused-result"] if env.ToolchainIs("clang") else [],
|
|
LINKFLAGS_WERROR=["-Wl,--fatal-warnings"] if not env.TargetOSIs("darwin") else [],
|
|
)
|
|
elif env.TargetOSIs("windows"):
|
|
env.Append(CCFLAGS_WERROR=["/WX"])
|
|
|
|
if env.ToolchainIs("clang"):
|
|
|
|
def assembler_with_cpp_gen(target, source, env, for_signature):
|
|
if source[0].get_suffix() == ".sx":
|
|
return "-x assembler-with-cpp"
|
|
|
|
env["CLANG_ASSEMBLER_WITH_CPP"] = assembler_with_cpp_gen
|
|
env.Append(ASFLAGS=["$CLANG_ASSEMBLER_WITH_CPP"])
|
|
|
|
env["CC_VERSION"] = mongo_toolchain.get_toolchain_ver(env, "CC")
|
|
env["CXX_VERSION"] = mongo_toolchain.get_toolchain_ver(env, "CXX")
|
|
|
|
if not env["HOST_ARCH"]:
|
|
env["HOST_ARCH"] = env["TARGET_ARCH"]
|
|
|
|
# In some places we have POSIX vs Windows cpp files, and so there's an additional
|
|
# env variable to interpolate their names in child sconscripts
|
|
|
|
env["TARGET_OS_FAMILY"] = "posix" if env.TargetOSIs("posix") else env.GetTargetOSName()
|
|
|
|
if env.TargetOSIs("linux") or "tcmalloc-google" == get_option("allocator"):
|
|
# tcmalloc from google has some requirements on the kernel version for rseq support
|
|
# here we check if it should be available
|
|
try:
|
|
kernel_version = platform.release().split(".")
|
|
kernel_major = int(kernel_version[0])
|
|
kernel_minor = int(kernel_version[1])
|
|
except (ValueError, IndexError):
|
|
print(
|
|
f"Failed to extract kernel major and minor versions, tcmalloc-google will not be available for use: {kernel_version}"
|
|
)
|
|
kernel_major = 0
|
|
kernel_minor = 0
|
|
|
|
# Normalize the allocator option and store it in the Environment. It
|
|
# would be nicer to use SetOption here, but you can't reset user
|
|
# options for some strange reason in SCons. Instead, we store this
|
|
# option as a new variable in the environment.
|
|
if get_option("allocator") == "auto":
|
|
if env.TargetOSIs("linux") and env["TARGET_ARCH"] in ("x86_64", "aarch64"):
|
|
env["MONGO_ALLOCATOR"] = "tcmalloc-google"
|
|
|
|
# googles tcmalloc uses the membarrier() system call which was added in Linux 4.3,
|
|
# so fall back to gperf implementation for older kernels
|
|
if kernel_major < 4 or (kernel_major == 4 and kernel_minor < 3):
|
|
env["MONGO_ALLOCATOR"] = "tcmalloc-gperf"
|
|
|
|
elif env.TargetOSIs("windows") or (
|
|
env.TargetOSIs("linux") and env["TARGET_ARCH"] in ("ppc64le", "s390x")
|
|
):
|
|
env["MONGO_ALLOCATOR"] = "tcmalloc-gperf"
|
|
else:
|
|
env["MONGO_ALLOCATOR"] = "system"
|
|
else:
|
|
env["MONGO_ALLOCATOR"] = get_option("allocator")
|
|
|
|
if env["MONGO_ALLOCATOR"] == "tcmalloc-google":
|
|
if kernel_major < 4 or (kernel_major == 4 and kernel_minor < 3):
|
|
env.ConfError(
|
|
f"tcmalloc-google allocator only supported on linux kernel 4.3 or greater: kenerl verison={platform.release()}"
|
|
)
|
|
|
|
if env["MONGO_ALLOCATOR"] == "tcmalloc-google":
|
|
env.Append(CPPDEFINES=["ABSL_ALLOCATOR_NOTHROW"])
|
|
|
|
if has_option("cache"):
|
|
if has_option("gcov"):
|
|
env.FatalError("Mixing --cache and --gcov doesn't work correctly yet. See SERVER-11084")
|
|
env.CacheDir(str(env.Dir(cacheDir)))
|
|
|
|
# Normalize the link model. If it is auto, then for now both developer and release builds
|
|
# use the "static" mode. Someday later, we probably want to make the developer build default
|
|
# dynamic.
|
|
link_model = get_option("link-model")
|
|
if link_model == "auto":
|
|
link_model = "static"
|
|
|
|
if link_model.startswith("dynamic") and get_option("install-action") == "symlink":
|
|
env.FatalError(
|
|
f"Options '--link-model={link_model}' not supported with '--install-action={get_option('install-action')}'."
|
|
)
|
|
|
|
if (
|
|
link_model == "dynamic"
|
|
and env.TargetOSIs("darwin")
|
|
and not get_option("force-macos-dynamic-link")
|
|
):
|
|
macos_version_message = textwrap.dedent("""\
|
|
link-model=dynamic us only supported on macos version 12 or higher.
|
|
This is due to a 512 dylib RUNTIME limit on older macos. See this post for
|
|
more information: https://developer.apple.com/forums//thread/708366?login=true&page=1#717495022
|
|
Use '--force-macos-dynamic-link' to bypass this check.
|
|
""")
|
|
|
|
try:
|
|
macos_version_major = int(platform.mac_ver()[0].split(".")[0])
|
|
if macos_version_major < 12:
|
|
env.FatalError(
|
|
textwrap.dedent(f"""\
|
|
Macos version detected: {macos_version_major}
|
|
""")
|
|
+ macos_version_message
|
|
)
|
|
except (IndexError, TypeError) as exc:
|
|
env.FatalError(
|
|
textwrap.dedent(f"""\
|
|
Failed to detect macos version: {exc}
|
|
""")
|
|
+ macos_version_message
|
|
)
|
|
|
|
# libunwind configuration.
|
|
# In which the following globals are set and normalized to bool:
|
|
# - use_libunwind
|
|
# - use_system_libunwind
|
|
# - use_vendored_libunwind
|
|
use_libunwind = get_option("use-libunwind")
|
|
use_system_libunwind = use_system_version_of_library("libunwind")
|
|
|
|
# Assume system libunwind works if it's installed and selected.
|
|
can_use_libunwind = (
|
|
use_system_libunwind
|
|
or env.TargetOSIs("linux")
|
|
and (env["TARGET_ARCH"] in ("x86_64", "aarch64", "ppc64le", "s390x"))
|
|
)
|
|
|
|
if use_libunwind == "off":
|
|
use_libunwind = False
|
|
use_system_libunwind = False
|
|
elif use_libunwind == "on":
|
|
use_libunwind = True
|
|
if not can_use_libunwind:
|
|
env.ConfError("libunwind not supported on target platform")
|
|
Exit(1)
|
|
elif use_libunwind == "auto":
|
|
use_libunwind = can_use_libunwind
|
|
|
|
use_vendored_libunwind = use_libunwind and not use_system_libunwind
|
|
env["USE_VENDORED_LIBUNWIND"] = use_vendored_libunwind
|
|
if use_system_libunwind and not use_libunwind:
|
|
print("Error: --use-system-libunwind requires --use-libunwind")
|
|
Exit(1)
|
|
|
|
if get_option("visibility-support") == "auto":
|
|
visibility_annotations_enabled = not env.TargetOSIs("windows") and link_model.startswith(
|
|
"dynamic"
|
|
)
|
|
else:
|
|
visibility_annotations_enabled = get_option("visibility-support") == "on"
|
|
|
|
# Windows can't currently support anything other than 'object' or 'static', until
|
|
# we have annotated functions for export.
|
|
if env.TargetOSIs("windows") and not visibility_annotations_enabled:
|
|
if link_model not in ["object", "static", "dynamic-sdk"]:
|
|
env.FatalError(
|
|
"Windows builds must use the 'object', 'dynamic-sdk', or 'static' link models"
|
|
)
|
|
|
|
# TODO(SERVER-85904): remove check when object mode & LTO are supported in bazel
|
|
if link_model == "object":
|
|
env.FatalError(
|
|
'Bazel-enabled builds currently do not support the "object" link model. Reffer to SERVER-85904 for more info.'
|
|
)
|
|
|
|
# The 'object' mode for libdeps is enabled by setting _LIBDEPS to $_LIBDEPS_OBJS. The other two
|
|
# modes operate in library mode, enabled by setting _LIBDEPS to $_LIBDEPS_LIBS.
|
|
env["_LIBDEPS"] = "$_LIBDEPS_OBJS" if link_model == "object" else "$_LIBDEPS_LIBS"
|
|
|
|
env["BUILDERS"]["ProgramObject"] = env["BUILDERS"]["StaticObject"]
|
|
env["BUILDERS"]["LibraryObject"] = env["BUILDERS"]["StaticObject"]
|
|
|
|
env["SHARPREFIX"] = "$LIBPREFIX"
|
|
env["SHARSUFFIX"] = "${SHLIBSUFFIX}${LIBSUFFIX}"
|
|
env["BUILDERS"]["SharedArchive"] = SCons.Builder.Builder(
|
|
action=env["BUILDERS"]["StaticLibrary"].action,
|
|
emitter="$SHAREMITTER",
|
|
prefix="$SHARPREFIX",
|
|
suffix="$SHARSUFFIX",
|
|
src_suffix=env["BUILDERS"]["SharedLibrary"].src_suffix,
|
|
)
|
|
|
|
# Teach object builders how to build underlying generated types
|
|
for builder in ["SharedObject", "StaticObject"]:
|
|
env["BUILDERS"][builder].add_src_builder("Protoc")
|
|
|
|
|
|
# These allow delayed evaluation of the AIB values for the default values of
|
|
# the corresponding command line variables
|
|
def protoc_var_gen(env, target, source, for_signature):
|
|
return env.File("$DESTDIR/$PREFIX_BINDIR/protobuf_compiler$PROGSUFFIX")
|
|
|
|
|
|
env["PROTOC_VAR_GEN"] = protoc_var_gen
|
|
|
|
|
|
def protoc_grpc_plugin_var_gen(env, target, source, for_signature):
|
|
return env.File("$DESTDIR/$PREFIX_BINDIR/grpc_cpp_plugin$PROGSUFFIX")
|
|
|
|
|
|
env["PROTOC_GRPC_PLUGIN_GEN"] = protoc_grpc_plugin_var_gen
|
|
|
|
if link_model.startswith("dynamic"):
|
|
if link_model == "dynamic" and visibility_annotations_enabled:
|
|
|
|
def visibility_cppdefines_generator(target, source, env, for_signature):
|
|
if "MONGO_API_NAME" not in env:
|
|
return None
|
|
return "MONGO_API_${MONGO_API_NAME}"
|
|
|
|
env["MONGO_VISIBILITY_CPPDEFINES_GENERATOR"] = visibility_cppdefines_generator
|
|
|
|
def visibility_shccflags_generator(target, source, env, for_signature):
|
|
if env.get("MONGO_API_NAME"):
|
|
return "-fvisibility=hidden"
|
|
return None
|
|
|
|
if not env.TargetOSIs("windows"):
|
|
env["MONGO_VISIBILITY_SHCCFLAGS_GENERATOR"] = visibility_shccflags_generator
|
|
|
|
env.AppendUnique(
|
|
CPPDEFINES=[
|
|
"MONGO_USE_VISIBILITY",
|
|
"$MONGO_VISIBILITY_CPPDEFINES_GENERATOR",
|
|
],
|
|
SHCCFLAGS=[
|
|
"$MONGO_VISIBILITY_SHCCFLAGS_GENERATOR",
|
|
],
|
|
)
|
|
|
|
def library(env, target, source, *args, **kwargs):
|
|
sharedLibrary = env.SharedLibrary(target, source, *args, **kwargs)
|
|
sharedArchive = env.SharedArchive(target, source=sharedLibrary[0].sources, *args, **kwargs)
|
|
sharedLibrary.extend(sharedArchive)
|
|
return sharedLibrary
|
|
|
|
env["BUILDERS"]["Library"] = library
|
|
env["BUILDERS"]["LibraryObject"] = env["BUILDERS"]["SharedObject"]
|
|
|
|
# TODO: Ideally, the conditions below should be based on a
|
|
# detection of what linker we are using, not the local OS, but I
|
|
# doubt very much that we will see the mach-o linker on anything
|
|
# other than Darwin, or a BFD/sun-esque linker elsewhere.
|
|
|
|
# On Darwin, we need to tell the linker that undefined symbols are
|
|
# resolved via dynamic lookup; otherwise we get build failures. On
|
|
# other unixes, we need to suppress as-needed behavior so that
|
|
# initializers are ensured present, even if there is no visible
|
|
# edge to the library in the symbol graph.
|
|
#
|
|
# NOTE: The darwin linker flag is only needed because the library
|
|
# graph is not a DAG. Once the graph is a DAG, we will require all
|
|
# edges to be expressed, and we should drop the flag. When that
|
|
# happens, we should also add -z,defs flag on ELF platforms to
|
|
# ensure that missing symbols due to unnamed dependency edges
|
|
# result in link errors.
|
|
#
|
|
# NOTE: The `illegal_cyclic_or_unresolved_dependencies_allowlisted`
|
|
# tag can be applied to a library to indicate that it does not (or
|
|
# cannot) completely express all of its required link dependencies.
|
|
# This can occur for four reasons:
|
|
#
|
|
# - No unique provider for the symbol: Some symbols do not have a
|
|
# unique dependency that provides a definition, in which case it
|
|
# is impossible for the library to express a dependency edge to
|
|
# resolve the symbol.
|
|
#
|
|
# - The library is part of a cycle: If library A depends on B,
|
|
# which depends on C, which depends on A, then it is impossible
|
|
# to express all three edges in SCons, since otherwise there is
|
|
# no way to sequence building the libraries. The cyclic
|
|
# libraries actually work at runtime, because some parent object
|
|
# links all of them.
|
|
#
|
|
# - The symbol is provided by an executable into which the library
|
|
# will be linked. The mongo::inShutdown symbol is a good
|
|
# example.
|
|
#
|
|
# - The symbol is provided by a third-party library, outside of our
|
|
# control.
|
|
#
|
|
# All of these are defects in the linking model. In an effort to
|
|
# eliminate these issues, we have begun tagging those libraries
|
|
# that are affected, and requiring that all non-tagged libraries
|
|
# correctly express all dependencies. As we repair each defective
|
|
# library, we can remove the tag. When all the tags are removed
|
|
# the graph will be acyclic. Libraries which are incomplete for the
|
|
# final reason, "libraries outside of our control", may remain for
|
|
# reasons beyond our control. Such libraries ideally should
|
|
# have no dependencies (and thus be leaves in our linking DAG).
|
|
# If that condition is met, then the graph will be acyclic.
|
|
|
|
if env.TargetOSIs("darwin"):
|
|
if link_model.startswith("dynamic"):
|
|
print(
|
|
"WARNING: Building MongoDB server with dynamic linking "
|
|
+ "on macOS is not supported. Static linking is recommended."
|
|
)
|
|
|
|
if link_model == "dynamic-strict":
|
|
# Darwin is strict by default
|
|
pass
|
|
else:
|
|
|
|
def libdeps_tags_expand_incomplete(source, target, env, for_signature):
|
|
# On darwin, since it is strict by default, we need to add a flag
|
|
# when libraries are tagged incomplete.
|
|
if "illegal_cyclic_or_unresolved_dependencies_allowlisted" in target[
|
|
0
|
|
].get_env().get("LIBDEPS_TAGS", []):
|
|
return ["-Wl,-undefined,dynamic_lookup"]
|
|
return []
|
|
|
|
env["LIBDEPS_TAG_EXPANSIONS"].append(libdeps_tags_expand_incomplete)
|
|
elif env.TargetOSIs("windows"):
|
|
if link_model == "dynamic-strict":
|
|
# Windows is strict by default
|
|
pass
|
|
else:
|
|
|
|
def libdeps_tags_expand_incomplete(source, target, env, for_signature):
|
|
# On windows, since it is strict by default, we need to add a flag
|
|
# when libraries are tagged incomplete.
|
|
if "illegal_cyclic_or_unresolved_dependencies_allowlisted" in target[
|
|
0
|
|
].get_env().get("LIBDEPS_TAGS", []):
|
|
return ["/FORCE:UNRESOLVED"]
|
|
return []
|
|
|
|
env["LIBDEPS_TAG_EXPANSIONS"].append(libdeps_tags_expand_incomplete)
|
|
else:
|
|
env.AppendUnique(LINKFLAGS=["-Wl,--no-as-needed"])
|
|
|
|
# Using zdefs doesn't work at all with the sanitizers
|
|
if not has_option("sanitize") and get_option("sanitize") != "":
|
|
if link_model == "dynamic-strict":
|
|
env.AppendUnique(SHLINKFLAGS=["-Wl,-z,defs"])
|
|
else:
|
|
# On BFD/gold linker environments, which are not strict by
|
|
# default, we need to add a flag when libraries are not
|
|
# tagged incomplete.
|
|
def libdeps_tags_expand_incomplete(source, target, env, for_signature):
|
|
if "illegal_cyclic_or_unresolved_dependencies_allowlisted" not in target[
|
|
0
|
|
].get_env().get("LIBDEPS_TAGS", []):
|
|
return ["-Wl,-z,defs"]
|
|
return []
|
|
|
|
env["LIBDEPS_TAG_EXPANSIONS"].append(libdeps_tags_expand_incomplete)
|
|
|
|
# Enable the fast decider if explicitly requested or if in 'auto' mode
|
|
# and not in conflict with other options like the ninja option which
|
|
# sets its own decider.
|
|
if (
|
|
get_option("ninja") == "disabled"
|
|
and get_option("build-fast-and-loose") == "on"
|
|
or (get_option("build-fast-and-loose") == "auto" and not has_option("release"))
|
|
):
|
|
# See http://www.scons.org/wiki/GoFastButton for details
|
|
env.Decider("MD5-timestamp")
|
|
env.SetOption("max_drift", 1)
|
|
|
|
# If the user has requested the git decider, enable it if it is available. We want to do this after
|
|
# we set the basic decider above, so that we 'chain' to that one.
|
|
if get_option("git-decider") == "on":
|
|
git_decider = Tool("git_decider")
|
|
if git_decider.exists(env):
|
|
git_decider(env)
|
|
|
|
# On non-windows platforms, we may need to differentiate between flags being used to target an
|
|
# executable (like -fPIE), vs those being used to target a (shared) library (like -fPIC). To do so,
|
|
# we inject a new family of SCons variables PROG*FLAGS, by reaching into the various COMs.
|
|
if not env.TargetOSIs("windows"):
|
|
env["CCCOM"] = env["CCCOM"].replace("$CCFLAGS", "$PROGCCFLAGS")
|
|
env["CXXCOM"] = env["CXXCOM"].replace("$CCFLAGS", "$PROGCCFLAGS")
|
|
env["PROGCCFLAGS"] = ["$CCFLAGS"]
|
|
|
|
env["CCCOM"] = env["CCCOM"].replace("$CFLAGS", "$PROGCFLAGS")
|
|
env["PROGCFLAGS"] = ["$CFLAGS"]
|
|
|
|
env["CXXCOM"] = env["CXXCOM"].replace("$CXXFLAGS", "$PROGCXXFLAGS")
|
|
env["PROGCXXFLAGS"] = ["$CXXFLAGS"]
|
|
|
|
env["LINKCOM"] = env["LINKCOM"].replace("$LINKFLAGS", "$PROGLINKFLAGS")
|
|
env["PROGLINKFLAGS"] = ["$LINKFLAGS"]
|
|
|
|
# ASPPFLAGS is used for assembler commands, this condition below assumes assembler files
|
|
# will be only directly assembled in librarys and not programs
|
|
if link_model.startswith("dynamic"):
|
|
env.Append(ASPPFLAGS=["-fPIC"])
|
|
else:
|
|
env.Append(ASPPFLAGS=["-fPIE"])
|
|
|
|
# When it is necessary to supply additional SHLINKFLAGS without modifying the toolset default,
|
|
# following appends contents of SHLINKFLAGS_EXTRA variable to the linker command
|
|
env.AppendUnique(SHLINKFLAGS=["$SHLINKFLAGS_EXTRA"])
|
|
|
|
|
|
class ForceVerboseConftest:
|
|
"""
|
|
This class allows for configurable substition calls to enable forcing
|
|
the conftest to use verbose logs even when verbose mode is not specified.
|
|
"""
|
|
|
|
def __init__(self, msg):
|
|
self.msg = msg
|
|
|
|
def __call__(self, target, source, env, for_signature):
|
|
for t in target:
|
|
# TODO: SERVER-60915 switch to SCons api conftest check
|
|
if "conftest" in str(t):
|
|
return None
|
|
return self.msg
|
|
|
|
|
|
if not env.Verbose():
|
|
# Even though we are not in Verbose mode, conftest logs should
|
|
# always be verbose, because they go to a file and not seen
|
|
# by the user anyways.
|
|
env.Append(CCCOMSTR=ForceVerboseConftest("Compiling $TARGET"))
|
|
env.Append(CXXCOMSTR=ForceVerboseConftest(env["CCCOMSTR"]))
|
|
env.Append(SHCCCOMSTR=ForceVerboseConftest("Compiling $TARGET"))
|
|
env.Append(SHCXXCOMSTR=ForceVerboseConftest(env["SHCCCOMSTR"]))
|
|
env.Append(LINKCOMSTR=ForceVerboseConftest("Linking $TARGET"))
|
|
env.Append(SHLINKCOMSTR=ForceVerboseConftest(env["LINKCOMSTR"]))
|
|
env.Append(ARCOMSTR=ForceVerboseConftest("Generating library $TARGET"))
|
|
|
|
# Link tools other than mslink don't setup TEMPFILE in LINKCOM,
|
|
# disabling SCons automatically falling back to a temp file when
|
|
# running link commands that are over MAXLINELENGTH. With our object
|
|
# file linking mode, we frequently hit even the large linux command
|
|
# line length, so we want it everywhere. If we aren't using mslink,
|
|
# add TEMPFILE in. For verbose builds when using a tempfile, we need
|
|
# some trickery so that we print the command we are running, and not
|
|
# just the invocation of the compiler being fed the command file.
|
|
if "mslink" not in env["TOOLS"]:
|
|
if env.Verbose():
|
|
env["LINKCOM"] = "${{TEMPFILE('{0}', '')}}".format(env["LINKCOM"])
|
|
env["SHLINKCOM"] = "${{TEMPFILE('{0}', '')}}".format(env["SHLINKCOM"])
|
|
if "libtool" not in env["TOOLS"]:
|
|
env["ARCOM"] = "${{TEMPFILE('{0}', '')}}".format(env["ARCOM"])
|
|
else:
|
|
env["LINKCOM"] = "${{TEMPFILE('{0}', 'LINKCOMSTR')}}".format(env["LINKCOM"])
|
|
env["SHLINKCOM"] = "${{TEMPFILE('{0}', 'SHLINKCOMSTR')}}".format(env["SHLINKCOM"])
|
|
if "libtool" not in env["TOOLS"]:
|
|
env["ARCOM"] = "${{TEMPFILE('{0}', 'ARCOMSTR')}}".format(env["ARCOM"])
|
|
|
|
if env["_LIBDEPS"] == "$_LIBDEPS_OBJS":
|
|
# The libraries we build in LIBDEPS_OBJS mode are just placeholders for tracking dependencies.
|
|
# This avoids wasting time and disk IO on them.
|
|
def write_uuid_to_file(env, target, source):
|
|
with open(env.File(target[0]).abspath, "w") as fake_lib:
|
|
fake_lib.write(str(uuid.uuid4()))
|
|
fake_lib.write("\n")
|
|
|
|
# We originally did this by setting ARCOM to write_uuid_to_file.
|
|
# This worked more or less by accident. It works when SCons is
|
|
# doing the action execution because when it would subst the
|
|
# command line subst would execute the function as part of string
|
|
# resolution which would have the side effect of writing the
|
|
# file. Since it returned None subst would do some special
|
|
# handling to make sure it never made it to the command line. This
|
|
# breaks Ninja however because we are taking that return value and
|
|
# trying to pass it to the command executor (/bin/sh or
|
|
# cmd.exe) and end up with the function name as a command. The
|
|
# resulting command looks something like `/bin/sh -c
|
|
# 'write_uuid_to_file(env, target, source)`. If we instead
|
|
# actually do what we want and that is make the StaticLibrary
|
|
# builder's action a FunctionAction the Ninja generator will
|
|
# correctly dispatch it and not generate an invalid command
|
|
# line. This also has the side benefit of being more clear that
|
|
# we're expecting a Python function to execute here instead of
|
|
# pretending to be a CommandAction that just happens to not run a
|
|
# command but instead runs a function.
|
|
env["BUILDERS"]["StaticLibrary"].action = SCons.Action.Action(
|
|
write_uuid_to_file, "Generating placeholder library $TARGET"
|
|
)
|
|
|
|
import libdeps_tool as libdeps
|
|
|
|
libdeps.setup_environment(
|
|
env,
|
|
emitting_shared=link_model,
|
|
debug=get_option("libdeps-debug"),
|
|
linting=get_option("libdeps-linting"),
|
|
)
|
|
|
|
# The abilink/tapilink tools and the thin archive tool must be loaded
|
|
# after libdeps, so that the scanners they inject can see the library
|
|
# dependencies added by libdeps. Neither abilink nor tapilink can work
|
|
# with the current Ninja generation because they rely on adding
|
|
# ListActions to builders.
|
|
if get_option("ninja") == "disabled" and link_model.startswith("dynamic"):
|
|
# Add in the abi linking tool if the user requested and it is
|
|
# supported on this platform.
|
|
#
|
|
# TODO: Can we unify the `abilink` and `tapilink` tools?
|
|
if env.get("ABIDW"):
|
|
abilink = Tool("abilink")
|
|
if abilink.exists(env):
|
|
abilink(env)
|
|
|
|
if env.get("TAPI"):
|
|
# TAPI is less useful when running with Bazel + Remote Execution. Disable since the initial implementation
|
|
# of the build system with Bazel will not support it.
|
|
# TODO(SERVER-88612): Remove fatal error we decide to implement TAPI support in Bazel
|
|
env.FatalError("TAPI is not supported with the hybrid build system.")
|
|
|
|
tapilink = Tool("tapilink")
|
|
if tapilink.exists(env):
|
|
tapilink(env)
|
|
|
|
if env["_LIBDEPS"] == "$_LIBDEPS_LIBS":
|
|
# The following platforms probably aren't using the binutils
|
|
# toolchain, or may be using it for the archiver but not the
|
|
# linker, and binutils currently is the only thing that supports
|
|
# thin archives. Don't even try on those platforms.
|
|
if not env.TargetOSIs("solaris", "darwin", "windows", "openbsd"):
|
|
env.Tool("thin_archive")
|
|
|
|
if env.TargetOSIs("linux", "freebsd", "openbsd"):
|
|
env["LINK_WHOLE_ARCHIVE_LIB_START"] = "-Wl,--whole-archive"
|
|
env["LINK_WHOLE_ARCHIVE_LIB_END"] = "-Wl,--no-whole-archive"
|
|
env["LINK_AS_NEEDED_LIB_START"] = "-Wl,--as-needed"
|
|
env["LINK_AS_NEEDED_LIB_END"] = "-Wl,--no-as-needed"
|
|
elif env.TargetOSIs("darwin"):
|
|
env["LINK_WHOLE_ARCHIVE_LIB_START"] = "-Wl,-force_load"
|
|
env["LINK_WHOLE_ARCHIVE_LIB_END"] = ""
|
|
env["LINK_AS_NEEDED_LIB_START"] = "-Wl,-mark_dead_strippable_dylib"
|
|
env["LINK_AS_NEEDED_LIB_END"] = ""
|
|
elif env.TargetOSIs("solaris"):
|
|
env["LINK_WHOLE_ARCHIVE_LIB_START"] = "-Wl,-z,allextract"
|
|
env["LINK_WHOLE_ARCHIVE_LIB_END"] = "-Wl,-z,defaultextract"
|
|
elif env.TargetOSIs("windows"):
|
|
env["LINK_WHOLE_ARCHIVE_LIB_START"] = "/WHOLEARCHIVE"
|
|
env["LINK_WHOLE_ARCHIVE_LIB_END"] = ""
|
|
env["LIBDEPS_FLAG_SEPARATORS"] = {env["LINK_WHOLE_ARCHIVE_LIB_START"]: {"suffix": ":"}}
|
|
|
|
if env.TargetOSIs("darwin") and link_model.startswith("dynamic"):
|
|
|
|
def init_no_global_libdeps_tag_expansion(source, target, env, for_signature):
|
|
"""
|
|
This callable will be expanded by scons and modify the environment by
|
|
adjusting the prefix and postfix flags to account for linking options
|
|
related to the use of global static initializers for any given libdep.
|
|
"""
|
|
|
|
if "init-no-global-side-effects" in env.get(libdeps.Constants.LibdepsTags, []):
|
|
# macos as-needed flag is used on the library directly when it is built
|
|
return env.get("LINK_AS_NEEDED_LIB_START", "")
|
|
|
|
env["LIBDEPS_TAG_EXPANSIONS"].append(init_no_global_libdeps_tag_expansion)
|
|
|
|
|
|
def init_no_global_add_flags(target, start_flag, end_flag):
|
|
"""Helper function for init_no_global_libdeps_tag_expand"""
|
|
|
|
setattr(target[0].attributes, "libdeps_prefix_flags", [start_flag])
|
|
setattr(target[0].attributes, "libdeps_postfix_flags", [end_flag])
|
|
if env.TargetOSIs("linux", "freebsd", "openbsd"):
|
|
setattr(
|
|
target[0].attributes,
|
|
"libdeps_switch_flags",
|
|
[
|
|
{
|
|
"on": start_flag,
|
|
"off": end_flag,
|
|
}
|
|
],
|
|
)
|
|
|
|
|
|
def init_no_global_libdeps_tag_emitter(target, source, env):
|
|
"""
|
|
This emitter will be attached the correct pre and post fix flags to
|
|
a given library to cause it to have certain flags before or after on the link
|
|
line.
|
|
"""
|
|
|
|
if link_model == "dynamic":
|
|
start_flag = env.get("LINK_AS_NEEDED_LIB_START", "")
|
|
end_flag = env.get("LINK_AS_NEEDED_LIB_END", "")
|
|
|
|
# In the dynamic case, any library that is known to not have global static
|
|
# initializers can supply the flag and be wrapped in --as-needed linking,
|
|
# allowing the linker to be smart about linking libraries it may not need.
|
|
if "init-no-global-side-effects" in env.get(
|
|
libdeps.Constants.LibdepsTags, []
|
|
) and not env.TargetOSIs("darwin"):
|
|
init_no_global_add_flags(target, start_flag, end_flag)
|
|
else:
|
|
init_no_global_add_flags(target, "", "")
|
|
|
|
else:
|
|
start_flag = env.get("LINK_WHOLE_ARCHIVE_LIB_START", "")
|
|
end_flag = env.get("LINK_WHOLE_ARCHIVE_LIB_END", "")
|
|
|
|
# In the static case, any library that is unknown to have global static
|
|
# initializers should supply the flag and be wrapped in --whole-archive linking,
|
|
# allowing the linker to bring in all those symbols which may not be directly needed
|
|
# at link time.
|
|
if "init-no-global-side-effects" not in env.get(libdeps.Constants.LibdepsTags, []):
|
|
init_no_global_add_flags(target, start_flag, end_flag)
|
|
else:
|
|
init_no_global_add_flags(target, "", "")
|
|
return target, source
|
|
|
|
|
|
for target_builder in ["SharedLibrary", "SharedArchive", "StaticLibrary"]:
|
|
builder = env["BUILDERS"][target_builder]
|
|
base_emitter = builder.emitter
|
|
new_emitter = SCons.Builder.ListEmitter([base_emitter, init_no_global_libdeps_tag_emitter])
|
|
builder.emitter = new_emitter
|
|
|
|
link_guard_rules = {
|
|
"test": [
|
|
"dist",
|
|
],
|
|
}
|
|
|
|
|
|
class LibdepsLinkGuard(SCons.Errors.UserError):
|
|
pass
|
|
|
|
|
|
def checkComponentType(target_comps, comp, target, lib):
|
|
"""
|
|
For a libdep and each AIB_COMPONENT its labeled as, check if its violates
|
|
any of the link gaurd rules.
|
|
"""
|
|
for target_comp in target_comps:
|
|
for link_guard_rule in link_guard_rules:
|
|
if target_comp in link_guard_rules[link_guard_rule] and link_guard_rule in comp:
|
|
raise LibdepsLinkGuard(
|
|
textwrap.dedent(f"""\n
|
|
LibdepsLinkGuard:
|
|
\tTarget '{target[0]}' links LIBDEP '{lib}'
|
|
\tbut is listed as AIB_COMPONENT '{target_comp}' which is not allowed link libraries
|
|
\twith AIB_COMPONENTS that include the word '{link_guard_rule}'\n""")
|
|
)
|
|
|
|
|
|
def get_comps(env):
|
|
"""util function for extracting all AIB_COMPONENTS as a list"""
|
|
comps = env.get("AIB_COMPONENTS_EXTRA", [])
|
|
comp = env.get("AIB_COMPONENT", None)
|
|
if comp:
|
|
comps += [comp]
|
|
return comps
|
|
|
|
|
|
def link_guard_libdeps_tag_expand(source, target, env, for_signature):
|
|
"""
|
|
Callback function called on all binaries to check if a certain binary
|
|
from a given component is linked to another binary of a given component,
|
|
the goal being to define rules that prevents test components from being
|
|
linked into production or releaseable components.
|
|
"""
|
|
for lib in libdeps.get_libdeps(source, target, env, for_signature):
|
|
if not lib.env:
|
|
continue
|
|
|
|
for comp in get_comps(lib.env):
|
|
checkComponentType(get_comps(env), comp, target, lib)
|
|
|
|
return []
|
|
|
|
|
|
env["LIBDEPS_TAG_EXPANSIONS"].append(link_guard_libdeps_tag_expand)
|
|
|
|
env.Tool("forceincludes")
|
|
|
|
# ---- other build setup -----
|
|
if not debugBuild:
|
|
env.AppendUnique(CPPDEFINES=["NDEBUG"])
|
|
|
|
# Normalize our experimental optimiation and hardening flags
|
|
selected_experimental_optimizations = set()
|
|
for suboption in get_option("experimental-optimization"):
|
|
if suboption == "*":
|
|
selected_experimental_optimizations.update(experimental_optimizations)
|
|
elif suboption.startswith("-"):
|
|
selected_experimental_optimizations.discard(suboption[1:])
|
|
elif suboption.startswith("+"):
|
|
selected_experimental_optimizations.add(suboption[1:])
|
|
|
|
selected_experimental_runtime_hardenings = set()
|
|
for suboption in get_option("experimental-runtime-hardening"):
|
|
if suboption == "*":
|
|
selected_experimental_runtime_hardenings.update(experimental_runtime_hardenings)
|
|
elif suboption.startswith("-"):
|
|
selected_experimental_runtime_hardenings.discard(suboption[1:])
|
|
elif suboption.startswith("+"):
|
|
selected_experimental_runtime_hardenings.add(suboption[1:])
|
|
|
|
# Disable floating-point contractions such as forming of fused multiply-add operations.
|
|
if env.ToolchainIs("clang", "gcc"):
|
|
env.Append(CCFLAGS=["-ffp-contract=off"])
|
|
else:
|
|
# msvc defaults to /fp:precise. Visual Studio 2022 does not emit floating-point contractions
|
|
# with /fp:precise, but previous versions can. Disable contractions altogether by using
|
|
# /fp:strict.
|
|
env.Append(CCFLAGS=["/fp:strict"])
|
|
|
|
if env.TargetOSIs("linux"):
|
|
env.Append(LIBS=["m"])
|
|
if not env.TargetOSIs("android"):
|
|
env.Append(LIBS=["resolv"])
|
|
|
|
elif env.TargetOSIs("solaris"):
|
|
env.Append(LIBS=["socket", "resolv", "lgrp"])
|
|
|
|
elif env.TargetOSIs("freebsd"):
|
|
env.Append(LIBS=["kvm"])
|
|
env.Append(CCFLAGS=["-fno-omit-frame-pointer"])
|
|
|
|
elif env.TargetOSIs("darwin"):
|
|
env.Append(LIBS=["resolv"])
|
|
|
|
elif env.TargetOSIs("openbsd"):
|
|
env.Append(LIBS=["kvm"])
|
|
|
|
elif env.TargetOSIs("windows"):
|
|
env["DIST_ARCHIVE_SUFFIX"] = ".zip"
|
|
|
|
# If tools configuration fails to set up 'cl' in the path, fall back to importing the whole
|
|
# shell environment and hope for the best. This will work, for instance, if you have loaded
|
|
# an SDK shell.
|
|
for pathdir in env["ENV"]["PATH"].split(os.pathsep):
|
|
if os.path.exists(os.path.join(pathdir, "cl.exe")):
|
|
break
|
|
else:
|
|
print("NOTE: Tool configuration did not find 'cl' compiler, falling back to os environment")
|
|
env["ENV"] = dict(os.environ)
|
|
|
|
env.Append(
|
|
CPPDEFINES=[
|
|
# This tells the Windows compiler not to link against the .lib files
|
|
# and to use boost as a bunch of header-only libraries
|
|
"BOOST_ALL_NO_LIB",
|
|
]
|
|
)
|
|
|
|
env.Append(CPPDEFINES=["_UNICODE"])
|
|
env.Append(CPPDEFINES=["UNICODE"])
|
|
|
|
# Temporary fixes to allow compilation with VS2017
|
|
env.Append(
|
|
CPPDEFINES=[
|
|
"_SILENCE_CXX17_ALLOCATOR_VOID_DEPRECATION_WARNING",
|
|
"_SILENCE_CXX17_OLD_ALLOCATOR_MEMBERS_DEPRECATION_WARNING",
|
|
"_SILENCE_CXX17_CODECVT_HEADER_DEPRECATION_WARNING",
|
|
# TODO(SERVER-60151): Until we are fully in C++20 mode, it is
|
|
# easier to simply suppress C++20 deprecations. After we have
|
|
# switched over we should address any actual deprecated usages
|
|
# and then remove this flag.
|
|
"_SILENCE_ALL_CXX20_DEPRECATION_WARNINGS",
|
|
]
|
|
)
|
|
|
|
# /EHsc exception handling style for visual studio
|
|
# /W3 warning level
|
|
env.Append(CCFLAGS=["/EHsc", "/W3"])
|
|
|
|
# Suppress some warnings we don't like, or find necessary to
|
|
# suppress. Please keep this list alphabetized and commented.
|
|
env.Append(
|
|
CCFLAGS=[
|
|
# C4068: unknown pragma. added so that we can specify unknown
|
|
# pragmas for other compilers.
|
|
"/wd4068",
|
|
# C4244: 'conversion' conversion from 'type1' to 'type2',
|
|
# possible loss of data. An integer type is converted to a
|
|
# smaller integer type.
|
|
"/wd4244",
|
|
# C4267: 'var' : conversion from 'size_t' to 'type', possible
|
|
# loss of data. When compiling with /Wp64, or when compiling
|
|
# on a 64-bit operating system, type is 32 bits but size_t is
|
|
# 64 bits when compiling for 64-bit targets. To fix this
|
|
# warning, use size_t instead of a type.
|
|
"/wd4267",
|
|
# C4290: C++ exception specification ignored except to
|
|
# indicate a function is not __declspec(nothrow). A function
|
|
# is declared using exception specification, which Visual C++
|
|
# accepts but does not implement.
|
|
"/wd4290",
|
|
# C4351: On extremely old versions of MSVC (pre 2k5), default
|
|
# constructing an array member in a constructor's
|
|
# initialization list would not zero the array members "in
|
|
# some cases". Since we don't target MSVC versions that old,
|
|
# this warning is safe to ignore.
|
|
"/wd4351",
|
|
# C4355: 'this' : used in base member initializer list. The
|
|
# this pointer is valid only within nonstatic member
|
|
# functions. It cannot be used in the initializer list for a
|
|
# base class.
|
|
"/wd4355",
|
|
# C4373: Older versions of MSVC would fail to make a function
|
|
# in a derived class override a virtual function in the
|
|
# parent, when defined inline and at least one of the
|
|
# parameters is made const. The behavior is incorrect under
|
|
# the standard. MSVC is fixed now, and the warning exists
|
|
# merely to alert users who may have relied upon the older,
|
|
# non-compliant behavior. Our code should not have any
|
|
# problems with the older behavior, so we can just disable
|
|
# this warning.
|
|
"/wd4373",
|
|
# C4800: 'type' : forcing value to bool 'true' or 'false'
|
|
# (performance warning). This warning is generated when a
|
|
# value that is not bool is assigned or coerced into type
|
|
# bool.
|
|
"/wd4800",
|
|
# C4251: This warning attempts to prevent usage of CRT (C++
|
|
# standard library) types in DLL interfaces. That is a good
|
|
# idea for DLLs you ship to others, but in our case, we know
|
|
# that all DLLs are built consistently. Suppress the warning.
|
|
"/wd4251",
|
|
]
|
|
)
|
|
|
|
# mozjs requires the following
|
|
# 'declaration' : no matching operator delete found; memory will not be freed if
|
|
# initialization throws an exception
|
|
env.Append(CCFLAGS=["/wd4291"])
|
|
|
|
# some warnings we should treat as errors:
|
|
# c4013
|
|
# 'function' undefined; assuming extern returning int
|
|
# This warning occurs when files compiled for the C language use functions not defined
|
|
# in a header file.
|
|
# c4099
|
|
# 'identifier' : type name first seen using 'objecttype1' now seen using 'objecttype2'
|
|
# This warning occurs when classes and structs are declared with a mix of struct and class
|
|
# which can cause linker failures
|
|
# c4930
|
|
# 'identifier': prototyped function not called (was a variable definition intended?)
|
|
# This warning indicates a most-vexing parse error, where a user declared a function that
|
|
# was probably intended as a variable definition. A common example is accidentally
|
|
# declaring a function called lock that takes a mutex when one meant to create a guard
|
|
# object called lock on the stack.
|
|
env.Append(CCFLAGS=["/we4013", "/we4099", "/we4930"])
|
|
|
|
env.Append(
|
|
CPPDEFINES=[
|
|
"_CONSOLE",
|
|
"_CRT_SECURE_NO_WARNINGS",
|
|
"_ENABLE_EXTENDED_ALIGNED_STORAGE",
|
|
"_SCL_SECURE_NO_WARNINGS",
|
|
]
|
|
)
|
|
|
|
# this would be for pre-compiled headers, could play with it later
|
|
# env.Append( CCFLAGS=['/Yu"pch.h"'] )
|
|
|
|
# Don't send error reports in case of internal compiler error
|
|
env.Append(CCFLAGS=["/errorReport:none"])
|
|
|
|
# Select debugging format. /Zi gives faster links but seems to use more memory.
|
|
if get_option("msvc-debugging-format") == "codeview" and debug_symbols:
|
|
env["CCPDBFLAGS"] = "/Z7"
|
|
elif get_option("msvc-debugging-format") == "pdb" and debug_symbols:
|
|
env["CCPDBFLAGS"] = "/Zi /Fd${TARGET}.pdb"
|
|
|
|
# The SCons built-in pdbGenerator always adds /DEBUG, but we would like
|
|
# control over that flag so that users can override with /DEBUG:fastlink
|
|
# for better local builds. So we overwrite the builtin.
|
|
def pdbGenerator(env, target, source, for_signature):
|
|
try:
|
|
return ["/PDB:%s" % target[0].attributes.pdb]
|
|
except (AttributeError, IndexError):
|
|
return None
|
|
|
|
env["_PDB"] = pdbGenerator
|
|
|
|
# SCons by default adds debug flags /Z7, /Zi and /Debug
|
|
# we want to remove these if debug_symbols are off
|
|
if not debug_symbols:
|
|
del env["CCPDBFLAGS"]
|
|
|
|
# /DEBUG will tell the linker to create a .pdb file
|
|
# which WinDbg and Visual Studio will use to resolve
|
|
# symbols if you want to debug a release-mode image.
|
|
# Note that this means we can't do parallel links in the build.
|
|
#
|
|
# Please also note that this has nothing to do with _DEBUG or optimization.
|
|
|
|
# If the user set a /DEBUG flag explicitly, don't add
|
|
# another. Otherwise use the standard /DEBUG flag, since we always
|
|
# want PDBs.
|
|
if not any(flag.startswith("/DEBUG") for flag in env["LINKFLAGS"]) and debug_symbols:
|
|
env.Append(LINKFLAGS=["/DEBUG"])
|
|
|
|
# /MD: use the multithreaded, DLL version of the run-time library (MSVCRT.lib/MSVCR###.DLL)
|
|
# /MDd: Defines _DEBUG, _MT, _DLL, and uses MSVCRTD.lib/MSVCRD###.DLL
|
|
env.Append(CCFLAGS=["/MDd" if debugBuild else "/MD"])
|
|
|
|
if optBuild == "off":
|
|
env.Append(
|
|
CCFLAGS=["/Od"],
|
|
# windows non optimized builds will cause the PDB to blow up in size,
|
|
# this allows a larger PDB. The flag is undocumented at the time of writing
|
|
# but the microsoft thread which brought about its creation can be found here:
|
|
# https://developercommunity.visualstudio.com/t/pdb-limit-of-4-gib-is-likely-to-be-a-problem-in-a/904784
|
|
#
|
|
# Without this flag MSVC will report a red herring error message, about disk space or invalid path.
|
|
LINKFLAGS=["/pdbpagesize:16384"],
|
|
)
|
|
|
|
if debugBuild:
|
|
# /RTC1: - Enable Stack Frame Run-Time Error Checking; Reports when a variable is used
|
|
# without having been initialized (implies /Od: no optimizations)
|
|
env.Append(CCFLAGS=["/RTC1"])
|
|
else:
|
|
# /O1: optimize for size
|
|
# /O2: optimize for speed (as opposed to size)
|
|
# /Oy-: disable frame pointer optimization (overrides /O2, only affects 32-bit)
|
|
# /INCREMENTAL: NO - disable incremental link - avoid the level of indirection for function
|
|
# calls
|
|
|
|
optFlags = []
|
|
if optBuild == "size":
|
|
optFlags += ["/Os"]
|
|
elif optBuild == "debug":
|
|
optFlags += ["/Ox", "/Zo"]
|
|
else:
|
|
optFlags += ["/O2"]
|
|
optFlags += ["/Oy-"]
|
|
|
|
env.Append(CCFLAGS=optFlags)
|
|
env.Append(LINKFLAGS=["/INCREMENTAL:NO"])
|
|
|
|
# Support large object files since some unit-test sources contain a lot of code
|
|
env.Append(CCFLAGS=["/bigobj"])
|
|
|
|
# Set Source and Executable character sets to UTF-8, this will produce a warning C4828 if the
|
|
# file contains invalid UTF-8.
|
|
env.Append(CCFLAGS=["/utf-8"])
|
|
|
|
# Specify standards conformance mode to the compiler.
|
|
env.Append(CCFLAGS=["/permissive-"])
|
|
|
|
# Enables the __cplusplus preprocessor macro to report an updated value for recent C++ language
|
|
# standards support.
|
|
env.Append(CCFLAGS=["/Zc:__cplusplus"])
|
|
|
|
# Tells the compiler to preferentially call global operator delete or operator delete[]
|
|
# functions that have a second parameter of type size_t when the size of the object is available.
|
|
env.Append(CCFLAGS=["/Zc:sizedDealloc"])
|
|
|
|
# Treat volatile according to the ISO standard and do not guarantee acquire/release semantics.
|
|
env.Append(CCFLAGS=["/volatile:iso"])
|
|
|
|
# Tell CL to produce more useful error messages.
|
|
env.Append(CCFLAGS=["/diagnostics:caret"])
|
|
|
|
# This gives 32-bit programs 4 GB of user address space in WOW64, ignored in 64-bit builds.
|
|
env.Append(LINKFLAGS=["/LARGEADDRESSAWARE"])
|
|
|
|
env.Append(
|
|
LIBS=[
|
|
"DbgHelp",
|
|
"Iphlpapi",
|
|
"Psapi",
|
|
"advapi32",
|
|
"bcrypt",
|
|
"crypt32",
|
|
"dnsapi",
|
|
"kernel32",
|
|
"shell32",
|
|
"pdh",
|
|
"version",
|
|
"winmm",
|
|
"ws2_32",
|
|
"secur32",
|
|
],
|
|
)
|
|
|
|
# When building on visual studio, this sets the name of the debug symbols file
|
|
if env.ToolchainIs("msvc") and debug_symbols:
|
|
env["PDB"] = "${TARGET.base}.pdb"
|
|
|
|
# Python uses APPDATA to determine the location of user installed
|
|
# site-packages. If we do not pass this variable down to Python
|
|
# subprocesses then anything installed with `pip install --user`
|
|
# will be inaccessible leading to import errors.
|
|
#
|
|
# Use env['PLATFORM'] instead of TargetOSIs since we always want this
|
|
# to run on Windows hosts but not always for Windows targets.
|
|
if env["PLATFORM"] == "win32":
|
|
appdata = os.getenv("APPDATA", None)
|
|
if appdata is not None:
|
|
env["ENV"]["APPDATA"] = appdata
|
|
|
|
if env.TargetOSIs("posix"):
|
|
# On linux, C code compiled with gcc/clang -std=c11 causes
|
|
# __STRICT_ANSI__ to be set, and that drops out all of the feature
|
|
# test definitions, resulting in confusing errors when we run C
|
|
# language configure checks and expect to be able to find newer
|
|
# POSIX things. Explicitly enabling _XOPEN_SOURCE fixes that, and
|
|
# should be mostly harmless as on Linux, these macros are
|
|
# cumulative. The C++ compiler already sets _XOPEN_SOURCE, and,
|
|
# notably, setting it again does not disable any other feature
|
|
# test macros, so this is safe to do. Other platforms like macOS
|
|
# and BSD have crazy rules, so don't try this there.
|
|
#
|
|
# Furthermore, as both C++ compilers appear to define _GNU_SOURCE
|
|
# unconditionally (because libstdc++ requires it), it seems
|
|
# prudent to explicitly add that too, so that C language checks
|
|
# see a consistent set of definitions.
|
|
if env.TargetOSIs("linux"):
|
|
env.AppendUnique(
|
|
CPPDEFINES=[
|
|
("_XOPEN_SOURCE", 700),
|
|
"_GNU_SOURCE",
|
|
],
|
|
)
|
|
|
|
# If shared and static object files stripped of their rightmost
|
|
# dot-delimited suffix would collide, modify the shared library
|
|
# ones so that they won't. We do this because if split dwarf is in
|
|
# play, static and dynamic builds would otherwise overwrite each
|
|
# other's .dwo files, because GCC strips the last suffix and adds
|
|
# .dwo, rather than simply appending .dwo to the full filename.
|
|
objsuffelts = env.subst("$OBJSUFFIX").split(".")
|
|
shobjsuffelts = env.subst("$SHOBJSUFFIX").split(".")
|
|
if objsuffelts[0:-1] == shobjsuffelts[0:-1]:
|
|
env["SHOBJSUFFIX"] = ".dyn${OBJSUFFIX}"
|
|
|
|
# Everything on OS X is position independent by default.
|
|
if not env.TargetOSIs("darwin"):
|
|
if get_option("runtime-hardening") == "on":
|
|
# If runtime hardening is requested, then build anything
|
|
# destined for an executable with the necessary flags for PIE.
|
|
env.AppendUnique(
|
|
PROGCFLAGS=["-fPIE"],
|
|
PROGCCFLAGS=["-fPIE"],
|
|
PROGCXXFLAGS=["-fPIE"],
|
|
PROGLINKFLAGS=["-pie"],
|
|
)
|
|
|
|
# -Winvalid-pch Warn if a precompiled header (see Precompiled Headers) is found in the search path but can't be used.
|
|
env.Append(
|
|
CCFLAGS=[
|
|
"-fasynchronous-unwind-tables",
|
|
"-g2" if not env.TargetOSIs("emscripten") else "-g",
|
|
"-Wall",
|
|
"-Wsign-compare",
|
|
"-Wno-unknown-pragmas",
|
|
"-Winvalid-pch",
|
|
],
|
|
)
|
|
|
|
if env.get("DWARF_VERSION"):
|
|
if env.TargetOSIs("darwin"):
|
|
env.FatalError("Setting DWARF_VERSION on darwin is not supported.")
|
|
env.AppendUnique(
|
|
CCFLAGS=["-gdwarf-$DWARF_VERSION"],
|
|
LINKFLAGS=["-gdwarf-$DWARF_VERSION"],
|
|
)
|
|
|
|
# TODO: At least on x86, glibc as of 2.3.4 will consult the
|
|
# .eh_frame info via _Unwind_Backtrace to do backtracing without
|
|
# needing the frame pointer, despite what the backtrace man page
|
|
# actually says. We should see if we can drop the requirement that
|
|
# we use libunwind here.
|
|
can_nofp = env.TargetOSIs("darwin") or use_libunwind
|
|
|
|
# For debug builds with tcmalloc, we need the frame pointer so it can
|
|
# record the stack of allocations.
|
|
can_nofp &= not (
|
|
debugBuild and (env["MONGO_ALLOCATOR"] in ["tcmalloc-google", "tcmalloc-gperf"])
|
|
)
|
|
|
|
# Only disable frame pointers if requested
|
|
can_nofp &= "nofp" in selected_experimental_optimizations
|
|
|
|
if not can_nofp:
|
|
env.Append(CCFLAGS=["-fno-omit-frame-pointer"])
|
|
|
|
if "tbaa" not in selected_experimental_optimizations:
|
|
env.Append(CCFLAGS=["-fno-strict-aliasing"])
|
|
|
|
# Enabling hidden visibility on non-darwin requires that we have
|
|
# libunwind in play, since glibc backtrace will not work
|
|
# correctly.
|
|
if "vishidden" in selected_experimental_optimizations and (
|
|
env.TargetOSIs("darwin") or use_libunwind
|
|
):
|
|
if link_model.startswith("dynamic"):
|
|
# In dynamic mode, we can't make the default visibility
|
|
# hidden because not all libraries have export tags. But
|
|
# we can at least make inlines hidden.
|
|
#
|
|
# TODO: Except on macOS, where we observe lots of crashes
|
|
# when we enable this. We should investigate further but
|
|
# it isn't relevant for the purpose of exploring these
|
|
# flags on linux, where they seem to work fine.
|
|
if not env.TargetOSIs("darwin"):
|
|
env.Append(CXXFLAGS=["-fvisibility-inlines-hidden"])
|
|
else:
|
|
# In static mode, we need an escape hatch for a few
|
|
# libraries that don't work correctly when built with
|
|
# hidden visiblity.
|
|
def conditional_visibility_generator(target, source, env, for_signature):
|
|
if "DISALLOW_VISHIDDEN" in env:
|
|
return
|
|
return "-fvisibility=hidden"
|
|
|
|
env.Append(
|
|
CCFLAGS_VISIBILITY_HIDDEN_GENERATOR=conditional_visibility_generator,
|
|
CCFLAGS="$CCFLAGS_VISIBILITY_HIDDEN_GENERATOR",
|
|
)
|
|
|
|
# env.Append( " -Wconversion" ) TODO: this doesn't really work yet
|
|
env.Append(CXXFLAGS=["-Woverloaded-virtual"])
|
|
|
|
# On OS X, clang doesn't want the pthread flag at link time, or it
|
|
# issues warnings which make it impossible for us to declare link
|
|
# warnings as errors. See http://stackoverflow.com/a/19382663.
|
|
if not (env.TargetOSIs("darwin") and env.ToolchainIs("clang")):
|
|
env.Append(LINKFLAGS=["-pthread"])
|
|
|
|
# SERVER-9761: Ensure early detection of missing symbols in dependent
|
|
# libraries at program startup. For non-release dynamic builds we disable
|
|
# this behavior in the interest of improved mongod startup times.
|
|
|
|
# Xcode15 removed bind_at_load functionality so we cannot have a selection for macosx here
|
|
# ld: warning: -bind_at_load is deprecated on macOS
|
|
if has_option("release") or get_option("link-model") != "dynamic":
|
|
if not env.TargetOSIs("macOS"):
|
|
env.Append(LINKFLAGS=["-Wl,-z,now"])
|
|
|
|
# We need to use rdynamic for backtraces with glibc unless we have libunwind.
|
|
nordyn = env.TargetOSIs("darwin") or use_libunwind
|
|
|
|
# And of course only do rdyn if the experimenter asked for it.
|
|
nordyn &= "nordyn" in selected_experimental_optimizations
|
|
|
|
if nordyn:
|
|
|
|
def export_symbol_generator(source, target, env, for_signature):
|
|
symbols = copy.copy(env.get("EXPORT_SYMBOLS", []))
|
|
for lib in libdeps.get_libdeps(source, target, env, for_signature):
|
|
if lib.env:
|
|
symbols.extend(lib.env.get("EXPORT_SYMBOLS", []))
|
|
export_expansion = "${EXPORT_SYMBOL_FLAG}"
|
|
return [f"-Wl,{export_expansion}{symbol}" for symbol in symbols]
|
|
|
|
env["EXPORT_SYMBOL_GEN"] = export_symbol_generator
|
|
|
|
# For darwin, we need the leading underscore but for others we
|
|
# don't. Hacky but it works to jam that distinction into the
|
|
# flag itself, since it already differs on darwin.
|
|
if env.TargetOSIs("darwin"):
|
|
env["EXPORT_SYMBOL_FLAG"] = "-exported_symbol,_"
|
|
else:
|
|
env["EXPORT_SYMBOL_FLAG"] = "--export-dynamic-symbol,"
|
|
|
|
env.Append(
|
|
PROGLINKFLAGS=[
|
|
"$EXPORT_SYMBOL_GEN",
|
|
],
|
|
)
|
|
elif not env.TargetOSIs("darwin"):
|
|
env.Append(
|
|
PROGLINKFLAGS=[
|
|
"-rdynamic",
|
|
],
|
|
)
|
|
|
|
# make scons colorgcc friendly
|
|
for key in ("HOME", "TERM"):
|
|
try:
|
|
env["ENV"][key] = os.environ[key]
|
|
except KeyError:
|
|
pass
|
|
|
|
if has_option("gcov"):
|
|
if not (env.TargetOSIs("linux") and (env.ToolchainIs("gcc", "clang"))):
|
|
# TODO: This should become supported under: https://jira.mongodb.org/browse/SERVER-49877
|
|
env.FatalError(
|
|
"Coverage option 'gcov' is currently only supported on linux with gcc and clang. See SERVER-49877."
|
|
)
|
|
|
|
env.AppendUnique(
|
|
CCFLAGS=["--coverage"],
|
|
LINKFLAGS=["--coverage"],
|
|
)
|
|
|
|
if optBuild == "off":
|
|
env.Append(CCFLAGS=["-O0"])
|
|
else:
|
|
if optBuild == "size":
|
|
env.Append(CCFLAGS=["-Os"])
|
|
elif optBuild == "debug":
|
|
env.Append(CCFLAGS=["-Og"])
|
|
else:
|
|
if "O3" in selected_experimental_optimizations:
|
|
env.Append(CCFLAGS=["-O3"])
|
|
else:
|
|
env.Append(CCFLAGS=["-O2"])
|
|
|
|
if "treevec" in selected_experimental_optimizations:
|
|
env.Append(CCFLAGS=["-ftree-vectorize"])
|
|
|
|
wiredtiger = False
|
|
if get_option("wiredtiger") == "on":
|
|
# Wiredtiger only supports 64-bit architecture, and will fail to compile on 32-bit
|
|
# so disable WiredTiger automatically on 32-bit since wiredtiger is on by default
|
|
if env["TARGET_ARCH"] == "i386":
|
|
env.FatalError(
|
|
"WiredTiger is not supported on 32-bit platforms\n"
|
|
"Re-run scons with --wiredtiger=off to build on 32-bit platforms"
|
|
)
|
|
else:
|
|
wiredtiger = True
|
|
|
|
if not env.TargetOSIs("windows", "macOS") and (env.ToolchainIs("GCC", "clang")):
|
|
# By default, apply our current microarchitecture minima. If the
|
|
# user has customized a flag of the same name in any of CCFLAGS,
|
|
# CFLAGS, or CXXFLAGS, we disable applying our default to
|
|
# CCFLAGS. We are assuming the user knows what they are doing,
|
|
# e.g. we don't try to be smart and notice that they applied it to
|
|
# CXXFLAGS and therefore still apply it to CFLAGS since they
|
|
# didn't customize that. Basically, don't send those flags in
|
|
# unless you a) mean it, and b) know what you are doing, and c)
|
|
# cover all your bases by either setting it via CCFLAGS, or
|
|
# setting it for both C and C++ by setting both of CFLAGS and
|
|
# CXXFLAGS.
|
|
|
|
default_targeting_flags_for_architecture = {
|
|
"aarch64": {"-march=": "armv8.2-a", "-mtune=": "generic"},
|
|
"i386": {"-march=": "nocona", "-mtune=": "generic"},
|
|
"ppc64le": {"-mcpu=": "power8", "-mtune=": "power8", "-mcmodel=": "medium"},
|
|
"s390x": {"-march=": "z196", "-mtune=": "zEC12"},
|
|
}
|
|
|
|
# If we are enabling vectorization in sandybridge mode, we'd
|
|
# rather not hit the 256 wide vector instructions because the
|
|
# heavy versions can cause clock speed reductions.
|
|
if "sandybridge" in selected_experimental_optimizations:
|
|
default_targeting_flags_for_architecture["x86_64"] = {
|
|
"-march=": "sandybridge",
|
|
"-mtune=": "generic",
|
|
"-mprefer-vector-width=": "128",
|
|
}
|
|
|
|
default_targeting_flags = default_targeting_flags_for_architecture.get(env["TARGET_ARCH"])
|
|
if default_targeting_flags:
|
|
search_variables = ["CCFLAGS", "CFLAGS", "CXXFLAGS"]
|
|
for targeting_flag, targeting_flag_value in default_targeting_flags.items():
|
|
if not any(
|
|
flag_value.startswith(targeting_flag)
|
|
for search_variable in search_variables
|
|
for flag_value in env[search_variable]
|
|
):
|
|
env.Append(CCFLAGS=[f"{targeting_flag}{targeting_flag_value}"])
|
|
|
|
# discover modules, and load the (python) module for each module's build.py
|
|
mongo_modules = moduleconfig.discover_modules("src/mongo/db/modules", get_option("modules"))
|
|
|
|
has_ninja_module = False
|
|
for module in mongo_modules:
|
|
if hasattr(module, "NinjaFile"):
|
|
has_ninja_module = True
|
|
break
|
|
|
|
if get_option("ninja") != "disabled" and has_ninja_module:
|
|
env.FatalError(
|
|
textwrap.dedent("""\
|
|
ERROR: Ninja tool option '--ninja' should not be used with the ninja module.
|
|
Using both options simultaneously may clobber build.ninja files.
|
|
Remove the ninja module directory or use '--modules= ' to select no modules.
|
|
If using enterprise module, explicitly set '--modules=<name-of-enterprise-module>' to exclude the ninja module.""")
|
|
)
|
|
|
|
if has_ninja_module:
|
|
print(
|
|
"WARNING: You are attempting to use the unsupported/legacy ninja module, instead of the integrated ninja generator. You are strongly encouraged to remove the ninja module from your module list and invoke scons with --ninja generate-ninja"
|
|
)
|
|
|
|
# --- check system ---
|
|
ssl_provider = None
|
|
http_client = get_option("enable-http-client")
|
|
|
|
|
|
def isSanitizerEnabled(self, sanitizerName):
|
|
if "SANITIZERS_ENABLED" not in self:
|
|
return False
|
|
if sanitizerName == "fuzzer":
|
|
return "fuzzer-no-link" in self["SANITIZERS_ENABLED"]
|
|
return sanitizerName in self["SANITIZERS_ENABLED"]
|
|
|
|
|
|
env.AddMethod(isSanitizerEnabled, "IsSanitizerEnabled")
|
|
|
|
|
|
def doConfigure(myenv):
|
|
global wiredtiger
|
|
global ssl_provider
|
|
global http_client
|
|
|
|
# Check that the compilers work.
|
|
#
|
|
# TODO: Currently, we have some flags already injected. Eventually, this should test the
|
|
# bare compilers, and we should re-check at the very end that TryCompile and TryLink still
|
|
# work with the flags we have selected.
|
|
if myenv.ToolchainIs("msvc"):
|
|
compiler_minimum_string = "Microsoft Visual Studio 2022 17.0"
|
|
compiler_test_body = textwrap.dedent(
|
|
"""
|
|
#if !defined(_MSC_VER)
|
|
#error
|
|
#endif
|
|
|
|
#if _MSC_VER < 1930
|
|
#error %s or newer is required to build MongoDB
|
|
#endif
|
|
|
|
int main(int argc, char* argv[]) {
|
|
return 0;
|
|
}
|
|
"""
|
|
% compiler_minimum_string
|
|
)
|
|
elif myenv.ToolchainIs("gcc"):
|
|
compiler_minimum_string = "GCC 11.3"
|
|
compiler_test_body = textwrap.dedent(
|
|
"""
|
|
#if !defined(__GNUC__) || defined(__clang__)
|
|
#error
|
|
#endif
|
|
|
|
#if (__GNUC__ < 11) || (__GNUC__ == 11 && __GNUC_MINOR__ < 3)
|
|
#error %s or newer is required to build MongoDB
|
|
#endif
|
|
|
|
int main(int argc, char* argv[]) {
|
|
return 0;
|
|
}
|
|
"""
|
|
% compiler_minimum_string
|
|
)
|
|
elif env.ToolchainIs("clang"):
|
|
compiler_minimum_string = "clang 12.0 (or Apple XCode 13.0)"
|
|
compiler_test_body = textwrap.dedent(
|
|
"""
|
|
#if !defined(__clang__)
|
|
#error
|
|
#endif
|
|
|
|
#if defined(__apple_build_version__)
|
|
#if __apple_build_version__ < 13000029
|
|
#error %s or newer is required to build MongoDB
|
|
#endif
|
|
#elif (__clang_major__ < 12) || (__clang_major__ == 12 && __clang_minor__ < 0)
|
|
#error %s or newer is required to build MongoDB
|
|
#endif
|
|
|
|
int main(int argc, char* argv[]) {
|
|
return 0;
|
|
}
|
|
"""
|
|
% (compiler_minimum_string, compiler_minimum_string)
|
|
)
|
|
else:
|
|
myenv.ConfError("Error: can't check compiler minimum; don't know this compiler...")
|
|
|
|
def CheckForMinimumCompiler(context, language):
|
|
extension_for = {
|
|
"C": ".c",
|
|
"C++": ".cpp",
|
|
}
|
|
context.Message(
|
|
"Checking if %s compiler is %s or newer..." % (language, compiler_minimum_string)
|
|
)
|
|
result = context.TryCompile(compiler_test_body, extension_for[language])
|
|
context.Result(result)
|
|
return result
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckForMinimumCompiler": CheckForMinimumCompiler,
|
|
},
|
|
)
|
|
|
|
c_compiler_validated = conf.CheckForMinimumCompiler("C")
|
|
cxx_compiler_validated = conf.CheckForMinimumCompiler("C++")
|
|
|
|
suppress_invalid = has_option("disable-minimum-compiler-version-enforcement")
|
|
if releaseBuild and suppress_invalid:
|
|
env.FatalError("--disable-minimum-compiler-version-enforcement is forbidden with --release")
|
|
|
|
if not (c_compiler_validated and cxx_compiler_validated):
|
|
if not suppress_invalid:
|
|
env.ConfError("ERROR: Refusing to build with compiler that does not meet requirements")
|
|
print("WARNING: Ignoring failed compiler version check per explicit user request.")
|
|
print("WARNING: The build may fail, binaries may crash, or may run but corrupt data...")
|
|
|
|
# Figure out what our minimum windows version is. If the user has specified, then use
|
|
# that.
|
|
if env.TargetOSIs("windows"):
|
|
if has_option("win-version-min"):
|
|
win_version_min = get_option("win-version-min")
|
|
else:
|
|
# If no minimum version has been specified, use our default.
|
|
win_version_min = "win10"
|
|
|
|
env["WIN_VERSION_MIN"] = win_version_min
|
|
win_version_min = win_version_min_choices[win_version_min]
|
|
env.Append(CPPDEFINES=[("_WIN32_WINNT", "0x" + win_version_min[0])])
|
|
env.Append(CPPDEFINES=[("BOOST_USE_WINAPI_VERSION", "0x" + win_version_min[0])])
|
|
env.Append(CPPDEFINES=[("NTDDI_VERSION", "0x" + win_version_min[0] + win_version_min[1])])
|
|
|
|
conf.Finish()
|
|
|
|
# We require macOS 10.14 or newer
|
|
if env.TargetOSIs("darwin"):
|
|
# TODO: Better error messages, mention the various -mX-version-min-flags in the error, and
|
|
# single source of truth for versions, plumbed through #ifdef ladder and error messages.
|
|
def CheckDarwinMinima(context):
|
|
test_body = """
|
|
#include <Availability.h>
|
|
#include <AvailabilityMacros.h>
|
|
#include <TargetConditionals.h>
|
|
|
|
#if TARGET_OS_OSX && (__MAC_OS_X_VERSION_MIN_REQUIRED < __MAC_10_14)
|
|
#error 1
|
|
#endif
|
|
"""
|
|
|
|
context.Message(
|
|
"Checking for sufficient {0} target version minimum... ".format(
|
|
context.env["TARGET_OS"]
|
|
)
|
|
)
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".c")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckDarwinMinima": CheckDarwinMinima,
|
|
},
|
|
)
|
|
|
|
if not conf.CheckDarwinMinima():
|
|
conf.env.ConfError("Required target minimum of macOS 10.14 not found")
|
|
|
|
conf.Finish()
|
|
|
|
def CheckFlag(env, flag, tool, extension, link, **mutation):
|
|
def CheckFlagTest(context, tool, extension, flag):
|
|
if link:
|
|
if tool == "C":
|
|
test_body = """
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
int main() {
|
|
printf("Hello, World!");
|
|
return EXIT_SUCCESS;
|
|
}"""
|
|
elif tool == "C++":
|
|
test_body = """
|
|
#include <iostream>
|
|
#include <cstdlib>
|
|
int main() {
|
|
std::cout << "Hello, World!" << std::endl;
|
|
return EXIT_SUCCESS;
|
|
}"""
|
|
context.Message("Checking if linker supports %s... " % (flag))
|
|
ret = context.TryLink(textwrap.dedent(test_body), extension)
|
|
else:
|
|
test_body = ""
|
|
context.Message("Checking if %s compiler supports %s... " % (tool, flag))
|
|
ret = context.TryCompile(textwrap.dedent(test_body), extension)
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
if env.ToolchainIs("msvc"):
|
|
env.FatalError("AddFlagIfSupported is not currently supported with MSVC")
|
|
|
|
test_mutation = mutation
|
|
if env.ToolchainIs("gcc"):
|
|
test_mutation = copy.deepcopy(mutation)
|
|
# GCC helpfully doesn't issue a diagnostic on unknown flags of the form -Wno-xxx
|
|
# unless other diagnostics are triggered. That makes it tough to check for support
|
|
# for -Wno-xxx. To work around, if we see that we are testing for a flag of the
|
|
# form -Wno-xxx (but not -Wno-error=xxx), we also add -Wxxx to the flags. GCC does
|
|
# warn on unknown -Wxxx style flags, so this lets us probe for availablity of
|
|
# -Wno-xxx.
|
|
for kw in list(test_mutation.keys()):
|
|
test_flags = test_mutation[kw]
|
|
for test_flag in test_flags:
|
|
if test_flag.startswith("-Wno-") and not test_flag.startswith("-Wno-error="):
|
|
test_flags.append(re.sub("^-Wno-", "-W", test_flag))
|
|
|
|
# If the user has selected ``configure` in
|
|
# `disable-warnings-as-errors`, the usual mechanisms that
|
|
# would inject Werror or similar are disabled for
|
|
# conftests. But AddFlagIfSupported requires that those flags
|
|
# be used. Disable the generators so we have explicit control.
|
|
cloned = env.Clone(
|
|
CCFLAGS_GENERATE_WERROR=[],
|
|
CXXFLAGS_GENERATE_WERROR=[],
|
|
LINKFLAGS_GENERATE_WERROR=[],
|
|
)
|
|
|
|
cloned.Append(**test_mutation)
|
|
|
|
# Add these *after* the test mutation, so that the mutation
|
|
# can't step on the warnings-as-errors state.
|
|
cloned.Append(
|
|
CCFLAGS=["$CCFLAGS_WERROR"],
|
|
CXXFLAGS=["$CXXFLAGS_WERROR"],
|
|
LINKFLAGS=["$LINKFLAGS_WERROR"],
|
|
)
|
|
|
|
conf = Configure(
|
|
cloned,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckFlag": lambda ctx: CheckFlagTest(ctx, tool, extension, flag),
|
|
},
|
|
)
|
|
available = conf.CheckFlag()
|
|
conf.Finish()
|
|
return available
|
|
|
|
def AddFlagIfSupported(env, flag, tool, extension, link, **mutation):
|
|
available = CheckFlag(env, flag, tool, extension, link, **mutation)
|
|
|
|
if available:
|
|
env.Append(**mutation)
|
|
return available
|
|
|
|
conf_check_vars = {
|
|
"CFLAGS": {"tool": "C", "extension": ".c", "link": False},
|
|
"CCFLAGS": {"tool": "C", "extension": ".c", "link": False},
|
|
"CXXFLAGS": {"tool": "C++", "extension": ".cpp", "link": False},
|
|
"LINKFLAGS": {"tool": "C", "extension": ".c", "link": True},
|
|
"SHLINKFLAGS": {"tool": "C", "extension": ".c", "link": True},
|
|
}
|
|
|
|
def var_func(env, flag, var, func):
|
|
kwargs = dict({var: [flag]}, **conf_check_vars[var])
|
|
return func(env, flag, **kwargs)
|
|
|
|
for var in conf_check_vars:
|
|
myenv.AddMethod(
|
|
functools.partial(var_func, var=var, func=AddFlagIfSupported), f"AddTo{var}IfSupported"
|
|
)
|
|
myenv.AddMethod(
|
|
functools.partial(var_func, var=var, func=CheckFlag), f"Check{var}Supported"
|
|
)
|
|
|
|
if myenv.ToolchainIs("gcc", "clang"):
|
|
# This tells clang/gcc to use the gold linker if it is available - we prefer the gold linker
|
|
# because it is much faster. Don't use it if the user has already configured another linker
|
|
# selection manually.
|
|
if any(flag.startswith("-fuse-ld=") for flag in env["LINKFLAGS"]):
|
|
myenv.FatalError(
|
|
"Use the '--linker' option instead of modifying the LINKFLAGS directly."
|
|
)
|
|
|
|
linker_ld = get_option("linker")
|
|
|
|
if linker_ld == "bfd":
|
|
myenv.FatalError("The linker 'bfd' is not supported.")
|
|
elif linker_ld == "auto":
|
|
if not env.TargetOSIs("darwin", "macOS"):
|
|
if not myenv.AddToLINKFLAGSIfSupported("-fuse-ld=lld"):
|
|
myenv.FatalError(
|
|
"The recommended linker 'lld' is not supported with the current compiler configuration, you can try the 'gold' linker with '--linker=gold'."
|
|
)
|
|
elif link_model.startswith("dynamic") and linker_ld == "bfd":
|
|
# BFD is not supported due to issues with it causing warnings from some of
|
|
# the third party libraries that mongodb is linked with:
|
|
# https://jira.mongodb.org/browse/SERVER-49465
|
|
myenv.FatalError(f"Linker {linker_ld} is not supported with dynamic link model builds.")
|
|
else:
|
|
if not myenv.AddToLINKFLAGSIfSupported(f"-fuse-ld={linker_ld}"):
|
|
myenv.FatalError(f"Linker {linker_ld} could not be configured.")
|
|
|
|
if has_option("gcov") and myenv.AddToCCFLAGSIfSupported("-fprofile-update=single"):
|
|
myenv.AppendUnique(LINKFLAGS=["-fprofile-update=single"])
|
|
|
|
detectCompiler = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckForCXXLink": CheckForCXXLink,
|
|
},
|
|
)
|
|
|
|
if not detectCompiler.CheckCC():
|
|
env.ConfError(
|
|
"C compiler {0} doesn't work",
|
|
detectEnv["CC"],
|
|
)
|
|
|
|
if not detectCompiler.CheckCXX():
|
|
env.ConfError(
|
|
"C++ compiler {0} doesn't work",
|
|
detectEnv["CXX"],
|
|
)
|
|
|
|
if not detectCompiler.CheckForCXXLink():
|
|
env.ConfError(
|
|
"C++ compiler {0} can't link C++ programs",
|
|
detectEnv["CXX"],
|
|
)
|
|
|
|
detectCompiler.Finish()
|
|
|
|
if myenv.ToolchainIs("clang", "gcc"):
|
|
# This warning was added in g++-4.8.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-unused-local-typedefs")
|
|
|
|
# Clang likes to warn about unused functions, which seems a tad aggressive and breaks
|
|
# -Werror, which we want to be able to use.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-unused-function")
|
|
|
|
# TODO: Note that the following two flags are added to CCFLAGS even though they are
|
|
# really C++ specific. We need to do this because SCons passes CXXFLAGS *before*
|
|
# CCFLAGS, but CCFLAGS contains -Wall, which re-enables the warnings we are trying to
|
|
# suppress. In the future, we should move all warning flags to CCWARNFLAGS and
|
|
# CXXWARNFLAGS and add these to CCOM and CXXCOM as appropriate.
|
|
#
|
|
# Clang likes to warn about unused private fields, but some of our third_party
|
|
# libraries have such things.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-unused-private-field")
|
|
|
|
# Prevents warning about using deprecated features (such as auto_ptr in c++11)
|
|
# Using -Wno-error=deprecated-declarations does not seem to work on some compilers,
|
|
# including at least g++-4.6.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-deprecated-declarations")
|
|
|
|
# As of clang-3.4, this warning appears in v8, and gets escalated to an error.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-tautological-constant-out-of-range-compare")
|
|
|
|
# As of clang in Android NDK 17, these warnings appears in boost and/or ICU, and get escalated to errors
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-tautological-constant-compare")
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-tautological-unsigned-zero-compare")
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-tautological-unsigned-enum-zero-compare")
|
|
|
|
# New in clang-3.4, trips up things mostly in third_party, but in a few places in the
|
|
# primary mongo sources as well.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-unused-const-variable")
|
|
|
|
# Prevents warning about unused but set variables found in boost version 1.49
|
|
# in boost/date_time/format_date_parser.hpp which does not work for compilers
|
|
# GCC >= 4.6. Error explained in https://svn.boost.org/trac/boost/ticket/6136 .
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-unused-but-set-variable")
|
|
|
|
# This has been suppressed in gcc 4.8, due to false positives, but not in clang. So
|
|
# we explicitly disable it here.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-missing-braces")
|
|
|
|
# Suppress warnings about not consistently using override everywhere in a class. It seems
|
|
# very pedantic, and we have a fair number of instances.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-inconsistent-missing-override")
|
|
|
|
# Don't issue warnings about potentially evaluated expressions
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-potentially-evaluated-expression")
|
|
|
|
# SERVER-76472 we don't try to maintain ABI so disable warnings about possible ABI issues.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-psabi")
|
|
|
|
# Warn about moves of prvalues, which can inhibit copy elision.
|
|
myenv.AddToCXXFLAGSIfSupported("-Wpessimizing-move")
|
|
|
|
# Disable warning about variables that may not be initialized
|
|
# Failures are triggered in the case of boost::optional in GCC 4.8.x
|
|
# TODO: re-evaluate when we move to GCC 5.3
|
|
# see: http://stackoverflow.com/questions/21755206/how-to-get-around-gcc-void-b-4-may-be-used-uninitialized-in-this-funct
|
|
myenv.AddToCXXFLAGSIfSupported("-Wno-maybe-uninitialized")
|
|
|
|
# Disable warning about templates that can't be implicitly instantiated. It is an attempt to
|
|
# make a link error into an easier-to-debug compiler failure, but it triggers false
|
|
# positives if explicit instantiation is used in a TU that can see the full definition. This
|
|
# is a problem at least for the S2 headers.
|
|
myenv.AddToCXXFLAGSIfSupported("-Wno-undefined-var-template")
|
|
|
|
# This warning was added in clang-4.0, but it warns about code that is required on some
|
|
# platforms. Since the warning just states that 'explicit instantiation of [a template] that
|
|
# occurs after an explicit specialization has no effect', it is harmless on platforms where
|
|
# it isn't required
|
|
myenv.AddToCXXFLAGSIfSupported("-Wno-instantiation-after-specialization")
|
|
|
|
# This warning was added in clang-5 and flags many of our lambdas. Since it isn't actively
|
|
# harmful to capture unused variables we are suppressing for now with a plan to fix later.
|
|
myenv.AddToCCFLAGSIfSupported("-Wno-unused-lambda-capture")
|
|
|
|
# Enable sized deallocation support.
|
|
myenv.AddToCXXFLAGSIfSupported("-fsized-deallocation")
|
|
|
|
# This warning was added in Apple clang version 11 and flags many explicitly defaulted move
|
|
# constructors and assignment operators for being implicitly deleted, which is not useful.
|
|
myenv.AddToCXXFLAGSIfSupported("-Wno-defaulted-function-deleted")
|
|
|
|
# SERVER-44856: Our windows builds complain about unused
|
|
# exception parameters, but GCC and clang don't seem to do
|
|
# that for us automatically. In the interest of making it more
|
|
# likely to catch these errors early, add the (currently clang
|
|
# only) flag that turns it on.
|
|
myenv.AddToCXXFLAGSIfSupported("-Wunused-exception-parameter")
|
|
|
|
# TODO SERVER-58675 - Remove this suppression after abseil is upgraded
|
|
myenv.AddToCXXFLAGSIfSupported("-Wno-deprecated-builtins")
|
|
|
|
# Check if we can set "-Wnon-virtual-dtor" when "-Werror" is set. The only time we can't set it is on
|
|
# clang 3.4, where a class with virtual function(s) and a non-virtual destructor throws a warning when
|
|
# it shouldn't.
|
|
def CheckNonVirtualDtor(context):
|
|
test_body = """
|
|
class Base {
|
|
public:
|
|
virtual void foo() const = 0;
|
|
protected:
|
|
~Base() {};
|
|
};
|
|
|
|
class Derived : public Base {
|
|
public:
|
|
virtual void foo() const {}
|
|
};
|
|
"""
|
|
|
|
context.Message("Checking if -Wnon-virtual-dtor works reasonably... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
myenvClone = myenv.Clone()
|
|
myenvClone.Append(
|
|
CCFLAGS=[
|
|
"$CCFLAGS_WERROR",
|
|
"-Wnon-virtual-dtor",
|
|
],
|
|
)
|
|
conf = Configure(
|
|
myenvClone,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckNonVirtualDtor": CheckNonVirtualDtor,
|
|
},
|
|
)
|
|
if conf.CheckNonVirtualDtor():
|
|
myenv.Append(CXXFLAGS=["-Wnon-virtual-dtor"])
|
|
conf.Finish()
|
|
|
|
# As of XCode 9, this flag must be present (it is not enabled
|
|
# by -Wall), in order to enforce that -mXXX-version-min=YYY
|
|
# will enforce that you don't use APIs from ZZZ.
|
|
if env.TargetOSIs("darwin"):
|
|
env.AddToCCFLAGSIfSupported("-Wunguarded-availability")
|
|
env.AddToCCFLAGSIfSupported("-Wno-enum-constexpr-conversion")
|
|
# TODO SERVER-54659 - ASIO depends on std::result_of which was removed in C++ 20
|
|
myenv.Append(CPPDEFINES=["ASIO_HAS_STD_INVOKE_RESULT"])
|
|
# This is needed to compile boost on the newer xcodes
|
|
myenv.Append(CPPDEFINES=["BOOST_NO_CXX98_FUNCTION_BASE"])
|
|
|
|
if get_option("runtime-hardening") == "on":
|
|
# Enable 'strong' stack protection preferentially, but fall back to 'all' if it is not
|
|
# available. Note that we need to add these to the LINKFLAGS as well, since otherwise we
|
|
# might not link libssp when we need to (see SERVER-12456).
|
|
if myenv.ToolchainIs("gcc", "clang"):
|
|
if myenv.AddToCCFLAGSIfSupported("-fstack-protector-strong"):
|
|
myenv.Append(
|
|
LINKFLAGS=[
|
|
"-fstack-protector-strong",
|
|
],
|
|
)
|
|
elif myenv.AddToCCFLAGSIfSupported("-fstack-protector-all"):
|
|
myenv.Append(
|
|
LINKFLAGS=[
|
|
"-fstack-protector-all",
|
|
],
|
|
)
|
|
|
|
if "cfex" in selected_experimental_runtime_hardenings:
|
|
myenv.Append(
|
|
CFLAGS=[
|
|
"-fexceptions",
|
|
],
|
|
)
|
|
|
|
if "stackclash" in selected_experimental_runtime_hardenings:
|
|
myenv.AddToCCFLAGSIfSupported("-fstack-clash-protection")
|
|
|
|
if "controlflow" in selected_experimental_runtime_hardenings:
|
|
myenv.AddToCCFLAGSIfSupported("-fcf-protection=full")
|
|
|
|
if myenv.ToolchainIs("clang"):
|
|
# TODO: There are several interesting things to try here, but they each have
|
|
# consequences we need to investigate.
|
|
#
|
|
# - fsanitize=bounds: This does static bounds checking. We can
|
|
# probably turn this on along with fsanitize-trap so that we
|
|
# don't depend on the ASAN runtime.
|
|
#
|
|
# - fsanitize=safestack: This looks very interesting, and is
|
|
# probably what we want. However there are a few problems:
|
|
#
|
|
# - It relies on having the RT library available, and it is
|
|
# unclear whether we can ship binaries that depend on
|
|
# that.
|
|
#
|
|
# - It is incompatible with a shared object build.
|
|
#
|
|
# - It may not work with SpiderMonkey due to needing to
|
|
# inform the GC about the stacks so that mark-sweep
|
|
#
|
|
# - fsanitize=cfi: Again, very interesting, however it
|
|
# requires LTO builds.
|
|
pass
|
|
|
|
if has_option("osx-version-min"):
|
|
message = """
|
|
The --osx-version-min option is no longer supported.
|
|
|
|
To specify a target minimum for Darwin platforms, please explicitly add the appropriate options
|
|
to CCFLAGS and LINKFLAGS on the command line:
|
|
|
|
macOS: scons CCFLAGS="-mmacosx-version-min=10.14" LINKFLAGS="-mmacosx-version-min=10.14" ..
|
|
|
|
Note that MongoDB requires macOS 10.14 or later.
|
|
"""
|
|
myenv.ConfError(textwrap.dedent(message))
|
|
|
|
usingLibStdCxx = False
|
|
if has_option("libc++"):
|
|
# TODO SERVER-54659 - ASIO depends on std::result_of which was removed in C++ 20
|
|
myenv.Append(CPPDEFINES=["ASIO_HAS_STD_INVOKE_RESULT"])
|
|
|
|
if not myenv.ToolchainIs("clang"):
|
|
myenv.FatalError("libc++ is currently only supported for clang")
|
|
if myenv.AddToCXXFLAGSIfSupported("-stdlib=libc++"):
|
|
myenv.Append(LINKFLAGS=["-stdlib=libc++"])
|
|
else:
|
|
myenv.ConfError("libc++ requested, but compiler does not support -stdlib=libc++")
|
|
else:
|
|
|
|
def CheckLibStdCxx(context):
|
|
test_body = """
|
|
#include <vector>
|
|
#if !defined(__GLIBCXX__)
|
|
#error
|
|
#endif
|
|
"""
|
|
|
|
context.Message("Checking if we are using libstdc++... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckLibStdCxx": CheckLibStdCxx,
|
|
},
|
|
)
|
|
usingLibStdCxx = conf.CheckLibStdCxx()
|
|
conf.Finish()
|
|
|
|
if myenv.ToolchainIs("msvc"):
|
|
if get_option("cxx-std") == "20":
|
|
myenv.AppendUnique(CCFLAGS=["/std:c++20"])
|
|
else:
|
|
if get_option("cxx-std") == "20":
|
|
if not myenv.AddToCXXFLAGSIfSupported("-std=c++20"):
|
|
myenv.ConfError("Compiler does not honor -std=c++20")
|
|
|
|
if not myenv.AddToCFLAGSIfSupported("-std=c11"):
|
|
myenv.ConfError("C++20 mode selected for C++ files, but can't enable C11 for C files")
|
|
|
|
if using_system_version_of_cxx_libraries():
|
|
print("WARNING: System versions of C++ libraries must be compiled with C++20 support")
|
|
|
|
def CheckCxx20(context):
|
|
test_body = """
|
|
#if __cplusplus < 202002L
|
|
#error
|
|
#endif
|
|
#include <compare>
|
|
[[maybe_unused]] constexpr auto spaceship_operator_is_a_cxx20_feature = 2 <=> 4;
|
|
"""
|
|
|
|
context.Message("Checking for C++20... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckCxx20": CheckCxx20,
|
|
},
|
|
)
|
|
|
|
if get_option("cxx-std") == "20" and not conf.CheckCxx20():
|
|
myenv.ConfError("C++20 support is required to build MongoDB")
|
|
|
|
conf.Finish()
|
|
|
|
# If we are using libstdc++, check to see if we are using a
|
|
# libstdc++ that is older than our GCC minimum of 5.3.0. This is
|
|
# primarly to help people using clang on OS X but forgetting to
|
|
# use --libc++ (or set the target OS X version high enough to get
|
|
# it as the default). We would, ideally, check the __GLIBCXX__
|
|
# version, but for various reasons this is not workable. Instead,
|
|
# we switch on the fact that the <experimental/filesystem> header
|
|
# wasn't introduced until libstdc++ 5.3.0. Yes, this is a terrible
|
|
# hack.
|
|
if usingLibStdCxx:
|
|
|
|
def CheckModernLibStdCxx(context):
|
|
test_body = """
|
|
#if !__has_include(<experimental/filesystem>)
|
|
#error "libstdc++ from GCC 5.3.0 or newer is required"
|
|
#endif
|
|
"""
|
|
|
|
context.Message("Checking for libstdc++ 5.3.0 or better... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckModernLibStdCxx": CheckModernLibStdCxx,
|
|
},
|
|
)
|
|
|
|
suppress_invalid = has_option("disable-minimum-compiler-version-enforcement")
|
|
if not conf.CheckModernLibStdCxx() and not suppress_invalid:
|
|
myenv.ConfError(
|
|
"When using libstdc++, MongoDB requires libstdc++ from GCC 5.3.0 or newer"
|
|
)
|
|
|
|
conf.Finish()
|
|
|
|
if has_option("use-glibcxx-debug"):
|
|
# If we are using a modern libstdc++ and this is a debug build and we control all C++
|
|
# dependencies, then turn on the debugging features in libstdc++.
|
|
# TODO: Need a new check here.
|
|
if not debugBuild:
|
|
myenv.FatalError("--use-glibcxx-debug requires --dbg=on")
|
|
if not usingLibStdCxx:
|
|
myenv.FatalError(
|
|
"--use-glibcxx-debug is only compatible with the GNU implementation "
|
|
"of the C++ standard libary"
|
|
)
|
|
if using_system_version_of_cxx_libraries():
|
|
myenv.FatalError(
|
|
"--use-glibcxx-debug not compatible with system versions of " "C++ libraries."
|
|
)
|
|
myenv.Append(CPPDEFINES=["_GLIBCXX_DEBUG"])
|
|
|
|
# Check if we have a modern Windows SDK
|
|
if env.TargetOSIs("windows"):
|
|
|
|
def CheckWindowsSDKVersion(context):
|
|
test_body = """
|
|
#include <windows.h>
|
|
#if !defined(NTDDI_WINBLUE)
|
|
#error Need Windows SDK Version 8.1 or higher
|
|
#endif
|
|
"""
|
|
|
|
context.Message("Checking Windows SDK is 8.1 or newer... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".c")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckWindowsSDKVersion": CheckWindowsSDKVersion,
|
|
},
|
|
)
|
|
|
|
if not conf.CheckWindowsSDKVersion():
|
|
myenv.ConfError("Windows SDK Version 8.1 or higher is required to build MongoDB")
|
|
|
|
conf.Finish()
|
|
|
|
# Check if we are on a POSIX system by testing if _POSIX_VERSION is defined.
|
|
def CheckPosixSystem(context):
|
|
test_body = """
|
|
// POSIX requires the existence of unistd.h, so if we can't include unistd.h, we
|
|
// are definitely not a POSIX system.
|
|
#include <unistd.h>
|
|
#if !defined(_POSIX_VERSION)
|
|
#error not a POSIX system
|
|
#endif
|
|
"""
|
|
|
|
context.Message("Checking if we are on a POSIX system... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".c")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckPosixSystem": CheckPosixSystem,
|
|
},
|
|
)
|
|
posix_system = conf.CheckPosixSystem()
|
|
|
|
conf.Finish()
|
|
|
|
# Check if we are on a system that support the POSIX clock_gettime function
|
|
# and the "monotonic" clock.
|
|
if posix_system:
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
)
|
|
# On 32-bit systems, we need to define this in order to get access to
|
|
# the 64-bit versions of fseek, etc.
|
|
# except on 32 bit android where it breaks boost
|
|
if not conf.CheckTypeSize("off_t", includes="#include <sys/types.h>", expect=8):
|
|
if not env.TargetOSIs("android"):
|
|
myenv.Append(CPPDEFINES=["_FILE_OFFSET_BITS=64"])
|
|
|
|
conf.Finish()
|
|
|
|
if has_option("sanitize") and get_option("sanitize") != "":
|
|
if not myenv.ToolchainIs("clang", "gcc"):
|
|
env.FatalError("sanitize is only supported with clang or gcc")
|
|
|
|
# sanitizer libs may inject undefined refs (for hooks) at link time, but
|
|
# the symbols will be available at runtime via the compiler runtime lib.
|
|
env.Append(LINKFLAGS="-Wl,--allow-shlib-undefined")
|
|
|
|
if myenv.ToolchainIs("gcc"):
|
|
# GCC's implementation of ASAN depends on libdl.
|
|
env.Append(LIBS=["dl"])
|
|
|
|
sanitizer_list = get_option("sanitize").split(",")
|
|
|
|
using_asan = "address" in sanitizer_list
|
|
using_fsan = "fuzzer" in sanitizer_list
|
|
using_lsan = "leak" in sanitizer_list
|
|
using_tsan = "thread" in sanitizer_list
|
|
using_ubsan = "undefined" in sanitizer_list
|
|
using_msan = "memory" in sanitizer_list
|
|
|
|
if get_option("shared-libsan") == "on" and len(sanitizer_list) > 0:
|
|
if not myenv.ToolchainIs("clang") or not myenv.TargetOSIs("linux"):
|
|
env.FatalError("Error: --shared-libsan is only supported with clang on linux")
|
|
|
|
def get_san_lib_path(sanitizer):
|
|
# TODO SERVER-83727: the v4 clang toolchain doesn't support shared TSAN. Add
|
|
# support here once the toolchain is upgraded.
|
|
san_to_lib = {
|
|
"address": "asan",
|
|
"undefined": "ubsan_standalone",
|
|
}
|
|
sanitizer_lib = san_to_lib.get(sanitizer)
|
|
if sanitizer_lib is None:
|
|
env.FatalError(
|
|
f"Error: --shared-libsan is not supported with {sanitizer} sanitizer"
|
|
)
|
|
arch = env["TARGET_ARCH"]
|
|
san_rt_name = f"libclang_rt.{sanitizer_lib}-{arch}.so"
|
|
p = subprocess.run(
|
|
[env["CXX"], f"-print-file-name={san_rt_name}"], capture_output=True, text=True
|
|
)
|
|
clang_rt_path = p.stdout.strip()
|
|
if not os.path.isfile(clang_rt_path):
|
|
env.FatalError(f"Error: couldn't find sanitizer runtime library {san_rt_name}")
|
|
return clang_rt_path
|
|
|
|
env["SANITIZER_RUNTIME_LIBS"] = [
|
|
get_san_lib_path(sanitizer) for sanitizer in sanitizer_list
|
|
]
|
|
|
|
if "thread" not in sanitizer_list:
|
|
env.Append(LINKFLAGS=["-rtlib=compiler-rt", "-unwindlib=libgcc"])
|
|
|
|
if using_lsan:
|
|
env.FatalError("Please use --sanitize=address instead of --sanitize=leak")
|
|
|
|
if (using_asan or using_msan) and env["MONGO_ALLOCATOR"] in [
|
|
"tcmalloc-google",
|
|
"tcmalloc-gperf",
|
|
]:
|
|
# There are multiply defined symbols between the sanitizer and
|
|
# our vendorized tcmalloc.
|
|
env.FatalError("Cannot use --sanitize=address or --sanitize=memory with tcmalloc")
|
|
|
|
if not myenv.ToolchainIs("clang") and using_msan:
|
|
env.FatalError("Memory Sanitizer (MSan) is only supported with clang.")
|
|
|
|
if using_fsan:
|
|
|
|
def CheckForFuzzerCompilerSupport(context):
|
|
test_body = """
|
|
#include <stddef.h>
|
|
#include <stdint.h>
|
|
|
|
// fuzz_target.cc
|
|
extern "C" int LLVMFuzzerTestOneInput(const uint8_t *Data, size_t Size) {
|
|
return 0;
|
|
}
|
|
"""
|
|
|
|
context.Message("Checking if libfuzzer is supported by the compiler... ")
|
|
|
|
context.env.AppendUnique(
|
|
LINKFLAGS=[
|
|
"-fprofile-instr-generate",
|
|
"-fcoverage-mapping",
|
|
"-fsanitize=fuzzer",
|
|
],
|
|
CCFLAGS=[
|
|
"-fprofile-instr-generate",
|
|
"-fcoverage-mapping",
|
|
],
|
|
)
|
|
|
|
ret = context.TryLink(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
confEnv = myenv.Clone()
|
|
fuzzerConf = Configure(
|
|
confEnv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckForFuzzerCompilerSupport": CheckForFuzzerCompilerSupport,
|
|
},
|
|
)
|
|
if not fuzzerConf.CheckForFuzzerCompilerSupport():
|
|
myenv.FatalError("libfuzzer is not supported by the compiler")
|
|
fuzzerConf.Finish()
|
|
|
|
# We can't include the fuzzer flag with the other sanitize flags
|
|
# The libfuzzer library already has a main function, which will cause the dependencies check
|
|
# to fail
|
|
sanitizer_list.remove("fuzzer")
|
|
sanitizer_list.append("fuzzer-no-link")
|
|
# These flags are needed to generate a coverage report
|
|
myenv.Append(
|
|
LINKFLAGS=[
|
|
"-fprofile-instr-generate",
|
|
"-fcoverage-mapping",
|
|
],
|
|
)
|
|
myenv.Append(
|
|
CCFLAGS=[
|
|
"-fprofile-instr-generate",
|
|
"-fcoverage-mapping",
|
|
],
|
|
)
|
|
|
|
sanitizer_option = "-fsanitize=" + ",".join(sanitizer_list)
|
|
|
|
if myenv.AddToCCFLAGSIfSupported(sanitizer_option):
|
|
myenv.Append(LINKFLAGS=[sanitizer_option])
|
|
myenv.Append(CCFLAGS=["-fno-omit-frame-pointer"])
|
|
else:
|
|
myenv.ConfError("Failed to enable sanitizers with flag: {0}", sanitizer_option)
|
|
|
|
if get_option("shared-libsan") == "on":
|
|
shared_libsan_option = "-shared-libsan"
|
|
if myenv.AddToCCFLAGSIfSupported(shared_libsan_option):
|
|
myenv.Append(LINKFLAGS=[shared_libsan_option])
|
|
|
|
myenv["SANITIZERS_ENABLED"] = sanitizer_list
|
|
|
|
if has_option("sanitize-coverage") and using_fsan:
|
|
sanitize_coverage_list = get_option("sanitize-coverage")
|
|
sanitize_coverage_option = "-fsanitize-coverage=" + sanitize_coverage_list
|
|
if myenv.AddToCCFLAGSIfSupported(sanitize_coverage_option):
|
|
myenv.Append(LINKFLAGS=[sanitize_coverage_option])
|
|
else:
|
|
myenv.ConfError(
|
|
"Failed to enable -fsanitize-coverage with flag: {0}", sanitize_coverage_option
|
|
)
|
|
|
|
denyfiles_map = {
|
|
"address": myenv.File("#etc/asan.denylist"),
|
|
"thread": myenv.File("#etc/tsan.denylist"),
|
|
"undefined": myenv.File("#etc/ubsan.denylist"),
|
|
"memory": myenv.File("#etc/msan.denylist"),
|
|
}
|
|
|
|
# Select those unique deny files that are associated with the
|
|
# currently enabled sanitizers, but filter out those that are
|
|
# zero length.
|
|
denyfiles = {v for (k, v) in denyfiles_map.items() if k in sanitizer_list}
|
|
denyfiles = [f for f in denyfiles if os.stat(f.path).st_size != 0]
|
|
|
|
# Filter out any denylist options that the toolchain doesn't support.
|
|
supportedDenyfiles = []
|
|
denyfilesTestEnv = myenv.Clone()
|
|
for denyfile in denyfiles:
|
|
if denyfilesTestEnv.AddToCCFLAGSIfSupported(f"-fsanitize-blacklist={denyfile}"):
|
|
supportedDenyfiles.append(denyfile)
|
|
denyfilesTestEnv = None
|
|
supportedDenyfiles = sorted(supportedDenyfiles)
|
|
|
|
# If we ended up with any denyfiles after the above filters,
|
|
# then expand them into compiler flag arguments, and use a
|
|
# generator to return at command line expansion time so that
|
|
# we can change the signature if the file contents change.
|
|
if supportedDenyfiles:
|
|
# Unconditionally using the full path can affect SCons cached builds, so we only do
|
|
# this in cases where we know it's going to matter.
|
|
denylist_options = [
|
|
f"-fsanitize-blacklist={denyfile.path}" for denyfile in supportedDenyfiles
|
|
]
|
|
|
|
if "ICECC" in env and env["ICECC"]:
|
|
# Make these files available to remote icecream builds if requested.
|
|
# These paths *must* be absolute to match the paths in the remote
|
|
# toolchain archive. Local builds remain relative.
|
|
local_denylist_options = denylist_options[:]
|
|
denylist_options = [
|
|
f"-fsanitize-blacklist={denyfile.abspath}" for denyfile in supportedDenyfiles
|
|
]
|
|
|
|
# Build a regex of all the regexes in the denylist
|
|
# the regex in the denylist are a shell wildcard format
|
|
# https://clang.llvm.org/docs/SanitizerSpecialCaseList.html#format
|
|
# so a bit of massaging (* -> .*) to get a python regex.
|
|
icecc_denylist_regexes = []
|
|
for denyfile in supportedDenyfiles:
|
|
for line in denyfile.get_contents().decode("utf-8").split("\n"):
|
|
if line.strip().startswith("src:"):
|
|
regex_line = line.replace("src:", "").strip()
|
|
regex_line = re.escape(regex_line)
|
|
icecc_denylist_regexes += [regex_line.replace("\\*", ".*")]
|
|
|
|
icecc_denylist_regex = re.compile("^(?:" + "|".join(icecc_denylist_regexes) + ")$")
|
|
|
|
def is_local_compile(env, target, source, for_signature):
|
|
return icecc_denylist_regex.match(str(source[0])) is not None
|
|
|
|
env["ICECC_LOCAL_COMPILATION_FILTER"] = is_local_compile
|
|
# If a sanitizer is in use with a denylist file, we have to ensure they get
|
|
# added to the toolchain package that gets sent to the remote hosts so they
|
|
# can be found by the remote compiler.
|
|
env.Append(ICECC_CREATE_ENV_ADDFILES=supportedDenyfiles)
|
|
|
|
if "CCACHE" in env and env["CCACHE"]:
|
|
# Work around the fact that some versions of ccache either don't yet support
|
|
# -fsanitize-blacklist at all or only support one instance of it. This will
|
|
# work on any version of ccache because the point is only to ensure that the
|
|
# resulting hash for any compiled object is guaranteed to take into account
|
|
# the effect of any sanitizer denylist files used as part of the build.
|
|
# TODO: This will no longer be required when the following pull requests/
|
|
# issues have been merged and deployed.
|
|
# https://github.com/ccache/ccache/pull/258
|
|
# https://github.com/ccache/ccache/issues/318
|
|
env.Append(CCACHE_EXTRAFILES=supportedDenyfiles)
|
|
env["CCACHE_EXTRAFILES_USE_SOURCE_PATHS"] = True
|
|
|
|
def CCSanitizerDenylistGenerator(source, target, env, for_signature):
|
|
# TODO: SERVER-60915 use new conftest API
|
|
if "conftest" in str(target[0]):
|
|
return ""
|
|
|
|
# TODO: SERVER-64620 use scanner instead of for_signature
|
|
if for_signature:
|
|
return [f.get_csig() for f in supportedDenyfiles]
|
|
|
|
# Check if the denylist gets a match and if so it will be local
|
|
# build and should use the non-abspath.
|
|
# NOTE: in non icecream builds denylist_options becomes relative paths.
|
|
if (
|
|
env.subst("$ICECC_LOCAL_COMPILATION_FILTER", target=target, source=source)
|
|
== "True"
|
|
):
|
|
return local_denylist_options
|
|
|
|
return denylist_options
|
|
|
|
def LinkSanitizerDenylistGenerator(source, target, env, for_signature):
|
|
# TODO: SERVER-60915 use new conftest API
|
|
if "conftest" in str(target[0]):
|
|
return ""
|
|
|
|
# TODO: SERVER-64620 use scanner instead of for_signature
|
|
if for_signature:
|
|
return [f.get_csig() for f in supportedDenyfiles]
|
|
|
|
return denylist_options
|
|
|
|
myenv.AppendUnique(
|
|
CC_SANITIZER_DENYLIST_GENERATOR=CCSanitizerDenylistGenerator,
|
|
LINK_SANITIZER_DENYLIST_GENERATOR=LinkSanitizerDenylistGenerator,
|
|
CCFLAGS="${CC_SANITIZER_DENYLIST_GENERATOR}",
|
|
LINKFLAGS="${LINK_SANITIZER_DENYLIST_GENERATOR}",
|
|
)
|
|
|
|
symbolizer_option = ""
|
|
if env.get("LLVM_SYMBOLIZER", False):
|
|
llvm_symbolizer = env["LLVM_SYMBOLIZER"]
|
|
|
|
if not os.path.isabs(llvm_symbolizer):
|
|
llvm_symbolizer = myenv.WhereIs(llvm_symbolizer)
|
|
|
|
if not myenv.File(llvm_symbolizer).exists():
|
|
myenv.FatalError(f"Symbolizer binary at path {llvm_symbolizer} does not exist")
|
|
|
|
symbolizer_option = f':external_symbolizer_path="{llvm_symbolizer}"'
|
|
|
|
elif using_asan or using_tsan or using_ubsan or using_msan:
|
|
myenv.FatalError(
|
|
"The address, thread, memory, and undefined behavior sanitizers require llvm-symbolizer for meaningful reports. Please set LLVM_SYMBOLIZER to the path to llvm-symbolizer in your SCons invocation"
|
|
)
|
|
|
|
if using_asan:
|
|
# Unfortunately, abseil requires that we make these macros
|
|
# (this, and THREAD_ and UNDEFINED_BEHAVIOR_ below) set,
|
|
# because apparently it is too hard to query the running
|
|
# compiler. We do this unconditionally because abseil is
|
|
# basically pervasive via the 'base' library.
|
|
myenv.AppendUnique(CPPDEFINES=["ADDRESS_SANITIZER"])
|
|
# If anything is changed, added, or removed in either asan_options or
|
|
# lsan_options, be sure to make the corresponding changes to the
|
|
# appropriate build variants in etc/evergreen.yml
|
|
asan_options_clear = [
|
|
"detect_leaks=1",
|
|
"check_initialization_order=true",
|
|
"strict_init_order=true",
|
|
"abort_on_error=1",
|
|
"disable_coredump=0",
|
|
"handle_abort=1",
|
|
"strict_string_checks=true",
|
|
"detect_invalid_pointer_pairs=1",
|
|
]
|
|
asan_options = ":".join(asan_options_clear)
|
|
lsan_options = (
|
|
f"report_objects=1:suppressions={myenv.File('#etc/lsan.suppressions').abspath}"
|
|
)
|
|
env["ENV"]["ASAN_OPTIONS"] = asan_options + symbolizer_option
|
|
env["ENV"]["LSAN_OPTIONS"] = lsan_options + symbolizer_option
|
|
|
|
if using_msan:
|
|
# Makes it easier to debug memory failures at the cost of some perf
|
|
myenv.Append(CCFLAGS=["-fsanitize-memory-track-origins"])
|
|
env["ENV"]["MSAN_OPTIONS"] = symbolizer_option
|
|
if using_tsan:
|
|
if use_libunwind:
|
|
# TODO: SERVER-48622
|
|
#
|
|
# See https://github.com/google/sanitizers/issues/943
|
|
# for why we disallow combining TSAN with
|
|
# libunwind. We could, atlernatively, have added logic
|
|
# to automate the decision about whether to enable
|
|
# libunwind based on whether TSAN is enabled, but that
|
|
# logic is already complex, and it feels better to
|
|
# make it explicit that using TSAN means you won't get
|
|
# the benefits of libunwind. Fixing this is:
|
|
env.FatalError(
|
|
"Cannot use libunwind with TSAN, please add --use-libunwind=off to your compile flags"
|
|
)
|
|
|
|
# We add supressions based on the library file in etc/tsan.suppressions
|
|
# so the link-model needs to be dynamic.
|
|
if not link_model.startswith("dynamic"):
|
|
env.FatalError("TSAN is only supported with dynamic link models")
|
|
|
|
# If anything is changed, added, or removed in
|
|
# tsan_options, be sure to make the corresponding changes
|
|
# to the appropriate build variants in etc/evergreen.yml
|
|
#
|
|
# TODO SERVER-49121: die_after_fork=0 is a temporary
|
|
# setting to allow tests to continue while we figure out
|
|
# why we're running afoul of it.
|
|
#
|
|
# TODO SERVER-65936: report_thread_leaks=0 suppresses
|
|
# reporting thread leaks, which we have because we don't
|
|
# do a clean shutdown of the ServiceContext.
|
|
#
|
|
tsan_options = f"abort_on_error=1:disable_coredump=0:handle_abort=1:halt_on_error=1:report_thread_leaks=0:die_after_fork=0:history_size=4:suppressions={myenv.File('#etc/tsan.suppressions').abspath}"
|
|
myenv["ENV"]["TSAN_OPTIONS"] = tsan_options + symbolizer_option
|
|
myenv.AppendUnique(CPPDEFINES=["THREAD_SANITIZER"])
|
|
|
|
if using_ubsan:
|
|
# By default, undefined behavior sanitizer doesn't stop on
|
|
# the first error. Make it so. Newer versions of clang
|
|
# have renamed the flag.
|
|
if not myenv.AddToCCFLAGSIfSupported("-fno-sanitize-recover"):
|
|
myenv.AddToCCFLAGSIfSupported("-fno-sanitize-recover=undefined")
|
|
myenv.AppendUnique(CPPDEFINES=["UNDEFINED_BEHAVIOR_SANITIZER"])
|
|
|
|
# If anything is changed, added, or removed in ubsan_options, be
|
|
# sure to make the corresponding changes to the appropriate build
|
|
# variants in etc/evergreen.yml
|
|
ubsan_options = "print_stacktrace=1"
|
|
myenv["ENV"]["UBSAN_OPTIONS"] = ubsan_options + symbolizer_option
|
|
|
|
# In dynamic builds, the `vptr` sanitizer check can
|
|
# require additional LIBDEPS edges. That is very
|
|
# inconvenient, because such builds can't use z,defs. The
|
|
# result is a very fragile link graph, where refactoring
|
|
# the link graph in one place can have surprising effects
|
|
# in others. Instead, we just disable the `vptr` sanitizer
|
|
# for dynamic builds. We tried some other approaches in
|
|
# SERVER-49798 of adding a new LIBDEPS_TYPEINFO type, but
|
|
# that didn't address the fundamental issue that the
|
|
# correct link graph for a dynamic+ubsan build isn't the
|
|
# same as the correct link graph for a regular dynamic
|
|
# build.
|
|
if link_model == "dynamic":
|
|
if myenv.AddToCCFLAGSIfSupported("-fno-sanitize=vptr"):
|
|
myenv.AppendUnique(LINKFLAGS=["-fno-sanitize=vptr"])
|
|
|
|
if myenv.ToolchainIs("msvc") and optBuild != "off":
|
|
# http://blogs.msdn.com/b/vcblog/archive/2013/09/11/introducing-gw-compiler-switch.aspx
|
|
#
|
|
myenv.Append(CCFLAGS=["/Gw", "/Gy"])
|
|
myenv.Append(LINKFLAGS=["/OPT:REF"])
|
|
|
|
# http://blogs.msdn.com/b/vcblog/archive/2014/03/25/linker-enhancements-in-visual-studio-2013-update-2-ctp2.aspx
|
|
#
|
|
myenv.Append(CCFLAGS=["/Zc:inline"])
|
|
|
|
if myenv.ToolchainIs("clang"):
|
|
# We add this flag to make clang emit debug info for c++ stl types so that our pretty
|
|
# printers will work with newer clang's which omit this debug info. This does increase
|
|
# the overall debug info size.
|
|
myenv.AddToCCFLAGSIfSupported("-fno-limit-debug-info")
|
|
|
|
if myenv.ToolchainIs("gcc", "clang"):
|
|
# Pass -gdwarf{32,64} if an explicit value was selected
|
|
# or defaulted. Fail the build if we can't honor the
|
|
# selection.
|
|
if myenv["DWARF_WIDTH"]:
|
|
if myenv.AddToCCFLAGSIfSupported("-gdwarf$DWARF_WIDTH"):
|
|
myenv.AppendUnique(LINKFLAGS=["-gdwarf$DWARF_WIDTH"])
|
|
else:
|
|
myenv.FatalError("Could not enable selected dwarf width")
|
|
|
|
# try to determine the if dwarf64 is viable, and fallback to dwarf32 if not
|
|
elif myenv.CheckCCFLAGSSupported("-gdwarf64"):
|
|
|
|
def CheckForDWARF64Support(context):
|
|
context.Message("Checking that DWARF64 format is viable... ")
|
|
try:
|
|
dwarf_version = int(myenv.get("DWARF_VERSION", 0))
|
|
except ValueError:
|
|
dwarf_version = None
|
|
|
|
if dwarf_version is None or dwarf_version <= 4:
|
|
result = False
|
|
else:
|
|
test_body = """
|
|
#include <iostream>
|
|
#include <cstdlib>
|
|
int main() {
|
|
std::cout << "Hello, World" << std::endl;
|
|
return EXIT_SUCCESS;
|
|
}
|
|
"""
|
|
original_ccflags = context.env.get("CCFLAGS")
|
|
original_linkflags = context.env.get("LINKFLAGS")
|
|
|
|
context.env.Append(CCFLAGS=["-gdwarf64"], LINKFLAGS=["-gdwarf64"])
|
|
|
|
ret = context.TryLink(textwrap.dedent(test_body), ".cpp")
|
|
|
|
context.env["CCFLAGS"] = original_ccflags
|
|
context.env["LINKFLAGS"] = original_linkflags
|
|
|
|
if not ret:
|
|
context.Result("unknown")
|
|
return False
|
|
|
|
regex = re.compile(r"^\s*Length:.*[64|32]-bit\)$", re.MULTILINE)
|
|
p = subprocess.run(
|
|
[context.env["READELF"], "-wi", context.lastTarget.path],
|
|
capture_output=True,
|
|
text=True,
|
|
)
|
|
matches = re.findall(regex, p.stdout)
|
|
address_types = set()
|
|
for match in matches:
|
|
address_types.add(match[-len("(XX-bit)") :])
|
|
result = len(address_types) == 1 and list(address_types)[0] == "(64-bit)"
|
|
|
|
context.Result(result)
|
|
return result
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckForDWARF64Support": CheckForDWARF64Support,
|
|
},
|
|
)
|
|
|
|
if conf.CheckForDWARF64Support():
|
|
myenv["DWARF_WIDTH"] = 64
|
|
myenv.AppendUnique(LINKFLAGS=["-gdwarf64"], CCFLAGS=["-gdwarf64"])
|
|
else:
|
|
myenv["DWARF_WIDTH"] = 32
|
|
myenv.AppendUnique(LINKFLAGS=["-gdwarf32"], CCFLAGS=["-gdwarf32"])
|
|
|
|
conf.Finish()
|
|
|
|
if myenv["DWARF_WIDTH"] == 32 and link_model != "dynamic":
|
|
# This will create an extra section where debug types can be referred from,
|
|
# reducing other section sizes. This helps most with big static links as there
|
|
# will be lots of duplicate debug type info.
|
|
if myenv.AddToCCFLAGSIfSupported("-fdebug-types-section"):
|
|
myenv.AppendUnique(LINKFLAGS=["-fdebug-types-section"])
|
|
|
|
# Turn off debug symbols. Due to g0 disabling any previously added debugging flags,
|
|
# it is easier to append g0 near the end rather than trying to not add all the other
|
|
# debug flags. This should be added after any debug flags.
|
|
if not debug_symbols:
|
|
myenv.AppendUnique(LINKFLAGS=["-g0"], CCFLAGS=["-g0"])
|
|
|
|
# Our build is already parallel.
|
|
if not myenv.AddToLINKFLAGSIfSupported("-Wl,--no-threads"):
|
|
myenv.AddToLINKFLAGSIfSupported("--Wl,--threads=1")
|
|
|
|
# Explicitly enable GNU build id's if the linker supports it.
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,--build-id")
|
|
|
|
# Explicitly use the new gnu hash section if the linker offers
|
|
# it, except on android since older runtimes seem to not
|
|
# support it. For that platform, use 'both'.
|
|
if env.TargetOSIs("android"):
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,--hash-style=both")
|
|
else:
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,--hash-style=gnu")
|
|
|
|
# Try to have the linker tell us about ODR violations. Don't
|
|
# use it when using clang with libstdc++, as libstdc++ was
|
|
# probably built with GCC. That combination appears to cause
|
|
# false positives for the ODR detector. See SERVER-28133 for
|
|
# additional details.
|
|
if has_option("detect-odr-violations"):
|
|
if myenv.ToolchainIs("clang") and usingLibStdCxx:
|
|
env.FatalError(
|
|
"The --detect-odr-violations flag does not work with clang and libstdc++"
|
|
)
|
|
if optBuild != "off":
|
|
env.FatalError(
|
|
"The --detect-odr-violations flag is expected to only be reliable with --opt=off"
|
|
)
|
|
if linker_ld != "gold":
|
|
myenv.FatalError(
|
|
"The --detect-odr-violations flag currently only works with --linker=gold"
|
|
)
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,--detect-odr-violations")
|
|
|
|
# Disallow an executable stack. Also, issue a warning if any files are found that would
|
|
# cause the stack to become executable if the noexecstack flag was not in play, so that we
|
|
# can find them and fix them. We do this here after we check for ld.gold because the
|
|
# --warn-execstack is currently only offered with gold.
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,-z,noexecstack")
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,--warn-execstack")
|
|
|
|
# If possible with the current linker, mark relocations as read-only.
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,-z,relro")
|
|
|
|
if has_option("thin-lto"):
|
|
if not myenv.AddToCCFLAGSIfSupported(
|
|
"-flto=thin"
|
|
) or not myenv.AddToLINKFLAGSIfSupported("-flto=thin"):
|
|
myenv.ConfError("Failed to enable thin LTO")
|
|
|
|
if linker_ld != "gold" and not env.TargetOSIs("darwin", "macOS") and optBuild != "off":
|
|
if has_option("pgo"):
|
|
print("WARNING: skipping symbol ordering as pgo is enabled")
|
|
else:
|
|
myenv.AppendUnique(
|
|
CCFLAGS=["-ffunction-sections"],
|
|
LINKFLAGS=[
|
|
"-Wl,--symbol-ordering-file=symbols.orderfile",
|
|
"-Wl,--no-warn-symbol-ordering",
|
|
],
|
|
)
|
|
else:
|
|
print("WARNING: lld linker is required to sort symbols")
|
|
|
|
if has_option("pgo-profile"):
|
|
if (
|
|
not myenv.ToolchainIs("clang")
|
|
or not myenv.TargetOSIs("linux")
|
|
or linker_ld == "gold"
|
|
):
|
|
myenv.FatalError("Error: pgo only works on linux with clang + lld")
|
|
myenv.AppendUnique(
|
|
CCFLAGS=["-fprofile-instr-generate"],
|
|
LINKFLAGS=["-fprofile-instr-generate"],
|
|
)
|
|
|
|
if has_option("pgo"):
|
|
if (
|
|
not myenv.ToolchainIs("clang")
|
|
or not myenv.TargetOSIs("linux")
|
|
or linker_ld == "gold"
|
|
):
|
|
myenv.FatalError("Error: pgo only works on linux with clang + lld")
|
|
myenv.AppendUnique(
|
|
_NON_CONF_CCFLAGS_GEN=["-fprofile-use=./default.profdata"],
|
|
)
|
|
myenv["CCFLAGS_WERROR"].remove("-Werror")
|
|
|
|
# As far as we know these flags only apply on posix-y systems,
|
|
# and not on Darwin.
|
|
if env.TargetOSIs("posix") and not env.TargetOSIs("darwin"):
|
|
# Disable debug compression in both the assembler and linker
|
|
# by default. If the user requested compression, only allow
|
|
# the zlib-gabi form.
|
|
debug_compress = get_option("debug-compress")
|
|
|
|
# If a value was provided on the command line for --debug-compress, it should
|
|
# inhibit the application of auto, so strip it out.
|
|
if "auto" in debug_compress and len(debug_compress) > 1:
|
|
debug_compress = debug_compress[1:]
|
|
|
|
# Disallow saying --debug-compress=off --debug-compress=ld and similar
|
|
if "off" in debug_compress and len(debug_compress) > 1:
|
|
env.FatalError("Cannot combine 'off' for --debug-compress with other values")
|
|
|
|
# Transform the 'auto' argument into a real value.
|
|
if "auto" in debug_compress:
|
|
debug_compress = []
|
|
|
|
# We only automatically enable ld compression for
|
|
# dynamic builds because it seems to use enormous
|
|
# amounts of memory in static builds.
|
|
if link_model.startswith("dynamic"):
|
|
debug_compress.append("ld")
|
|
|
|
compress_type = "zlib-gabi"
|
|
compress_flag = "compress-debug-sections"
|
|
|
|
myenv.AddToCCFLAGSIfSupported(
|
|
f"-Wa,--{compress_flag}={compress_type}"
|
|
if "as" in debug_compress
|
|
else f"-Wa,--no{compress_flag}",
|
|
)
|
|
|
|
# We shouldn't enable debug compression in the linker
|
|
# (meaning our final binaries contain compressed debug
|
|
# info) unless our local elf environment appears to at
|
|
# least be aware of SHF_COMPRESSED. This seems like a
|
|
# necessary precondition, but is it sufficient?
|
|
#
|
|
# https://gnu.wildebeest.org/blog/mjw/2016/01/13/elf-libelf-compressed-sections-and-elfutils/
|
|
|
|
def CheckElfHForSHF_COMPRESSED(context):
|
|
test_body = """
|
|
#include <elf.h>
|
|
#if !defined(SHF_COMPRESSED)
|
|
#error
|
|
#endif
|
|
"""
|
|
|
|
context.Message("Checking elf.h for SHF_COMPRESSED... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".c")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckElfHForSHF_COMPRESSED": CheckElfHForSHF_COMPRESSED,
|
|
},
|
|
)
|
|
|
|
have_shf_compressed = conf.CheckElfHForSHF_COMPRESSED()
|
|
conf.Finish()
|
|
|
|
if have_shf_compressed and "ld" in debug_compress:
|
|
myenv.AddToLINKFLAGSIfSupported(
|
|
f"-Wl,--{compress_flag}={compress_type}",
|
|
)
|
|
else:
|
|
myenv.AddToLINKFLAGSIfSupported(
|
|
f"-Wl,--{compress_flag}=none",
|
|
)
|
|
|
|
if "fnsi" in selected_experimental_optimizations:
|
|
myenv.AddToCCFLAGSIfSupported("-fno-semantic-interposition")
|
|
|
|
# Avoid deduping symbols on OS X debug builds, as it takes a long time.
|
|
if optBuild == "off" and myenv.ToolchainIs("clang") and env.TargetOSIs("darwin"):
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,-no_deduplicate")
|
|
|
|
# Apply any link time optimization settings as selected by the 'lto' option.
|
|
if has_option("lto"):
|
|
if myenv.ToolchainIs("msvc"):
|
|
# Note that this is actually more aggressive than LTO, it is whole program
|
|
# optimization due to /GL. However, this is historically what we have done for
|
|
# windows, so we are keeping it.
|
|
#
|
|
# /GL implies /LTCG, so no need to say it in CCFLAGS, but we do need /LTCG on the
|
|
# link flags.
|
|
myenv.Append(CCFLAGS=["/GL"])
|
|
myenv.Append(LINKFLAGS=["/LTCG"])
|
|
myenv.Append(ARFLAGS=["/LTCG"])
|
|
elif myenv.ToolchainIs("gcc", "clang"):
|
|
# For GCC and clang, the flag is -flto, and we need to pass it both on the compile
|
|
# and link lines.
|
|
if not myenv.AddToCCFLAGSIfSupported("-flto") or not myenv.AddToLINKFLAGSIfSupported(
|
|
"-flto"
|
|
):
|
|
myenv.ConfError(
|
|
"Link time optimization requested, "
|
|
"but selected compiler does not honor -flto"
|
|
)
|
|
|
|
if myenv.TargetOSIs("darwin"):
|
|
myenv.AddToLINKFLAGSIfSupported("-Wl,-object_path_lto,${TARGET}.lto")
|
|
else:
|
|
# According to intel benchmarks -fno-plt increases perf
|
|
# See PM-2215
|
|
if linker_ld != "gold":
|
|
myenv.ConfError("lto compilation currently only works with the --linker=gold")
|
|
if link_model != "object":
|
|
myenv.ConfError(
|
|
"lto compilation currently only works with the --link-model=object"
|
|
)
|
|
if not myenv.AddToCCFLAGSIfSupported(
|
|
"-fno-plt"
|
|
) or not myenv.AddToLINKFLAGSIfSupported("-fno-plt"):
|
|
myenv.ConfError("-fno-plt is not supported by the compiler")
|
|
|
|
else:
|
|
myenv.ConfError("Don't know how to enable --lto on current toolchain")
|
|
|
|
if get_option("runtime-hardening") == "on" and optBuild != "off":
|
|
# Older glibc doesn't work well with _FORTIFY_SOURCE=2. Selecting 2.11 as the minimum was an
|
|
# emperical decision, as that is the oldest non-broken glibc we seem to require. It is possible
|
|
# that older glibc's work, but we aren't trying.
|
|
#
|
|
# https://gforge.inria.fr/tracker/?func=detail&group_id=131&atid=607&aid=14070
|
|
# https://github.com/jedisct1/libsodium/issues/202
|
|
def CheckForGlibcKnownToSupportFortify(context):
|
|
test_body = """
|
|
#include <features.h>
|
|
#if !__GLIBC_PREREQ(2, 11)
|
|
#error
|
|
#endif
|
|
"""
|
|
context.Message("Checking for glibc with non-broken _FORTIFY_SOURCE...")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".c")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckForFortify": CheckForGlibcKnownToSupportFortify,
|
|
},
|
|
)
|
|
|
|
# Fortify only possibly makes sense on POSIX systems, and we know that clang is not a valid
|
|
# combination:
|
|
#
|
|
# http://lists.llvm.org/pipermail/cfe-dev/2015-November/045852.html
|
|
#
|
|
if env.TargetOSIs("posix") and not env.ToolchainIs("clang") and conf.CheckForFortify():
|
|
conf.env.Append(
|
|
CPPDEFINES=[
|
|
("_FORTIFY_SOURCE", 2),
|
|
],
|
|
)
|
|
|
|
myenv = conf.Finish()
|
|
|
|
# Our build generally assumes that we have C11-compliant libc headers for
|
|
# C++ source. On most systems, that will be the case. However, on systems
|
|
# using glibc older than 2.18 (or other libc implementations that have
|
|
# stubbornly refused to update), we need to add some preprocessor defines.
|
|
#
|
|
# See: https://sourceware.org/bugzilla/show_bug.cgi?id=15366
|
|
#
|
|
# These headers are only fully standards-compliant on POSIX platforms. Windows
|
|
# in particular doesn't implement inttypes.h
|
|
if env.TargetOSIs("posix"):
|
|
|
|
def NeedStdCLimitMacros(context):
|
|
test_body = """
|
|
#undef __STDC_LIMIT_MACROS
|
|
#include <stdint.h>
|
|
#if defined(INT64_MAX)
|
|
# error
|
|
#endif
|
|
"""
|
|
context.Message("Checking whether to define __STDC_LIMIT_MACROS... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
def NeedStdCConstantMacros(context):
|
|
test_body = """
|
|
#undef __STDC_CONSTANT_MACROS
|
|
#include <stdint.h>
|
|
#if defined(INTMAX_C)
|
|
# error
|
|
#endif
|
|
"""
|
|
context.Message("Checking whether to define __STDC_CONSTANT_MACROS... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
def NeedStdCFormatMacros(context):
|
|
test_body = """
|
|
#undef __STDC_FORMAT_MACROS
|
|
#include <inttypes.h>
|
|
#if defined(PRIx64)
|
|
# error
|
|
#endif
|
|
"""
|
|
context.Message("Checking whether to define __STDC_FORMAT_MACROS... ")
|
|
ret = context.TryCompile(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
help=False,
|
|
custom_tests={
|
|
"NeedStdCLimitMacros": NeedStdCLimitMacros,
|
|
"NeedStdCConstantMacros": NeedStdCConstantMacros,
|
|
"NeedStdCFormatMacros": NeedStdCFormatMacros,
|
|
},
|
|
)
|
|
|
|
conf.env.AppendUnique(
|
|
CPPDEFINES=[
|
|
"__STDC_LIMIT_MACROS" if conf.NeedStdCLimitMacros() else "",
|
|
"__STDC_CONSTANT_MACROS" if conf.NeedStdCConstantMacros() else "",
|
|
"__STDC_FORMAT_MACROS" if conf.NeedStdCFormatMacros() else "",
|
|
]
|
|
)
|
|
|
|
myenv = conf.Finish()
|
|
|
|
# We set this with GCC on x86 platforms to work around
|
|
# https://gcc.gnu.org/bugzilla/show_bug.cgi?id=43052
|
|
if myenv.ToolchainIs("gcc") and (env["TARGET_ARCH"] in ["i386", "x86_64"]):
|
|
if "builtin-memcmp" not in selected_experimental_optimizations:
|
|
myenv.AddToCCFLAGSIfSupported("-fno-builtin-memcmp")
|
|
|
|
def CheckBoostMinVersion(context):
|
|
compile_test_body = textwrap.dedent("""
|
|
#include <boost/version.hpp>
|
|
|
|
#if BOOST_VERSION < 104900
|
|
#error
|
|
#endif
|
|
""")
|
|
|
|
context.Message("Checking if system boost version is 1.49 or newer...")
|
|
result = context.TryCompile(compile_test_body, ".cpp")
|
|
context.Result(result)
|
|
return result
|
|
|
|
conf = Configure(
|
|
myenv,
|
|
custom_tests={
|
|
"CheckBoostMinVersion": CheckBoostMinVersion,
|
|
},
|
|
)
|
|
|
|
libdeps.setup_conftests(conf)
|
|
|
|
### --ssl checks
|
|
def checkOpenSSL(conf):
|
|
sslLibName = "ssl"
|
|
cryptoLibName = "crypto"
|
|
sslLinkDependencies = ["crypto", "dl"]
|
|
if conf.env.TargetOSIs("freebsd"):
|
|
sslLinkDependencies = ["crypto"]
|
|
|
|
if conf.env.TargetOSIs("windows"):
|
|
sslLibName = "ssleay32"
|
|
cryptoLibName = "libeay32"
|
|
sslLinkDependencies = ["libeay32"]
|
|
|
|
# Used to import system certificate keychains
|
|
if conf.env.TargetOSIs("darwin"):
|
|
conf.env.AppendUnique(
|
|
FRAMEWORKS=[
|
|
"CoreFoundation",
|
|
"Security",
|
|
]
|
|
)
|
|
|
|
def maybeIssueDarwinSSLAdvice(env):
|
|
if env.TargetOSIs("macOS"):
|
|
advice = textwrap.dedent("""\
|
|
NOTE: Recent versions of macOS no longer ship headers for the system OpenSSL libraries.
|
|
NOTE: Either build without the --ssl flag, or describe how to find OpenSSL.
|
|
NOTE: Set the include path for the OpenSSL headers with the CPPPATH SCons variable.
|
|
NOTE: Set the library path for OpenSSL libraries with the LIBPATH SCons variable.
|
|
NOTE: If you are using HomeBrew, and have installed OpenSSL, this might look like:
|
|
\tscons CPPPATH=/usr/local/opt/openssl/include LIBPATH=/usr/local/opt/openssl/lib ...
|
|
NOTE: Consult the output of 'brew info openssl' for details on the correct paths.""")
|
|
print(advice)
|
|
brew = env.WhereIs("brew")
|
|
if brew:
|
|
try:
|
|
# TODO: If we could programmatically extract the paths from the info output
|
|
# we could give a better message here, but brew info's machine readable output
|
|
# doesn't seem to include the whole 'caveats' section.
|
|
message = subprocess.check_output([brew, "info", "openssl"]).decode("utf-8")
|
|
advice = textwrap.dedent("""\
|
|
NOTE: HomeBrew installed to {0} appears to have OpenSSL installed.
|
|
NOTE: Consult the output from '{0} info openssl' to determine CPPPATH and LIBPATH.""").format(
|
|
brew, message
|
|
)
|
|
|
|
print(advice)
|
|
except:
|
|
pass
|
|
|
|
if not conf.CheckLibWithHeader(
|
|
cryptoLibName,
|
|
["openssl/crypto.h"],
|
|
"C",
|
|
"SSLeay_version(0);",
|
|
autoadd=True,
|
|
):
|
|
maybeIssueDarwinSSLAdvice(conf.env)
|
|
conf.env.ConfError("Couldn't find OpenSSL crypto.h header and library")
|
|
|
|
def CheckLibSSL(context):
|
|
res = SCons.Conftest.CheckLib(
|
|
context,
|
|
libs=[sslLibName],
|
|
extra_libs=sslLinkDependencies,
|
|
header='#include "openssl/ssl.h"',
|
|
language="C",
|
|
call="SSL_version(NULL);",
|
|
autoadd=True,
|
|
)
|
|
context.did_show_result = 1
|
|
return not res
|
|
|
|
conf.AddTest("CheckLibSSL", CheckLibSSL)
|
|
|
|
if not conf.CheckLibSSL():
|
|
maybeIssueDarwinSSLAdvice(conf.env)
|
|
conf.env.ConfError("Couldn't find OpenSSL ssl.h header and library")
|
|
|
|
def CheckLinkSSL(context):
|
|
test_body = """
|
|
#include <openssl/err.h>
|
|
#include <openssl/ssl.h>
|
|
#include <stdlib.h>
|
|
|
|
int main() {
|
|
SSL_library_init();
|
|
SSL_load_error_strings();
|
|
ERR_load_crypto_strings();
|
|
|
|
OpenSSL_add_all_algorithms();
|
|
ERR_free_strings();
|
|
|
|
return EXIT_SUCCESS;
|
|
}
|
|
"""
|
|
context.Message("Checking that linking to OpenSSL works...")
|
|
ret = context.TryLink(textwrap.dedent(test_body), ".c")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf.AddTest("CheckLinkSSL", CheckLinkSSL)
|
|
|
|
if not conf.CheckLinkSSL():
|
|
maybeIssueDarwinSSLAdvice(conf.env)
|
|
conf.env.ConfError("SSL is enabled, but is unavailable")
|
|
|
|
# We require ssl by default unless the user has specified --ssl=off
|
|
require_ssl = get_option("ssl") != "off"
|
|
|
|
# The following platform checks setup both
|
|
# ssl_provider for TLS implementation
|
|
# and MONGO_CRYPTO for hashing support.
|
|
#
|
|
# MONGO_CRYPTO is always enabled regardless of --ssl=on/off
|
|
# However, ssl_provider will be rewritten to 'none' if --ssl=off
|
|
if conf.env.TargetOSIs("windows"):
|
|
# SChannel on Windows
|
|
ssl_provider = "windows"
|
|
conf.env.Append(MONGO_CRYPTO=["windows"])
|
|
|
|
elif conf.env.TargetOSIs("darwin", "macOS"):
|
|
# SecureTransport on macOS
|
|
ssl_provider = "apple"
|
|
conf.env.Append(MONGO_CRYPTO=["apple"])
|
|
conf.env.AppendUnique(FRAMEWORKS=["CoreFoundation", "Security"])
|
|
|
|
elif require_ssl:
|
|
checkOpenSSL(conf)
|
|
# Working OpenSSL available, use it.
|
|
conf.env.Append(MONGO_CRYPTO=["openssl"])
|
|
ssl_provider = "openssl"
|
|
|
|
else:
|
|
# If we don't need an SSL build, we can get by with TomCrypt.
|
|
conf.env.Append(MONGO_CRYPTO=["tom"])
|
|
|
|
if require_ssl:
|
|
# Either crypto engine is native,
|
|
# or it's OpenSSL and has been checked to be working.
|
|
print("Using SSL Provider: {0}".format(ssl_provider))
|
|
else:
|
|
ssl_provider = "none"
|
|
|
|
def checkHTTPLib(required=False):
|
|
# WinHTTP available on Windows
|
|
if env.TargetOSIs("windows"):
|
|
return True
|
|
|
|
# libcurl on all other platforms
|
|
if conf.CheckLibWithHeader(
|
|
"curl",
|
|
["curl/curl.h"],
|
|
"C",
|
|
"curl_global_init(0);",
|
|
autoadd=False,
|
|
):
|
|
return True
|
|
|
|
if required:
|
|
env.ConfError("Could not find <curl/curl.h> and curl lib")
|
|
|
|
return False
|
|
|
|
if use_system_version_of_library("pcre2"):
|
|
conf.FindSysLibDep("pcre2", ["pcre2-8"])
|
|
else:
|
|
conf.env.Prepend(CPPDEFINES=["PCRE2_STATIC"])
|
|
|
|
if use_system_version_of_library("snappy"):
|
|
conf.FindSysLibDep("snappy", ["snappy"])
|
|
|
|
if use_system_version_of_library("zlib"):
|
|
conf.FindSysLibDep("zlib", ["zdll" if conf.env.TargetOSIs("windows") else "z"])
|
|
|
|
if use_system_version_of_library("zstd"):
|
|
conf.FindSysLibDep("zstd", ["libzstd" if conf.env.TargetOSIs("windows") else "zstd"])
|
|
|
|
if use_system_version_of_library("stemmer"):
|
|
conf.FindSysLibDep("stemmer", ["stemmer"])
|
|
|
|
if use_system_version_of_library("yaml"):
|
|
conf.FindSysLibDep("yaml", ["yaml-cpp"])
|
|
|
|
if use_system_version_of_library("fmt"):
|
|
conf.FindSysLibDep("fmt", ["fmt"])
|
|
|
|
if use_system_version_of_library("tomcrypt"):
|
|
conf.FindSysLibDep("tomcrypt", ["tomcrypt"])
|
|
|
|
if use_system_version_of_library("libunwind"):
|
|
conf.FindSysLibDep("unwind", ["unwind"])
|
|
if not conf.FindSysLibDep("lzma", ["lzma"]):
|
|
myenv.ConfError(
|
|
"Cannot find system library 'lzma' required for use with system libunwind"
|
|
)
|
|
|
|
if use_system_version_of_library("intel_decimal128"):
|
|
conf.FindSysLibDep("intel_decimal128", ["bid"])
|
|
|
|
if use_system_version_of_library("icu"):
|
|
conf.FindSysLibDep("icudata", ["icudata"])
|
|
# We can't use FindSysLibDep() for icui18n and icuuc below, since SConf.CheckLib() (which
|
|
# FindSysLibDep() relies on) doesn't expose an 'extra_libs' parameter to indicate that the
|
|
# library being tested has additional dependencies (icuuc depends on icudata, and icui18n
|
|
# depends on both). As a workaround, we skip the configure check for these two libraries and
|
|
# manually assign the library name. We hope that if the user has icudata installed on their
|
|
# system, then they also have icu18n and icuuc installed.
|
|
conf.env["LIBDEPS_ICUI18N_SYSLIBDEP"] = "icui18n"
|
|
conf.env["LIBDEPS_ICUUC_SYSLIBDEP"] = "icuuc"
|
|
|
|
if wiredtiger and use_system_version_of_library("wiredtiger"):
|
|
if not conf.CheckCXXHeader("wiredtiger.h"):
|
|
myenv.ConfError("Cannot find wiredtiger headers")
|
|
conf.FindSysLibDep("wiredtiger", ["wiredtiger"])
|
|
|
|
conf.env.Append(
|
|
CPPDEFINES=[
|
|
"ABSL_FORCE_ALIGNED_ACCESS",
|
|
"BOOST_ENABLE_ASSERT_DEBUG_HANDLER",
|
|
# TODO: Ideally, we could not set this define in C++20
|
|
# builds, but at least our current Xcode 12 doesn't offer
|
|
# std::atomic_ref, so we cannot.
|
|
"BOOST_FILESYSTEM_NO_CXX20_ATOMIC_REF",
|
|
"BOOST_LOG_NO_SHORTHAND_NAMES",
|
|
"BOOST_LOG_USE_NATIVE_SYSLOG",
|
|
"BOOST_LOG_WITHOUT_THREAD_ATTR",
|
|
"BOOST_MATH_NO_LONG_DOUBLE_MATH_FUNCTIONS",
|
|
"BOOST_SYSTEM_NO_DEPRECATED",
|
|
"BOOST_THREAD_USES_DATETIME",
|
|
("BOOST_THREAD_VERSION", "5"),
|
|
]
|
|
)
|
|
|
|
if link_model.startswith("dynamic") and not link_model == "dynamic-sdk":
|
|
conf.env.AppendUnique(
|
|
CPPDEFINES=[
|
|
"BOOST_LOG_DYN_LINK",
|
|
]
|
|
)
|
|
|
|
if use_system_version_of_library("protobuf"):
|
|
conf.FindSysLibDep("protobuf", ["protobuf"])
|
|
conf.FindSysLibDep("protoc", ["protoc"])
|
|
|
|
if use_system_version_of_library("grpc"):
|
|
conf.FindSysLibDep("grpc", ["grpc"])
|
|
conf.FindSysLibDep("grpcxx", ["grpc++"])
|
|
conf.FindSysLibDep("grpcxx_reflection", ["grpc++_reflection"])
|
|
|
|
if posix_system:
|
|
conf.CheckLib("rt")
|
|
conf.CheckLib("dl")
|
|
|
|
conf.env["_HAVEPCAP"] = conf.CheckLib(["pcap", "wpcap"], autoadd=False)
|
|
|
|
if env.TargetOSIs("solaris"):
|
|
conf.CheckLib("nsl")
|
|
|
|
conf.env["MONGO_BUILD_SASL_CLIENT"] = bool(has_option("use-sasl-client"))
|
|
if conf.env["MONGO_BUILD_SASL_CLIENT"] and not conf.CheckLibWithHeader(
|
|
"sasl2",
|
|
["stddef.h", "sasl/sasl.h"],
|
|
"C",
|
|
"sasl_version_info(0, 0, 0, 0, 0, 0);",
|
|
autoadd=False,
|
|
):
|
|
myenv.ConfError("Couldn't find SASL header/libraries")
|
|
|
|
# requires ports devel/libexecinfo to be installed
|
|
if env.TargetOSIs("freebsd", "openbsd"):
|
|
if not conf.CheckLib("execinfo"):
|
|
myenv.ConfError("Cannot find libexecinfo, please install devel/libexecinfo.")
|
|
|
|
# 'tcmalloc' needs to be the last library linked. Please, add new libraries before this
|
|
# point.
|
|
if myenv["MONGO_ALLOCATOR"] == "tcmalloc-google":
|
|
if use_system_version_of_library("tcmalloc-google"):
|
|
conf.FindSysLibDep("tcmalloc-google", ["tcmalloc"])
|
|
elif myenv["MONGO_ALLOCATOR"] == "tcmalloc-gperf":
|
|
if use_system_version_of_library("tcmalloc-gperf"):
|
|
conf.FindSysLibDep("tcmalloc-gperf", ["tcmalloc"])
|
|
elif myenv["MONGO_ALLOCATOR"] in ["system"]:
|
|
pass
|
|
else:
|
|
myenv.FatalError(f"Invalid --allocator parameter: {env['MONGO_ALLOCATOR']}")
|
|
|
|
def CheckStdAtomic(context, base_type, extra_message):
|
|
test_body = """
|
|
#include <atomic>
|
|
|
|
int main(int argc, char* argv[]) {{
|
|
std::atomic<{0}> x;
|
|
|
|
x.store(0);
|
|
// Use argc to ensure we can't optimize everything away.
|
|
{0} y = argc;
|
|
x.fetch_add(y);
|
|
x.fetch_sub(y);
|
|
x.exchange(y);
|
|
if (x.compare_exchange_strong(y, x) && x.is_lock_free())
|
|
return 0;
|
|
return x.load();
|
|
}}
|
|
""".format(base_type)
|
|
|
|
context.Message(
|
|
"Checking if std::atomic<{0}> works{1}... ".format(base_type, extra_message)
|
|
)
|
|
|
|
ret = context.TryLink(textwrap.dedent(test_body), ".cpp")
|
|
context.Result(ret)
|
|
return ret
|
|
|
|
conf.AddTest("CheckStdAtomic", CheckStdAtomic)
|
|
|
|
def check_all_atomics(extra_message=""):
|
|
for t in ("int64_t", "uint64_t", "int32_t", "uint32_t"):
|
|
if not conf.CheckStdAtomic(t, extra_message):
|
|
return False
|
|
return True
|
|
|
|
if not check_all_atomics():
|
|
if not conf.CheckLib("atomic", symbol=None, header=None, language="C", autoadd=1):
|
|
myenv.ConfError(
|
|
"Some atomic ops are not intrinsically supported, but " "no libatomic found"
|
|
)
|
|
if not check_all_atomics(" with libatomic"):
|
|
myenv.ConfError("The toolchain does not support std::atomic, cannot continue")
|
|
|
|
def CheckMongoCMinVersion(context):
|
|
compile_test_body = textwrap.dedent("""
|
|
#include <mongoc/mongoc.h>
|
|
|
|
#if !MONGOC_CHECK_VERSION(1,13,0)
|
|
#error
|
|
#endif
|
|
""")
|
|
|
|
context.Message("Checking if mongoc version is 1.13.0 or newer...")
|
|
result = context.TryCompile(compile_test_body, ".cpp")
|
|
context.Result(result)
|
|
return result
|
|
|
|
conf.AddTest("CheckMongoCMinVersion", CheckMongoCMinVersion)
|
|
|
|
conf.env["MONGO_HAVE_LIBMONGOC"] = False
|
|
|
|
# ask each module to configure itself and the build environment.
|
|
moduleconfig.configure_modules(mongo_modules, conf)
|
|
|
|
# Resolve --enable-http-client
|
|
if http_client == "auto":
|
|
if checkHTTPLib():
|
|
http_client = "on"
|
|
else:
|
|
print("Disabling http-client as libcurl was not found")
|
|
http_client = "off"
|
|
elif http_client == "on":
|
|
checkHTTPLib(required=True)
|
|
|
|
if get_option("xray") == "on":
|
|
if not (myenv.ToolchainIs("clang") and env.TargetOSIs("linux")):
|
|
env.FatalError("LLVM Xray is only supported with clang on linux")
|
|
|
|
myenv.AppendUnique(
|
|
CCFLAGS=[
|
|
"-fxray-instrument",
|
|
"-fxray-instruction-threshold=" + str(get_option("xray-instruction-threshold")),
|
|
],
|
|
LINKFLAGS=["-fxray-instrument"],
|
|
)
|
|
|
|
if "ldap" in myenv.get("MONGO_ENTERPRISE_FEATURES", []):
|
|
if myenv.TargetOSIs("windows"):
|
|
conf.env["MONGO_LDAP_LIB"] = ["Wldap32"]
|
|
else:
|
|
have_ldap_h = conf.CheckLibWithHeader(
|
|
"ldap",
|
|
["ldap.h"],
|
|
"C",
|
|
'ldap_is_ldap_url("ldap://127.0.0.1");',
|
|
autoadd=False,
|
|
)
|
|
|
|
have_lber_h = conf.CheckLibWithHeader(
|
|
"lber",
|
|
["lber.h"],
|
|
"C",
|
|
"ber_free(NULL, 0);",
|
|
autoadd=False,
|
|
)
|
|
|
|
if have_ldap_h:
|
|
conf.env.AppendUnique(MONGO_LDAP_LIB=["ldap"])
|
|
else:
|
|
myenv.ConfError(
|
|
"Could not find <ldap.h> and ldap library from OpenLDAP, "
|
|
"required for LDAP authorization in the enterprise build"
|
|
)
|
|
|
|
if have_lber_h:
|
|
conf.env.AppendUnique(MONGO_LDAP_LIB=["lber"])
|
|
else:
|
|
myenv.ConfError(
|
|
"Could not find <lber.h> and lber library from OpenLDAP, "
|
|
"required for LDAP authorizaton in the enterprise build"
|
|
)
|
|
|
|
if "sasl" in myenv.get("MONGO_ENTERPRISE_FEATURES", []):
|
|
if conf.env.TargetOSIs("windows"):
|
|
conf.env["MONGO_GSSAPI_IMPL"] = "sspi"
|
|
conf.env["MONGO_GSSAPI_LIB"] = ["secur32"]
|
|
else:
|
|
if conf.CheckLib(library="gssapi_krb5", autoadd=False):
|
|
conf.env["MONGO_GSSAPI_IMPL"] = "gssapi"
|
|
if conf.env.TargetOSIs("freebsd"):
|
|
conf.env.AppendUnique(MONGO_GSSAPI_LIB=["gssapi"])
|
|
conf.env.AppendUnique(MONGO_GSSAPI_LIB=["gssapi_krb5"])
|
|
|
|
myenv = conf.Finish()
|
|
|
|
return myenv
|
|
|
|
|
|
env = doConfigure(env)
|
|
saslconf = Configure(env)
|
|
|
|
have_sasl_lib = saslconf.CheckLibWithHeader(
|
|
"sasl2",
|
|
["stddef.h", "sasl/sasl.h"],
|
|
"C",
|
|
"sasl_version_info(0, 0, 0, 0, 0, 0);",
|
|
autoadd=False,
|
|
)
|
|
|
|
saslconf.Finish()
|
|
|
|
env["NINJA_SYNTAX"] = "#site_scons/third_party/ninja_syntax.py"
|
|
|
|
if env.ToolchainIs("clang"):
|
|
env["ICECC_COMPILER_TYPE"] = "clang"
|
|
elif env.ToolchainIs("gcc"):
|
|
env["ICECC_COMPILER_TYPE"] = "gcc"
|
|
|
|
# Now that we are done with configure checks, enable ccache and
|
|
# icecream if requested.
|
|
if "CCACHE" in env and env["CCACHE"]:
|
|
ccache = Tool("ccache")
|
|
if not ccache.exists(env):
|
|
env.FatalError(f"Failed to load ccache tool with CCACHE={env['CCACHE']}")
|
|
ccache(env)
|
|
if "ICECC" in env and env["ICECC"]:
|
|
env["ICECREAM_VERBOSE"] = env.Verbose()
|
|
env["ICECREAM_TARGET_DIR"] = "$BUILD_ROOT/scons/icecream"
|
|
|
|
# Posssibly multiple ninja files are in play, and there are cases where ninja will
|
|
# use the wrong icecc run script, so we must create a unique script per ninja variant
|
|
# for ninja to track separately. We will use the variant dir which contains the each
|
|
# separate ninja builds meta files. This has to be under an additional flag then just
|
|
# ninja disabled, because the run icecc script is generated under a context where ninja
|
|
# is always disabled via the scons callback mechanism. The __NINJA_NO flag is intended
|
|
# to differentiate this particular context.
|
|
if env.get("__NINJA_NO") or get_option("ninja") != "disabled":
|
|
env["ICECREAM_RUN_SCRIPT_SUBPATH"] = "$VARIANT_DIR"
|
|
|
|
icecream = Tool("icecream")
|
|
if not icecream.exists(env):
|
|
# SERVER-70648: Need to revert on how to update icecream
|
|
if "ICECREAM_VERSION" in env and env["ICECREAM_VERSION"] < parse_version("1.3"):
|
|
env.FatalError(
|
|
textwrap.dedent("""\
|
|
Please refer to the following commands to update your icecream:
|
|
sudo add-apt-repository ppa:mongodb-dev-prod/mongodb-build
|
|
sudo apt update
|
|
sudo apt-get --only-upgrade install icecc
|
|
""")
|
|
)
|
|
|
|
env.FatalError(f"Failed to load icecream tool with ICECC={env['ICECC']}")
|
|
icecream(env)
|
|
|
|
# Defaults for SCons provided flags. SetOption only sets the option to our value
|
|
# if the user did not provide it. So for any flag here if it's explicitly passed
|
|
# the values below set with SetOption will be overwritten.
|
|
#
|
|
# Default j to the number of CPUs on the system. Note: in containers this
|
|
# reports the number of CPUs for the host system. Perhaps in a future version of
|
|
# psutil it will instead report the correct number when in a container.
|
|
#
|
|
# The presence of the variable ICECC means the icecream tool is
|
|
# enabled and so the default j value should scale accordingly. In this
|
|
# scenario multiply the cpu count by 8 to set a reasonable default since the
|
|
# cluster can handle many more jobs than your local machine but is
|
|
# still throttled by your cpu count in the sense that you can only
|
|
# handle so many python threads sending out jobs.
|
|
#
|
|
# Capitalize on the weird way SCons handles arguments to determine if
|
|
# the user configured it or not. If not, it is under our control. Try
|
|
# to set some helpful defaults.
|
|
initial_num_jobs = env.GetOption("num_jobs")
|
|
altered_num_jobs = initial_num_jobs + 1
|
|
env.SetOption("num_jobs", altered_num_jobs)
|
|
cpu_count = psutil.cpu_count()
|
|
if env.GetOption("num_jobs") == altered_num_jobs:
|
|
# psutil.cpu_count returns None when it can't determine the
|
|
# number. This always fails on BSD's for example. If the user
|
|
# didn't specify, and we can't determine for a parallel build, it
|
|
# is better to make the user restart and be explicit, rather than
|
|
# give them a very slow build.
|
|
if cpu_count is None:
|
|
if get_option("ninja") != "disabled":
|
|
env.FatalError(
|
|
"Cannot auto-determine the appropriate size for the Ninja local_job pool. Please regenerate with an explicit -j argument to SCons"
|
|
)
|
|
else:
|
|
env.FatalError(
|
|
"Cannot auto-determine the appropriate build parallelism on this platform. Please build with an explicit -j argument to SCons"
|
|
)
|
|
|
|
if "ICECC" in env and env["ICECC"] and get_option("ninja") == "disabled":
|
|
# If SCons is driving and we are using icecream, scale up the
|
|
# number of jobs. The icerun integration will prevent us from
|
|
# overloading the local system.
|
|
env.SetOption("num_jobs", 8 * cpu_count)
|
|
else:
|
|
# Otherwise, either icecream isn't in play, so just use local
|
|
# concurrency for SCons builds, or we are generating for
|
|
# Ninja, in which case num_jobs controls the size of the local
|
|
# pool. Scale that up to the number of local CPUs.
|
|
env.SetOption("num_jobs", cpu_count)
|
|
else:
|
|
if (
|
|
not has_option("force-jobs")
|
|
and ("ICECC" not in env or not env["ICECC"])
|
|
and env.GetOption("num_jobs") > cpu_count
|
|
):
|
|
env.FatalError(
|
|
"ERROR: Icecream not enabled while using -j higher than available cpu's. "
|
|
+ "Use --force-jobs to override."
|
|
)
|
|
|
|
if (
|
|
get_option("ninja") != "disabled"
|
|
and ("ICECC" not in env or not env["ICECC"])
|
|
and not has_option("force-jobs")
|
|
):
|
|
print(
|
|
f"WARNING: Icecream not enabled - Ninja concurrency will be capped at {cpu_count} jobs "
|
|
+ "without regard to the -j value passed to it. "
|
|
+ "Generate your ninja file with --force-jobs to disable this behavior."
|
|
)
|
|
env["NINJA_MAX_JOBS"] = cpu_count
|
|
|
|
if get_option("ninja") != "disabled":
|
|
if env.ToolchainIs("gcc", "clang"):
|
|
env.AppendUnique(CCFLAGS=["-fdiagnostics-color"])
|
|
|
|
ninja_builder = Tool("ninja")
|
|
|
|
env["NINJA_BUILDDIR"] = env.Dir("$NINJA_BUILDDIR")
|
|
ninja_builder.generate(env)
|
|
|
|
ninjaConf = Configure(
|
|
env,
|
|
help=False,
|
|
custom_tests={
|
|
"CheckNinjaCompdbExpand": env.CheckNinjaCompdbExpand,
|
|
},
|
|
)
|
|
env["NINJA_COMPDB_EXPAND"] = ninjaConf.CheckNinjaCompdbExpand()
|
|
ninjaConf.Finish()
|
|
|
|
# TODO: API for getting the sconscripts programmatically
|
|
# exists upstream: https://github.com/SCons/scons/issues/3625
|
|
def ninja_generate_deps(env, target, source, for_signature):
|
|
# TODO SERVER-72851 add api for vars files to exec other vars files
|
|
# this would allow us to get rid of this regex here
|
|
def find_nested_variable_files(variables_file):
|
|
variable_files = [variables_file]
|
|
|
|
with open(variables_file, "r") as file:
|
|
data = file.read()
|
|
pattern = "exec\\(open\\(['\"](.*)['\"]\, ['\"][a-z]+['\"]\\).read\\(\\)\\)"
|
|
nested_files = re.findall(pattern, data)
|
|
for file_name in nested_files:
|
|
variable_files.extend(find_nested_variable_files(file_name))
|
|
|
|
return variable_files
|
|
|
|
# vars files can be from outside of the repo dir and can exec other vars files
|
|
# so we cannot just glob them
|
|
variables_files = []
|
|
for variable_file in variables_files_args:
|
|
variables_files.extend(find_nested_variable_files(variable_file))
|
|
|
|
dependencies = env.Flatten(
|
|
[
|
|
"SConstruct",
|
|
"WORKSPACE.bazel",
|
|
"BUILD.bazel",
|
|
".bazelrc",
|
|
".bazelignore",
|
|
glob(os.path.join("src", "**", "SConscript"), recursive=True),
|
|
glob(os.path.join("src", "**", "BUILD.bazel"), recursive=True),
|
|
glob(os.path.join("buildscripts", "**", "*.py"), recursive=True),
|
|
glob(os.path.join("bazel", "**", "*.bzl"), recursive=True),
|
|
glob(os.path.join(os.path.expanduser("~/.scons/"), "**", "*.py"), recursive=True),
|
|
glob(os.path.join("site_scons", "**", "*.py"), recursive=True),
|
|
glob(os.path.join("buildscripts", "**", "*.py"), recursive=True),
|
|
glob(os.path.join("src/third_party/scons-*", "**", "*.py"), recursive=True),
|
|
glob(os.path.join("src/mongo/db/modules", "**", "*.py"), recursive=True),
|
|
variables_files,
|
|
]
|
|
)
|
|
|
|
return dependencies
|
|
|
|
env["NINJA_REGENERATE_DEPS"] = ninja_generate_deps
|
|
|
|
if env.TargetOSIs("windows"):
|
|
# The /b option here will make sure that windows updates the mtime
|
|
# when copying the file. This allows to not need to use restat for windows
|
|
# copy commands.
|
|
copy_install_cmd = "cmd.exe /c copy /b $in $out 1>NUL"
|
|
else:
|
|
copy_install_cmd = "install $in $out"
|
|
|
|
if env.GetOption("install-action") == "hardlink":
|
|
if env.TargetOSIs("windows"):
|
|
install_cmd = f"cmd.exe /c mklink /h $out $in 1>nul || {copy_install_cmd}"
|
|
else:
|
|
install_cmd = f"ln $in $out || {copy_install_cmd}"
|
|
|
|
elif env.GetOption("install-action") == "symlink":
|
|
# macOS's ln and Windows mklink command do not support relpaths
|
|
# out of the box so we will precompute during generation in a
|
|
# custom handler.
|
|
def symlink_install_action_function(_env, node):
|
|
# should only be one output and input for this case
|
|
output_file = _env.NinjaGetOutputs(node)[0]
|
|
input_file = _env.NinjaGetDependencies(node)[0]
|
|
try:
|
|
relpath = os.path.relpath(input_file, os.path.dirname(output_file))
|
|
except ValueError:
|
|
relpath = os.path.abspath(input_file)
|
|
|
|
return {
|
|
"outputs": [output_file],
|
|
"rule": "INSTALL",
|
|
"inputs": [input_file],
|
|
"implicit": _env.NinjaGetDependencies(node),
|
|
"variables": {"precious": node.precious, "relpath": relpath},
|
|
}
|
|
|
|
env.NinjaRegisterFunctionHandler("installFunc", symlink_install_action_function)
|
|
|
|
if env.TargetOSIs("windows"):
|
|
install_cmd = "cmd.exe /c mklink $out $relpath 1>nul"
|
|
else:
|
|
install_cmd = "ln -s $relpath $out"
|
|
|
|
else:
|
|
install_cmd = copy_install_cmd
|
|
|
|
env.NinjaRule("INSTALL", install_cmd, description="Installed $out", pool="install_pool")
|
|
|
|
if env.TargetOSIs("windows"):
|
|
# This is a workaround on windows for SERVER-48691 where the line length
|
|
# in response files is too long:
|
|
# https://developercommunity.visualstudio.com/content/problem/441978/fatal-error-lnk1170-line-in-command-file-contains.html
|
|
#
|
|
# Ninja currently does not support
|
|
# storing a newline in the ninja file, and therefore you can not
|
|
# easily generate it to the response files. The only documented
|
|
# way to get newlines into the response file is to use the $in_newline
|
|
# variable in the rule.
|
|
#
|
|
# This workaround will move most of the object or lib links into the
|
|
# inputs and then make the respone file consist of the inputs plus
|
|
# whatever options are left in the original response content
|
|
# more info can be found here:
|
|
# https://github.com/ninja-build/ninja/pull/1223/files/e71bcceefb942f8355aab83ab447d702354ba272#r179526824
|
|
# https://github.com/ninja-build/ninja/issues/1000
|
|
|
|
# we are making a new special rule which will leverage
|
|
# the $in_newline to get newlines into our response file
|
|
env.NinjaRule(
|
|
"WINLINK",
|
|
"$env$WINLINK @$out.rsp",
|
|
description="Linked $out",
|
|
deps=None,
|
|
pool="link_pool",
|
|
use_depfile=False,
|
|
use_response_file=True,
|
|
response_file_content="$rspc $in_newline",
|
|
)
|
|
|
|
# Setup the response file content generation to use our workaround rule
|
|
# for LINK commands.
|
|
provider = env.NinjaGenResponseFileProvider(
|
|
"WINLINK",
|
|
"$LINK",
|
|
)
|
|
env.NinjaRuleMapping("${LINKCOM}", provider)
|
|
env.NinjaRuleMapping(env["LINKCOM"], provider)
|
|
|
|
# The workaround function will move some of the content from the rspc
|
|
# variable into the nodes inputs. We only want to move build nodes because
|
|
# inputs must be files, so we make sure the the option in the rspc
|
|
# file starts with the build directory.
|
|
def winlink_workaround(env, node, ninja_build):
|
|
if ninja_build and "rspc" in ninja_build["variables"]:
|
|
rsp_content = []
|
|
inputs = []
|
|
for opt in ninja_build["variables"]["rspc"].split():
|
|
# if its a candidate to go in the inputs add it, else keep it in the non-newline
|
|
# rsp_content list
|
|
if opt.startswith(str(env.Dir("$BUILD_DIR"))) and opt != str(node):
|
|
inputs.append(opt)
|
|
else:
|
|
rsp_content.append(opt)
|
|
|
|
ninja_build["variables"]["rspc"] = " ".join(rsp_content)
|
|
ninja_build["inputs"] += [
|
|
infile for infile in inputs if infile not in ninja_build["inputs"]
|
|
]
|
|
|
|
# We apply the workaround to all Program nodes as they have potential
|
|
# response files that have lines that are too long.
|
|
# This will setup a callback function for a node
|
|
# so that when its been processed, we can make some final adjustments before
|
|
# its generated to the ninja file.
|
|
def winlink_workaround_emitter(target, source, env):
|
|
env.NinjaSetBuildNodeCallback(target[0], winlink_workaround)
|
|
return target, source
|
|
|
|
builder = env["BUILDERS"]["Program"]
|
|
base_emitter = builder.emitter
|
|
new_emitter = SCons.Builder.ListEmitter([base_emitter, winlink_workaround_emitter])
|
|
builder.emitter = new_emitter
|
|
|
|
# We can create empty files for FAKELIB in Ninja because it
|
|
# does not care about content signatures. We have to
|
|
# write_uuid_to_file for FAKELIB in SCons because SCons does.
|
|
env.NinjaRule(
|
|
rule="FAKELIB",
|
|
command="cmd /c copy 1>NUL NUL $out" if env["PLATFORM"] == "win32" else "touch $out",
|
|
)
|
|
|
|
def fakelib_in_ninja(env, node):
|
|
"""Generates empty .a files"""
|
|
return {
|
|
"outputs": [node.get_path()],
|
|
"rule": "FAKELIB",
|
|
"implicit": [str(s) for s in node.sources],
|
|
}
|
|
|
|
env.NinjaRegisterFunctionHandler("write_uuid_to_file", fakelib_in_ninja)
|
|
|
|
def ninja_test_list_builder(env, node):
|
|
test_files = [test_file.path for test_file in env["MONGO_TEST_REGISTRY"][node.path]]
|
|
files = " ".join(test_files)
|
|
return {
|
|
"outputs": [node.get_path()],
|
|
"rule": "TEST_LIST",
|
|
"implicit": test_files,
|
|
"variables": {
|
|
"files": files,
|
|
},
|
|
}
|
|
|
|
if env["PLATFORM"] == "win32":
|
|
cmd = 'cmd.exe /c del "$out" && for %a in ($files) do (echo %a >> "$out")'
|
|
else:
|
|
cmd = 'rm -f "$out"; for i in $files; do echo "$$i" >> "$out"; done;'
|
|
|
|
env.NinjaRule(
|
|
rule="TEST_LIST",
|
|
description="Compiled test list: $out",
|
|
command=cmd,
|
|
)
|
|
env.NinjaRegisterFunctionHandler("test_list_builder_action", ninja_test_list_builder)
|
|
|
|
env["NINJA_GENERATED_SOURCE_ALIAS_NAME"] = "generated-sources"
|
|
|
|
gdb_index_enabled = env.get("GDB_INDEX")
|
|
if gdb_index_enabled == "auto" and link_model == "dynamic":
|
|
gdb_index_enabled = True
|
|
|
|
if gdb_index_enabled is True:
|
|
gdb_index = Tool("gdb_index")
|
|
if gdb_index.exists(env):
|
|
gdb_index.generate(env)
|
|
elif env.get("GDB_INDEX") != "auto":
|
|
env.FatalError("Could not enable explicit request for gdb index generation.")
|
|
|
|
if env.get("ENABLE_GRPC_BUILD"):
|
|
env.Tool("protobuf_compiler")
|
|
|
|
if (get_option("separate-debug") == "on" or env.TargetOSIs("windows")) and debug_symbols:
|
|
separate_debug = Tool("separate_debug")
|
|
if not separate_debug.exists(env):
|
|
env.FatalError(
|
|
"Cannot honor --separate-debug because the separate_debug.py Tool reported as nonexistent"
|
|
)
|
|
separate_debug(env)
|
|
|
|
if env["SPLIT_DWARF"] == "auto":
|
|
# For static builds, splitting out the dwarf info reduces memory requirments, link time
|
|
# and binary size significantly. It's affect is less prominent in dynamic builds. The downside
|
|
# is .dwo files use absolute paths in the debug info, so it's not relocatable.
|
|
# We also found the running splitdwarf with dwarf5 failed to compile
|
|
# so unless we set DWARF_VERSION = 4 we are going to turn off split dwarf
|
|
env["SPLIT_DWARF"] = (
|
|
not link_model == "dynamic"
|
|
and env.ToolchainIs("gcc", "clang")
|
|
and not env.TargetOSIs("darwin")
|
|
and env.CheckCCFLAGSSupported("-gsplit-dwarf")
|
|
and env.get("DWARF_VERSION") == 4
|
|
and not gdb_index_enabled
|
|
)
|
|
|
|
if env["SPLIT_DWARF"]:
|
|
if env.TargetOSIs("darwin"):
|
|
env.FatalError("Setting SPLIT_DWARF=1 on darwin is not supported.")
|
|
if env.get("DWARF_VERSION") != 4:
|
|
env.FatalError(
|
|
"Running split dwarf outside of DWARF4 has shown compilation issues when using DWARF5 and gdb index. Disabling this functionality for now. Use SPLIT_DWARF=0 to disable building with split dwarf or use DWARF_VERSION=4 to pin to DWARF version 4."
|
|
)
|
|
if gdb_index_enabled:
|
|
env.FatalError(
|
|
"SPLIT_DWARF is not supported when using GDB_INDEX. Use GDB_INDEX=0 to allow enabling SPLIT_DWARF"
|
|
)
|
|
env.Tool("split_dwarf")
|
|
|
|
env["AUTO_ARCHIVE_TARBALL_SUFFIX"] = "tgz"
|
|
|
|
env["AIB_META_COMPONENT"] = "all"
|
|
env["AIB_BASE_COMPONENT"] = "common"
|
|
env["AIB_DEFAULT_COMPONENT"] = "mongodb"
|
|
|
|
env.Tool("auto_install_binaries")
|
|
env.Tool("auto_archive")
|
|
|
|
env.DeclareRoles(
|
|
roles=[
|
|
env.Role(
|
|
name="base",
|
|
),
|
|
env.Role(
|
|
name="debug",
|
|
),
|
|
env.Role(
|
|
name="dev",
|
|
dependencies=[
|
|
"runtime",
|
|
],
|
|
),
|
|
env.Role(
|
|
name="meta",
|
|
),
|
|
env.Role(
|
|
name="runtime",
|
|
dependencies=[
|
|
# On windows, we want the runtime role to depend
|
|
# on the debug role so that PDBs end in the
|
|
# runtime package.
|
|
"debug" if env.TargetOSIs("windows") else None,
|
|
],
|
|
silent=True,
|
|
),
|
|
],
|
|
base_role="base",
|
|
meta_role="meta",
|
|
)
|
|
|
|
|
|
def _aib_debugdir(source, target, env, for_signature):
|
|
for s in source:
|
|
origin = getattr(s.attributes, "debug_file_for", None)
|
|
oentry = env.Entry(origin)
|
|
osuf = oentry.get_suffix()
|
|
map_entry = env["AIB_SUFFIX_MAP"].get(osuf)
|
|
if map_entry:
|
|
return map_entry[0]
|
|
env.FatalError(
|
|
"Unable to find debuginfo file in _aib_debugdir: (source='{}')".format(str(source))
|
|
)
|
|
|
|
|
|
if debug_symbols:
|
|
env["PREFIX_DEBUGDIR"] = _aib_debugdir
|
|
|
|
env.AddSuffixMapping(
|
|
{
|
|
"$PROGSUFFIX": env.SuffixMap(
|
|
directory="$PREFIX_BINDIR",
|
|
default_role="runtime",
|
|
),
|
|
"$SHLIBSUFFIX": env.SuffixMap(
|
|
directory="$PREFIX_BINDIR"
|
|
if mongo_platform.get_running_os_name() == "windows"
|
|
else "$PREFIX_LIBDIR",
|
|
default_role="runtime",
|
|
),
|
|
".debug": env.SuffixMap(
|
|
directory="$PREFIX_DEBUGDIR",
|
|
default_role="debug",
|
|
),
|
|
".dSYM": env.SuffixMap(
|
|
directory="$PREFIX_DEBUGDIR",
|
|
default_role="debug",
|
|
),
|
|
".pdb": env.SuffixMap(
|
|
directory="$PREFIX_DEBUGDIR",
|
|
default_role="debug",
|
|
),
|
|
}
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="dist",
|
|
role="runtime",
|
|
name="mongodb-dist",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="dist",
|
|
role="debug",
|
|
name="mongodb-dist-debugsymbols",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="dist-test",
|
|
role="runtime",
|
|
name="mongodb-binaries",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="dist-test",
|
|
role="debug",
|
|
name="mongo-debugsymbols",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="dbtest",
|
|
role="runtime",
|
|
name="dbtest-binary",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="dbtest",
|
|
role="debug",
|
|
name="dbtest-debugsymbols",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="jstestshell",
|
|
role="runtime",
|
|
name="mongodb-jstestshell",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="jstestshell",
|
|
role="debug",
|
|
name="mongodb-jstestshell-debugsymbols",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="mongocryptd",
|
|
role="runtime",
|
|
name="mongodb-cryptd",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="mongocryptd",
|
|
role="debug",
|
|
name="mongodb-cryptd-debugsymbols",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="benchmarks",
|
|
role="runtime",
|
|
name="benchmarks",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="benchmarks",
|
|
role="debug",
|
|
name="benchmarks-debugsymbols",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="mh",
|
|
role="runtime",
|
|
# TODO: we should be able to move this to where the mqlrun binary is
|
|
# defined when AIB correctly uses environments instead of hooking into
|
|
# the first environment used.
|
|
name="mh-binaries",
|
|
)
|
|
|
|
env.AddPackageNameAlias(
|
|
component="mh",
|
|
role="debug",
|
|
# TODO: we should be able to move this to where the mqlrun binary is
|
|
# defined when AIB correctly uses environments instead of hooking into
|
|
# the first environment used.
|
|
name="mh-debugsymbols",
|
|
)
|
|
|
|
env.AutoInstall(
|
|
target="$PREFIX",
|
|
source="$PRETTY_PRINTER_TEST_LIST",
|
|
AIB_ROLE="runtime",
|
|
AIB_COMPONENT="pretty-printer-tests",
|
|
AIB_COMPONENTS_EXTRA=["dist-test"],
|
|
)
|
|
|
|
env.AutoInstall(
|
|
".",
|
|
"$BENCHMARK_LIST",
|
|
AIB_COMPONENT="benchmarks",
|
|
AIB_ROLE="runtime",
|
|
)
|
|
|
|
if "SANITIZER_RUNTIME_LIBS" in env:
|
|
env.AutoInstall(
|
|
target="$PREFIX_LIBDIR",
|
|
source=[env.File(path) for path in env["SANITIZER_RUNTIME_LIBS"]],
|
|
AIB_COMPONENT="dist",
|
|
AIB_ROLE="runtime",
|
|
AIB_COMPONENTS_EXTRA=["dist-test"],
|
|
)
|
|
|
|
env["RPATH_ESCAPED_DOLLAR_ORIGIN"] = "\\$$$$ORIGIN"
|
|
|
|
|
|
def isSupportedStreamsPlatform(thisEnv):
|
|
# TODO https://jira.mongodb.org/browse/SERVER-74961: Support other platforms.
|
|
# linux x86 and ARM64 are supported.
|
|
return (
|
|
thisEnv.TargetOSIs("linux")
|
|
and thisEnv["TARGET_ARCH"] in ("x86_64", "aarch64")
|
|
and ssl_provider == "openssl"
|
|
)
|
|
|
|
|
|
def shouldBuildStreams(thisEnv):
|
|
if releaseBuild:
|
|
# The streaming enterprise module and dependencies are only included in release builds.
|
|
# when streams-release-build is set.
|
|
return get_option("streams-release-build") and isSupportedStreamsPlatform(thisEnv)
|
|
else:
|
|
return isSupportedStreamsPlatform(thisEnv)
|
|
|
|
|
|
env.AddMethod(shouldBuildStreams, "ShouldBuildStreams")
|
|
|
|
|
|
def prefix_libdir_rpath_generator(env, source, target, for_signature):
|
|
# If the PREFIX_LIBDIR has an absolute path, we will use that directly as
|
|
# RPATH because that indicates the final install destination of the libraries.
|
|
prefix_libdir = env.subst("$PREFIX_LIBDIR")
|
|
if os.path.isabs(prefix_libdir):
|
|
return ["$PREFIX_LIBDIR"]
|
|
|
|
# If the PREFIX_LIBDIR is not an absolute path, we will use a relative path
|
|
# from the bin to the lib dir.
|
|
lib_rel = os.path.relpath(prefix_libdir, env.subst("$PREFIX_BINDIR"))
|
|
|
|
if env["PLATFORM"] == "posix":
|
|
return f"$RPATH_ESCAPED_DOLLAR_ORIGIN/{lib_rel}"
|
|
|
|
if env["PLATFORM"] == "darwin":
|
|
return f"@loader_path/{lib_rel}"
|
|
|
|
|
|
if get_option("link-model").startswith("dynamic"):
|
|
env["PREFIX_LIBDIR_RPATH_GENERATOR"] = prefix_libdir_rpath_generator
|
|
|
|
if env["PLATFORM"] == "posix":
|
|
env.AppendUnique(
|
|
RPATH=["$PREFIX_LIBDIR_RPATH_GENERATOR"],
|
|
LINKFLAGS=[
|
|
# Most systems *require* -z,origin to make origin work, but android
|
|
# blows up at runtime if it finds DF_ORIGIN_1 in DT_FLAGS_1.
|
|
# https://android.googlesource.com/platform/bionic/+/cbc80ba9d839675a0c4891e2ab33f39ba51b04b2/linker/linker.h#68
|
|
# https://android.googlesource.com/platform/bionic/+/cbc80ba9d839675a0c4891e2ab33f39ba51b04b2/libc/include/elf.h#215
|
|
"-Wl,-z,origin" if not env.TargetOSIs("android") else [],
|
|
"-Wl,--enable-new-dtags",
|
|
],
|
|
SHLINKFLAGS=[
|
|
# -h works for both the sun linker and the gnu linker.
|
|
"-Wl,-h,${TARGET.file}",
|
|
],
|
|
)
|
|
elif env["PLATFORM"] == "darwin":
|
|
# The darwin case uses an adhoc implementation of RPATH for SCons
|
|
# since SCons does not support RPATH directly for macOS:
|
|
# https://github.com/SCons/scons/issues/2127
|
|
# so we setup RPATH and LINKFLAGS ourselves.
|
|
env["RPATHPREFIX"] = "-Wl,-rpath,"
|
|
env["RPATHSUFFIX"] = ""
|
|
env.AppendUnique(
|
|
LINKFLAGS="${_concat(RPATHPREFIX, RPATH, RPATHSUFFIX, __env__)}",
|
|
SHLINKFLAGS=[
|
|
"-Wl,-install_name,@rpath/${TARGET.file}",
|
|
],
|
|
RPATH=["$PREFIX_LIBDIR_RPATH_GENERATOR"],
|
|
)
|
|
|
|
env.Default(env.Alias("install-default"))
|
|
|
|
# Load the compilation_db tool. We want to do this after configure so we don't end up with
|
|
# compilation database entries for the configure tests, which is weird.
|
|
# We also set a few tools we know will not work with compilationdb, these
|
|
# wrapper tools get appended on the front of the command and in most
|
|
# cases don't want that in the compilation database.
|
|
env["_COMPILATIONDB_IGNORE_WRAPPERS"] = env.get("COMPILATIONDB_IGNORE_WRAPPERS", "").split(",")
|
|
if get_option("ninja") == "disabled":
|
|
env.Tool("compilation_db")
|
|
|
|
incremental_link = Tool("incremental_link")
|
|
if incremental_link.exists(env):
|
|
incremental_link(env)
|
|
|
|
|
|
# Resource Files are Windows specific
|
|
def env_windows_resource_file(env, path):
|
|
if env.TargetOSIs("windows"):
|
|
return [env.RES(path)]
|
|
else:
|
|
return []
|
|
|
|
|
|
env.AddMethod(env_windows_resource_file, "WindowsResourceFile")
|
|
|
|
# --- lint ----
|
|
|
|
if get_option("lint-scope") == "changed":
|
|
patch_file = env.Command(
|
|
target="$BUILD_DIR/current.git.patch",
|
|
source=[env.WhereIs("git")],
|
|
action="${SOURCES[0]} diff $GITDIFFFLAGS > $TARGET",
|
|
)
|
|
|
|
env.AlwaysBuild(patch_file)
|
|
|
|
clang_format = env.Command(
|
|
target="#lint-clang-format",
|
|
source=[
|
|
"buildscripts/clang_format.py",
|
|
patch_file,
|
|
],
|
|
action="REVISION=$REVISION $PYTHON ${SOURCES[0]} lint-git-diff",
|
|
)
|
|
|
|
eslint = env.Command(
|
|
target="#lint-eslint",
|
|
source=[
|
|
"buildscripts/eslint.py",
|
|
patch_file,
|
|
],
|
|
action="REVISION=$REVISION $PYTHON ${SOURCES[0]} lint-git-diff",
|
|
)
|
|
|
|
else:
|
|
clang_format = env.Command(
|
|
target="#lint-clang-format",
|
|
source=[
|
|
"buildscripts/clang_format.py",
|
|
],
|
|
action="$PYTHON ${SOURCES[0]} lint-all",
|
|
)
|
|
|
|
eslint = env.Command(
|
|
target="#lint-eslint",
|
|
source=[
|
|
"buildscripts/eslint.py",
|
|
],
|
|
action="$PYTHON ${SOURCES[0]} --dirmode lint jstests/ src/mongo",
|
|
)
|
|
|
|
pylinters = env.Command(
|
|
target="#lint-pylinters",
|
|
source=[
|
|
"buildscripts/pylinters.py",
|
|
],
|
|
action="$PYTHON ${SOURCES[0]} lint",
|
|
)
|
|
|
|
quickmongolint = env.Command(
|
|
target="#lint-lint.py",
|
|
source=["buildscripts/quickmongolint.py"],
|
|
action="$PYTHON ${SOURCES[0]} lint",
|
|
)
|
|
|
|
lint_errorcodes = env.Command(
|
|
target="#lint-errorcodes",
|
|
source=["buildscripts/errorcodes.py"],
|
|
action="$PYTHON ${SOURCES[0]} --quiet",
|
|
)
|
|
|
|
env.Alias("lint", [quickmongolint, eslint, clang_format, pylinters, lint_errorcodes])
|
|
env.Alias("lint-fast", [eslint, clang_format, pylinters, lint_errorcodes])
|
|
env.AlwaysBuild("lint")
|
|
env.AlwaysBuild("lint-fast")
|
|
|
|
# ---- INSTALL -------
|
|
|
|
|
|
def getSystemInstallName():
|
|
arch_name = env.subst("$MONGO_DISTARCH")
|
|
|
|
# We need to make sure the directory names inside dist tarballs are permanently
|
|
# consistent, even if the target OS name used in scons is different. Any differences
|
|
# between the names used by env.TargetOSIs/env.GetTargetOSName should be added
|
|
# to the translation dictionary below.
|
|
os_name_translations = {
|
|
"windows": "win32",
|
|
"macOS": "macos",
|
|
}
|
|
os_name = env.GetTargetOSName()
|
|
os_name = os_name_translations.get(os_name, os_name)
|
|
n = os_name + "-" + arch_name
|
|
|
|
if len(mongo_modules):
|
|
n += "-" + "-".join(m.name for m in mongo_modules)
|
|
|
|
dn = env.subst("$MONGO_DISTMOD")
|
|
if len(dn) > 0:
|
|
n = n + "-" + dn
|
|
|
|
return n
|
|
|
|
|
|
# This function will add the version.txt file to the source tarball
|
|
# so that versioning will work without having the git repo available.
|
|
def add_version_to_distsrc(env, archive):
|
|
version_file_path = env.subst("$MONGO_DIST_SRC_PREFIX") + "version.json"
|
|
if version_file_path not in archive:
|
|
version_data = {
|
|
"version": env["MONGO_VERSION"],
|
|
"githash": env["MONGO_GIT_HASH"],
|
|
}
|
|
archive.append_file_contents(
|
|
version_file_path,
|
|
json.dumps(
|
|
version_data,
|
|
sort_keys=True,
|
|
indent=4,
|
|
separators=(",", ": "),
|
|
),
|
|
)
|
|
|
|
|
|
env.AddDistSrcCallback(add_version_to_distsrc)
|
|
|
|
env["SERVER_DIST_BASENAME"] = env.subst("mongodb-%s-$MONGO_DISTNAME" % (getSystemInstallName()))
|
|
env["MH_DIST_BASENAME"] = "mh"
|
|
if get_option("legacy-tarball") == "true":
|
|
if (
|
|
"tar-dist" not in COMMAND_LINE_TARGETS
|
|
and "zip-dist" not in COMMAND_LINE_TARGETS
|
|
and "archive-dist" not in COMMAND_LINE_TARGETS
|
|
):
|
|
env.FatalError("option --legacy-tarball only valid with an archive-dist target")
|
|
env["PREFIX"] = "$SERVER_DIST_BASENAME"
|
|
|
|
module_sconscripts = moduleconfig.get_module_sconscripts(mongo_modules)
|
|
|
|
# This generates a numeric representation of the version string so that
|
|
# you can easily compare versions of MongoDB without having to parse
|
|
# the version string.
|
|
#
|
|
# Examples:
|
|
# 5.1.1-123 => ['5', '1', '1', '123', None, None] => [5, 1, 2, -100]
|
|
# 5.1.1-rc2 => ['5', '1', '1', 'rc2', 'rc', '2'] => [5, 1, 1, -23]
|
|
# 5.1.1-rc2-123 => ['5', '1', '1', 'rc2-123', 'rc', '2'] => [5, 1, 1, -23]
|
|
# 5.1.0-alpha-123 => ['5', '1', '0', 'alpha-123', 'alpha', ''] => [5, 1, 0, -50]
|
|
# 5.1.0-alpha1-123 => ['5', '1', '0', 'alpha1-123', 'alpha', '1'] => [5, 1, 0, -49]
|
|
# 5.1.1 => ['5', '1', '1', '', None, None] => [5, 1, 1, 0]
|
|
|
|
version_parts = [
|
|
x
|
|
for x in re.match(
|
|
r"^(\d+)\.(\d+)\.(\d+)-?((?:(rc|alpha)(\d?))?.*)?", env["MONGO_VERSION"]
|
|
).groups()
|
|
]
|
|
version_extra = version_parts[3] if version_parts[3] else ""
|
|
if version_parts[4] == "rc":
|
|
version_parts[3] = int(version_parts[5]) + -25
|
|
elif version_parts[4] == "alpha":
|
|
if version_parts[5] == "":
|
|
version_parts[3] = -50
|
|
else:
|
|
version_parts[3] = int(version_parts[5]) + -50
|
|
elif version_parts[3]:
|
|
version_parts[2] = int(version_parts[2]) + 1
|
|
version_parts[3] = -100
|
|
else:
|
|
version_parts[3] = 0
|
|
version_parts = [int(x) for x in version_parts[:4]]
|
|
|
|
# The following symbols are exported for use in subordinate SConscript files.
|
|
# Ideally, the SConscript files would be purely declarative. They would only
|
|
# import build environment objects, and would contain few or no conditional
|
|
# statements or branches.
|
|
#
|
|
# Currently, however, the SConscript files do need some predicates for
|
|
# conditional decision making that hasn't been moved up to this SConstruct file,
|
|
# and they are exported here, as well.
|
|
Export(
|
|
[
|
|
"debugBuild",
|
|
"endian",
|
|
"get_option",
|
|
"have_sasl_lib",
|
|
"has_option",
|
|
"http_client",
|
|
"jsEngine",
|
|
"module_sconscripts",
|
|
"optBuild",
|
|
"releaseBuild",
|
|
"selected_experimental_optimizations",
|
|
"serverJs",
|
|
"ssl_provider",
|
|
"use_libunwind",
|
|
"use_system_libunwind",
|
|
"use_system_version_of_library",
|
|
"use_vendored_libunwind",
|
|
"version_extra",
|
|
"version_parts",
|
|
"wiredtiger",
|
|
]
|
|
)
|
|
|
|
|
|
def injectModule(env, module, **kwargs):
|
|
injector = env["MODULE_INJECTORS"].get(module)
|
|
if injector:
|
|
return injector(env, **kwargs)
|
|
return env
|
|
|
|
|
|
env.AddMethod(injectModule, "InjectModule")
|
|
|
|
if get_option("ninja") == "disabled":
|
|
compileCommands = env.CompilationDatabase("compile_commands.json")
|
|
# Initialize generated-sources Alias as a placeholder so that it can be used as a
|
|
# dependency for compileCommands. This Alias will be properly updated in other SConscripts.
|
|
env.Depends(compileCommands, env.Alias("generated-sources"))
|
|
compileDb = env.Alias("compiledb", compileCommands)
|
|
|
|
msvc_version = ""
|
|
if "MSVC_VERSION" in env and env["MSVC_VERSION"]:
|
|
msvc_version = "--version " + env["MSVC_VERSION"] + " "
|
|
|
|
# Microsoft Visual Studio Project generation for code browsing
|
|
if get_option("ninja") == "disabled":
|
|
vcxprojFile = env.Command(
|
|
"mongodb.vcxproj",
|
|
compileCommands,
|
|
r"$PYTHON buildscripts\make_vcxproj.py " + msvc_version + "mongodb",
|
|
)
|
|
vcxproj = env.Alias("vcxproj", vcxprojFile)
|
|
|
|
# TODO: maybe make these work like the other archive- aliases
|
|
# even though they aren't piped through AIB?
|
|
distSrc = env.DistSrc("distsrc.tar", NINJA_SKIP=True)
|
|
env.NoCache(distSrc)
|
|
env.Alias("distsrc-tar", distSrc)
|
|
|
|
distSrcGzip = env.GZip(
|
|
target="distsrc.tgz",
|
|
source=[distSrc],
|
|
NINJA_SKIP=True,
|
|
)
|
|
env.NoCache(distSrcGzip)
|
|
env.Alias("distsrc-tgz", distSrcGzip)
|
|
|
|
distSrcZip = env.DistSrc("distsrc.zip", NINJA_SKIP=True)
|
|
env.NoCache(distSrcZip)
|
|
env.Alias("distsrc-zip", distSrcZip)
|
|
|
|
env.Alias("distsrc", "distsrc-tgz")
|
|
|
|
# Do this as close to last as possible before reading SConscripts, so
|
|
# that any tools that may have injected other things via emitters are included
|
|
# among the side effect adornments.
|
|
env.Tool("task_limiter")
|
|
if has_option("jlink"):
|
|
link_jobs = env.SetupTaskLimiter(
|
|
name="jlink",
|
|
concurrency_ratio=get_option("jlink"),
|
|
builders=["Program", "SharedLibrary", "LoadableModule"],
|
|
)
|
|
if get_option("ninja") != "disabled":
|
|
env["NINJA_LINK_JOBS"] = link_jobs
|
|
|
|
if env.get("UNITTESTS_COMPILE_CONCURRENCY"):
|
|
if hasattr(SCons.Tool, "cxx"):
|
|
c_suffixes = SCons.Tool.cxx.CXXSuffixes
|
|
else:
|
|
c_suffixes = SCons.Tool.msvc.CXXSuffixes
|
|
|
|
env.SetupTaskLimiter(
|
|
name="unit_cc",
|
|
concurrency_ratio=float(env.get("UNITTESTS_COMPILE_CONCURRENCY")),
|
|
builders={"Object": c_suffixes, "SharedObject": c_suffixes},
|
|
source_file_regex=r"^.*_test\.cpp$",
|
|
)
|
|
|
|
first_half_flag = False
|
|
|
|
|
|
def half_source_emitter(target, source, env):
|
|
global first_half_flag
|
|
if first_half_flag:
|
|
first_half_flag = False
|
|
if "conftest" not in str(target[0]) and not str(source[0]).endswith("_test.cpp"):
|
|
env.Alias("compile_first_half_non_test_source", target)
|
|
else:
|
|
first_half_flag = True
|
|
return target, source
|
|
|
|
|
|
# Cribbed from Tool/cc.py and Tool/c++.py. It would be better if
|
|
# we could obtain this from SCons.
|
|
_CSuffixes = [".c"]
|
|
if not SCons.Util.case_sensitive_suffixes(".c", ".C"):
|
|
_CSuffixes.append(".C")
|
|
|
|
_CXXSuffixes = [".cpp", ".cc", ".cxx", ".c++", ".C++"]
|
|
if SCons.Util.case_sensitive_suffixes(".c", ".C"):
|
|
_CXXSuffixes.append(".C")
|
|
|
|
for object_builder in SCons.Tool.createObjBuilders(env):
|
|
emitterdict = object_builder.builder.emitter
|
|
for suffix in emitterdict.keys():
|
|
if suffix not in _CSuffixes + _CXXSuffixes:
|
|
continue
|
|
base = emitterdict[suffix]
|
|
emitterdict[suffix] = SCons.Builder.ListEmitter(
|
|
[
|
|
base,
|
|
half_source_emitter,
|
|
]
|
|
)
|
|
|
|
# Keep this late in the game so that we can investigate attributes set by all the tools that have run.
|
|
if has_option("cache"):
|
|
if get_option("cache") == "nolinked":
|
|
|
|
def noCacheEmitter(target, source, env):
|
|
for t in target:
|
|
try:
|
|
if getattr(t.attributes, "thin_archive", False):
|
|
continue
|
|
except AttributeError:
|
|
pass
|
|
env.NoCache(t)
|
|
return target, source
|
|
|
|
def addNoCacheEmitter(builder):
|
|
origEmitter = builder.emitter
|
|
if SCons.Util.is_Dict(origEmitter):
|
|
for k, v in origEmitter:
|
|
origEmitter[k] = SCons.Builder.ListEmitter([v, noCacheEmitter])
|
|
elif SCons.Util.is_List(origEmitter):
|
|
origEmitter.append(noCacheEmitter)
|
|
else:
|
|
builder.emitter = SCons.Builder.ListEmitter([origEmitter, noCacheEmitter])
|
|
|
|
addNoCacheEmitter(env["BUILDERS"]["Program"])
|
|
addNoCacheEmitter(env["BUILDERS"]["StaticLibrary"])
|
|
addNoCacheEmitter(env["BUILDERS"]["SharedLibrary"])
|
|
addNoCacheEmitter(env["BUILDERS"]["SharedArchive"])
|
|
addNoCacheEmitter(env["BUILDERS"]["LoadableModule"])
|
|
|
|
if env.GetOption("patch-build-mongot-url"):
|
|
binary_url = env.GetOption("patch-build-mongot-url")
|
|
|
|
env.Command(
|
|
target="mongot-localdev",
|
|
source=[],
|
|
action=[
|
|
f"curl {binary_url} | tar xvz",
|
|
],
|
|
)
|
|
|
|
env.AutoInstall(
|
|
target="$PREFIX_BINDIR",
|
|
source=["mongot-localdev"],
|
|
AIB_COMPONENT="mongot",
|
|
AIB_ROLE="runtime",
|
|
AIB_COMPONENTS_EXTRA=["dist-test"],
|
|
)
|
|
|
|
# mongot is a MongoDB-specific process written as a wrapper around Lucene. Using Lucene, mongot
|
|
# indexes MongoDB databases to provide our customers with full text search capabilities.
|
|
#
|
|
# --build-mongot is utilized as a compile flag by the evergreen build variants that run end-to-end
|
|
# search suites. It downloads & bundles mongot with the other mongo binaries. These binaries become
|
|
# available to the build variants in question when the binaries are extracted via archive_dist_test
|
|
# during compilation.
|
|
elif env.GetOption("build-mongot"):
|
|
# '--build-mongot` can be 'latest' or'release'
|
|
# - 'latest' describes the binaries created by the most recent commit merged to 10gen/mongot.
|
|
# - 'release' refers to the mongot binaries running in atlas prod.
|
|
binary_ver_str = env.GetOption("build-mongot")
|
|
|
|
platform_str = ""
|
|
if mongo_platform.is_running_os("linux"):
|
|
platform_str = "linux"
|
|
elif mongo_platform.is_running_os("darwin"):
|
|
platform_str = "macos"
|
|
else:
|
|
print("mongot is only supported on macOS and linux")
|
|
Exit(1)
|
|
|
|
arch_str = "x86_64"
|
|
# macos arm64 is not supported by mongot, but macos x86_64 runs on it successfully
|
|
if mongo_platform.is_arm_processor() and platform_str != "macos":
|
|
arch_str = "aarch64"
|
|
|
|
db_contrib_tool = env.Command(
|
|
target=["$BUILD_ROOT/db_contrib_tool_venv/bin/db-contrib-tool"],
|
|
source=[],
|
|
action=[
|
|
"rm -rf $BUILD_ROOT/db_contrib_tool_venv",
|
|
f"{sys.executable} -m virtualenv -p {sys.executable} $BUILD_ROOT/db_contrib_tool_venv",
|
|
"$BUILD_ROOT/db_contrib_tool_venv/bin/python3 -m pip install db-contrib-tool",
|
|
],
|
|
BUILD_ROOT=env.Dir("$BUILD_ROOT").path,
|
|
)
|
|
|
|
env.Command(
|
|
target=["mongot-localdev"],
|
|
source=db_contrib_tool,
|
|
action=[
|
|
f"$SOURCE setup-mongot-repro-env {binary_ver_str} --platform={platform_str} --architecture={arch_str}",
|
|
"mv build/mongot-localdev mongot-localdev",
|
|
],
|
|
ENV=os.environ,
|
|
)
|
|
|
|
env.AutoInstall(
|
|
target="$PREFIX_BINDIR",
|
|
source=["mongot-localdev"],
|
|
AIB_COMPONENT="mongot",
|
|
AIB_ROLE="runtime",
|
|
AIB_COMPONENTS_EXTRA=["dist-test"],
|
|
)
|
|
|
|
|
|
# load the tool late to make sure we can copy over any new
|
|
# emitters/scanners we may have created in the SConstruct when
|
|
# we go to make stand in bazel builders for the various scons builders
|
|
|
|
# __NINJA_NO is ninja callback to scons signal, in that case we care about
|
|
# scons only targets not thin targets.
|
|
env.Tool("integrate_bazel")
|
|
if env.get("__NINJA_NO") == "1":
|
|
env.LoadBazelBuilders()
|
|
|
|
def noop(*args, **kwargs):
|
|
pass
|
|
|
|
env.AddMethod(noop, "WaitForBazel")
|
|
env.AddMethod(noop, "BazelAutoInstall")
|
|
|
|
if env.get("__NINJA_NO") != "1":
|
|
BAZEL_AUTOINSTALLED_LIBDEPS = set()
|
|
|
|
# the next emitters will read link lists
|
|
# to determine dependencies in order for scons
|
|
# to handle the install
|
|
def bazel_auto_install_emitter(target, source, env):
|
|
for libdep in env.Flatten(env.get("LIBDEPS", [])) + env.Flatten(
|
|
env.get("LIBDEPS_PRIVATE", [])
|
|
):
|
|
libdep_node = libdeps._get_node_with_ixes(
|
|
env, env.Entry(libdep).abspath, "SharedLibrary"
|
|
)
|
|
try:
|
|
shlib_suffix = env.subst("$SHLIBSUFFIX")
|
|
bazel_libdep = env.File(
|
|
f"#/{env['SCONS2BAZEL_TARGETS'].bazel_output(libdep_node.path)}"
|
|
)
|
|
if str(bazel_libdep).endswith(shlib_suffix):
|
|
if bazel_libdep not in BAZEL_AUTOINSTALLED_LIBDEPS:
|
|
env.BazelAutoInstall(bazel_libdep, shlib_suffix)
|
|
BAZEL_AUTOINSTALLED_LIBDEPS.add(bazel_libdep)
|
|
env.Depends(
|
|
env.GetAutoInstalledFiles(target[0]),
|
|
env.GetAutoInstalledFiles(bazel_libdep),
|
|
)
|
|
except KeyError:
|
|
if env.Verbose():
|
|
print(f"BazelAutoInstall not processing non bazel target:\n{libdep_node}")
|
|
|
|
return target, source
|
|
|
|
for builder_name in ["Program", "SharedLibrary"]:
|
|
builder = env["BUILDERS"][builder_name]
|
|
base_emitter = builder.emitter
|
|
new_emitter = SCons.Builder.ListEmitter([base_emitter, bazel_auto_install_emitter])
|
|
builder.emitter = new_emitter
|
|
|
|
def bazel_program_auto_install_emitter(target, source, env):
|
|
if env.GetOption("link-model") == "dynamic-sdk":
|
|
return target, source
|
|
|
|
bazel_target = env["SCONS2BAZEL_TARGETS"].bazel_target(target[0].path)
|
|
|
|
linkfile = bazel_target.replace("//src/", "bazel-bin/src/") + "_links.list"
|
|
linkfile = "/".join(linkfile.rsplit(":", 1))
|
|
|
|
with open(os.path.join(env.Dir("#").abspath, linkfile)) as f:
|
|
query_results = f.read()
|
|
|
|
filtered_results = ""
|
|
for lib in query_results.splitlines():
|
|
bazel_out_path = lib.replace("\\", "/").replace(
|
|
f"{env['BAZEL_OUT_DIR']}/src", "bazel-bin/src"
|
|
)
|
|
if os.path.exists(
|
|
env.File("#/" + bazel_out_path + ".exclude_lib").abspath.replace("\\", "/")
|
|
):
|
|
continue
|
|
filtered_results += lib + "\n"
|
|
query_results = filtered_results
|
|
|
|
t = target[0]
|
|
suffix = getattr(t.attributes, "aib_effective_suffix", t.get_suffix())
|
|
bazel_node = env.File(
|
|
t.abspath.replace(
|
|
f"{env.Dir('#').abspath}/{env['BAZEL_OUT_DIR']}/src", env.Dir("$BUILD_DIR").path
|
|
)
|
|
)
|
|
|
|
prog_output = env.BazelAutoInstallSingleTarget(bazel_node, suffix, bazel_node)
|
|
|
|
libs = []
|
|
debugs = []
|
|
for lib in query_results.splitlines():
|
|
libdep = env.File(
|
|
lib.replace("\\", "/").replace(f"{env['BAZEL_OUT_DIR']}/src", "$BUILD_DIR")
|
|
)
|
|
libdep_node = libdeps._get_node_with_ixes(
|
|
env, env.Entry(libdep).abspath, "SharedLibrary"
|
|
)
|
|
bazel_libdep = env.File(
|
|
f"#/{env['SCONS2BAZEL_TARGETS'].bazel_output(libdep_node.path)}"
|
|
)
|
|
shlib_suffix = env.subst("$SHLIBSUFFIX")
|
|
if str(bazel_libdep).endswith(shlib_suffix):
|
|
if bazel_libdep not in BAZEL_AUTOINSTALLED_LIBDEPS:
|
|
env.BazelAutoInstall(bazel_libdep, shlib_suffix)
|
|
BAZEL_AUTOINSTALLED_LIBDEPS.add(bazel_libdep)
|
|
libs.append(env.GetAutoInstalledFiles(bazel_libdep)[0])
|
|
if hasattr(bazel_libdep.attributes, "separate_debug_files"):
|
|
debugs.append(
|
|
env.GetAutoInstalledFiles(
|
|
getattr(bazel_libdep.attributes, "separate_debug_files")[0]
|
|
)[0]
|
|
)
|
|
|
|
env.Depends(prog_output[0], libs)
|
|
if len(prog_output) == 2:
|
|
env.Depends(prog_output[1], debugs)
|
|
return target, source
|
|
|
|
builder = env["BUILDERS"]["BazelProgram"]
|
|
base_emitter = builder.emitter
|
|
new_emitter = SCons.Builder.ListEmitter([bazel_program_auto_install_emitter, base_emitter])
|
|
builder.emitter = new_emitter
|
|
|
|
|
|
def injectMongoIncludePaths(thisEnv):
|
|
if thisEnv.get("BAZEL_OUT_DIR"):
|
|
thisEnv.AppendUnique(CPPPATH=["#$BAZEL_OUT_DIR/src"])
|
|
thisEnv.AppendUnique(CPPPATH=["$BUILD_DIR"])
|
|
|
|
|
|
env.AddMethod(injectMongoIncludePaths, "InjectMongoIncludePaths")
|
|
|
|
gen_header_paths = [
|
|
(pathlib.Path(env.Dir("$BAZEL_OUT_DIR").path) / "src" / "mongo").as_posix(),
|
|
(pathlib.Path(env.Dir("$BUILD_DIR").path) / "mongo").as_posix(),
|
|
]
|
|
|
|
replacements = {
|
|
"@MONGO_BUILD_DIR@": ("|".join([path + "/.*" for path in gen_header_paths])),
|
|
"@MONGO_BRACKET_BUILD_DIR@": (";".join(gen_header_paths)),
|
|
}
|
|
|
|
clang_tidy_config = env.Substfile(
|
|
target=".clang-tidy",
|
|
source=[
|
|
".clang-tidy.in",
|
|
],
|
|
SUBST_DICT=replacements,
|
|
)
|
|
env.Alias("generated-sources", clang_tidy_config)
|
|
|
|
if get_option("bazel-includes-info"):
|
|
env.Tool("bazel_includes_info")
|
|
|
|
env.WaitForBazel()
|
|
|
|
if str(env["LIBDEPS_GRAPH_ALIAS"]) in COMMAND_LINE_TARGETS:
|
|
# The find_symbols binary is a small fast C binary which will extract the missing
|
|
# symbols from the target library, and discover what linked libraries supply it. This
|
|
# setups the binary to be built.
|
|
find_symbols_env = env.Clone()
|
|
find_symbols_env.VariantDir("${BUILD_DIR}/libdeps", "buildscripts/libdeps", duplicate=0)
|
|
find_symbols_env.Program(
|
|
target="${BUILD_DIR}/libdeps/find_symbols",
|
|
source=["${BUILD_DIR}/libdeps/find_symbols.c"],
|
|
CFLAGS=["-O3"],
|
|
)
|
|
|
|
env.SConscript(
|
|
must_exist=1,
|
|
dirs=[
|
|
"src",
|
|
],
|
|
duplicate=False,
|
|
exports=[
|
|
"env",
|
|
],
|
|
variant_dir="$BUILD_DIR",
|
|
)
|
|
|
|
# TODO: find a way to consolidate SConscript calls to one call in
|
|
# SConstruct so they all use variant_dir
|
|
env.SConscript(
|
|
must_exist=1,
|
|
dirs=[
|
|
"jstests",
|
|
],
|
|
duplicate=False,
|
|
exports=[
|
|
"env",
|
|
],
|
|
)
|
|
|
|
# Critically, this approach is technically incorrect. While all MongoDB
|
|
# SConscript files use our add_option wrapper, builtin tools can
|
|
# access SCons's GetOption/AddOption methods directly, causing their options
|
|
# to not be validated by this block.
|
|
(_, leftover) = _parser.parse_args(sys.argv)
|
|
# leftover contains unrecognized options, including environment variables,and
|
|
# the argv[0]. If we only look at flags starting with --, and we skip the first
|
|
# leftover value (argv[0]), anything that remains is an invalid option
|
|
invalid_options = list(filter(lambda x: x.startswith("--"), leftover[1:]))
|
|
if len(invalid_options) > 0:
|
|
# users frequently misspell "variables-files" (note two `s`s) as
|
|
# "variable-files" or "variables-file". Detect and help them out.
|
|
for opt in invalid_options:
|
|
bad_var_file_opts = ["--variable-file", "--variables-file", "--variable-files"]
|
|
if opt in bad_var_file_opts or any(
|
|
[opt.startswith(f"{bad_opt}=") for bad_opt in bad_var_file_opts]
|
|
):
|
|
print(
|
|
f"WARNING: You supplied the invalid parameter '{opt}' to SCons. Did you mean --variables-files (both words plural)?"
|
|
)
|
|
fatal_error(None, f"ERROR: unknown options supplied to scons: {invalid_options}")
|
|
|
|
# Declare the cache prune target
|
|
cachePrune = env.Command(
|
|
target="#cache-prune",
|
|
source=[
|
|
"#buildscripts/scons_cache_prune.py",
|
|
],
|
|
action="$PYTHON ${SOURCES[0]} --cache-dir=${CACHE_DIR.abspath} --cache-size=${CACHE_SIZE} --prune-ratio=${CACHE_PRUNE_TARGET/100.00}",
|
|
CACHE_DIR=env.Dir(cacheDir),
|
|
)
|
|
|
|
env.AlwaysBuild(cachePrune)
|
|
|
|
# Add a trivial Alias called `configure`. This makes it simple to run,
|
|
# or re-run, the SConscript reading and conf tests, but not build any
|
|
# real targets. This can be helpful when you are planning a dry-run
|
|
# build, or simply want to validate your changes to SConstruct, tools,
|
|
# and all the other setup that happens before we begin a real graph
|
|
# walk.
|
|
env.Alias("configure", None)
|
|
|
|
env.CreateConsolidatedTargets()
|
|
|
|
# We have finished all SConscripts and targets, so we can ask
|
|
# auto_install_binaries to finalize the installation setup.
|
|
env.FinalizeInstallDependencies()
|
|
|
|
# Create a install-all-meta alias that excludes unittests. This is most useful in
|
|
# static builds where the resource requirements of linking 100s of static unittest
|
|
# binaries is prohibitive.
|
|
candidate_nodes = set()
|
|
for child in env.Alias("install-all-meta")[0].all_children():
|
|
candidate_nodes.add(child)
|
|
for gchild in child.all_children():
|
|
candidate_nodes.add(gchild)
|
|
|
|
names = [
|
|
f'install-{env["AIB_META_COMPONENT"]}',
|
|
"install-tests",
|
|
env["UNITTEST_ALIAS"],
|
|
"install-first-quarter-unittests",
|
|
"install-second-quarter-unittests",
|
|
"install-third-quarter-unittests",
|
|
"install-fourth-quarter-unittests",
|
|
]
|
|
|
|
env.Alias(
|
|
"install-all-meta-but-not-unittests",
|
|
[
|
|
node
|
|
for node in candidate_nodes
|
|
if str(node) not in names
|
|
and not str(node).startswith(tuple([prefix_name + "-" for prefix_name in names]))
|
|
],
|
|
)
|
|
|
|
# prove prefix only operates on real AIB_COMPONENTS, unittests is now a meta component
|
|
# made up of all the quarter components combined. We create the prove alias for this meta
|
|
# component for compatibility with the past and ease of use.
|
|
env.Alias(
|
|
"prove-unittests",
|
|
[
|
|
"prove-first-quarter-unittests",
|
|
"prove-second-quarter-unittests",
|
|
"prove-third-quarter-unittests",
|
|
"prove-fourth-quarter-unittests",
|
|
],
|
|
)
|
|
|
|
# We don't want installing files to cause them to flow into the cache,
|
|
# since presumably we can re-install them from the origin if needed.
|
|
env.NoCache(env.FindInstalledFiles())
|
|
|
|
# Substitute environment variables in any build targets so that we can
|
|
# say, for instance:
|
|
#
|
|
# > scons --prefix=/foo/bar '$DESTDIR'
|
|
# or
|
|
# > scons \$BUILD_DIR/mongo/base
|
|
#
|
|
# That way, you can reference targets under the variant dir or install
|
|
# path via an invariant name.
|
|
#
|
|
# We need to replace the values in the BUILD_TARGETS object in-place
|
|
# because SCons wants it to be a particular object.
|
|
for i, s in enumerate(BUILD_TARGETS):
|
|
BUILD_TARGETS[i] = env.subst(s)
|
|
|
|
# Do any final checks the Libdeps linter may need to do once all
|
|
# SConscripts have been read but before building begins.
|
|
libdeps.LibdepLinter(env).final_checks()
|
|
libdeps.generate_libdeps_graph(env)
|
|
|
|
# We put this next section at the end of the SConstruct since all the targets
|
|
# have been declared, and we know all possible bazel targets so
|
|
# we can now generate this info into a file for the ninja build to consume.
|
|
if env.GetOption("ninja") != "disabled" and env.get("__NINJA_NO") != "1":
|
|
# convert the SCons FunctioAction into a format that ninja can understand
|
|
env.NinjaRegisterFunctionHandler("bazel_builder_action", env.NinjaBazelBuilder)
|
|
|
|
# we generate the list of all targets that were labeled Bazel* builder targets
|
|
# via the emitter, this outputs a json file which will be read during the ninja
|
|
# build.
|
|
env.GenerateBazelInfoForNinja()
|