2.5.2. Global Configuration
The keys in this section affect the operations of the buildmaster globally.
2.5.2.1. Database Specification
Buildbot requires a connection to a database to maintain certain state information, such as tracking pending build requests.
In the default configuration Buildbot uses a file-based SQLite database, stored in the state.sqlite
file of the master’s base directory.
Important
SQLite3 is perfectly suitable for small setups with a few users. However, it does not scale well with large numbers of builders, workers and users. If you expect your Buildbot to grow over time, it is strongly advised to use a real database server (e.g., MySQL or Postgres).
A SQLite3 database may be migrated to a real database server using buildbot copy-db
script.
See the Using A Database Server section for more details.
Override this configuration with the db_url
parameter.
Buildbot accepts a database configuration in a dictionary named db
.
All keys are optional:
c['db'] = {
'db_url' : 'sqlite:///state.sqlite',
}
The db_url
key indicates the database engine to use.
The format of this parameter is completely documented at http://www.sqlalchemy.org/docs/dialects/, but is generally of the form:
"driver://[username:password@]host:port/database[?args]"
This parameter can be specified directly in the configuration dictionary, as c['db_url']
, although this method is deprecated.
The following sections give additional information for particular database backends:
SQLite
For sqlite databases, since there is no host and port, relative paths are specified with sqlite:///
and absolute paths with sqlite:////
.
For example:
c['db_url'] = "sqlite:///state.sqlite"
SQLite requires no special configuration.
MySQL
c['db_url'] = "mysql://username:password@example.com/database_name?max_idle=300"
The max_idle
argument for MySQL connections is unique to Buildbot and should be set to something less than the wait_timeout
configured for your server.
This controls the SQLAlchemy pool_recycle
parameter, which defaults to no timeout.
Setting this parameter ensures that connections are closed and re-opened after the configured amount of idle time.
If you see errors such as _mysql_exceptions.OperationalError: (2006, 'MySQL server has gone away')
, this means your max_idle
setting is probably too high.
show global variables like 'wait_timeout';
will show what the currently configured wait_timeout
is on your MySQL server.
Buildbot requires use_unique=True
and charset=utf8
, and will add them automatically, so they do not need to be specified in db_url
.
MySQL defaults to the MyISAM storage engine, but this can be overridden with the storage_engine
URL argument.
Postgres
c['db_url'] = "postgresql://username:password@hostname/dbname"
PosgreSQL requires no special configuration.
2.5.2.2. MQ Specification
Buildbot uses a message-queueing system to handle communication within the master. Messages are used to indicate events within the master, and components that are interested in those events arrange to receive them.
The message queueing implementation is configured as a dictionary in the mq
option.
The type
key describes the type of MQ implementation to be used.
Note that the implementation type cannot be changed in a reconfig.
The available implementation types are described in the following sections.
Simple
c['mq'] = {
'type' : 'simple',
'debug' : False,
}
This is the default MQ implementation. Similar to SQLite, it has no additional software dependencies, but does not support multi-master mode.
Note that this implementation also does not support message persistence across a restart of the master. For example, if a change is received, but the master shuts down before the schedulers can create build requests for it, then those schedulers will not be notified of the change when the master starts again.
The debug
key, which defaults to False, can be used to enable logging of every message produced on this master.
Wamp
Note
At the moment, wamp is the only message queue implementation for multimaster.
It has been privileged as this is the only message queue that has very solid support for Twisted.
Other more common message queue systems like RabbitMQ
(using the AMQP
protocol) do not have a convincing driver for twisted, and this would require to run on threads, which will add an important performance overhead.
c['mq'] = {
'type' : 'wamp',
'router_url': 'ws://localhost:8080/ws',
'realm': 'realm1',
# valid are: none, critical, error, warn, info, debug, trace
'wamp_debug_level' : 'error'
}
This is a MQ implementation using the wamp protocol. This implementation uses Python Autobahn wamp client library, and is fully asynchronous (no use of threads). To use this implementation, you need a wamp router like Crossbar.
The implementation does not yet support wamp authentication. This MQ allows buildbot to run in multi-master mode.
Note that this implementation also does not support message persistence across a restart of the master. For example, if a change is received, but the master shuts down before the schedulers can create build requests for it, then those schedulers will not be notified of the change when the master starts again.
router_url
(mandatory): points to your router websocket url.Buildbot is only supporting wamp over websocket, which is a sub-protocol of http. SSL is supported using
wss://
instead ofws://
.
realm
(optional, defaults to buildbot
): defines the wamp realm to use for your buildbot messages.
wamp_debug_level
(optional, defaults to error
): defines the log level of autobahn.
You must use a router with very reliable connection to the master. If for some reason, the wamp connection is lost, then the master will stop, and should be restarted via a process manager.
Crossbar
The default Crossbar setup will just work with Buildbot, provided you use the example mq
configuration below, and start Crossbar with:
# of course, you should work in a virtualenv...
pip install crossbar
crossbar init
crossbar start
.crossbar/config.json:
{
"version": 2,
"controller": {},
"workers": [
{
"type": "router",
"realms": [
{
"name": "test_realm",
"roles": [
{
"name": "anonymous",
"permissions": [
{
"uri": "",
"match": "prefix",
"allow": {
"call": true,
"register": true,
"publish": true,
"subscribe": true
},
"disclose": {
"caller": false,
"publisher": false
},
"cache": true
}
]
}
]
}
],
"transports": [
{
"type": "web",
"endpoint": {
"type": "tcp",
"port": 1245
},
"paths": {
"ws": {
"type": "websocket"
}
}
}
]
}
]
}
Buildbot can be configured to use Crossbar by the following:
c["mq"] = {
"type" : "wamp",
"router_url": "ws://localhost:1245/ws",
"realm": "test_realm",
"wamp_debug_level" : "warn"
}
Please refer to Crossbar documentation for more details.
2.5.2.3. Multi-master mode
See Multimaster for details on the multi-master mode in Buildbot Nine.
By default, Buildbot makes coherency checks that prevent typos in your master.cfg
.
It makes sure schedulers are not referencing unknown builders, and enforces there is at least one builder.
In the case of an asymmetric multimaster, those coherency checks can be harmful and prevent you to implement what you want. For example, you might want to have one master dedicated to the UI, so that a big load generated by builds will not impact page load times.
To enable multi-master mode in this configuration, you will need to set the multiMaster
option so that buildbot doesn’t warn about missing schedulers or builders.
# Enable multiMaster mode; disables warnings about unknown builders and
# schedulers
c['multiMaster'] = True
c['db'] = {
'db_url' : 'mysql://...',
}
c['mq'] = { # Need to enable multimaster aware mq. Wamp is the only option for now.
'type' : 'wamp',
'router_url': 'ws://localhost:8080',
'realm': 'realm1',
# valid are: none, critical, error, warn, info, debug, trace
'wamp_debug_level' : 'error'
}
2.5.2.4. Site Definitions
Three basic settings describe the buildmaster in status reports:
c['title'] = "Buildbot"
c['titleURL'] = "http://buildbot.sourceforge.net/"
title
is a short string that will appear at the top of this buildbot installation’s home page (linked to the titleURL
).
titleURL
is a URL string.
HTML status displays will show title
as a link to titleURL
.
This URL is often used to provide a link from buildbot HTML pages to your project’s home page.
The buildbotURL
string should point to the location where the buildbot’s internal web server is visible.
When status notices are sent to users (e.g., by email or over IRC), buildbotURL
will be used to create a URL to the specific build or problem that they are being notified about.
2.5.2.5. Log Handling
c['logCompressionMethod'] = 'gz'
c['logMaxSize'] = 1024*1024 # 1M
c['logMaxTailSize'] = 32768
c['logEncoding'] = 'utf-8'
The logCompressionLimit
enables compression of build logs on disk for logs that are bigger than the given size, or disables that completely if set to False
.
The default value is 4096, which should be a reasonable default on most file systems.
This setting has no impact on status plugins, and merely affects the required disk space on the master for build logs.
The logCompressionMethod
controls what type of compression is used for build logs.
Valid option are ‘raw’ (no compression), ‘gz’, ‘lz4’ (required lz4 package), ‘br’ (requires buildbot[brotli] extra) or ‘zstd’ (requires buildbot[zstd] extra).
The default is ‘zstd’ if the buildbot[zstd]
is installed, otherwise defaults to ‘gz’.
Please find below some stats extracted from 50x “trial Pyflakes” runs (results may differ according to log type).
compression |
raw log size |
compressed log size |
space saving |
compression speed |
---|---|---|---|---|
bz2 |
2.981 MB |
0.603 MB |
79.77% |
3.433 MB/s |
gz |
2.981 MB |
0.568 MB |
80.95% |
6.604 MB/s |
lz4 |
2.981 MB |
0.844 MB |
71.68% |
77.668 MB/s |
The logMaxSize
parameter sets an upper limit (in bytes) to how large logs from an individual build step can be.
The default value is None, meaning no upper limit to the log size.
Any output exceeding logMaxSize
will be truncated, and a message to this effect will be added to the log’s HEADER channel.
If logMaxSize
is set, and the output from a step exceeds the maximum, the logMaxTailSize
parameter controls how much of the end of the build log will be kept.
The effect of setting this parameter is that the log will contain the first logMaxSize
bytes and the last logMaxTailSize
bytes of output.
Don’t set this value too high, as the tail of the log is kept in memory.
The logEncoding
parameter specifies the character encoding to use to decode bytestrings provided as logs.
It defaults to utf-8
, which should work in most cases, but can be overridden if necessary.
In extreme cases, a callable can be specified for this parameter.
It will be called with byte strings, and should return the corresponding Unicode string.
This setting can be overridden for a single build step with the logEncoding
step parameter.
It can also be overridden for a single log file by passing the logEncoding
parameter to addLog
.
2.5.2.6. Data Lifetime
Horizons
Previously Buildbot implemented a global configuration for horizons.
Now it is implemented as a utility Builder, and shall be configured via the JanitorConfigurator
.
Caches
c['caches'] = {
'Changes' : 100, # formerly c['changeCacheSize']
'Builds' : 500, # formerly c['buildCacheSize']
'chdicts' : 100,
'BuildRequests' : 10,
'SourceStamps' : 20,
'ssdicts' : 20,
'objectids' : 10,
'usdicts' : 100,
}
The caches
configuration key contains the configuration for Buildbot’s in-memory caches.
These caches keep frequently-used objects in memory to avoid unnecessary trips to the database.
Caches are divided by object type, and each has a configurable maximum size.
The default size for each cache is 1, except where noted below. A value of 1 allows Buildbot to make a number of optimizations without consuming much memory. Larger, busier installations will likely want to increase these values.
The available caches are:
Changes
the number of change objects to cache in memory. This should be larger than the number of changes that typically arrive in the span of a few minutes, otherwise your schedulers will be reloading changes from the database every time they run. For distributed version control systems, like Git or Hg, several thousand changes may arrive at once, so setting this parameter to something like 10000 isn’t unreasonable.
This parameter is the same as the deprecated global parameter
changeCacheSize
. Its default value is 10.Builds
The
buildCacheSize
parameter gives the number of builds for each builder which are cached in memory. This number should be larger than the number of builds required for commonly-used status displays (the waterfall or grid views), so that those displays do not miss the cache on a refresh.This parameter is the same as the deprecated global parameter
buildCacheSize
. Its default value is 15.chdicts
The number of rows from the
changes
table to cache in memory. This value should be similar to the value forChanges
.BuildRequests
The number of BuildRequest objects kept in memory. This number should be higher than the typical number of outstanding build requests. If the master ordinarily finds jobs for BuildRequests immediately, you may set a lower value.
SourceStamps
the number of SourceStamp objects kept in memory. This number should generally be similar to the number
BuildRequesets
.ssdicts
The number of rows from the
sourcestamps
table to cache in memory. This value should be similar to the value forSourceStamps
.objectids
The number of object IDs - a means to correlate an object in the Buildbot configuration with an identity in the database–to cache. In this version, object IDs are not looked up often during runtime, so a relatively low value such as 10 is fine.
usdicts
The number of rows from the
users
table to cache in memory. Note that for a given user there will be a row for each attribute that user has.c[‘buildCacheSize’] = 15
2.5.2.7. Merging Build Requests
c['collapseRequests'] = True
This is a global default value for builders’ collapseRequests
parameter, and controls the merging of build requests.
This parameter can be overridden on a per-builder basis. See Collapsing Build Requests for the allowed values for this parameter.
2.5.2.8. Prioritizing Builders
def prioritizeBuilders(buildmaster, builders):
...
c['prioritizeBuilders'] = prioritizeBuilders
By default, buildbot will attempt to start builds on builders in order, beginning with the builder with the oldest pending request.
Customize this behavior with the prioritizeBuilders
configuration key, which takes a callable.
See Builder Priority Functions for details on this callable.
This parameter controls the order that the buildmaster can start builds, and is useful in situations where there is resource contention between builders, e.g., for a test database. It does not affect the order in which a builder processes the build requests in its queue. For that purpose, see Prioritizing Builds.
2.5.2.9. Prioritizing Workers
By default Buildbot will select worker for a build randomly from available workers. This can be
adjusted by select_next_worker
function in global master configuration and additionally by
nextWorker
per-builder configuration parameter. These two functions work exactly the same:
The function is passed three arguments, the Builder
object which is assigning a new job,
a list of WorkerForBuilder
objects and the BuildRequest
.
The function should return one of the WorkerForBuilder
objects, or None
if none of the
available workers should be used. The function can optionally return a Deferred, which should fire
with the same results.
def select_next_worker(builder, workers, buildrequest):
...
c["select_next_worker"] = select_next_worker
2.5.2.10. Configuring worker protocols
The protocols
key defines how buildmaster listens to connections from workers. The value of
the key is dictionary with keys being protocol names and values being per-protocol configuration.
The following protocols are supported:
pb
- Perspective Broker protocol. This protocol supports not only connections from workers, but also remote Change Sources, status clients and debug tools. It supports the following configuration:port
- specifies the listening port configuration. This may be a numeric port, or a connection string, as defined in the ConnectionStrings guide.
msgpack_experimental_v7
- (experimental) MessagePack-based protocol. It supports the following configuration:port
- specifies the listening port configuration. This may be a numeric port, or a connection string, as defined in the ConnectionStrings guide.
Note
Note, that the master host must be visible to all workers that would attempt to connect to it. The firewall (if any) must be configured to allow external connections. Additionally, the configured listen port must be larger than 1024 in most cases, as lower ports are usually restricted to root processes only.
The following is a minimal example of protocol configuration:
c['protocols'] = {"pb": {"port": 10000}}
The following example only allows connections from localhost. This might be useful in cases workers
are run on the same machine as master (e.g. in very small Buildbot installations). The workers would
need to be configured to contact the buildmaster at localhost:10000
.
c['protocols'] = {"pb": {"port": "tcp:10000:interface=127.0.0.1"}}
The following example shows how to configure worker connections via TLS:
c['protocols'] = {"pb": {"port":
"ssl:9989:privateKey=master.key:certKey=master.crt"}}
Please note that IPv6 addresses with : must be escaped with as well as : in paths and in paths. Read more about the connection strings format in ConnectionStrings documentation.
See also Worker TLS Configuration
2.5.2.11. Defining Global Properties
The properties
configuration key defines a dictionary of properties that will be available to all builds started by the buildmaster:
c['properties'] = {
'Widget-version' : '1.2',
'release-stage' : 'alpha'
}
2.5.2.12. Manhole
Manhole is an interactive Python shell which allows full access to the Buildbot master instance. It is probably only useful for buildbot developers.
See documentation on Manhole implementations for available authentication and connection methods.
The manhole
configuration key accepts a single instance of a Manhole class.
For example:
from buildbot import manhole
c['manhole'] = manhole.PasswordManhole("tcp:1234:interface=127.0.0.1",
"admin", "passwd",
ssh_hostkey_dir="data/ssh_host_keys")
2.5.2.13. Metrics Options
c['metrics'] = {
"log_interval": 10,
"periodic_interval": 10
}
metrics
can be a dictionary that configures various aspects of the metrics subsystem.
If metrics
is None
, then metrics collection, logging and reporting will be disabled.
log_interval
determines how often metrics should be logged to twistd.log.
It defaults to 60s.
If set to 0 or None
, then logging of metrics will be disabled.
This value can be changed via a reconfig.
periodic_interval
determines how often various non-event based metrics are collected, such as memory usage, uncollectable garbage, reactor delay.
This defaults to 10s.
If set to 0 or None
, then periodic collection of this data is disabled.
This value can also be changed via a reconfig.
Read more about metrics in the Metrics section in the developer documentation.
2.5.2.14. Statistics Service
The Statistics Service (stats service for short) supports the collection of arbitrary data from within a running Buildbot instance and the export to a number of storage backends.
Currently, only InfluxDB is supported as a storage backend.
Also, InfluxDB (or any other storage backend) is not a mandatory dependency.
Buildbot can run without it, although StatsService
will be of no use in such a case.
At present, StatsService
can keep track of build properties, build times (start, end, duration) and arbitrary data produced inside Buildbot (more on this later).
Example usage:
captures = [stats.CaptureProperty('Builder1', 'tree-size-KiB'),
stats.CaptureBuildDuration('Builder2')]
c['services'] = []
c['services'].append(stats.StatsService(
storage_backends=[
stats.InfluxStorageService('localhost', 8086, 'root', 'root', 'test', captures)
], name="StatsService"))
The services
configuration value should be initialized as a list and a StatsService
instance should be appended to it as shown in the example above.
Statistics Service
- class buildbot.statistics.stats_service.StatsService
This is the main class for statistics services. It is initialized in the master configuration as shown in the example above. It takes two arguments:
storage_backends
A list of storage backends (see Storage Backends). In the example above,
stats.InfluxStorageService
is an instance of a storage backend. Each storage backend is an instance of subclasses ofstatsStorageBase
.name
The name of this service.
yieldMetricsValue
: This method can be used to send arbitrary data for storage. (See Using StatsService.yieldMetricsValue for more information.)
Capture Classes
- class buildbot.statistics.capture.CaptureProperty
Instance of this class declares which properties must be captured and sent to the Storage Backends. It takes the following arguments:
builder_name
The name of builder in which the property is recorded.
property_name
The name of property needed to be recorded as a statistic.
callback=None
(Optional) A custom callback function for this class. This callback function should take in two arguments - build_properties (dict) and property_name (str) and return a string that will be sent for storage in the storage backends.
regex=False
If this is set to
True
, then the property name can be a regular expression. All properties matching this regular expression will be sent for storage.
- class buildbot.statistics.capture.CapturePropertyAllBuilders
Instance of this class declares which properties must be captured on all builders and sent to the Storage Backends. It takes the following arguments:
property_name
The name of property needed to be recorded as a statistic.
callback=None
(Optional) A custom callback function for this class. This callback function should take in two arguments - build_properties (dict) and property_name (str) and return a string that will be sent for storage in the storage backends.
regex=False
If this is set to
True
, then the property name can be a regular expression. All properties matching this regular expression will be sent for storage.
- class buildbot.statistics.capture.CaptureBuildStartTime
Instance of this class declares which builders’ start times are to be captured and sent to Storage Backends. It takes the following arguments:
builder_name
The name of builder whose times are to be recorded.
callback=None
(Optional) A custom callback function for this class. This callback function should take in a Python datetime object and return a string that will be sent for storage in the storage backends.
- class buildbot.statistics.capture.CaptureBuildStartTimeAllBuilders
Instance of this class declares start times of all builders to be captured and sent to Storage Backends. It takes the following arguments:
callback=None
(Optional) A custom callback function for this class. This callback function should take in a Python datetime object and return a string that will be sent for storage in the storage backends.
- class buildbot.statistics.capture.CaptureBuildEndTime
Exactly like
CaptureBuildStartTime
except it declares the builders whose end time is to be recorded. The arguments are same asCaptureBuildStartTime
.
- class buildbot.statistics.capture.CaptureBuildEndTimeAllBuilders
Exactly like
CaptureBuildStartTimeAllBuilders
except it declares all builders’ end time to be recorded. The arguments are same asCaptureBuildStartTimeAllBuilders
.
- class buildbot.statistics.capture.CaptureBuildDuration
Instance of this class declares the builders whose build durations are to be recorded. It takes the following arguments:
builder_name
The name of builder whose times are to be recorded.
report_in='seconds'
Can be one of three:
'seconds'
,'minutes'
, or'hours'
. This is the units in which the build time will be reported.callback=None
(Optional) A custom callback function for this class. This callback function should take in two Python datetime objects - a
start_time
and anend_time
and return a string that will be sent for storage in the storage backends.
- class buildbot.statistics.capture.CaptureBuildDurationAllBuilders
Instance of this class declares build durations to be recorded for all builders. It takes the following arguments:
report_in='seconds'
Can be one of three:
'seconds'
,'minutes'
, or'hours'
. This is the units in which the build time will be reported.callback=None
(Optional) A custom callback function for this class. This callback function should take in two Python datetime objects - a
start_time
and anend_time
and return a string that will be sent for storage in the storage backends.
- class buildbot.statistics.capture.CaptureData
Instance of this capture class is for capturing arbitrary data that is not stored as build-data. Needs to be used in combination with
yieldMetricsValue
(see Using StatsService.yieldMetricsValue). Takes the following arguments:data_name
The name of data to be captured. Same as in
yieldMetricsValue
.builder_name
The name of builder whose times are to be recorded.
callback=None
The callback function for this class. This callback receives the data sent to
yieldMetricsValue
aspost_data
(see Using StatsService.yieldMetricsValue). It must return a string that is to be sent to the storage backends for storage.
- class buildbot.statistics.capture.CaptureDataAllBuilders
Instance of this capture class for capturing arbitrary data that is not stored as build-data on all builders. Needs to be used in combination with
yieldMetricsValue
(see Using StatsService.yieldMetricsValue). Takes the following arguments:data_name
The name of data to be captured. Same as in
yieldMetricsValue
.callback=None
The callback function for this class. This callback receives the data sent to
yieldMetricsValue
aspost_data
(see Using StatsService.yieldMetricsValue). It must return a string that is to be sent to the storage backends for storage.
Using StatsService.yieldMetricsValue
Advanced users can modify BuildSteps
to use StatsService.yieldMetricsValue
which will send arbitrary data for storage to the StatsService
.
It takes the following arguments:
data_name
The name of the data being sent or storage.
post_data
A dictionary of key value pair that is sent for storage. The keys will act as columns in a database and the value is stored under that column.
buildid
The integer build id of the current build. Obtainable in all
BuildSteps
.
Along with using yieldMetricsValue
, the user will also need to use the CaptureData
capture class.
As an example, we can add the following to a build step:
yieldMetricsValue('test_data_name', {'some_data': 'some_value'}, buildid)
Then, we can add in the master configuration a capture class like this:
captures = [CaptureBuildData('test_data_name', 'Builder1')]
Pass this captures
list to a storage backend (as shown in the example at the top of this section) for capturing this data.
Storage Backends
Storage backends are responsible for storing any statistics data sent to them.
A storage backend will generally be some sort of a database-server running on a machine.
(Note: This machine may be different from the one running BuildMaster
)
Currently, only InfluxDB is supported as a storage backend.
- class buildbot.statistics.storage_backends.influxdb_client.InfluxStorageService
This class is a Buildbot client to the InfluxDB storage backend. InfluxDB is a distributed, time series database that employs a key-value pair storage system.
It requires the following arguments:
url
The URL where the service is running.
port
The port on which the service is listening.
user
Username of a InfluxDB user.
password
Password for
user
.db
The name of database to be used.
captures
A list of objects of Capture Classes. This tells which statistics are to be stored in this storage backend.
name=None
(Optional) The name of this storage backend.
2.5.2.15. secretsProviders
See Secret Management for details on secret concepts.
Example usage:
c['secretsProviders'] = [ .. ]
secretsProviders
is a list of secrets storage.
See Secret Management to configure a secret storage provider.
2.5.2.16. BuildbotNetUsageData
Since buildbot 0.9.0, buildbot has a simple feature which sends usage analysis info to buildbot.net. This is very important for buildbot developers to understand how the community is using the tools. This allows to better prioritize issues, and understand what plugins are actually being used. This will also be a tool to decide whether to keep support for very old tools. For example buildbot contains support for the venerable CVS, but we have no information whether it actually works beyond the unit tests. We rely on the community to test and report issues with the old features.
With BuildbotNetUsageData, we can know exactly what combination of plugins are working together, how much people are customizing plugins, what versions of the main dependencies people run.
We take your privacy very seriously.
BuildbotNetUsageData will never send information specific to your Code or Intellectual Property. No repository url, shell command values, host names, ip address or custom class names. If it does, then this is a bug, please report.
We still need to track unique number for installation. This is done via doing a sha1 hash of master’s hostname, installation path and fqdn. Using a secure hash means there is no way of knowing hostname, path and fqdn given the hash, but still there is a different hash for each master.
You can see exactly what is sent in the master’s twisted.log. Usage data is sent every time the master is started.
BuildbotNetUsageData can be configured with 4 values:
c['buildbotNetUsageData'] = None
disables the featurec['buildbotNetUsageData'] = 'basic'
sends the basic information to buildbot including:versions of buildbot, python and twisted
platform information (CPU, OS, distribution, python flavor (i.e CPython vs PyPy))
mq and database type (mysql or sqlite?)
www plugins usage
Plugins usages: This counts the number of time each class of buildbot is used in your configuration. This counts workers, builders, steps, schedulers, change sources. If the plugin is subclassed, then it will be prefixed with a >
example of basic report (for the metabuildbot):
{ 'versions': { 'Python': '2.7.6', 'Twisted': '15.5.0', 'Buildbot': '0.9.0rc2-176-g5fa9dbf' }, 'platform': { 'machine': 'x86_64', 'python_implementation': 'CPython', 'version': '#140-Ubuntu SMP Mon Jul', 'processor': 'x86_64', 'distro:': ('Ubuntu', '14.04', 'trusty') }, 'db': 'sqlite', 'mq': 'simple', 'plugins': { 'buildbot.schedulers.forcesched.ForceScheduler': 2, 'buildbot.schedulers.triggerable.Triggerable': 1, 'buildbot.config.BuilderConfig': 4, 'buildbot.schedulers.basic.AnyBranchScheduler': 2, 'buildbot.steps.source.git.Git': 4, '>>buildbot.steps.trigger.Trigger': 2, '>>>buildbot.worker.base.Worker': 4, 'buildbot.reporters.irc.IRC': 1}, 'www_plugins': ['buildbot_travis', 'waterfall_view'] }
c['buildbotNetUsageData'] = 'full'
sends the basic information plus additional information:configuration of each builders: how the steps are arranged together. for example:
{ 'builders': [ ['buildbot.steps.source.git.Git', '>>>buildbot.process.buildstep.BuildStep'], ['buildbot.steps.source.git.Git', '>>buildbot.steps.trigger.Trigger'], ['buildbot.steps.source.git.Git', '>>>buildbot.process.buildstep.BuildStep'], ['buildbot.steps.source.git.Git', '>>buildbot.steps.trigger.Trigger'] ] }
c['buildbotNetUsageData'] = myCustomFunction
declares a callback to use to specify exactly what to send.This custom function takes the generated data from full report in the form of a dictionary, and returns a customized report as a jsonable dictionary. You can use this to filter any information you don’t want to disclose. You can also use a custom http_proxy environment variable in order to not send any data while developing your callback.
2.5.2.17. Users Options
from buildbot.plugins import util
c['user_managers'] = []
c['user_managers'].append(util.CommandlineUserManager(username="user",
passwd="userpw",
port=9990))
user_managers
contains a list of ways to manually manage User Objects within Buildbot (see User Objects).
Currently implemented is a commandline tool buildbot user, described at length in user
.
In the future, a web client will also be able to manage User Objects and their attributes.
As shown above, to enable the buildbot user tool, you must initialize a CommandlineUserManager instance in your master.cfg. CommandlineUserManager instances require the following arguments:
username
This is the username that will be registered on the PB connection and need to be used when calling buildbot user.
passwd
This is the passwd that will be registered on the PB connection and need to be used when calling buildbot user.
port
The PB connection port must be different than c[‘protocols’][‘pb’][‘port’] and be specified when calling buildbot user
2.5.2.18. Input Validation
import re
c['validation'] = {
'branch' : re.compile(r'^[\w.+/~-]*$'),
'revision' : re.compile(r'^[ \w\.\-\/]*$'),
'property_name' : re.compile(r'^[\w\.\-\/\~:]*$'),
'property_value' : re.compile(r'^[\w\.\-\/\~:]*$'),
}
This option configures the validation applied to user inputs of various types. This validation is important since these values are often included in command-line arguments executed on workers. Allowing arbitrary input from untrusted users may raise security concerns.
The keys describe the type of input validated; the values are compiled regular expressions against which the input will be matched. The defaults for each type of input are those given in the example, above.
2.5.2.19. Revision Links
The revlink
parameter is used to create links from revision IDs in the web status to a web-view of your source control system.
The parameter’s value must be a callable.
By default, Buildbot is configured to generate revlinks for a number of open source hosting platforms (https://github.com, https://sourceforge.net and https://bitbucket.org).
The callable takes the revision id and repository argument, and should return a URL to the revision. Note that the revision id may not always be in the form you expect, so code defensively. In particular, a revision of “??” may be supplied when no other information is available.
Note that SourceStamp
s that are not created from version-control changes (e.g., those created by a Nightly
or Periodic
scheduler) may have an empty repository string if the repository is not known to the scheduler.
Revision Link Helpers
Buildbot provides two helpers for generating revision links.
buildbot.revlinks.RevlinkMatcher
takes a list of regular expressions and a replacement text.
The regular expressions should all have the same number of capture groups.
The replacement text should have sed-style references to that capture groups (i.e. ‘1’ for the first capture group), and a single ‘%s’ reference for the revision ID.
The repository given is tried against each regular expression in turn.
The results are then substituted into the replacement text, along with the revision ID, to obtain the revision link.
from buildbot.plugins import util
c['revlink'] = util.RevlinkMatch([r'git://notmuchmail.org/git/(.*)'],
r'http://git.notmuchmail.org/git/\1/commit/%s')
buildbot.revlinks.RevlinkMultiplexer
takes a list of revision link callables, and tries each in turn, returning the first successful match.
2.5.2.20. Codebase Generator
all_repositories = {
r'https://hg/hg/mailsuite/mailclient': 'mailexe',
r'https://hg/hg/mailsuite/mapilib': 'mapilib',
r'https://hg/hg/mailsuite/imaplib': 'imaplib',
r'https://github.com/mailinc/mailsuite/mailclient': 'mailexe',
r'https://github.com/mailinc/mailsuite/mapilib': 'mapilib',
r'https://github.com/mailinc/mailsuite/imaplib': 'imaplib',
}
def codebaseGenerator(chdict):
return all_repositories[chdict['repository']]
c['codebaseGenerator'] = codebaseGenerator
For any incoming change, the codebase is set to ‘’. This codebase value is sufficient if all changes come from the same repository (or clones). If changes come from different repositories, extra processing will be needed to determine the codebase for the incoming change. This codebase will then be a logical name for the combination of repository and or branch etc.
The codebaseGenerator accepts a change dictionary as produced by the buildbot.db.changes.ChangesConnectorComponent
, with a changeid equal to None.