AI prompts
base on A suite of data generation and load testing tools # `lading` - A tool for measuring the performance of long-running programs.
The `lading` project is a tool for measuring the performance behavior of
long-running programs -- daemons, often -- using synthetic, repeatable load
generation across a variety of protocols. The ambition is to be a worry-free
component of a larger performance testing strategy for complex programs. The
[Vector][vector] project uses lading in their 'soak' tests.
## Operating Model
`lading` operates on three conceptual components:
* generators,
* target and
* blackholes
A "generator" in `lading` is responsible for creating load and "pushing" it into
the target. The "target" in is the program that `lading` runs as a sub-process,
collecting resource consumption data about the target from the operating system
but also self-telemetry regarding generation. The "blackhole" exists for targets
to push load into, when necessary, allowing for a target to be run and examined
by `lading` without including external programs or APIs into the setup.
As much as possible `lading` pre-computes any generator and blackhole payloads,
intending to minimize runtime overhead compared to the target program. Users
must provide a seed to `lading` -- see below -- ensuring that payloads are
generated in a repeatable manner across `lading` runs. While pre-computation
does mean that `lading` is capable of outpacing many targets with minimal
runtime interference it does also mean higher memory use compared to other load
generation tools.
## Configuration
The configuration of `lading` is split between a configuration file in YAML
format and command-line options. Generators and blackholes are configured in the
config file, targets on the command line. This is, at first glance, awkward but
does allow for `lading` to be used in dynamic environments like CI without
foreknowledge of the target.
That "push" can be as direct as network IO into the target or as indirect as
doing file operations for a target who's tracing with BPF.
Consider the following `lading.yaml`:
```yaml
generator:
- http:
seed: [2, 3, 5, 7, 11, 13, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 101, 103, 107, 109, 113, 127, 131, 137]
target_uri: "http://localhost:8282/"
bytes_per_second: "100 Mb"
parallel_connections: 10
method:
post:
variant: "fluent"
maximum_prebuild_cache_size_bytes: "256 Mb"
headers: {}
blackhole:
- http:
binding_addr: "0.0.0.0:8080"
```
In this setup `lading` is configured to run one generator and one blackhole. In
general, at least one generator is required and zero or more blackholes are. The
generator here, named "http", uses a fixed seed to repeatably produce [fluent's
forward protocol][fluent] instances at 100 Mb per second into the target, with a
pre-built size of 256 Mb. That is, `lading` will _attempt_ to push at a fixed
throughput, although the target might not be able to cope with that load and
`lading` will consume 256 Mb of RAM to accommodate pre-build payloads. The
blackhole in this configuration responds with an empty body 200 OK.
`lading` supports three types of targets, binary launch mode, PID watch mode,
and container watch mode.
In binary launch mode, `lading` acts like a wrapper around the target. To use
this mode, one specifies where on disk the configuration is, the path to the
target and its arguments. `--target-stderr-path` and `--target-stdout-path`
allow the target's stdout, stderr to be forwarded into files. At runtime
`lading` will output diagnostic information on its stdout about the current
state of the target.
In PID watch mode, `lading` monitors a process that is already running on the
system. `lading` does not handle any target orchestration in this mode. The
target should be launched before running `lading` and terminated once `lading`
exits.
Container watch mode is identical to PID watch mode except that the docker socket
is polled to find the container's PID instead of the user needing to specify it manually.
The data `lading` captures at runtime can be pulled by polling `lading`'s
prometheus endpoint -- configurable with `--prometheus-addr` / `--prometheus-path`
-- or can be written to disk by lading by specifying `--capture-path`. The
captured data, when written to disk, is newline delimited json payloads.
Logs about lading's operation can be controlled using the standard `RUST_LOG`
environment variable. eg `RUST_LOG=debug ./lading` will emit logs at `debug`
and above (ie, `info`, `warn`, `error`).
## Development Setup
`lading` requires the protobuf compiler to build. See
[installation instructions](https://grpc.io/docs/protoc-installation/) from the
protobuf docs.
For criterion benchmarks, you can run them via `cargo bench`.
[`cargo-criterion`](https://github.com/bheisler/cargo-criterion)
is a more advanced cargo extension that provides
historical (ie baseline) tracking functionality.
When using `lading` to monitor a linux process that is more privileged than `lading` is
(eg, a container), some metrics are unavailable due to security restrictions.
To give `lading` extra permissions needed, `setcap` can be used to grant the binary
`cap_sys_ptrace`. Eg: `sudo setcap cap_sys_ptrace=ep ./target/debug/lading`
## Binaries in-tree
- `lading` -- the main tool capable of generating load, blackholes, observers,
etc. Everything discussed above.
- `captool` -- Useful to inspect the resulting `capture.json` file which
contains metrics generated by a lading run.
- `payloadtool` -- Useful when writing/changing experiments to validate that a
specified payload behaves as expected.
## How to get a build (besides building from source)
### Container Releases
https://github.com/DataDog/lading/pkgs/container/lading
### Docker
For arm (replace platform if needed):
```
docker build -f arm64.Dockerfile . --tag [TAG NAME] --platform linux/arm64
```
For x86_64:
```
docker build -f amd64.Dockerfile . --tag [TAG NAME] --platform linux/amd64
```
## Contributing
See [Contributing][contributing].
## License
Licensed under the [MIT license][mit-license].
[contributing]: CONTRIBUTING.md
[mit-license]: LICENSE
[vector]: github.com/vectordotdev/vector
[fluent]: https://github.com/fluent/fluentd/wiki/Forward-Protocol-Specification-v1
", Assign "at most 3 tags" to the expected json: {"id":"7437","tags":[]} "only from the tags list I provide: [{"id":77,"name":"3d"},{"id":89,"name":"agent"},{"id":17,"name":"ai"},{"id":54,"name":"algorithm"},{"id":24,"name":"api"},{"id":44,"name":"authentication"},{"id":3,"name":"aws"},{"id":27,"name":"backend"},{"id":60,"name":"benchmark"},{"id":72,"name":"best-practices"},{"id":39,"name":"bitcoin"},{"id":37,"name":"blockchain"},{"id":1,"name":"blog"},{"id":45,"name":"bundler"},{"id":58,"name":"cache"},{"id":21,"name":"chat"},{"id":49,"name":"cicd"},{"id":4,"name":"cli"},{"id":64,"name":"cloud-native"},{"id":48,"name":"cms"},{"id":61,"name":"compiler"},{"id":68,"name":"containerization"},{"id":92,"name":"crm"},{"id":34,"name":"data"},{"id":47,"name":"database"},{"id":8,"name":"declarative-gui "},{"id":9,"name":"deploy-tool"},{"id":53,"name":"desktop-app"},{"id":6,"name":"dev-exp-lib"},{"id":59,"name":"dev-tool"},{"id":13,"name":"ecommerce"},{"id":26,"name":"editor"},{"id":66,"name":"emulator"},{"id":62,"name":"filesystem"},{"id":80,"name":"finance"},{"id":15,"name":"firmware"},{"id":73,"name":"for-fun"},{"id":2,"name":"framework"},{"id":11,"name":"frontend"},{"id":22,"name":"game"},{"id":81,"name":"game-engine "},{"id":23,"name":"graphql"},{"id":84,"name":"gui"},{"id":91,"name":"http"},{"id":5,"name":"http-client"},{"id":51,"name":"iac"},{"id":30,"name":"ide"},{"id":78,"name":"iot"},{"id":40,"name":"json"},{"id":83,"name":"julian"},{"id":38,"name":"k8s"},{"id":31,"name":"language"},{"id":10,"name":"learning-resource"},{"id":33,"name":"lib"},{"id":41,"name":"linter"},{"id":28,"name":"lms"},{"id":16,"name":"logging"},{"id":76,"name":"low-code"},{"id":90,"name":"message-queue"},{"id":42,"name":"mobile-app"},{"id":18,"name":"monitoring"},{"id":36,"name":"networking"},{"id":7,"name":"node-version"},{"id":55,"name":"nosql"},{"id":57,"name":"observability"},{"id":46,"name":"orm"},{"id":52,"name":"os"},{"id":14,"name":"parser"},{"id":74,"name":"react"},{"id":82,"name":"real-time"},{"id":56,"name":"robot"},{"id":65,"name":"runtime"},{"id":32,"name":"sdk"},{"id":71,"name":"search"},{"id":63,"name":"secrets"},{"id":25,"name":"security"},{"id":85,"name":"server"},{"id":86,"name":"serverless"},{"id":70,"name":"storage"},{"id":75,"name":"system-design"},{"id":79,"name":"terminal"},{"id":29,"name":"testing"},{"id":12,"name":"ui"},{"id":50,"name":"ux"},{"id":88,"name":"video"},{"id":20,"name":"web-app"},{"id":35,"name":"web-server"},{"id":43,"name":"webassembly"},{"id":69,"name":"workflow"},{"id":87,"name":"yaml"}]" returns me the "expected json"