This Python script is designed to compare bandwidth data rates from different sources and output a uniform unit format.
❯ python3 network-rate-convert.py 400MB/s
3200.00 Mb/s
#!/bin/bash | |
set -euo pipefail # Exit on error, unset variables, and prevent errors in pipelines | |
# Usage information | |
usage() { | |
echo "Usage: $0 <path-to-git-repo> <output-file>" | |
exit 1 | |
} | |
# Check for mandatory arguments |
version: "2" | |
services: | |
kafka-0: | |
container_name: kafka-0 | |
image: docker.io/bitnami/kafka:3.5 | |
ports: | |
- "9092:9092" | |
environment: | |
- KAFKA_PROCESS_ROLES=broker,controller |
job "http_servers" { | |
datacenters = ["dc1"] | |
group "group" { | |
network { | |
mode = "bridge" | |
port "taskA" { | |
static = 6000 | |
} | |
port "taskB" { |
This issue is described in hashicorp/nomad#5459 but this is a really simple to reproduce version of the same.
nomad_sighup
is a really simple Go program, which listens for SIGHUP
signal to be sent and prints "Received SIGHUP signal". It listens for this signal in an infinite for-loop.
The nomad deployment specs, deploys the same binary with raw_exec
driver. There's a template
block defined, which is a dummy configuration template. It watches for a Nomad service object doggo-web
and updates whenever the Address/IP of this service updates. When that happens, Nomad is configured to send a SIGHUP
signal to the underlying process.
Query id: 583d15b7-e488-4876-88bf-c8b0b4246f5c | |
┌─symbol────────┬──────return_percent─┬─────────new.close─┬──────────────close─┐ | |
│ CGCL.NS │ 1.5731874145006763 │ 742.5 │ 731 │ | |
│ CHOLAHLDNG.NS │ -5.385943910536817 │ 604.2999877929688 │ 638.7000122070312 │ | |
│ GUJGASLTD.NS │ 7.777898149642426 │ 513.4000244140625 │ 476.3500061035156 │ | |
│ NAM-INDIA.NS │ -1.7775422037903055 │ 260.6499938964844 │ 265.36700439453125 │ | |
│ NCC.NS │ 9.978915420737767 │ 78.25 │ 71.1500015258789 │ | |
└───────────────┴─────────────────────┴───────────────────┴────────────────────┘ | |
┌─symbol────────┬───────return_percent─┬──────────new.close─┬──────────────close─┐ |
settings.json
: "files.associations": {
"*.nomad": "hcl",
"*.nomad.tpl": "hcl",
"*.tf": "terraform",
job "hello-world" { | |
datacenters = ["dc1"] | |
namespace = "default" | |
type = "service" | |
group "redis" { | |
# Specify number of replicas of redis needed. | |
count = 1 | |
# Specify networking for the group, port allocs. |
thread 'vector-worker' panicked at 'Tried to ack beyond read offset', /project/lib/vector-core/buffers/src/disk/leveldb_buffer/reader.rs:184:13 | |
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace | |
thread 'vector-worker' panicked at 'Tried to ack beyond read offset', /project/lib/vector-core/buffers/src/disk/leveldb_buffer/reader.rs:184:13 | |
stack backtrace: | |
thread 'vector-worker' panicked at 'Tried to ack beyond read offset', /project/lib/vector-core/buffers/src/disk/leveldb_buffer/reader.rs:184:13 | |
thread 'vector-worker' panicked at 'Tried to ack beyond read offset', /project/lib/vector-core/buffers/src/disk/leveldb_buffer/reader.rs:184:13 | |
0: 0x55df1f0de430 - std::backtrace_rs::backtrace::libunwind::trace::h34055254b57d8e79 | |
at /rustc/a178d0322ce20e33eac124758e837cbd80a6f633/library/std/src/../../backtrace/src/backtrace/libunwind.rs:90:5 | |
1: 0x55df1f0de430 - std::backtrace_rs::backtrace::trace_unsynchronized::h8f1e3fbd9afff6ec | |
FROM alpine:latest | |
COPY ./arg.sh / | |
ENTRYPOINT ["/arg.sh"] |