summaryrefslogtreecommitdiffstats
path: root/fluent-bit/lib/librdkafka-2.1.0/packaging
diff options
context:
space:
mode:
Diffstat (limited to 'fluent-bit/lib/librdkafka-2.1.0/packaging')
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/RELEASE.md311
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/alpine/build-alpine.sh38
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/archlinux/PKGBUILD30
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Config.cmake.in37
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindLZ4.cmake38
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindZSTD.cmake27
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/LICENSE.FindZstd178
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/README.md38
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/config.h.in52
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/parseversion.cmake60
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/rdkafka.pc.in12
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_32_test.c8
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_64_test.c8
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/c11threads_test.c14
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/crc32c_hw_test.c27
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/dlopen_test.c11
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/libsasl2_test.c7
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_darwin_test.c6
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_freebsd_test.c7
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_gnu_test.c5
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rand_r_test.c7
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rdkafka_setup.cmake122
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/regex_test.c10
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/strndup_test.c5
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_32_test.c8
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_64_test.c8
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cp/README.md14
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/cp/check_features.c64
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-deb.sh34
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-packages.sh43
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-rpm.sh38
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/.gitignore6
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/changelog66
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/compat1
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/control49
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/copyright84
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/docs5
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/gbp.conf9
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.dirs2
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.examples2
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.install6
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.substvars1
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka.dsc16
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1-dbg.substvars1
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.dirs1
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.install2
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postinst.debhelper5
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postrm.debhelper5
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.symbols64
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/debian/rules19
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/source/format1
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/debian/watch2
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/get_version.py21
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/README.md15
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/brew-update-pr.sh31
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw-static.sh52
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw.sh21
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/run-tests.sh6
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/semaphoreci-build.sh38
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/travis-before-install.sh20
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/.gitignore7
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/README.md78
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/nuget/artifact.py177
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/nuget/cleanup-s3.py143
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr120.zipbin679055 -> 0 bytes
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr140.zipbin516022 -> 0 bytes
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr120.zipbin662837 -> 0 bytes
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr140.zipbin621912 -> 0 bytes
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nuget.sh21
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nugetpackage.py286
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/nuget/packaging.py448
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/nuget/push-to-nuget.sh21
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/nuget/release.py167
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/requirements.txt3
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/staticpackage.py178
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.nuspec21
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.props18
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.targets19
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/__init__.py0
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/zfile.py98
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/.gitignore7
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/Makefile92
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/README.md23
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/el7-x86_64.cfg40
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/librdkafka.spec118
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/rpm/mock-on-docker.sh97
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/.gitignore2
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/Makefile25
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/README.md8
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/run-test.sh49
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test-on-docker.sh56
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.c77
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.cpp34
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-deb-package.sh64
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-debian.sh65
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-manylinux.sh68
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-release-artifacts.sh138
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/distro-build.sh38
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/gh-release-checksums.py39
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/rdutcoverage.sh25
-rw-r--r--fluent-bit/lib/librdkafka-2.1.0/packaging/tools/requirements.txt2
-rwxr-xr-xfluent-bit/lib/librdkafka-2.1.0/packaging/tools/style-format.sh148
102 files changed, 0 insertions, 4718 deletions
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/RELEASE.md b/fluent-bit/lib/librdkafka-2.1.0/packaging/RELEASE.md
deleted file mode 100644
index 930636db4..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/RELEASE.md
+++ /dev/null
@@ -1,311 +0,0 @@
-# librdkafka release process
-
-This guide outlines the steps needed to release a new version of librdkafka
-and publish packages to channels (NuGet, Homebrew, etc,..).
-
-Releases are done in two phases:
- * release-candidate(s) - RC1 will be the first release candidate, and any
- changes to the repository will require a new RC.
- * final release - the final release is based directly on the last RC tag
- followed by a single version-bump commit (see below).
-
-Release tag and version format:
- * tagged release builds to verify CI release builders: vA.B.C-PREn
- * release-candidate: vA.B.C-RCn
- * final release: vA.B.C
-
-
-## Update protocol requests and error codes
-
-Check out the latest version of Apache Kafka (not trunk, needs to be a released
-version since protocol may change on trunk).
-
-### Protocol request types
-
-Generate protocol request type codes with:
-
- $ src/generate_proto.sh ~/src/your-kafka-dir
-
-Cut'n'paste the new defines and strings to `rdkafka_protocol.h` and
-`rdkafka_proto.h`.
-
-### Error codes
-
-Error codes must currently be parsed manually, open
-`clients/src/main/java/org/apache/kafka/common/protocol/Errors.java`
-in the Kafka source directory and update the `rd_kafka_resp_err_t` and
-`RdKafka::ErrorCode` enums in `rdkafka.h` and `rdkafkacpp.h`
-respectively.
-Add the error strings to `rdkafka.c`.
-The Kafka error strings are sometimes a bit too verbose for our taste,
-so feel free to rewrite them (usually removing a couple of 'the's).
-Error strings must not contain a trailing period.
-
-**NOTE**: Only add **new** error codes, do not alter existing ones since that
- will be a breaking API change.
-
-
-## Run regression tests
-
-**Build tests:**
-
- $ cd tests
- $ make -j build
-
-**Run the full regression test suite:** (requires Linux and the trivup python package)
-
- $ make full
-
-
-If all tests pass, carry on, otherwise identify and fix bug and start over.
-
-
-
-## Write release notes / changelog
-
-All relevant PRs should also include an update to [CHANGELOG.md](../CHANGELOG.md)
-that in a user-centric fashion outlines what changed.
-It might not be practical for all contributors to write meaningful changelog
-entries, so it is okay to add them separately later after the PR has been
-merged (make sure to credit community contributors for their work).
-
-The changelog should include:
- * What type of release (maintenance or feature release)
- * A short intro to the release, describing the type of release: maintenance
- or feature release, as well as fix or feature high-lights.
- * A section of **New features**, if any.
- * A section of **Upgrade considerations**, if any, to outline important changes
- that require user attention.
- * A section of **Enhancements**, if any.
- * A section of **Fixes**, if any, preferably with Consumer, Producer, and
- Generic sub-sections.
-
-
-## Pre-release code tasks
-
-**Switch to the release branch which is of the format `A.B.C.x` or `A.B.x`.**
-
- $ git checkout -b 0.11.1.x
-
-
-**Update in-code versions.**
-
-The last octet in the version hex number is the pre-build/release-candidate
-number, where 0xAABBCCff is the final release for version 0xAABBCC.
-Release candidates start at 200, thus 0xAABBCCc9 is RC1, 0xAABBCCca is RC2, etc.
-
-Change the `RD_KAFKA_VERSION` defines in both `src/rdkafka.h` and
-`src-cpp/rdkafkacpp.h` to the version to build, such as 0x000b01c9
-for v0.11.1-RC1, or 0x000b01ff for the final v0.11.1 release.
-Update the librdkafka version in `vcpkg.json`.
-
- # Update defines
- $ $EDITOR src/rdkafka.h src-cpp/rdkafkacpp.h vcpkg.json
-
- # Reconfigure and build
- $ ./configure
- $ make
-
- # Check git diff for correctness
- $ git diff
-
- # Commit
- $ git commit -m "Version v0.11.1-RC1" src/rdkafka.h src-cpp/rdkafkacpp.h
-
-
-**Create tag.**
-
- $ git tag v0.11.1-RC1 # for an RC
- # or for the final release:
- $ git tag v0.11.1 # for the final release
-
-
-**Push branch and commit to github**
-
- # Dry-run first to make sure things look correct
- $ git push --dry-run origin 0.11.1.x
-
- # Live
- $ git push origin 0.11.1.x
-**Push tags and commit to github**
-
- # Dry-run first to make sure things look correct.
- $ git push --dry-run --tags origin v0.11.1-RC1
-
- # Live
- $ git push --tags origin v0.11.1-RC1
-
-
-## Creating packages
-
-As soon as a tag is pushed the CI system (SemaphoreCI) will start its
-build pipeline and eventually upload packaging artifacts to the SemaphoreCI
-project artifact store.
-
-Monitor the Semaphore CI project page to know when the build pipeline
-is finished, then download the relevant artifacts for further use, see
-*The artifact pipeline* chapter below.
-
-
-## Publish release on github
-
-Create a release on github by going to https://github.com/edenhill/librdkafka/releases
-and Draft a new release.
-Name the release the same as the final release tag (e.g., `v1.9.0`) and set
-the tag to the same.
-Paste the CHANGELOG.md section for this release into the release description,
-look at the preview and fix any formatting issues.
-
-Run the following command to get checksums of the github release assets:
-
- $ packaging/tools/gh-release-checksums.py <the-tag>
-
-It will take some time for the script to download the files, when done
-paste the output to the end of the release page.
-
-Make sure the release page looks okay, is still correct (check for new commits),
-and has the correct tag, then click Publish release.
-
-
-
-### Homebrew recipe update
-
-**Note**: This is typically not needed since homebrew seems to pick up new
- release versions quickly enough. Recommend you skip this step.
-
-The brew-update-pr.sh script automatically pushes a PR to homebrew-core
-with a patch to update the librdkafka version of the formula.
-This should only be done for final releases and not release candidates.
-
-On a MacOSX host with homebrew installed:
-
- $ cd package/homebrew
- # Dry-run first to see that things are okay.
- $ ./brew-update-pr.sh v0.11.1
- # If everything looks good, do the live push:
- $ ./brew-update-pr.sh --upload v0.11.1
-
-
-### Deb and RPM packaging
-
-Debian and RPM packages are generated by Confluent packaging, called
-Independent client releases, which is a separate non-public process and the
-resulting packages are made available on Confluent's client deb and rpm
-repositories.
-
-That process is outside the scope of this document.
-
-See the Confluent docs for instructions how to access these packages:
-https://docs.confluent.io/current/installation.html
-
-
-
-
-## Build and release artifacts
-
-The following chapter explains what, how, and where artifacts are built.
-It also outlines where these artifacts are used.
-
-### So what is an artifact?
-
-An artifact is a build of the librdkafka library, dynamic/shared and/or static,
-with a certain set of external or built-in dependencies, for a specific
-architecture and operating system (and sometimes even operating system version).
-
-If you build librdkafka from source with no special `./configure` arguments
-you will end up with:
-
- * a dynamically linked library (e.g., `librdkafka.so.1`)
- with a set of dynamically linked external dependencies (OpenSSL, zlib, etc),
- all depending on what dependencies are available on the build host.
-
- * a static library (`librdkafka.a`) that will have external dependencies
- that needs to be linked dynamically. There is no way for a static library
- to express link dependencies, so there will also be `rdkafka-static.pc`
- pkg-config file generated that contains linker flags for the external
- dependencies.
- Those external dependencies are however most likely only available on the
- build host, so this static library is not particularily useful for
- repackaging purposes (such as for high-level clients using librdkafka).
-
- * a self-contained static-library (`librdkafka-static.a`) which attempts
- to contain static versions of all external dependencies, effectively making
- it possible to link just with `librdkafka-static.a` to get all
- dependencies needed.
- Since the state of static libraries in the various distro and OS packaging
- systems is of varying quality and availability, it is usually not possible
- for the librdkafka build system (mklove) to generate this completely
- self-contained static library simply using dependencies available on the
- build system, and the make phase of the build will emit warnings when it
- can't bundle all external dependencies due to this.
- To circumvent this problem it is possible for the build system (mklove)
- to download and build static libraries of all needed external dependencies,
- which in turn allows it to create a complete bundle of all dependencies.
- This results in a `librdkafka-static.a` that has no external dependecies
- other than the system libraries (libc, pthreads, rt, etc).
- To achieve this you will need to pass
- `--install-deps --source-deps-only --enable-static` to
- librdkafka's `./configure`.
-
- * `rdkafka.pc` and `rdkafka-static.pc` pkg-config files that tells
- applications and libraries that depend on librdkafka what external
- dependencies are needed to successfully link with librdkafka.
- This is mainly useful for the dynamic librdkafka librdkafka
- (`librdkafka.so.1` or `librdkafka.1.dylib` on OSX).
-
-
-**NOTE**: Due to libsasl2/cyrus-sasl's dynamically loaded plugins, it is
-not possible for us to provide a self-contained static library with
-GSSAPI/Kerberos support.
-
-
-
-### The artifact pipeline
-
-We rely solely on CI systems to build our artifacts; no artifacts must be built
-on a non-CI system (e.g., someones work laptop, some random ec2 instance, etc).
-
-The reasons for this are:
-
- 1. Reproducible builds: we want a well-defined environment that doesn't change
- (too much) without notice and that we can rebuild artifacts on at a later
- time if required.
- 2. Security; these CI systems provide at least some degree of security
- guarantees, and they're managed by people who knows what they're doing
- most of the time. This minimizes the risk for an artifact to be silently
- compromised due to the developer's laptop being hacked.
- 3. Logs; we have build logs for all artifacts, which contains checksums.
- This way we can know how an artifact was built, what features were enabled
- and what versions of dependencies were used, as well as know that an
- artifact has not been tampered with after leaving the CI system.
-
-
-By default the CI jobs are triggered by branch pushes and pull requests
-and contain a set of jobs to validate that the changes that were pushed does
-not break compilation or functionality (by running parts of the test suite).
-These jobs do not produce any artifacts.
-
-
-For the artifact pipeline there's tag builds, which are triggered by pushing a
-tag to the git repository.
-These tag builds will generate artifacts which are used by the same pipeline
-to create NuGet and static library packages, which are then uploaded to
-SemaphoreCI's project artifact store.
-
-Once a tag build pipeline is done, you can download the relevant packages
-from the Semaphore CI project artifact store.
-
-The NuGet package, `librdkafka.redist.<version>.nupkg`, needs to be
-manually uploaded to NuGet.
-
-The `librdkafka-static-bundle-<version>.tgz` static library bundle
-needs to be manually imported into the confluent-kafka-go client using the
-import script that resides in the Go client repository.
-
-
-**Note**: You will need a NuGet API key to upload nuget packages.
-
-
-See [nuget/nugetpackaging.py] and [nuget/staticpackaging.py] to see how
-packages are assembled from build artifacts.
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/alpine/build-alpine.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/alpine/build-alpine.sh
deleted file mode 100755
index e6d2471c9..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/alpine/build-alpine.sh
+++ /dev/null
@@ -1,38 +0,0 @@
-#!/bin/sh
-#
-# Build librdkafka on Alpine.
-#
-
-set -x
-
-if [ "$1" = "--in-docker" ]; then
- # Runs in docker, performs the actual build.
- shift
-
- apk add bash curl gcc g++ make musl-dev linux-headers bsd-compat-headers git python3 perl patch
-
- git clone /v /librdkafka
-
- cd /librdkafka
- ./configure --install-deps --disable-gssapi --disable-lz4-ext --enable-static $*
- make -j
- examples/rdkafka_example -X builtin.features
- CI=true make -C tests run_local_quick
-
- # Create a tarball in artifacts/
- cd src
- ldd librdkafka.so.1
- tar cvzf /v/artifacts/alpine-librdkafka.tgz librdkafka.so.1 librdkafka*.a rdkafka-static.pc
- cd ../..
-
-else
- # Runs on the host, simply spins up the in-docker build.
- if [ ! -f configure.self ]; then
- echo "Must be run from the top-level librdkafka dir"
- exit 1
- fi
-
- mkdir -p artifacts
-
- exec docker run -v $PWD:/v alpine:3.12 /v/packaging/alpine/$(basename $0) --in-docker $*
-fi
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/archlinux/PKGBUILD b/fluent-bit/lib/librdkafka-2.1.0/packaging/archlinux/PKGBUILD
deleted file mode 100644
index 7063d5cef..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/archlinux/PKGBUILD
+++ /dev/null
@@ -1,30 +0,0 @@
-pkgname=librdkafka
-pkgver=1.0.0.RC5.r11.g3cf68480
-pkgrel=1
-pkgdesc='The Apache Kafka C/C++ client library'
-url='https://github.com/edenhill/librdkafka'
-license=('BSD')
-arch=('x86_64')
-source=('git+https://github.com/edenhill/librdkafka#branch=master')
-sha256sums=('SKIP')
-depends=(glibc libsasl lz4 openssl zlib zstd)
-makedepends=(bash git python3)
-
-pkgver() {
- cd "$pkgname"
- git describe --long --tags --match "v[0-9]*" | sed 's/^v//;s/\([^-]*-g\)/r\1/;s/-/./g'
-}
-
-build() {
- cd "$pkgname"
- ./configure --prefix=/usr
- make
-}
-
-package() {
- cd "$pkgname"
- make install DESTDIR="$pkgdir"
- for f in $(find -type f -name 'LICENSE*'); do
- install -D -m0644 "$f" "$pkgdir/usr/share/licenses/$pkgname/$f"
- done
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Config.cmake.in b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Config.cmake.in
deleted file mode 100644
index 8a6522b06..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Config.cmake.in
+++ /dev/null
@@ -1,37 +0,0 @@
-@PACKAGE_INIT@
-
-include(CMakeFindDependencyMacro)
-
-if(@WITH_ZLIB@)
- find_dependency(ZLIB)
-endif()
-
-if(@WITH_CURL@)
- find_dependency(CURL)
-endif()
-
-if(@WITH_ZSTD@)
- find_library(ZSTD zstd)
- if(NOT ZSTD)
- message(ERROR "ZSTD library not found!")
- else()
- message(STATUS "Found ZSTD: " ${ZSTD})
- endif()
-endif()
-
-if(@WITH_SSL@)
- if(@WITH_BUNDLED_SSL@)
- # TODO: custom SSL library should be installed
- else()
- find_dependency(OpenSSL)
- endif()
-endif()
-
-if(@WITH_LZ4_EXT@)
- find_dependency(LZ4)
-endif()
-
-find_dependency(Threads)
-
-include("${CMAKE_CURRENT_LIST_DIR}/@targets_export_name@.cmake")
-check_required_components("@PROJECT_NAME@")
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindLZ4.cmake b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindLZ4.cmake
deleted file mode 100644
index 594c4290c..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindLZ4.cmake
+++ /dev/null
@@ -1,38 +0,0 @@
-find_path(LZ4_INCLUDE_DIR
- NAMES lz4.h
- DOC "lz4 include directory")
-mark_as_advanced(LZ4_INCLUDE_DIR)
-find_library(LZ4_LIBRARY
- NAMES lz4
- DOC "lz4 library")
-mark_as_advanced(LZ4_LIBRARY)
-
-if (LZ4_INCLUDE_DIR)
- file(STRINGS "${LZ4_INCLUDE_DIR}/lz4.h" _lz4_version_lines
- REGEX "#define[ \t]+LZ4_VERSION_(MAJOR|MINOR|RELEASE)")
- string(REGEX REPLACE ".*LZ4_VERSION_MAJOR *\([0-9]*\).*" "\\1" _lz4_version_major "${_lz4_version_lines}")
- string(REGEX REPLACE ".*LZ4_VERSION_MINOR *\([0-9]*\).*" "\\1" _lz4_version_minor "${_lz4_version_lines}")
- string(REGEX REPLACE ".*LZ4_VERSION_RELEASE *\([0-9]*\).*" "\\1" _lz4_version_release "${_lz4_version_lines}")
- set(LZ4_VERSION "${_lz4_version_major}.${_lz4_version_minor}.${_lz4_version_release}")
- unset(_lz4_version_major)
- unset(_lz4_version_minor)
- unset(_lz4_version_release)
- unset(_lz4_version_lines)
-endif ()
-
-include(FindPackageHandleStandardArgs)
-find_package_handle_standard_args(LZ4
- REQUIRED_VARS LZ4_LIBRARY LZ4_INCLUDE_DIR
- VERSION_VAR LZ4_VERSION)
-
-if (LZ4_FOUND)
- set(LZ4_INCLUDE_DIRS "${LZ4_INCLUDE_DIR}")
- set(LZ4_LIBRARIES "${LZ4_LIBRARY}")
-
- if (NOT TARGET LZ4::LZ4)
- add_library(LZ4::LZ4 UNKNOWN IMPORTED)
- set_target_properties(LZ4::LZ4 PROPERTIES
- IMPORTED_LOCATION "${LZ4_LIBRARY}"
- INTERFACE_INCLUDE_DIRECTORIES "${LZ4_INCLUDE_DIR}")
- endif ()
-endif ()
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindZSTD.cmake b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindZSTD.cmake
deleted file mode 100644
index 7de137e0f..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/FindZSTD.cmake
+++ /dev/null
@@ -1,27 +0,0 @@
-#
-# - Try to find Facebook zstd library
-# This will define
-# ZSTD_FOUND
-# ZSTD_INCLUDE_DIR
-# ZSTD_LIBRARY
-#
-
-find_path(ZSTD_INCLUDE_DIR NAMES zstd.h)
-
-find_library(ZSTD_LIBRARY_DEBUG NAMES zstdd zstd_staticd)
-find_library(ZSTD_LIBRARY_RELEASE NAMES zstd zstd_static)
-
-include(SelectLibraryConfigurations)
-SELECT_LIBRARY_CONFIGURATIONS(ZSTD)
-
-include(FindPackageHandleStandardArgs)
-FIND_PACKAGE_HANDLE_STANDARD_ARGS(
- ZSTD DEFAULT_MSG
- ZSTD_LIBRARY ZSTD_INCLUDE_DIR
-)
-
-if (ZSTD_FOUND)
- message(STATUS "Found Zstd: ${ZSTD_LIBRARY}")
-endif()
-
-mark_as_advanced(ZSTD_INCLUDE_DIR ZSTD_LIBRARY)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/LICENSE.FindZstd b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/LICENSE.FindZstd
deleted file mode 100644
index 9561f469b..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/Modules/LICENSE.FindZstd
+++ /dev/null
@@ -1,178 +0,0 @@
-FindZstd.cmake: git@github.com:facebook/folly.git 87f1a403b49552dae75ae94c8610dd5979913477
-
- Apache License
- Version 2.0, January 2004
- http://www.apache.org/licenses/
-
- TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
-
- 1. Definitions.
-
- "License" shall mean the terms and conditions for use, reproduction,
- and distribution as defined by Sections 1 through 9 of this document.
-
- "Licensor" shall mean the copyright owner or entity authorized by
- the copyright owner that is granting the License.
-
- "Legal Entity" shall mean the union of the acting entity and all
- other entities that control, are controlled by, or are under common
- control with that entity. For the purposes of this definition,
- "control" means (i) the power, direct or indirect, to cause the
- direction or management of such entity, whether by contract or
- otherwise, or (ii) ownership of fifty percent (50%) or more of the
- outstanding shares, or (iii) beneficial ownership of such entity.
-
- "You" (or "Your") shall mean an individual or Legal Entity
- exercising permissions granted by this License.
-
- "Source" form shall mean the preferred form for making modifications,
- including but not limited to software source code, documentation
- source, and configuration files.
-
- "Object" form shall mean any form resulting from mechanical
- transformation or translation of a Source form, including but
- not limited to compiled object code, generated documentation,
- and conversions to other media types.
-
- "Work" shall mean the work of authorship, whether in Source or
- Object form, made available under the License, as indicated by a
- copyright notice that is included in or attached to the work
- (an example is provided in the Appendix below).
-
- "Derivative Works" shall mean any work, whether in Source or Object
- form, that is based on (or derived from) the Work and for which the
- editorial revisions, annotations, elaborations, or other modifications
- represent, as a whole, an original work of authorship. For the purposes
- of this License, Derivative Works shall not include works that remain
- separable from, or merely link (or bind by name) to the interfaces of,
- the Work and Derivative Works thereof.
-
- "Contribution" shall mean any work of authorship, including
- the original version of the Work and any modifications or additions
- to that Work or Derivative Works thereof, that is intentionally
- submitted to Licensor for inclusion in the Work by the copyright owner
- or by an individual or Legal Entity authorized to submit on behalf of
- the copyright owner. For the purposes of this definition, "submitted"
- means any form of electronic, verbal, or written communication sent
- to the Licensor or its representatives, including but not limited to
- communication on electronic mailing lists, source code control systems,
- and issue tracking systems that are managed by, or on behalf of, the
- Licensor for the purpose of discussing and improving the Work, but
- excluding communication that is conspicuously marked or otherwise
- designated in writing by the copyright owner as "Not a Contribution."
-
- "Contributor" shall mean Licensor and any individual or Legal Entity
- on behalf of whom a Contribution has been received by Licensor and
- subsequently incorporated within the Work.
-
- 2. Grant of Copyright License. Subject to the terms and conditions of
- this License, each Contributor hereby grants to You a perpetual,
- worldwide, non-exclusive, no-charge, royalty-free, irrevocable
- copyright license to reproduce, prepare Derivative Works of,
- publicly display, publicly perform, sublicense, and distribute the
- Work and such Derivative Works in Source or Object form.
-
- 3. Grant of Patent License. Subject to the terms and conditions of
- this License, each Contributor hereby grants to You a perpetual,
- worldwide, non-exclusive, no-charge, royalty-free, irrevocable
- (except as stated in this section) patent license to make, have made,
- use, offer to sell, sell, import, and otherwise transfer the Work,
- where such license applies only to those patent claims licensable
- by such Contributor that are necessarily infringed by their
- Contribution(s) alone or by combination of their Contribution(s)
- with the Work to which such Contribution(s) was submitted. If You
- institute patent litigation against any entity (including a
- cross-claim or counterclaim in a lawsuit) alleging that the Work
- or a Contribution incorporated within the Work constitutes direct
- or contributory patent infringement, then any patent licenses
- granted to You under this License for that Work shall terminate
- as of the date such litigation is filed.
-
- 4. Redistribution. You may reproduce and distribute copies of the
- Work or Derivative Works thereof in any medium, with or without
- modifications, and in Source or Object form, provided that You
- meet the following conditions:
-
- (a) You must give any other recipients of the Work or
- Derivative Works a copy of this License; and
-
- (b) You must cause any modified files to carry prominent notices
- stating that You changed the files; and
-
- (c) You must retain, in the Source form of any Derivative Works
- that You distribute, all copyright, patent, trademark, and
- attribution notices from the Source form of the Work,
- excluding those notices that do not pertain to any part of
- the Derivative Works; and
-
- (d) If the Work includes a "NOTICE" text file as part of its
- distribution, then any Derivative Works that You distribute must
- include a readable copy of the attribution notices contained
- within such NOTICE file, excluding those notices that do not
- pertain to any part of the Derivative Works, in at least one
- of the following places: within a NOTICE text file distributed
- as part of the Derivative Works; within the Source form or
- documentation, if provided along with the Derivative Works; or,
- within a display generated by the Derivative Works, if and
- wherever such third-party notices normally appear. The contents
- of the NOTICE file are for informational purposes only and
- do not modify the License. You may add Your own attribution
- notices within Derivative Works that You distribute, alongside
- or as an addendum to the NOTICE text from the Work, provided
- that such additional attribution notices cannot be construed
- as modifying the License.
-
- You may add Your own copyright statement to Your modifications and
- may provide additional or different license terms and conditions
- for use, reproduction, or distribution of Your modifications, or
- for any such Derivative Works as a whole, provided Your use,
- reproduction, and distribution of the Work otherwise complies with
- the conditions stated in this License.
-
- 5. Submission of Contributions. Unless You explicitly state otherwise,
- any Contribution intentionally submitted for inclusion in the Work
- by You to the Licensor shall be under the terms and conditions of
- this License, without any additional terms or conditions.
- Notwithstanding the above, nothing herein shall supersede or modify
- the terms of any separate license agreement you may have executed
- with Licensor regarding such Contributions.
-
- 6. Trademarks. This License does not grant permission to use the trade
- names, trademarks, service marks, or product names of the Licensor,
- except as required for reasonable and customary use in describing the
- origin of the Work and reproducing the content of the NOTICE file.
-
- 7. Disclaimer of Warranty. Unless required by applicable law or
- agreed to in writing, Licensor provides the Work (and each
- Contributor provides its Contributions) on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
- implied, including, without limitation, any warranties or conditions
- of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
- PARTICULAR PURPOSE. You are solely responsible for determining the
- appropriateness of using or redistributing the Work and assume any
- risks associated with Your exercise of permissions under this License.
-
- 8. Limitation of Liability. In no event and under no legal theory,
- whether in tort (including negligence), contract, or otherwise,
- unless required by applicable law (such as deliberate and grossly
- negligent acts) or agreed to in writing, shall any Contributor be
- liable to You for damages, including any direct, indirect, special,
- incidental, or consequential damages of any character arising as a
- result of this License or out of the use or inability to use the
- Work (including but not limited to damages for loss of goodwill,
- work stoppage, computer failure or malfunction, or any and all
- other commercial damages or losses), even if such Contributor
- has been advised of the possibility of such damages.
-
- 9. Accepting Warranty or Additional Liability. While redistributing
- the Work or Derivative Works thereof, You may choose to offer,
- and charge a fee for, acceptance of support, warranty, indemnity,
- or other liability obligations and/or rights consistent with this
- License. However, in accepting such obligations, You may act only
- on Your own behalf and on Your sole responsibility, not on behalf
- of any other Contributor, and only if You agree to indemnify,
- defend, and hold each Contributor harmless for any liability
- incurred by, or claims asserted against, such Contributor by reason
- of your accepting any such warranty or additional liability.
-
- END OF TERMS AND CONDITIONS
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/README.md b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/README.md
deleted file mode 100644
index 47ad2cb63..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/README.md
+++ /dev/null
@@ -1,38 +0,0 @@
-# Build librdkafka with cmake
-
-The cmake build mode is experimental and not officially supported,
-the community is asked to maintain and support this mode through PRs.
-
-Set up build environment (from top-level librdkafka directory):
-
- $ cmake -H. -B_cmake_build
-
-On MacOSX and OpenSSL from Homebrew you might need to do:
-
- $ cmake -H. -B_cmake_build -DOPENSSL_ROOT_DIR=/usr/local/opt/openssl
-
-
-Build the library:
-
- $ cmake --build _cmake_build
-
-If you want to build static library:
-
- $ cmake --build _cmake_build -DRDKAFKA_BUILD_STATIC=1
-
-
-Run (local) tests:
-
- $ (cd _cmake_build && ctest -VV -R RdKafkaTestBrokerLess)
-
-
-Install library:
-
- $ cmake --build _cmake_build --target install
-
-
-If you use librdkafka as submodule in cmake project and want static link of librdkafka:
-
- set(RDKAFKA_BUILD_STATIC ON CACHE BOOL "")
- add_subdirectory(librdkafka)
- target_link_libraries(your_library_or_executable rdkafka)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/config.h.in b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/config.h.in
deleted file mode 100644
index 9e356c5f9..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/config.h.in
+++ /dev/null
@@ -1,52 +0,0 @@
-#cmakedefine01 WITHOUT_OPTIMIZATION
-#cmakedefine01 ENABLE_DEVEL
-#cmakedefine01 ENABLE_REFCNT_DEBUG
-
-#cmakedefine01 HAVE_ATOMICS_32
-#cmakedefine01 HAVE_ATOMICS_32_SYNC
-
-#if (HAVE_ATOMICS_32)
-# if (HAVE_ATOMICS_32_SYNC)
-# define ATOMIC_OP32(OP1,OP2,PTR,VAL) __sync_ ## OP1 ## _and_ ## OP2(PTR, VAL)
-# else
-# define ATOMIC_OP32(OP1,OP2,PTR,VAL) __atomic_ ## OP1 ## _ ## OP2(PTR, VAL, __ATOMIC_SEQ_CST)
-# endif
-#endif
-
-#cmakedefine01 HAVE_ATOMICS_64
-#cmakedefine01 HAVE_ATOMICS_64_SYNC
-
-#if (HAVE_ATOMICS_64)
-# if (HAVE_ATOMICS_64_SYNC)
-# define ATOMIC_OP64(OP1,OP2,PTR,VAL) __sync_ ## OP1 ## _and_ ## OP2(PTR, VAL)
-# else
-# define ATOMIC_OP64(OP1,OP2,PTR,VAL) __atomic_ ## OP1 ## _ ## OP2(PTR, VAL, __ATOMIC_SEQ_CST)
-# endif
-#endif
-
-#cmakedefine01 WITH_PKGCONFIG
-#cmakedefine01 WITH_HDRHISTOGRAM
-#cmakedefine01 WITH_ZLIB
-#cmakedefine01 WITH_CURL
-#cmakedefine01 WITH_OAUTHBEARER_OIDC
-#cmakedefine01 WITH_ZSTD
-#cmakedefine01 WITH_LIBDL
-#cmakedefine01 WITH_PLUGINS
-#define WITH_SNAPPY 1
-#define WITH_SOCKEM 1
-#cmakedefine01 WITH_SSL
-#cmakedefine01 WITH_SASL
-#cmakedefine01 WITH_SASL_SCRAM
-#cmakedefine01 WITH_SASL_OAUTHBEARER
-#cmakedefine01 WITH_SASL_CYRUS
-#cmakedefine01 WITH_LZ4_EXT
-#cmakedefine01 HAVE_REGEX
-#cmakedefine01 HAVE_STRNDUP
-#cmakedefine01 HAVE_RAND_R
-#cmakedefine01 HAVE_PTHREAD_SETNAME_GNU
-#cmakedefine01 HAVE_PTHREAD_SETNAME_DARWIN
-#cmakedefine01 HAVE_PTHREAD_SETNAME_FREEBSD
-#cmakedefine01 WITH_C11THREADS
-#cmakedefine01 WITH_CRC32C_HW
-#define SOLIB_EXT "${CMAKE_SHARED_LIBRARY_SUFFIX}"
-#define BUILT_WITH "${BUILT_WITH}"
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/parseversion.cmake b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/parseversion.cmake
deleted file mode 100644
index 592e8df54..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/parseversion.cmake
+++ /dev/null
@@ -1,60 +0,0 @@
-# hex2dec(<out-var> <input>):
-# Convert a hexadecimal value <input> to decimal and write the result
-# to <out-var>.
-macro(hex2dec var val)
- set(${var} 0)
-
- set(hex2dec_idx 0)
- string(LENGTH "${val}" hex2dec_len)
-
- while(hex2dec_idx LESS hex2dec_len)
- string(SUBSTRING ${val} ${hex2dec_idx} 1 hex2dec_char)
-
- if(hex2dec_char MATCHES "[0-9]")
- set(hex2dec_char ${hex2dec_char})
- elseif(hex2dec_char MATCHES "[aA]")
- set(hex2dec_char 10)
- elseif(hex2dec_char MATCHES "[bB]")
- set(hex2dec_char 11)
- elseif(hex2dec_char MATCHES "[cC]")
- set(hex2dec_char 12)
- elseif(hex2dec_char MATCHES "[dD]")
- set(hex2dec_char 13)
- elseif(hex2dec_char MATCHES "[eE]")
- set(hex2dec_char 14)
- elseif(hex2dec_char MATCHES "[fF]")
- set(hex2dec_char 15)
- else()
- message(FATAL_ERROR "Invalid format for hexidecimal character: " ${hex2dec_char})
- endif()
-
- math(EXPR hex2dec_char "${hex2dec_char} << ((${hex2dec_len}-${hex2dec_idx}-1)*4)")
- math(EXPR ${var} "${${var}}+${hex2dec_char}")
- math(EXPR hex2dec_idx "${hex2dec_idx}+1")
- endwhile()
-endmacro(hex2dec)
-
-# parseversion(<filepath>):
-# Parse the file given by <filepath> for the RD_KAFKA_VERSION constant
-# and convert the hex value to decimal version numbers.
-# Creates the following CMake variables:
-# * RDKAFKA_VERSION
-# * RDKAFKA_VERSION_MAJOR
-# * RDKAFKA_VERSION_MINOR
-# * RDKAFKA_VERSION_REVISION
-# * RDKAFKA_VERSION_PRERELEASE
-macro(parseversion path)
- file(STRINGS ${path} rdkafka_version_def REGEX "#define *RD_KAFKA_VERSION *\(0x[a-f0-9]*\)\.*")
- string(REGEX REPLACE "#define *RD_KAFKA_VERSION *0x" "" rdkafka_version_hex ${rdkafka_version_def})
-
- string(SUBSTRING ${rdkafka_version_hex} 0 2 rdkafka_version_major_hex)
- string(SUBSTRING ${rdkafka_version_hex} 2 2 rdkafka_version_minor_hex)
- string(SUBSTRING ${rdkafka_version_hex} 4 2 rdkafka_version_revision_hex)
- string(SUBSTRING ${rdkafka_version_hex} 6 2 rdkafka_version_prerelease_hex)
-
- hex2dec(RDKAFKA_VERSION_MAJOR ${rdkafka_version_major_hex})
- hex2dec(RDKAFKA_VERSION_MINOR ${rdkafka_version_minor_hex})
- hex2dec(RDKAFKA_VERSION_REVISION ${rdkafka_version_revision_hex})
- hex2dec(RDKAFKA_VERSION_PRERELEASE ${rdkafka_version_prerelease_hex})
- set(RDKAFKA_VERSION "${RDKAFKA_VERSION_MAJOR}.${RDKAFKA_VERSION_MINOR}.${RDKAFKA_VERSION_REVISION}")
-endmacro(parseversion)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/rdkafka.pc.in b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/rdkafka.pc.in
deleted file mode 100644
index 0eb17e856..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/rdkafka.pc.in
+++ /dev/null
@@ -1,12 +0,0 @@
-prefix=@CMAKE_INSTALL_PREFIX@
-exec_prefix=${prefix}
-includedir=${prefix}/include
-libdir=${prefix}/lib
-
-Name: @PKG_CONFIG_NAME@
-Description: @PKG_CONFIG_DESCRIPTION@
-Version: @PKG_CONFIG_VERSION@
-Requires: @PKG_CONFIG_REQUIRES@
-Cflags: @PKG_CONFIG_CFLAGS@
-Libs: @PKG_CONFIG_LIBS@
-Libs.private: @PKG_CONFIG_LIBS_PRIVATE@
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_32_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_32_test.c
deleted file mode 100644
index b3373bb8b..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_32_test.c
+++ /dev/null
@@ -1,8 +0,0 @@
-#include <inttypes.h>
-
-int32_t foo(int32_t i) {
- return __atomic_add_fetch(&i, 1, __ATOMIC_SEQ_CST);
-}
-
-int main() {
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_64_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_64_test.c
deleted file mode 100644
index 31922b85c..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/atomic_64_test.c
+++ /dev/null
@@ -1,8 +0,0 @@
-#include <inttypes.h>
-
-int64_t foo(int64_t i) {
- return __atomic_add_fetch(&i, 1, __ATOMIC_SEQ_CST);
-}
-
-int main() {
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/c11threads_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/c11threads_test.c
deleted file mode 100644
index 31681ae61..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/c11threads_test.c
+++ /dev/null
@@ -1,14 +0,0 @@
-#include <threads.h>
-
-static int start_func(void *arg) {
- int iarg = *(int *)arg;
- return iarg;
-}
-
-void main(void) {
- thrd_t thr;
- int arg = 1;
- if (thrd_create(&thr, start_func, (void *)&arg) != thrd_success) {
- ;
- }
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/crc32c_hw_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/crc32c_hw_test.c
deleted file mode 100644
index e80097803..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/crc32c_hw_test.c
+++ /dev/null
@@ -1,27 +0,0 @@
-#include <inttypes.h>
-#include <stdio.h>
-#define LONGx1 "8192"
-#define LONGx2 "16384"
-void main(void) {
- const char *n = "abcdefghijklmnopqrstuvwxyz0123456789";
- uint64_t c0 = 0, c1 = 1, c2 = 2;
- uint64_t s;
- uint32_t eax = 1, ecx;
- __asm__("cpuid" : "=c"(ecx) : "a"(eax) : "%ebx", "%edx");
- __asm__(
- "crc32b\t"
- "(%1), %0"
- : "=r"(c0)
- : "r"(n), "0"(c0));
- __asm__(
- "crc32q\t"
- "(%3), %0\n\t"
- "crc32q\t" LONGx1
- "(%3), %1\n\t"
- "crc32q\t" LONGx2 "(%3), %2"
- : "=r"(c0), "=r"(c1), "=r"(c2)
- : "r"(n), "0"(c0), "1"(c1), "2"(c2));
- s = c0 + c1 + c2;
- printf("avoiding unused code removal by printing %d, %d, %d\n", (int)s,
- (int)eax, (int)ecx);
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/dlopen_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/dlopen_test.c
deleted file mode 100644
index ecb478994..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/dlopen_test.c
+++ /dev/null
@@ -1,11 +0,0 @@
-#include <string.h>
-#include <dlfcn.h>
-
-int main() {
- void *h;
- /* Try loading anything, we don't care if it works */
- h = dlopen("__nothing_rdkafka.so", RTLD_NOW | RTLD_LOCAL);
- if (h)
- dlclose(h);
- return 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/libsasl2_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/libsasl2_test.c
deleted file mode 100644
index 3f3ab3409..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/libsasl2_test.c
+++ /dev/null
@@ -1,7 +0,0 @@
-#include <string.h>
-#include <sasl/sasl.h>
-
-int main() {
- sasl_done();
- return 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_darwin_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_darwin_test.c
deleted file mode 100644
index 73e31e069..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_darwin_test.c
+++ /dev/null
@@ -1,6 +0,0 @@
-#include <pthread.h>
-
-int main() {
- pthread_setname_np("abc");
- return 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_freebsd_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_freebsd_test.c
deleted file mode 100644
index 329ace08e..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_freebsd_test.c
+++ /dev/null
@@ -1,7 +0,0 @@
-#include <pthread.h>
-#include <pthread_np.h>
-
-int main() {
- pthread_set_name_np(pthread_self(), "abc");
- return 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_gnu_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_gnu_test.c
deleted file mode 100644
index 3be1b21bc..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/pthread_setname_gnu_test.c
+++ /dev/null
@@ -1,5 +0,0 @@
-#include <pthread.h>
-
-int main() {
- return pthread_setname_np(pthread_self(), "abc");
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rand_r_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rand_r_test.c
deleted file mode 100644
index be722d0a0..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rand_r_test.c
+++ /dev/null
@@ -1,7 +0,0 @@
-#include <stdlib.h>
-
-int main() {
- unsigned int seed = 0xbeaf;
- (void)rand_r(&seed);
- return 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rdkafka_setup.cmake b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rdkafka_setup.cmake
deleted file mode 100644
index 5ea7f7dc6..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/rdkafka_setup.cmake
+++ /dev/null
@@ -1,122 +0,0 @@
-try_compile(
- HAVE_REGEX
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/regex_test.c"
-)
-
-try_compile(
- HAVE_STRNDUP
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/strndup_test.c"
-)
-
-try_compile(
- HAVE_RAND_R
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/rand_r_test.c"
-)
-
-try_compile(
- HAVE_PTHREAD_SETNAME_GNU
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/pthread_setname_gnu_test.c"
- COMPILE_DEFINITIONS "-D_GNU_SOURCE"
- LINK_LIBRARIES "-lpthread"
-)
-
-try_compile(
- HAVE_PTHREAD_SETNAME_DARWIN
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/pthread_setname_darwin_test.c"
- COMPILE_DEFINITIONS "-D_DARWIN_C_SOURCE"
- LINK_LIBRARIES "-lpthread"
-)
-
-try_compile(
- HAVE_PTHREAD_SETNAME_FREEBSD
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/pthread_setname_freebsd_test.c"
- LINK_LIBRARIES "-lpthread"
-)
-
-# Atomic 32 tests {
-set(LINK_ATOMIC NO)
-set(HAVE_ATOMICS_32 NO)
-set(HAVE_ATOMICS_32_SYNC NO)
-
-try_compile(
- _atomics_32
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/atomic_32_test.c"
-)
-
-if(_atomics_32)
- set(HAVE_ATOMICS_32 YES)
-else()
- try_compile(
- _atomics_32_lib
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/atomic_32_test.c"
- LINK_LIBRARIES "-latomic"
- )
- if(_atomics_32_lib)
- set(HAVE_ATOMICS_32 YES)
- set(LINK_ATOMIC YES)
- else()
- try_compile(
- HAVE_ATOMICS_32_SYNC
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/sync_32_test.c"
- )
- endif()
-endif()
-# }
-
-# Atomic 64 tests {
-set(HAVE_ATOMICS_64 NO)
-set(HAVE_ATOMICS_64_SYNC NO)
-
-try_compile(
- _atomics_64
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/atomic_64_test.c"
-)
-
-if(_atomics_64)
- set(HAVE_ATOMICS_64 YES)
-else()
- try_compile(
- _atomics_64_lib
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/atomic_64_test.c"
- LINK_LIBRARIES "-latomic"
- )
- if(_atomics_64_lib)
- set(HAVE_ATOMICS_64 YES)
- set(LINK_ATOMIC YES)
- else()
- try_compile(
- HAVE_ATOMICS_64_SYNC
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/sync_64_test.c"
- )
- endif()
-endif()
-# }
-
-# C11 threads
-try_compile(
- WITH_C11THREADS
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/c11threads_test.c"
- LINK_LIBRARIES "-pthread"
-)
-# }
-
-# CRC32C {
-try_compile(
- WITH_CRC32C_HW
- "${CMAKE_CURRENT_BINARY_DIR}/try_compile"
- "${TRYCOMPILE_SRC_DIR}/crc32c_hw_test.c"
-)
-# }
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/regex_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/regex_test.c
deleted file mode 100644
index 329098d20..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/regex_test.c
+++ /dev/null
@@ -1,10 +0,0 @@
-#include <stddef.h>
-#include <regex.h>
-
-int main() {
- regcomp(NULL, NULL, 0);
- regexec(NULL, NULL, 0, NULL, 0);
- regerror(0, NULL, NULL, 0);
- regfree(NULL);
- return 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/strndup_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/strndup_test.c
deleted file mode 100644
index a10b74526..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/strndup_test.c
+++ /dev/null
@@ -1,5 +0,0 @@
-#include <string.h>
-
-int main() {
- return strndup("hi", 2) ? 0 : 1;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_32_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_32_test.c
deleted file mode 100644
index 2bc80ab4c..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_32_test.c
+++ /dev/null
@@ -1,8 +0,0 @@
-#include <inttypes.h>
-
-int32_t foo(int32_t i) {
- return __sync_add_and_fetch(&i, 1);
-}
-
-int main() {
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_64_test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_64_test.c
deleted file mode 100644
index 4b6ad6d38..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cmake/try_compile/sync_64_test.c
+++ /dev/null
@@ -1,8 +0,0 @@
-#include <inttypes.h>
-
-int64_t foo(int64_t i) {
- return __sync_add_and_fetch(&i, 1);
-}
-
-int main() {
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/README.md b/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/README.md
deleted file mode 100644
index 24a82f142..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/README.md
+++ /dev/null
@@ -1,14 +0,0 @@
-# Confluent Platform package verification
-
-This small set of scripts verifies the librdkafka packages that
-are part of the Confluent Platform.
-
-The base_url is the http S3 bucket path to the a PR job, or similar.
-
-## How to use
-
- $ ./verify-packages.sh 5.3 https://thes3bucketpath/X/Y
-
-
-Requires docker and patience.
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/check_features.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/check_features.c
deleted file mode 100644
index 4229402fd..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/check_features.c
+++ /dev/null
@@ -1,64 +0,0 @@
-#include <stdio.h>
-#include <string.h>
-#include <librdkafka/rdkafka.h>
-
-int main(int argc, char **argv) {
- rd_kafka_conf_t *conf;
- char buf[512];
- size_t sz = sizeof(buf);
- rd_kafka_conf_res_t res;
- static const char *expected_features = "ssl,sasl_gssapi,lz4,zstd";
- char errstr[512];
- int i;
- int failures = 0;
-
- printf("librdkafka %s (0x%x, define: 0x%x)\n", rd_kafka_version_str(),
- rd_kafka_version(), RD_KAFKA_VERSION);
-
- if (argc > 1 && !(argc & 1)) {
- printf("Usage: %s [config.property config-value ..]\n",
- argv[0]);
- return 1;
- }
-
- conf = rd_kafka_conf_new();
- res = rd_kafka_conf_get(conf, "builtin.features", buf, &sz);
-
- if (res != RD_KAFKA_CONF_OK) {
- printf("ERROR: conf_get failed: %d\n", res);
- return 1;
- }
-
- printf("builtin.features: %s\n", buf);
-
- /* librdkafka allows checking for expected features
- * by setting the corresponding feature flags in builtin.features,
- * which will return an error if one or more flags are not enabled. */
- if (rd_kafka_conf_set(conf, "builtin.features", expected_features,
- errstr, sizeof(errstr)) != RD_KAFKA_CONF_OK) {
- printf(
- "ERROR: expected at least features: %s\n"
- "got error: %s\n",
- expected_features, errstr);
- failures++;
- }
-
- printf("all expected features matched: %s\n", expected_features);
-
- /* Apply config from argv key value pairs */
- for (i = 1; i + 1 < argc; i += 2) {
- printf("verifying config %s=%s\n", argv[i], argv[i + 1]);
- if (rd_kafka_conf_set(conf, argv[i], argv[i + 1], errstr,
- sizeof(errstr)) != RD_KAFKA_CONF_OK) {
- printf("ERROR: failed to set %s=%s: %s\n", argv[i],
- argv[i + 1], errstr);
- failures++;
- }
- }
-
- rd_kafka_conf_destroy(conf);
-
- printf("%d failures\n", failures);
-
- return !!failures;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-deb.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-deb.sh
deleted file mode 100755
index 1350d0655..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-deb.sh
+++ /dev/null
@@ -1,34 +0,0 @@
-#!/bin/bash
-#
-
-set -e
-
-cpver=$1
-base_url=$2
-
-if [[ -z $base_url ]]; then
- echo "Usage: $0 <cp-base-ver> <base_url>"
- exit 1
-fi
-
-apt-get update
-apt-get install -y apt-transport-https wget
-
-wget -qO - ${base_url}/deb/${cpver}/archive.key | apt-key add -
-
-
-cat >/etc/apt/sources.list.d/Confluent.list <<EOF
-deb [arch=amd64] $base_url/deb/${cpver} stable main
-EOF
-
-apt-get update
-apt-get install -y librdkafka-dev gcc
-
-gcc /v/check_features.c -o /tmp/check_features -lrdkafka
-
-/tmp/check_features
-
-# Verify plugins
-apt-get install -y confluent-librdkafka-plugins
-
-/tmp/check_features plugin.library.paths monitoring-interceptor
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-packages.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-packages.sh
deleted file mode 100755
index ecddbd558..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-packages.sh
+++ /dev/null
@@ -1,43 +0,0 @@
-#!/bin/bash
-#
-# Verifies RPM and DEB packages from Confluent Platform
-#
-
-cpver=$1
-base_url=$2
-
-if [[ -z $base_url ]]; then
- echo "Usage: $0 <CP-M.m-version> <base-url>"
- echo ""
- echo " <CP-M.m-version> is the Major.minor version of CP, e.g., 5.3"
- echo " <base-url> is the release base bucket URL"
- exit 1
-fi
-
-thisdir="$( cd "$(dirname "$0")" ; pwd -P )"
-
-echo "#### Verifying RPM packages ####"
-docker run -v $thisdir:/v centos:7 /v/verify-rpm.sh $cpver $base_url
-rpm_status=$?
-
-echo "#### Verifying Debian packages ####"
-docker run -v $thisdir:/v ubuntu:16.04 /v/verify-deb.sh $cpver $base_url
-deb_status=$?
-
-
-if [[ $rpm_status == 0 ]]; then
- echo "SUCCESS: RPM packages verified"
-else
- echo "ERROR: RPM package verification failed"
-fi
-
-if [[ $deb_status == 0 ]]; then
- echo "SUCCESS: Debian packages verified"
-else
- echo "ERROR: Debian package verification failed"
-fi
-
-if [[ $deb_status != 0 || $rpm_status != 0 ]]; then
- exit 1
-fi
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-rpm.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-rpm.sh
deleted file mode 100755
index d7b3b1a14..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/cp/verify-rpm.sh
+++ /dev/null
@@ -1,38 +0,0 @@
-#!/bin/bash
-#
-
-set -e
-
-cpver=$1
-base_url=$2
-
-if [[ -z $base_url ]]; then
- echo "Usage: $0 <cp-base-ver> <base_url>"
- exit 1
-fi
-
-cat >/etc/yum.repos.d/Confluent.repo <<EOF
-[Confluent.dist]
-name=Confluent repository (dist)
-baseurl=$base_url/rpm/${cpver}/7
-gpgcheck=0
-gpgkey=$base_url/rpm/${cpver}/archive.key
-enabled=1
-[Confluent]
-name=Confluent repository
-baseurl=$base_url/rpm/${cpver}
-gpgcheck=1
-gpgkey=$base_url/rpm/${cpver}/archive.key
-enabled=1
-EOF
-
-yum install -y librdkafka-devel gcc
-
-gcc /v/check_features.c -o /tmp/check_features -lrdkafka
-
-/tmp/check_features
-
-# Verify plugins
-yum install -y confluent-librdkafka-plugins
-
-/tmp/check_features plugin.library.paths monitoring-interceptor
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/.gitignore b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/.gitignore
deleted file mode 100644
index eb66d4d31..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/.gitignore
+++ /dev/null
@@ -1,6 +0,0 @@
-*.log
-files
-librdkafka-dev
-librdkafka1-dbg
-librdkafka1
-tmp
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/changelog b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/changelog
deleted file mode 100644
index c50cb5aa8..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/changelog
+++ /dev/null
@@ -1,66 +0,0 @@
-librdkafka (0.8.6-1) unstable; urgency=medium
-
- * New upstream release.
- * Backport upstream commit f6fd0da, adding --disable-silent-rules
- compatibility support to mklove. (Closes: #788742)
-
- -- Faidon Liambotis <paravoid@debian.org> Sun, 19 Jul 2015 01:36:18 +0300
-
-librdkafka (0.8.5-2) unstable; urgency=medium
-
- * Install rdkafka.pc in the right, multiarch location. (Closes: #766759)
-
- -- Faidon Liambotis <paravoid@debian.org> Sun, 26 Oct 2014 06:47:07 +0200
-
-librdkafka (0.8.5-1) unstable; urgency=medium
-
- * New upstream release.
- - Fixes kFreeBSD FTBFS.
- * Ship rdkafka.pc pkg-config in librdkafka-dev.
-
- -- Faidon Liambotis <paravoid@debian.org> Fri, 24 Oct 2014 18:03:22 +0300
-
-librdkafka (0.8.4-1) unstable; urgency=medium
-
- * New upstream release, including a new build system.
- - Add Build-Depends on perl, required by configure.
- - Support multiarch library paths.
- - Better detection of architecture atomic builtins, supporting more
- architectures. (Closes: #739930)
- - Various portability bugs fixed. (Closes: #730506)
- - Update debian/librdkafka1.symbols.
- * Convert to a multiarch package.
- * Switch to Architecture: any, because of renewed upstream portability.
- * Update debian/copyright to add src/ before Files: paths.
- * Update Standards-Version to 3.9.6, no changes needed.
- * Ship only the C library for now, not the new C++ library; the latter is
- still in flux in some ways and will probably be shipped in a separate
- package in a future release.
-
- -- Faidon Liambotis <paravoid@debian.org> Wed, 22 Oct 2014 23:57:24 +0300
-
-librdkafka (0.8.3-1) unstable; urgency=medium
-
- * New upstream release.
- - Multiple internal symbols hidden; breaks ABI without a SONAME bump, but
- these were internal and should not break any applications, packaged or
- not.
- * Update Standards-Version to 3.9.5, no changes needed.
-
- -- Faidon Liambotis <paravoid@debian.org> Tue, 18 Feb 2014 02:21:43 +0200
-
-librdkafka (0.8.1-1) unstable; urgency=medium
-
- * New upstream release.
- - Multiple fixes to FTBFS on various architectures. (Closes: #730506)
- - Remove dh_auto_clean override, fixed upstream.
- * Limit the set of architectures: upstream currently relies on 64-bit atomic
- operations that several Debian architectures do not support.
-
- -- Faidon Liambotis <paravoid@debian.org> Thu, 05 Dec 2013 16:53:28 +0200
-
-librdkafka (0.8.0-1) unstable; urgency=low
-
- * Initial release. (Closes: #710271)
-
- -- Faidon Liambotis <paravoid@debian.org> Mon, 04 Nov 2013 16:50:07 +0200
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/compat b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/compat
deleted file mode 100644
index ec635144f..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/compat
+++ /dev/null
@@ -1 +0,0 @@
-9
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/control b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/control
deleted file mode 100644
index 510db8f23..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/control
+++ /dev/null
@@ -1,49 +0,0 @@
-Source: librdkafka
-Priority: optional
-Maintainer: Faidon Liambotis <paravoid@debian.org>
-Build-Depends: debhelper (>= 9), zlib1g-dev, libssl-dev, libsasl2-dev, python3
-Standards-Version: 3.9.6
-Section: libs
-Homepage: https://github.com/edenhill/librdkafka
-Vcs-Git: git://github.com/edenhill/librdkafka.git -b debian
-Vcs-Browser: https://github.com/edenhill/librdkafka/tree/debian
-
-Package: librdkafka1
-Architecture: any
-Multi-Arch: same
-Depends: ${shlibs:Depends}, ${misc:Depends}
-Description: library implementing the Apache Kafka protocol
- librdkafka is a C implementation of the Apache Kafka protocol. It currently
- implements the 0.8 version of the protocol and can be used to develop both
- Producers and Consumers.
- .
- More information about Apache Kafka can be found at http://kafka.apache.org/
-
-Package: librdkafka-dev
-Section: libdevel
-Architecture: any
-Multi-Arch: same
-Depends: librdkafka1 (= ${binary:Version}), ${misc:Depends}
-Description: library implementing the Apache Kafka protocol (development headers)
- librdkafka is a C implementation of the Apache Kafka protocol. It currently
- implements the 0.8 version of the protocol and can be used to develop both
- Producers and Consumers.
- .
- More information about Apache Kafka can be found at http://kafka.apache.org/
- .
- This package contains the development headers.
-
-Package: librdkafka1-dbg
-Section: debug
-Priority: extra
-Architecture: any
-Multi-Arch: same
-Depends: librdkafka1 (= ${binary:Version}), ${misc:Depends}
-Description: library implementing the Apache Kafka protocol (debugging symbols)
- librdkafka is a C implementation of the Apache Kafka protocol. It currently
- implements the 0.8 version of the protocol and can be used to develop both
- Producers and Consumers.
- .
- More information about Apache Kafka can be found at http://kafka.apache.org/
- .
- This package contains the debugging symbols.
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/copyright b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/copyright
deleted file mode 100644
index 20885d9f3..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/copyright
+++ /dev/null
@@ -1,84 +0,0 @@
-Format: http://www.debian.org/doc/packaging-manuals/copyright-format/1.0/
-Upstream-Name: librdkafka
-Source: https://github.com/edenhill/librdkafka
-
-License: BSD-2-clause
- Redistribution and use in source and binary forms, with or without
- modification, are permitted provided that the following conditions are met:
- .
- 1. Redistributions of source code must retain the above copyright notice,
- this list of conditions and the following disclaimer.
- 2. Redistributions in binary form must reproduce the above copyright notice,
- this list of conditions and the following disclaimer in the documentation
- and/or other materials provided with the distribution.
- .
- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
- AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
- LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
- CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
- SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
- INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
- CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
- ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
- POSSIBILITY OF SUCH DAMAGE.
-
-Files: *
-Copyright: 2012-2015, Magnus Edenhill
-License: BSD-2-clause
-
-Files: src/rdcrc32.c src/rdcrc32.h
-Copyright: 2006-2012, Thomas Pircher <tehpeh@gmx.net>
-License: MIT
- Permission is hereby granted, free of charge, to any person obtaining a copy
- of this software and associated documentation files (the "Software"), to deal
- in the Software without restriction, including without limitation the rights
- to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
- copies of the Software, and to permit persons to whom the Software is
- furnished to do so, subject to the following conditions:
- .
- The above copyright notice and this permission notice shall be included in
- all copies or substantial portions of the Software.
- .
- THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
- IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
- FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
- AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
- LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
- OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
- THE SOFTWARE.
-
-Files: src/snappy.c src/snappy.h src/snappy_compat.h
-Copyright: 2005, Google Inc.
- 2011, Intel Corporation
-License: BSD-3-clause
- Redistribution and use in source and binary forms, with or without
- modification, are permitted provided that the following conditions are
- met:
- .
- * Redistributions of source code must retain the above copyright
- notice, this list of conditions and the following disclaimer.
- * Redistributions in binary form must reproduce the above
- copyright notice, this list of conditions and the following disclaimer
- in the documentation and/or other materials provided with the
- distribution.
- * Neither the name of Google Inc. nor the names of its
- contributors may be used to endorse or promote products derived from
- this software without specific prior written permission.
- .
- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
- OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
- SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
- LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
- DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
- THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
- (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
- OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-Files: debian/*
-Copyright: 2013 Faidon Liambotis <paravoid@debian.org>
-License: BSD-2-clause
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/docs b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/docs
deleted file mode 100644
index 0b76c34c4..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/docs
+++ /dev/null
@@ -1,5 +0,0 @@
-README.md
-INTRODUCTION.md
-CONFIGURATION.md
-STATISTICS.md
-CHANGELOG.md \ No newline at end of file
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/gbp.conf b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/gbp.conf
deleted file mode 100644
index b2a0f02e3..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/gbp.conf
+++ /dev/null
@@ -1,9 +0,0 @@
-[buildpackage]
-upstream-tree=tag
-upstream-branch=master
-debian-branch=debian
-upstream-tag=%(version)s
-debian-tag=debian/%(version)s
-no-create-orig = True
-tarball-dir = ../tarballs
-export-dir = ../build-area
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.dirs b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.dirs
deleted file mode 100644
index 44188162e..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.dirs
+++ /dev/null
@@ -1,2 +0,0 @@
-usr/lib
-usr/include
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.examples b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.examples
deleted file mode 100644
index b45032efe..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.examples
+++ /dev/null
@@ -1,2 +0,0 @@
-examples/rdkafka_example.c
-examples/rdkafka_performance.c
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.install b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.install
deleted file mode 100644
index 478f660f5..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.install
+++ /dev/null
@@ -1,6 +0,0 @@
-usr/include/*/rdkafka.h
-usr/include/*/rdkafkacpp.h
-usr/lib/*/librdkafka.a
-usr/lib/*/librdkafka.so
-usr/lib/*/librdkafka++.a
-usr/lib/*/librdkafka++.so
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.substvars b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.substvars
deleted file mode 100644
index abd3ebebc..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka-dev.substvars
+++ /dev/null
@@ -1 +0,0 @@
-misc:Depends=
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka.dsc b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka.dsc
deleted file mode 100644
index 447b9e656..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka.dsc
+++ /dev/null
@@ -1,16 +0,0 @@
-Format: 3.0 (quilt)
-Source: librdkafka
-Binary: librdkafka1, librdkafka-dev, librdkafka1-dbg
-Architecture: any
-Version: 0.9.1-1pre1
-Maintainer: Magnus Edenhill <librdkafka@edenhill.se>
-Homepage: https://github.com/edenhill/librdkafka
-Standards-Version: 3.9.6
-Vcs-Browser: https://github.com/edenhill/librdkafka/tree/master
-Vcs-Git: git://github.com/edenhill/librdkafka.git -b master
-Build-Depends: debhelper (>= 9), zlib1g-dev, libssl-dev, libsasl2-dev, python3
-Package-List:
- librdkafka-dev deb libdevel optional arch=any
- librdkafka1 deb libs optional arch=any
- librdkafka1-dbg deb debug extra arch=any
-Original-Maintainer: Faidon Liambotis <paravoid@debian.org>
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1-dbg.substvars b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1-dbg.substvars
deleted file mode 100644
index abd3ebebc..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1-dbg.substvars
+++ /dev/null
@@ -1 +0,0 @@
-misc:Depends=
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.dirs b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.dirs
deleted file mode 100644
index 68457717b..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.dirs
+++ /dev/null
@@ -1 +0,0 @@
-usr/lib
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.install b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.install
deleted file mode 100644
index 7e86e5f18..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.install
+++ /dev/null
@@ -1,2 +0,0 @@
-usr/lib/*/librdkafka.so.*
-usr/lib/*/librdkafka++.so.*
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postinst.debhelper b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postinst.debhelper
deleted file mode 100644
index 3d89d3ef6..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postinst.debhelper
+++ /dev/null
@@ -1,5 +0,0 @@
-# Automatically added by dh_makeshlibs
-if [ "$1" = "configure" ]; then
- ldconfig
-fi
-# End automatically added section
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postrm.debhelper b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postrm.debhelper
deleted file mode 100644
index 7f4404727..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.postrm.debhelper
+++ /dev/null
@@ -1,5 +0,0 @@
-# Automatically added by dh_makeshlibs
-if [ "$1" = "remove" ]; then
- ldconfig
-fi
-# End automatically added section
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.symbols b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.symbols
deleted file mode 100644
index 0ef576eb1..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/librdkafka1.symbols
+++ /dev/null
@@ -1,64 +0,0 @@
-librdkafka.so.1 librdkafka1 #MINVER#
-* Build-Depends-Package: librdkafka-dev
- rd_kafka_brokers_add@Base 0.8.0
- rd_kafka_conf_destroy@Base 0.8.0
- rd_kafka_conf_dump@Base 0.8.3
- rd_kafka_conf_dump_free@Base 0.8.3
- rd_kafka_conf_dup@Base 0.8.3
- rd_kafka_conf_new@Base 0.8.0
- rd_kafka_conf_properties_show@Base 0.8.0
- rd_kafka_conf_set@Base 0.8.0
- rd_kafka_conf_set_dr_cb@Base 0.8.0
- rd_kafka_conf_set_dr_msg_cb@Base 0.8.4
- rd_kafka_conf_set_error_cb@Base 0.8.0
- rd_kafka_conf_set_log_cb@Base 0.8.4
- rd_kafka_conf_set_opaque@Base 0.8.0
- rd_kafka_conf_set_open_cb@Base 0.8.4
- rd_kafka_conf_set_socket_cb@Base 0.8.4
- rd_kafka_conf_set_stats_cb@Base 0.8.0
- rd_kafka_consume@Base 0.8.0
- rd_kafka_consume_batch@Base 0.8.0
- rd_kafka_consume_batch_queue@Base 0.8.4
- rd_kafka_consume_callback@Base 0.8.0
- rd_kafka_consume_callback_queue@Base 0.8.4
- rd_kafka_consume_queue@Base 0.8.4
- rd_kafka_consume_start@Base 0.8.0
- rd_kafka_consume_start_queue@Base 0.8.4
- rd_kafka_consume_stop@Base 0.8.0
- rd_kafka_destroy@Base 0.8.0
- rd_kafka_dump@Base 0.8.0
- rd_kafka_err2str@Base 0.8.0
- rd_kafka_errno2err@Base 0.8.3
- rd_kafka_log_print@Base 0.8.0
- rd_kafka_log_syslog@Base 0.8.0
- rd_kafka_message_destroy@Base 0.8.0
- rd_kafka_metadata@Base 0.8.4
- rd_kafka_metadata_destroy@Base 0.8.4
- rd_kafka_msg_partitioner_random@Base 0.8.0
- rd_kafka_name@Base 0.8.0
- rd_kafka_new@Base 0.8.0
- rd_kafka_offset_store@Base 0.8.3
- rd_kafka_opaque@Base 0.8.4
- rd_kafka_outq_len@Base 0.8.0
- rd_kafka_poll@Base 0.8.0
- rd_kafka_produce@Base 0.8.0
- rd_kafka_produce_batch@Base 0.8.4
- rd_kafka_queue_destroy@Base 0.8.4
- rd_kafka_queue_new@Base 0.8.4
- rd_kafka_set_log_level@Base 0.8.0
- rd_kafka_set_logger@Base 0.8.0
- rd_kafka_thread_cnt@Base 0.8.0
- rd_kafka_topic_conf_destroy@Base 0.8.0
- rd_kafka_topic_conf_dump@Base 0.8.3
- rd_kafka_topic_conf_dup@Base 0.8.3
- rd_kafka_topic_conf_new@Base 0.8.0
- rd_kafka_topic_conf_set@Base 0.8.0
- rd_kafka_topic_conf_set_opaque@Base 0.8.0
- rd_kafka_topic_conf_set_partitioner_cb@Base 0.8.0
- rd_kafka_topic_destroy@Base 0.8.0
- rd_kafka_topic_name@Base 0.8.0
- rd_kafka_topic_new@Base 0.8.0
- rd_kafka_topic_partition_available@Base 0.8.0
- rd_kafka_version@Base 0.8.1
- rd_kafka_version_str@Base 0.8.1
- rd_kafka_wait_destroyed@Base 0.8.0
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/rules b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/rules
deleted file mode 100755
index a18c40d98..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/rules
+++ /dev/null
@@ -1,19 +0,0 @@
-#!/usr/bin/make -f
-
-# Uncomment this to turn on verbose mode.
-#export DH_VERBOSE=1
-
-%:
- dh $@
-
-override_dh_strip:
- dh_strip --dbg-package=librdkafka1-dbg
-
-override_dh_auto_install:
- dh_auto_install
- install -D -m 0644 rdkafka.pc \
- debian/librdkafka-dev/usr/lib/${DEB_HOST_MULTIARCH}/pkgconfig/rdkafka.pc
- install -D -m 0644 rdkafka-static.pc \
- debian/librdkafka-dev/usr/lib/${DEB_HOST_MULTIARCH}/pkgconfig/rdkafka-static.pc
-
-.PHONY: override_dh_strip override_dh_auth_install
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/source/format b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/source/format
deleted file mode 100644
index 163aaf8d8..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/source/format
+++ /dev/null
@@ -1 +0,0 @@
-3.0 (quilt)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/watch b/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/watch
deleted file mode 100644
index fc9aec86f..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/debian/watch
+++ /dev/null
@@ -1,2 +0,0 @@
-version=3
-http://github.com/edenhill/librdkafka/tags .*/(\d[\d\.]*)\.tar\.gz
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/get_version.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/get_version.py
deleted file mode 100755
index fad1d9718..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/get_version.py
+++ /dev/null
@@ -1,21 +0,0 @@
-#!/usr/bin/env python3
-
-import sys
-
-if len(sys.argv) != 2:
- raise Exception('Usage: %s path/to/rdkafka.h' % sys.argv[0])
-
-kafka_h_file = sys.argv[1]
-f = open(kafka_h_file)
-for line in f:
- if '#define RD_KAFKA_VERSION' in line:
- version = line.split()[-1]
- break
-f.close()
-
-major = int(version[2:4], 16)
-minor = int(version[4:6], 16)
-patch = int(version[6:8], 16)
-version = '.'.join(str(item) for item in (major, minor, patch))
-
-print(version)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/README.md b/fluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/README.md
deleted file mode 100644
index a23a08537..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/README.md
+++ /dev/null
@@ -1,15 +0,0 @@
-# Update the Homebrew librdkafka package version
-
-The `./brew-update-pr.sh` script in this directory updates the
-brew formula for librdkafka and pushes a PR to the homebrew-core repository.
-
-You should run it in two steps, first an implicit dry-run mode
-to check that things seem correct, and if that checks out a
-live upload mode which actually pushes the PR.
-
- # Do a dry-run first, v0.11.0 is the librdkafka tag:
- $ ./brew-update-pr.sh v0.11.0
-
- # If everything looks okay, run the live upload mode:
- $ ./brew-update-pr.sh --upload v0.11.0
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/brew-update-pr.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/brew-update-pr.sh
deleted file mode 100755
index f756159cd..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/homebrew/brew-update-pr.sh
+++ /dev/null
@@ -1,31 +0,0 @@
-#!/bin/bash
-#
-# Automatically pushes a PR to homebrew-core to update
-# the librdkafka version.
-#
-# Usage:
-# # Dry-run:
-# ./brew-update-pr.sh v0.11.0
-# # if everything looks good:
-# ./brew-update-pr.sh --upload v0.11.0
-#
-
-
-DRY_RUN="--dry-run"
-if [[ $1 == "--upload" ]]; then
- DRY_RUN=
- shift
-fi
-
-TAG=$1
-
-if [[ -z $TAG ]]; then
- echo "Usage: $0 [--upload] <librdkafka-tag>"
- exit 1
-fi
-
-set -eu
-
-brew bump-formula-pr $DRY_RUN --strict \
- --url=https://github.com/edenhill/librdkafka/archive/${TAG}.tar.gz \
- librdkafka
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw-static.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw-static.sh
deleted file mode 100644
index a5162caad..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw-static.sh
+++ /dev/null
@@ -1,52 +0,0 @@
-#!/bin/bash
-
-set -e
-
-cmake \
- -G "MinGW Makefiles" \
- -D CMAKE_INSTALL_PREFIX="$PWD/dest/" \
- -D RDKAFKA_BUILD_STATIC=ON \
- .
-
-$mingw64 mingw32-make
-$mingw64 mingw32-make install
-
-# Bundle all the static dependencies with the static lib we just built
-mkdir mergescratch
-pushd mergescratch
-cp /C/msys64/mingw64/lib/libzstd.a ./
-cp /C/msys64/mingw64/lib/libcrypto.a ./
-cp /C/msys64/mingw64/lib/liblz4.a ./
-cp /C/msys64/mingw64/lib/libssl.a ./
-cp /C/msys64/mingw64/lib/libz.a ./
-cp ../src/librdkafka.a ./
-
-# Have to rename because ar won't work with + in the name
-cp ../src-cpp/librdkafka++.a ./librdkafkacpp.a
-ar -M << EOF
-create librdkafka-static.a
-addlib librdkafka.a
-addlib libzstd.a
-addlib libcrypto.a
-addlib liblz4.a
-addlib libssl.a
-addlib libz.a
-save
-end
-EOF
-
-ar -M << EOF
-create librdkafkacpp-static.a
-addlib librdkafka-static.a
-addlib librdkafkacpp.a
-save
-end
-EOF
-
-strip -g ./librdkafka-static.a
-strip -g ./librdkafkacpp-static.a
-cp ./librdkafka-static.a ../dest/lib/
-cp ./librdkafkacpp-static.a ../dest/lib/librdkafka++-static.a
-popd
-rm -rf ./mergescratch
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw.sh
deleted file mode 100644
index b0b81fe0a..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/configure-build-msys2-mingw.sh
+++ /dev/null
@@ -1,21 +0,0 @@
-#!/bin/bash
-
-set -e
-
-cmake \
- -G "MinGW Makefiles" \
- -D CMAKE_INSTALL_PREFIX="$PWD/dest/" \
- -D WITHOUT_WIN32_CONFIG=ON \
- -D RDKAFKA_BUILD_EXAMPLES=ON \
- -D RDKAFKA_BUILD_TESTS=ON \
- -D RDKAFKA_BUILD_STATIC=OFF \
- -D CMAKE_WINDOWS_EXPORT_ALL_SYMBOLS=TRUE .
-
-$mingw64 mingw32-make
-$mingw64 mingw32-make install
-
-cd tests
-cp ../dest/bin/librdkafka.dll ./
-cp ../dest/bin/librdkafka++.dll ./
-CI=true ./test-runner.exe -l -Q
-cd ..
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/run-tests.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/run-tests.sh
deleted file mode 100644
index 6749add5d..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/run-tests.sh
+++ /dev/null
@@ -1,6 +0,0 @@
-#!/bin/bash
-
-set -e
-
-cd tests
-./test-runner.exe -l -Q -p1 0000
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/semaphoreci-build.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/semaphoreci-build.sh
deleted file mode 100644
index 378545b44..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/semaphoreci-build.sh
+++ /dev/null
@@ -1,38 +0,0 @@
-#!/bin/bash
-#
-
-set -ex
-
-if [[ $1 == "--static" ]]; then
- linkage="static"
- shift
-else
-linkage="dynamic"
-fi
-
-if [[ -z $1 ]]; then
- echo "Usage: $0 [--static] <relative-path-to-output-librdkafka.tgz>"
- exit 1
-fi
-
-archive="${PWD}/$1"
-
-source ./packaging/mingw-w64/travis-before-install.sh
-
-if [[ $linkage == "static" ]]; then
- ./packaging/mingw-w64/configure-build-msys2-mingw-static.sh
-else
- ./packaging/mingw-w64/configure-build-msys2-mingw.sh
-fi
-
-
-./packaging/mingw-w64/run-tests.sh
-
-pushd dest
-tar cvzf $archive .
-sha256sum $archive
-popd
-
-
-
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/travis-before-install.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/travis-before-install.sh
deleted file mode 100644
index e75507f93..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/mingw-w64/travis-before-install.sh
+++ /dev/null
@@ -1,20 +0,0 @@
-#!/bin/bash
-
-set -e
-
-export msys2='cmd //C RefreshEnv.cmd '
-export msys2+='& set MSYS=winsymlinks:nativestrict '
-export msys2+='& C:\\msys64\\msys2_shell.cmd -defterm -no-start'
-export mingw64="$msys2 -mingw64 -full-path -here -c "\"\$@"\" --"
-export msys2+=" -msys2 -c "\"\$@"\" --"
-
-# Have to update pacman first or choco upgrade will failure due to migration
-# to zstd instead of xz compression
-$msys2 pacman -Sy --noconfirm pacman
-
-## Install more MSYS2 packages from https://packages.msys2.org/base here
-$msys2 pacman --sync --noconfirm --needed mingw-w64-x86_64-gcc mingw-w64-x86_64-make mingw-w64-x86_64-cmake mingw-w64-x86_64-openssl mingw-w64-x86_64-lz4 mingw-w64-x86_64-zstd
-
-taskkill //IM gpg-agent.exe //F || true # https://travis-ci.community/t/4967
-export PATH=/C/msys64/mingw64/bin:$PATH
-export MAKE=mingw32-make # so that Autotools can find it
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/.gitignore b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/.gitignore
deleted file mode 100644
index 56919a155..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/.gitignore
+++ /dev/null
@@ -1,7 +0,0 @@
-dl-*
-out-*
-*.nupkg
-*.tgz
-*.key
-*.pyc
-__pycache__
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/README.md b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/README.md
deleted file mode 100644
index 87b176930..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/README.md
+++ /dev/null
@@ -1,78 +0,0 @@
-# Package assembly
-
-This set of scripts collect CI artifacts from a local directory or S3, and
-assembles them into a package structure defined by a packaging class in a
-staging directory.
-For the NugetPackage class the NuGet tool is then run (from within docker) on
-this staging directory to create a proper NuGet package (with all the metadata).
-While the StaticPackage class creates a tarball.
-
-The finalized nuget package maybe uploaded manually to NuGet.org
-
-## Requirements
-
- * Requires Python 3
- * Requires Docker
- * (if --s3) Requires private S3 access keys for the librdkafka-ci-packages bucket.
-
-
-
-## Usage
-
-1. Trigger CI builds by creating and pushing a new release (candidate) tag
- in the librdkafka repo. Make sure the tag is created on the correct branch.
-
- $ git tag v0.11.0-RC3
- $ git push origin v0.11.0-RC3
-
-2. Wait for CI builds to finish, monitor the builds here:
-
- * https://travis-ci.org/edenhill/librdkafka
- * https://ci.appveyor.com/project/edenhill/librdkafka
-
-Or if using SemaphoreCI, just have the packaging job depend on prior build jobs
-in the same pipeline.
-
-3. On a Linux host, run the release.py script to assemble the NuGet package
-
- $ cd packaging/nuget
- # Specify the tag
- $ ./release.py v0.11.0-RC3
- # Optionally, if the tag was moved and an exact sha is also required:
- # $ ./release.py --sha <the-full-git-sha> v0.11.0-RC3
-
-4. If all artifacts were available the NuGet package will be built
- and reside in the current directory as librdkafka.redist.<v-less-tag>.nupkg
-
-5. Test the package manually
-
-6. Upload the package to NuGet
-
- * https://www.nuget.org/packages/manage/upload
-
-7. If you trust this process you can have release.py upload the package
- automatically to NuGet after building it:
-
- $ ./release.py --retries 100 --upload your-nuget-api.key v0.11.0-RC3
-
-
-
-## Other uses
-
-### Create static library bundles
-
-To create a bundle (tarball) of librdkafka self-contained static library
-builds, use the following command:
-
- $ ./release.py --class StaticPackage v1.1.0
-
-
-### Clean up S3 bucket
-
-To clean up old non-release/non-RC builds from the S3 bucket, first check with:
-
- $ AWS_PROFILE=.. ./cleanup-s3.py --age 360
-
-Verify that the listed objects should really be deleted, then delete:
-
- $ AWS_PROFILE=.. ./cleanup-s3.py --age 360 --delete
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/artifact.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/artifact.py
deleted file mode 100755
index c58e0c9c7..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/artifact.py
+++ /dev/null
@@ -1,177 +0,0 @@
-#!/usr/bin/env python3
-#
-#
-# Collects CI artifacts from S3 storage, downloading them
-# to a local directory.
-#
-# The artifacts' folder in the S3 bucket must have the following token
-# format:
-# <token>-[<value>]__ (repeat)
-#
-# Recognized tokens (unrecognized tokens are ignored):
-# p - project (e.g., "confluent-kafka-python")
-# bld - builder (e.g., "travis")
-# plat - platform ("osx", "linux", ..)
-# arch - arch ("x64", ..)
-# tag - git tag
-# sha - git sha
-# bid - builder's build-id
-# bldtype - Release, Debug (appveyor)
-#
-# Example:
-# p-confluent-kafka-python__bld-travis__plat-linux__tag-__sha-112130ce297656ea1c39e7c94c99286f95133a24__bid-271588764__/confluent_kafka-0.11.0-cp35-cp35m-manylinux1_x86_64.whl
-
-
-import re
-import os
-import boto3
-
-import packaging
-
-s3_bucket = 'librdkafka-ci-packages'
-dry_run = False
-
-
-class Artifact (object):
- def __init__(self, arts, path, info=None):
- self.path = path
- # Remove unexpanded AppVeyor $(..) tokens from filename
- self.fname = re.sub(r'\$\([^\)]+\)', '', os.path.basename(path))
- slpath = os.path.join(os.path.dirname(path), self.fname)
- if os.path.isfile(slpath):
- # Already points to local file in correct location
- self.lpath = slpath
- else:
- # Prepare download location in dlpath
- self.lpath = os.path.join(arts.dlpath, slpath)
-
- if info is None:
- self.info = dict()
- else:
- # Assign the map and convert all keys to lower case
- self.info = {k.lower(): v for k, v in info.items()}
- # Rename values, e.g., 'plat':'linux' to 'plat':'debian'
- for k, v in self.info.items():
- rdict = packaging.rename_vals.get(k, None)
- if rdict is not None:
- self.info[k] = rdict.get(v, v)
-
- # Score value for sorting
- self.score = 0
-
- # AppVeyor symbol builds are of less value
- if self.fname.find('.symbols.') != -1:
- self.score -= 10
-
- self.arts = arts
- arts.artifacts.append(self)
-
- def __repr__(self):
- return self.path
-
- def __lt__(self, other):
- return self.score < other.score
-
- def download(self):
- """ Download artifact from S3 and store in local directory .lpath.
- If the artifact is already downloaded nothing is done. """
- if os.path.isfile(self.lpath) and os.path.getsize(self.lpath) > 0:
- return
- print('Downloading %s -> %s' % (self.path, self.lpath))
- if dry_run:
- return
- ldir = os.path.dirname(self.lpath)
- if not os.path.isdir(ldir):
- os.makedirs(ldir, 0o755)
- self.arts.s3_bucket.download_file(self.path, self.lpath)
-
-
-class Artifacts (object):
- def __init__(self, match, dlpath):
- super(Artifacts, self).__init__()
- self.match = match
- self.artifacts = list()
- # Download directory (make sure it ends with a path separator)
- if not dlpath.endswith(os.path.sep):
- dlpath = os.path.join(dlpath, '')
- self.dlpath = dlpath
- if not os.path.isdir(self.dlpath):
- if not dry_run:
- os.makedirs(self.dlpath, 0o755)
-
- def collect_single(self, path, req_tag=True):
- """ Collect single artifact, be it in S3 or locally.
- :param: path string: S3 or local (relative) path
- :param: req_tag bool: Require tag to match.
- """
-
- print('? %s' % path)
-
- # For local files, strip download path.
- # Also ignore any parent directories.
- if path.startswith(self.dlpath):
- folder = os.path.basename(os.path.dirname(path[len(self.dlpath):]))
- else:
- folder = os.path.basename(os.path.dirname(path))
-
- # The folder contains the tokens needed to perform
- # matching of project, gitref, etc.
- rinfo = re.findall(r'(?P<tag>[^-]+)-(?P<val>.*?)__', folder)
- if rinfo is None or len(rinfo) == 0:
- print('Incorrect folder/file name format for %s' % folder)
- return None
-
- info = dict(rinfo)
-
- # Ignore AppVeyor Debug builds
- if info.get('bldtype', '').lower() == 'debug':
- print('Ignoring debug artifact %s' % folder)
- return None
-
- tag = info.get('tag', None)
- if tag is not None and (len(tag) == 0 or tag.startswith('$(')):
- # AppVeyor doesn't substite $(APPVEYOR_REPO_TAG_NAME)
- # with an empty value when not set, it leaves that token
- # in the string - so translate that to no tag.
- del info['tag']
-
- # Match tag or sha to gitref
- unmatched = list()
- for m, v in self.match.items():
- if m not in info or info[m] != v:
- unmatched.append(m)
-
- # Make sure all matches were satisfied, unless this is a
- # common artifact.
- if info.get('p', '') != 'common' and len(unmatched) > 0:
- print(info)
- print('%s: %s did not match %s' %
- (info.get('p', None), folder, unmatched))
- return None
-
- return Artifact(self, path, info)
-
- def collect_s3(self):
- """ Collect and download build-artifacts from S3 based on
- git reference """
- print(
- 'Collecting artifacts matching %s from S3 bucket %s' %
- (self.match, s3_bucket))
- self.s3 = boto3.resource('s3')
- self.s3_bucket = self.s3.Bucket(s3_bucket)
- self.s3_client = boto3.client('s3')
- for item in self.s3_client.list_objects(
- Bucket=s3_bucket, Prefix='librdkafka/').get('Contents'):
- self.collect_single(item.get('Key'))
-
- for a in self.artifacts:
- a.download()
-
- def collect_local(self, path, req_tag=True):
- """ Collect artifacts from a local directory possibly previously
- collected from s3 """
- for f in [os.path.join(dp, f) for dp, dn,
- filenames in os.walk(path) for f in filenames]:
- if not os.path.isfile(f):
- continue
- self.collect_single(f, req_tag)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/cleanup-s3.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/cleanup-s3.py
deleted file mode 100755
index 2093af0c1..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/cleanup-s3.py
+++ /dev/null
@@ -1,143 +0,0 @@
-#!/usr/bin/env python3
-#
-# Clean up test builds from librdkafka's S3 bucket.
-# This also covers python builds.
-
-import re
-from datetime import datetime, timezone
-import boto3
-import argparse
-
-# Collects CI artifacts from S3 storage, downloading them
-# to a local directory, or collecting already downloaded artifacts from
-# local directory.
-#
-# The artifacts' folder in the S3 bucket must have the following token
-# format:
-# <token>-[<value>]__ (repeat)
-#
-# Recognized tokens (unrecognized tokens are ignored):
-# p - project (e.g., "confluent-kafka-python")
-# bld - builder (e.g., "travis")
-# plat - platform ("osx", "linux", ..)
-# arch - arch ("x64", ..)
-# tag - git tag
-# sha - git sha
-# bid - builder's build-id
-# bldtype - Release, Debug (appveyor)
-# lnk - std, static
-#
-# Example:
-# librdkafka/p-librdkafka__bld-travis__plat-linux__arch-x64__tag-v0.0.62__sha-d051b2c19eb0c118991cd8bc5cf86d8e5e446cde__bid-1562.1/librdkafka.tar.gz
-
-
-s3_bucket = 'librdkafka-ci-packages'
-
-
-def may_delete(path):
- """ Returns true if S3 object path is eligible for deletion, e.g.
- has a non-release/rc tag. """
-
- # The path contains the tokens needed to perform
- # matching of project, gitref, etc.
- rinfo = re.findall(r'(?P<tag>[^-]+)-(?P<val>.*?)(?:__|$)', path)
- if rinfo is None or len(rinfo) == 0:
- print(f"Incorrect folder/file name format for {path}")
- return False
-
- info = dict(rinfo)
-
- tag = info.get('tag', None)
- if tag is not None and (len(tag) == 0 or tag.startswith('$(')):
- # AppVeyor doesn't substite $(APPVEYOR_REPO_TAG_NAME)
- # with an empty value when not set, it leaves that token
- # in the string - so translate that to no tag.
- del info['tag']
- tag = None
-
- if tag is None:
- return True
-
- if re.match(r'^v?\d+\.\d+\.\d+(-?RC\d+)?$', tag,
- flags=re.IGNORECASE) is None:
- return True
-
- return False
-
-
-def collect_s3(s3, min_age_days=60):
- """ Collect artifacts from S3 """
- now = datetime.now(timezone.utc)
- eligible = []
- totcnt = 0
- # note: list_objects will return at most 1000 objects per call,
- # use continuation token to read full list.
- cont_token = None
- more = True
- while more:
- if cont_token is not None:
- res = s3.list_objects_v2(Bucket=s3_bucket,
- ContinuationToken=cont_token)
- else:
- res = s3.list_objects_v2(Bucket=s3_bucket)
-
- if res.get('IsTruncated') is True:
- cont_token = res.get('NextContinuationToken')
- else:
- more = False
-
- for item in res.get('Contents'):
- totcnt += 1
- age = (now - item.get('LastModified')).days
- path = item.get('Key')
- if age >= min_age_days and may_delete(path):
- eligible.append(path)
-
- return (eligible, totcnt)
-
-
-def chunk_list(lst, cnt):
- """ Split list into lists of cnt """
- for i in range(0, len(lst), cnt):
- yield lst[i:i + cnt]
-
-
-if __name__ == '__main__':
-
- parser = argparse.ArgumentParser()
- parser.add_argument("--delete",
- help="WARNING! Don't just check, actually delete "
- "S3 objects.",
- action="store_true")
- parser.add_argument("--age", help="Minimum object age in days.",
- type=int, default=360)
-
- args = parser.parse_args()
- dry_run = args.delete is not True
- min_age_days = args.age
-
- if dry_run:
- op = "Eligible for deletion"
- else:
- op = "Deleting"
-
- s3 = boto3.client('s3')
-
- # Collect eligible artifacts
- eligible, totcnt = collect_s3(s3, min_age_days=min_age_days)
- print(f"{len(eligible)}/{totcnt} eligible artifacts to delete")
-
- # Delete in chunks of 1000 (max what the S3 API can do)
- for chunk in chunk_list(eligible, 1000):
- print(op + ":\n" + '\n'.join(chunk))
- if dry_run:
- continue
-
- res = s3.delete_objects(Bucket=s3_bucket,
- Delete={
- 'Objects': [{'Key': x} for x in chunk],
- 'Quiet': True
- })
- errors = res.get('Errors', [])
- if len(errors) > 0:
- raise Exception(f"Delete failed: {errors}")
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr120.zip b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr120.zip
deleted file mode 100644
index 9bc5e9fbc..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr120.zip
+++ /dev/null
Binary files differ
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr140.zip b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr140.zip
deleted file mode 100644
index 152938138..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-win32__bldtype-Release/msvcr140.zip
+++ /dev/null
Binary files differ
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr120.zip b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr120.zip
deleted file mode 100644
index 3609c0385..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr120.zip
+++ /dev/null
Binary files differ
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr140.zip b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr140.zip
deleted file mode 100644
index b99e5ae5b..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/common/p-common__plat-windows__arch-x64__bldtype-Release/msvcr140.zip
+++ /dev/null
Binary files differ
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nuget.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nuget.sh
deleted file mode 100755
index 032371231..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nuget.sh
+++ /dev/null
@@ -1,21 +0,0 @@
-#!/bin/bash
-#
-#
-# Front-end for nuget that runs nuget in a docker image.
-
-set -ex
-
-if [[ -f /.dockerenv ]]; then
- echo "Inside docker"
-
- pushd $(dirname $0)
-
- nuget $*
-
- popd
-
-else
- echo "Running docker image"
- docker run -v $(pwd):/io mono:latest /io/$0 $*
-fi
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nugetpackage.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nugetpackage.py
deleted file mode 100644
index aea05ade0..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/nugetpackage.py
+++ /dev/null
@@ -1,286 +0,0 @@
-#!/usr/bin/env python3
-#
-# Create NuGet package
-#
-
-import os
-import tempfile
-import shutil
-import subprocess
-from packaging import Package, Mapping
-
-
-class NugetPackage (Package):
- """ All platforms, archs, et.al, are bundled into one set of
- NuGet output packages: "main", redist and symbols """
-
- # See .semamphore/semaphore.yml for where these are built.
- mappings = [
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'lnk': 'std'},
- 'librdkafka.tgz',
- './usr/local/include/librdkafka/rdkafka.h',
- 'build/native/include/librdkafka/rdkafka.h'),
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'lnk': 'std'},
- 'librdkafka.tgz',
- './usr/local/include/librdkafka/rdkafkacpp.h',
- 'build/native/include/librdkafka/rdkafkacpp.h'),
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'lnk': 'std'},
- 'librdkafka.tgz',
- './usr/local/include/librdkafka/rdkafka_mock.h',
- 'build/native/include/librdkafka/rdkafka_mock.h'),
-
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'lnk': 'std'},
- 'librdkafka.tgz',
- './usr/local/share/doc/librdkafka/README.md',
- 'README.md'),
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'lnk': 'std'},
- 'librdkafka.tgz',
- './usr/local/share/doc/librdkafka/CONFIGURATION.md',
- 'CONFIGURATION.md'),
- Mapping({'arch': 'x64',
- 'plat': 'osx',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/share/doc/librdkafka/LICENSES.txt',
- 'LICENSES.txt'),
-
- # OSX x64
- Mapping({'arch': 'x64',
- 'plat': 'osx'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka.dylib',
- 'runtimes/osx-x64/native/librdkafka.dylib'),
- # OSX arm64
- Mapping({'arch': 'arm64',
- 'plat': 'osx'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka.1.dylib',
- 'runtimes/osx-arm64/native/librdkafka.dylib'),
-
- # Linux glibc centos6 x64 with GSSAPI
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'centos6',
- 'lnk': 'std'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka.so.1',
- 'runtimes/linux-x64/native/librdkafka.so'),
- # Linux glibc centos6 x64 without GSSAPI (no external deps)
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'centos6',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka.so.1',
- 'runtimes/linux-x64/native/centos6-librdkafka.so'),
- # Linux glibc centos7 x64 with GSSAPI
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'centos7',
- 'lnk': 'std'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka.so.1',
- 'runtimes/linux-x64/native/centos7-librdkafka.so'),
- # Linux glibc centos7 arm64 without GSSAPI (no external deps)
- Mapping({'arch': 'arm64',
- 'plat': 'linux',
- 'dist': 'centos7',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka.so.1',
- 'runtimes/linux-arm64/native/librdkafka.so'),
-
- # Linux musl alpine x64 without GSSAPI (no external deps)
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'alpine',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka.so.1',
- 'runtimes/linux-x64/native/alpine-librdkafka.so'),
-
- # Common Win runtime
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'msvcr140.zip',
- 'vcruntime140.dll',
- 'runtimes/win-x64/native/vcruntime140.dll'),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'msvcr140.zip',
- 'msvcp140.dll', 'runtimes/win-x64/native/msvcp140.dll'),
-
- # matches x64 librdkafka.redist.zip
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/x64/Release/librdkafka.dll',
- 'runtimes/win-x64/native/librdkafka.dll'),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/x64/Release/librdkafkacpp.dll',
- 'runtimes/win-x64/native/librdkafkacpp.dll'),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/x64/Release/libcrypto-3-x64.dll',
- 'runtimes/win-x64/native/libcrypto-3-x64.dll'),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/x64/Release/libssl-3-x64.dll',
- 'runtimes/win-x64/native/libssl-3-x64.dll'),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/x64/Release/zlib1.dll',
- 'runtimes/win-x64/native/zlib1.dll'),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/x64/Release/zstd.dll',
- 'runtimes/win-x64/native/zstd.dll'),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/x64/Release/libcurl.dll',
- 'runtimes/win-x64/native/libcurl.dll'),
- # matches x64 librdkafka.redist.zip, lib files
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/lib/v142/x64/Release/librdkafka.lib',
- 'build/native/lib/win/x64/win-x64-Release/v142/librdkafka.lib' # noqa: E501
- ),
- Mapping({'arch': 'x64',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/lib/v142/x64/Release/librdkafkacpp.lib',
- 'build/native/lib/win/x64/win-x64-Release/v142/librdkafkacpp.lib' # noqa: E501
- ),
-
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'msvcr140.zip',
- 'vcruntime140.dll',
- 'runtimes/win-x86/native/vcruntime140.dll'),
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'msvcr140.zip',
- 'msvcp140.dll', 'runtimes/win-x86/native/msvcp140.dll'),
-
- # matches Win32 librdkafka.redist.zip
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/Win32/Release/librdkafka.dll',
- 'runtimes/win-x86/native/librdkafka.dll'),
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/Win32/Release/librdkafkacpp.dll',
- 'runtimes/win-x86/native/librdkafkacpp.dll'),
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/Win32/Release/libcrypto-3.dll',
- 'runtimes/win-x86/native/libcrypto-3.dll'),
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/Win32/Release/libssl-3.dll',
- 'runtimes/win-x86/native/libssl-3.dll'),
-
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/Win32/Release/zlib1.dll',
- 'runtimes/win-x86/native/zlib1.dll'),
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/Win32/Release/zstd.dll',
- 'runtimes/win-x86/native/zstd.dll'),
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/bin/v142/Win32/Release/libcurl.dll',
- 'runtimes/win-x86/native/libcurl.dll'),
-
- # matches Win32 librdkafka.redist.zip, lib files
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/lib/v142/Win32/Release/librdkafka.lib',
- 'build/native/lib/win/x86/win-x86-Release/v142/librdkafka.lib' # noqa: E501
- ),
- Mapping({'arch': 'x86',
- 'plat': 'win'},
- 'librdkafka.redist*',
- 'build/native/lib/v142/Win32/Release/librdkafkacpp.lib',
- 'build/native/lib/win/x86/win-x86-Release/v142/librdkafkacpp.lib' # noqa: E501
- )
- ]
-
- def __init__(self, version, arts):
- if version.startswith('v'):
- version = version[1:] # Strip v prefix
- super(NugetPackage, self).__init__(version, arts)
-
- def cleanup(self):
- if os.path.isdir(self.stpath):
- shutil.rmtree(self.stpath)
-
- def build(self, buildtype):
- """ Build single NuGet package for all its artifacts. """
-
- # NuGet removes the prefixing v from the version.
- vless_version = self.kv['version']
- if vless_version[0] == 'v':
- vless_version = vless_version[1:]
-
- self.stpath = tempfile.mkdtemp(prefix="out-", suffix="-%s" % buildtype,
- dir=".")
-
- self.render('librdkafka.redist.nuspec')
- self.copy_template('librdkafka.redist.targets',
- destpath=os.path.join('build', 'native'))
- self.copy_template('librdkafka.redist.props',
- destpath='build')
-
- # Generate template tokens for artifacts
- for a in self.arts.artifacts:
- if 'bldtype' not in a.info:
- a.info['bldtype'] = 'release'
-
- a.info['variant'] = '%s-%s-%s' % (a.info.get('plat'),
- a.info.get('arch'),
- a.info.get('bldtype'))
- if 'toolset' not in a.info:
- a.info['toolset'] = 'v142'
-
- # Apply mappings and extract files
- self.apply_mappings()
-
- print('Tree extracted to %s' % self.stpath)
-
- # After creating a bare-bone nupkg layout containing the artifacts
- # and some spec and props files, call the 'nuget' utility to
- # make a proper nupkg of it (with all the metadata files).
- subprocess.check_call("./nuget.sh pack %s -BasePath '%s' -NonInteractive" % # noqa: E501
- (os.path.join(self.stpath,
- 'librdkafka.redist.nuspec'),
- self.stpath), shell=True)
-
- return 'librdkafka.redist.%s.nupkg' % vless_version
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/packaging.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/packaging.py
deleted file mode 100755
index c4dab806d..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/packaging.py
+++ /dev/null
@@ -1,448 +0,0 @@
-#!/usr/bin/env python3
-#
-# Packaging script.
-# Assembles packages using CI artifacts.
-#
-
-import sys
-import re
-import os
-import shutil
-from fnmatch import fnmatch
-from string import Template
-from zfile import zfile
-import boto3
-import magic
-
-if sys.version_info[0] < 3:
- from urllib import unquote as _unquote
-else:
- from urllib.parse import unquote as _unquote
-
-
-def unquote(path):
- # Removes URL escapes, and normalizes the path by removing ./.
- path = _unquote(path)
- if path[:2] == './':
- return path[2:]
- return path
-
-
-# Rename token values
-rename_vals = {'plat': {'windows': 'win'},
- 'arch': {'x86_64': 'x64',
- 'amd64': 'x64',
- 'i386': 'x86',
- 'win32': 'x86'}}
-
-# Filemagic arch mapping.
-# key is (plat, arch, file_extension), value is a compiled filemagic regex.
-# This is used to verify that an artifact has the expected file type.
-magic_patterns = {
- ('win', 'x64', '.dll'): re.compile('PE32.*DLL.* x86-64, for MS Windows'),
- ('win', 'x86', '.dll'):
- re.compile('PE32.*DLL.* Intel 80386, for MS Windows'),
- ('win', 'x64', '.lib'): re.compile('current ar archive'),
- ('win', 'x86', '.lib'): re.compile('current ar archive'),
- ('linux', 'x64', '.so'): re.compile('ELF 64.* x86-64'),
- ('linux', 'arm64', '.so'): re.compile('ELF 64.* ARM aarch64'),
- ('osx', 'x64', '.dylib'): re.compile('Mach-O 64.* x86_64'),
- ('osx', 'arm64', '.dylib'): re.compile('Mach-O 64.*arm64')}
-
-magic = magic.Magic()
-
-
-def magic_mismatch(path, a):
- """ Verify that the filemagic for \\p path matches for artifact \\p a.
- Returns True if the magic file info does NOT match.
- Returns False if no matching is needed or the magic matches. """
- k = (a.info.get('plat', None), a.info.get('arch', None),
- os.path.splitext(path)[1])
- pattern = magic_patterns.get(k, None)
- if pattern is None:
- return False
-
- minfo = magic.id_filename(path)
- if not pattern.match(minfo):
- print(
- f"Warning: {path} magic \"{minfo}\" "
- f"does not match expected {pattern} for key {k}")
- return True
-
- return False
-
-
-# Collects CI artifacts from S3 storage, downloading them
-# to a local directory, or collecting already downloaded artifacts from
-# local directory.
-#
-# The artifacts' folder in the S3 bucket must have the following token
-# format:
-# <token>-[<value>]__ (repeat)
-#
-# Recognized tokens (unrecognized tokens are ignored):
-# p - project (e.g., "confluent-kafka-python")
-# bld - builder (e.g., "travis")
-# plat - platform ("osx", "linux", ..)
-# dist - distro or runtime ("centos6", "mingw", "msvcr", "alpine", ..).
-# arch - arch ("x64", ..)
-# tag - git tag
-# sha - git sha
-# bid - builder's build-id
-# bldtype - Release, Debug (appveyor)
-# lnk - Linkage ("std", "static", "all" (both std and static))
-# extra - Extra build options, typically "gssapi" (for cyrus-sasl linking).
-
-#
-# Example:
-# librdkafka/p-librdkafka__bld-travis__plat-linux__arch-x64__tag-v0.0.62__sha-d051b2c19eb0c118991cd8bc5cf86d8e5e446cde__bid-1562.1/librdkafka.tar.gz
-
-
-class MissingArtifactError(Exception):
- pass
-
-
-s3_bucket = 'librdkafka-ci-packages'
-dry_run = False
-
-
-class Artifact (object):
- def __init__(self, arts, path, info=None):
- self.path = path
- # Remove unexpanded AppVeyor $(..) tokens from filename
- self.fname = re.sub(r'\$\([^\)]+\)', '', os.path.basename(path))
- slpath = os.path.join(os.path.dirname(path), self.fname)
- if os.path.isfile(slpath):
- # Already points to local file in correct location
- self.lpath = slpath
- else:
- # Prepare download location in dlpath
- self.lpath = os.path.join(arts.dlpath, slpath)
-
- if info is None:
- self.info = dict()
- else:
- # Assign the map and convert all keys to lower case
- self.info = {k.lower(): v for k, v in info.items()}
- # Rename values, e.g., 'plat':'windows' to 'plat':'win'
- for k, v in self.info.items():
- rdict = rename_vals.get(k, None)
- if rdict is not None:
- self.info[k] = rdict.get(v, v)
-
- # Score value for sorting
- self.score = 0
-
- # AppVeyor symbol builds are of less value
- if self.fname.find('.symbols.') != -1:
- self.score -= 10
-
- self.arts = arts
- arts.artifacts.append(self)
-
- def __repr__(self):
- return self.path
-
- def __lt__(self, other):
- return self.score < other.score
-
- def download(self):
- """ Download artifact from S3 and store in local directory .lpath.
- If the artifact is already downloaded nothing is done. """
- if os.path.isfile(self.lpath) and os.path.getsize(self.lpath) > 0:
- return
- print('Downloading %s' % self.path)
- if dry_run:
- return
- ldir = os.path.dirname(self.lpath)
- if not os.path.isdir(ldir):
- os.makedirs(ldir, 0o755)
- self.arts.s3_bucket.download_file(self.path, self.lpath)
-
-
-class Artifacts (object):
- def __init__(self, match, dlpath):
- super(Artifacts, self).__init__()
- self.match = match
- self.artifacts = list()
- # Download directory (make sure it ends with a path separator)
- if not dlpath.endswith(os.path.sep):
- dlpath = os.path.join(dlpath, '')
- self.dlpath = dlpath
- if not os.path.isdir(self.dlpath):
- if not dry_run:
- os.makedirs(self.dlpath, 0o755)
-
- def collect_single(self, path, req_tag=True):
- """ Collect single artifact, be it in S3 or locally.
- :param: path string: S3 or local (relative) path
- :param: req_tag bool: Require tag to match.
- """
-
- # For local files, strip download path.
- # Also ignore any parent directories.
- if path.startswith(self.dlpath):
- folder = os.path.basename(os.path.dirname(path[len(self.dlpath):]))
- else:
- folder = os.path.basename(os.path.dirname(path))
-
- # The folder contains the tokens needed to perform
- # matching of project, gitref, etc.
- rinfo = re.findall(r'(?P<tag>[^-]+)-(?P<val>.*?)(?:__|$)', folder)
- if rinfo is None or len(rinfo) == 0:
- print('Incorrect folder/file name format for %s' % folder)
- return None
-
- info = dict(rinfo)
-
- # Ignore AppVeyor Debug builds
- if info.get('bldtype', '').lower() == 'debug':
- print('Ignoring debug artifact %s' % folder)
- return None
-
- tag = info.get('tag', None)
- if tag is not None and (len(tag) == 0 or tag.startswith('$(')):
- # AppVeyor doesn't substite $(APPVEYOR_REPO_TAG_NAME)
- # with an empty value when not set, it leaves that token
- # in the string - so translate that to no tag.
- del info['tag']
-
- # Perform matching
- unmatched = list()
- for m, v in self.match.items():
- if m not in info or info[m] != v:
- unmatched.append(f"{m} = {v}")
-
- # Make sure all matches were satisfied, unless this is a
- # common artifact.
- if info.get('p', '') != 'common' and len(unmatched) > 0:
- return None
-
- return Artifact(self, path, info)
-
- def collect_s3(self):
- """ Collect and download build-artifacts from S3 based on
- git reference """
- print(
- 'Collecting artifacts matching %s from S3 bucket %s' %
- (self.match, s3_bucket))
- self.s3 = boto3.resource('s3')
- self.s3_bucket = self.s3.Bucket(s3_bucket)
- self.s3_client = boto3.client('s3')
-
- # note: list_objects will return at most 1000 objects per call,
- # use continuation token to read full list.
- cont_token = None
- more = True
- while more:
- if cont_token is not None:
- res = self.s3_client.list_objects_v2(
- Bucket=s3_bucket,
- Prefix='librdkafka/',
- ContinuationToken=cont_token)
- else:
- res = self.s3_client.list_objects_v2(Bucket=s3_bucket,
- Prefix='librdkafka/')
-
- if res.get('IsTruncated') is True:
- cont_token = res.get('NextContinuationToken')
- else:
- more = False
-
- for item in res.get('Contents'):
- self.collect_single(item.get('Key'))
-
- for a in self.artifacts:
- a.download()
-
- def collect_local(self, path, req_tag=True):
- """ Collect artifacts from a local directory possibly previously
- collected from s3 """
- for f in [os.path.join(dp, f) for dp, dn,
- filenames in os.walk(path) for f in filenames]:
- if not os.path.isfile(f):
- continue
- self.collect_single(f, req_tag)
-
-
-class Mapping (object):
- """ Maps/matches a file in an input release artifact to
- the output location of the package, based on attributes and paths. """
-
- def __init__(self, attributes, artifact_fname_glob, path_in_artifact,
- output_pkg_path=None, artifact_fname_excludes=[]):
- """
- @param attributes A dict of artifact attributes that must match.
- If an attribute name (dict key) is prefixed
- with "!" (e.g., "!plat") then the attribute
- must not match.
- @param artifact_fname_glob Match artifacts with this filename glob.
- @param path_in_artifact On match, extract this file in the artifact,..
- @param output_pkg_path ..and write it to this location in the package.
- Defaults to path_in_artifact.
- @param artifact_fname_excludes Exclude artifacts matching these
- filenames.
-
- Pass a list of Mapping objects to FIXME to perform all mappings.
- """
- super(Mapping, self).__init__()
- self.attributes = attributes
- self.fname_glob = artifact_fname_glob
- self.input_path = path_in_artifact
- if output_pkg_path is None:
- self.output_path = self.input_path
- else:
- self.output_path = output_pkg_path
- self.name = self.output_path
- self.fname_excludes = artifact_fname_excludes
-
- def __str__(self):
- return self.name
-
-
-class Package (object):
- """ Generic Package class
- A Package is a working container for one or more output
- packages for a specific package type (e.g., nuget) """
-
- def __init__(self, version, arts):
- super(Package, self).__init__()
- self.version = version
- self.arts = arts
- # These may be overwritten by specific sub-classes:
- self.artifacts = arts.artifacts
- # Staging path, filled in later.
- self.stpath = None
- self.kv = {'version': version}
- self.files = dict()
-
- def add_file(self, file):
- self.files[file] = True
-
- def build(self):
- """ Build package output(s), return a list of paths "
- to built packages """
- raise NotImplementedError
-
- def cleanup(self):
- """ Optional cleanup routine for removing temporary files, etc. """
- pass
-
- def render(self, fname, destpath='.'):
- """ Render template in file fname and save to destpath/fname,
- where destpath is relative to stpath """
-
- outf = os.path.join(self.stpath, destpath, fname)
-
- if not os.path.isdir(os.path.dirname(outf)):
- os.makedirs(os.path.dirname(outf), 0o0755)
-
- with open(os.path.join('templates', fname), 'r') as tf:
- tmpl = Template(tf.read())
- with open(outf, 'w') as of:
- of.write(tmpl.substitute(self.kv))
-
- self.add_file(outf)
-
- def copy_template(self, fname, target_fname=None, destpath='.'):
- """ Copy template file to destpath/fname
- where destpath is relative to stpath """
-
- if target_fname is None:
- target_fname = fname
- outf = os.path.join(self.stpath, destpath, target_fname)
-
- if not os.path.isdir(os.path.dirname(outf)):
- os.makedirs(os.path.dirname(outf), 0o0755)
-
- shutil.copy(os.path.join('templates', fname), outf)
-
- self.add_file(outf)
-
- def apply_mappings(self):
- """ Applies a list of Mapping to match and extract files from
- matching artifacts. If any of the listed Mappings can not be
- fulfilled an exception is raised. """
-
- assert self.mappings
- assert len(self.mappings) > 0
-
- for m in self.mappings:
-
- artifact = None
- for a in self.arts.artifacts:
- found = True
-
- for attr in m.attributes:
- if attr[0] == '!':
- # Require attribute NOT to match
- origattr = attr
- attr = attr[1:]
-
- if attr in a.info and \
- a.info[attr] != m.attributes[origattr]:
- found = False
- break
- else:
- # Require attribute to match
- if attr not in a.info or \
- a.info[attr] != m.attributes[attr]:
- found = False
- break
-
- if not fnmatch(a.fname, m.fname_glob):
- found = False
-
- for exclude in m.fname_excludes:
- if exclude in a.fname:
- found = False
- break
-
- if found:
- artifact = a
- break
-
- if artifact is None:
- raise MissingArtifactError(
- '%s: unable to find artifact with tags %s matching "%s"' %
- (m, str(m.attributes), m.fname_glob))
-
- output_path = os.path.join(self.stpath, m.output_path)
-
- try:
- zfile.ZFile.extract(artifact.lpath, m.input_path, output_path)
-# except KeyError:
-# continue
- except Exception as e:
- raise Exception(
- '%s: file not found in archive %s: %s. Files in archive are:\n%s' % # noqa: E501
- (m, artifact.lpath, e, '\n'.join(zfile.ZFile(
- artifact.lpath).getnames())))
-
- # Check that the file type matches.
- if magic_mismatch(output_path, a):
- os.unlink(output_path)
- continue
-
- # All mappings found and extracted.
-
- def verify(self, path):
- """ Verify package content based on the previously defined mappings """
-
- missing = list()
- with zfile.ZFile(path, 'r') as zf:
- print('Verifying %s:' % path)
-
- # Zipfiles may url-encode filenames, unquote them before matching.
- pkgd = [unquote(x) for x in zf.getnames()]
- missing = [x for x in self.mappings if x.output_path not in pkgd]
-
- if len(missing) > 0:
- print(
- 'Missing files in package %s:\n%s' %
- (path, '\n'.join([str(x) for x in missing])))
- print('Actual: %s' % '\n'.join(pkgd))
- return False
-
- print('OK - %d expected files found' % len(self.mappings))
- return True
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/push-to-nuget.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/push-to-nuget.sh
deleted file mode 100755
index 598dd4cd7..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/push-to-nuget.sh
+++ /dev/null
@@ -1,21 +0,0 @@
-#!/bin/bash
-#
-# Upload NuGet package to NuGet.org using provided NuGet API key
-#
-
-set -e
-
-key=$1
-pkg=$2
-
-if [[ -z $pkg ]]; then
- echo "Usage: $0 <nuget.org-api-key> <nuget-package>"
- exit 1
-fi
-
-set -u
-
-docker run -t -v $PWD/$pkg:/$pkg mcr.microsoft.com/dotnet/sdk:3.1 \
- dotnet nuget push /$pkg -n -s https://api.nuget.org/v3/index.json \
- -k $key --source https://api.nuget.org/v3/index.json
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/release.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/release.py
deleted file mode 100755
index f230a580c..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/release.py
+++ /dev/null
@@ -1,167 +0,0 @@
-#!/usr/bin/env python3
-#
-#
-# NuGet release packaging tool.
-# Creates a NuGet package from CI artifacts on S3.
-#
-
-
-import os
-import sys
-import argparse
-import time
-import packaging
-import nugetpackage
-import staticpackage
-
-
-dry_run = False
-
-
-if __name__ == '__main__':
-
- parser = argparse.ArgumentParser()
- parser.add_argument(
- "--s3",
- help="Collect artifacts from S3 bucket",
- action="store_true")
- parser.add_argument("--dry-run",
- help="Locate artifacts but don't actually "
- "download or do anything",
- action="store_true")
- parser.add_argument(
- "--directory",
- help="Download directory (default: dl-<tag>)",
- default=None)
- parser.add_argument(
- "--no-cleanup",
- help="Don't clean up temporary folders",
- action="store_true")
- parser.add_argument(
- "--sha",
- help="Also match on this git sha1",
- default=None)
- parser.add_argument(
- "--ignore-tag",
- help="Ignore the artifacts' tag attribute (for devel use only)",
- action="store_true",
- default=False)
- parser.add_argument(
- "--nuget-version",
- help="The nuget package version (defaults to same as tag)",
- default=None)
- parser.add_argument("--upload", help="Upload package to after building, "
- "using provided NuGet API key "
- "(either file or the key itself)",
- default=None,
- type=str)
- parser.add_argument(
- "--class",
- help="Packaging class (either NugetPackage or StaticPackage)",
- default="NugetPackage",
- dest="pkgclass")
- parser.add_argument(
- "--retries",
- help="Number of retries to collect artifacts",
- default=0,
- type=int)
- parser.add_argument("tag", help="Git tag to collect")
-
- args = parser.parse_args()
- dry_run = args.dry_run
- retries = args.retries
- if not args.directory:
- args.directory = 'dl-%s' % args.tag
-
- match = {}
- if not args.ignore_tag:
- match['tag'] = args.tag
-
- if args.sha is not None:
- match['sha'] = args.sha
-
- if args.pkgclass == "NugetPackage":
- pkgclass = nugetpackage.NugetPackage
- elif args.pkgclass == "StaticPackage":
- pkgclass = staticpackage.StaticPackage
- else:
- raise ValueError(f'Unknown packaging class {args.pkgclass}: '
- 'should be one of NugetPackage or StaticPackage')
-
- try:
- match.update(getattr(pkgclass, 'match'))
- except BaseException:
- pass
-
- arts = packaging.Artifacts(match, args.directory)
-
- # Collect common local artifacts, such as support files.
- arts.collect_local('common', req_tag=False)
-
- while True:
- if args.s3:
- arts.collect_s3()
-
- arts.collect_local(arts.dlpath)
-
- if len(arts.artifacts) == 0:
- raise ValueError('No artifacts found for %s' % match)
-
- print('Collected artifacts (%s):' % (arts.dlpath))
- for a in arts.artifacts:
- print(' %s' % a.lpath)
- print('')
-
- if args.nuget_version is not None:
- package_version = args.nuget_version
- else:
- package_version = args.tag
-
- print('')
-
- if dry_run:
- sys.exit(0)
-
- print('Building packages:')
-
- try:
- p = pkgclass(package_version, arts)
- pkgfile = p.build(buildtype='release')
- break
- except packaging.MissingArtifactError as e:
- if retries <= 0 or not args.s3:
- if not args.no_cleanup:
- p.cleanup()
- raise e
-
- p.cleanup()
- retries -= 1
- print(e)
- print('Retrying in 30 seconds')
- time.sleep(30)
-
- if not args.no_cleanup:
- p.cleanup()
- else:
- print(' --no-cleanup: leaving %s' % p.stpath)
-
- print('')
-
- if not p.verify(pkgfile):
- print('Package failed verification.')
- sys.exit(1)
-
- print('Created package: %s' % pkgfile)
-
- if args.upload is not None:
- if os.path.isfile(args.upload):
- with open(args.upload, 'r') as f:
- nuget_key = f.read().replace('\n', '')
- else:
- nuget_key = args.upload
-
- print('Uploading %s to NuGet' % pkgfile)
- r = os.system("./push-to-nuget.sh '%s' %s" % (nuget_key, pkgfile))
- assert int(r) == 0, \
- f"NuGet upload failed with exit code {r}, see previous errors"
- print('%s successfully uploaded to NuGet' % pkgfile)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/requirements.txt b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/requirements.txt
deleted file mode 100644
index 0fa2fd19c..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/requirements.txt
+++ /dev/null
@@ -1,3 +0,0 @@
-boto3==1.18.45
-rpmfile==1.0.8
-filemagic==1.6
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/staticpackage.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/staticpackage.py
deleted file mode 100644
index 38567bb60..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/staticpackage.py
+++ /dev/null
@@ -1,178 +0,0 @@
-#!/usr/bin/env python3
-#
-# Create self-contained static-library tar-ball package
-#
-
-import os
-import tempfile
-import shutil
-import subprocess
-from packaging import Package, Mapping
-
-
-class StaticPackage (Package):
- """ Create a tar-ball with self-contained static libraries.
- These are later imported into confluent-kafka-go. """
-
- # Make sure gssapi (cyrus-sasl) is not linked, since that is a
- # dynamic linkage, by specifying negative match '!extra': 'gssapi'.
- # Except for on OSX where cyrus-sasl is always available, and
- # Windows where it is never linked.
- #
- # Match statically linked artifacts (which are included in 'all' builds)
- mappings = [
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'centos6',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/include/librdkafka/rdkafka.h',
- 'rdkafka.h'),
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'centos6',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/share/doc/librdkafka/LICENSES.txt',
- 'LICENSES.txt'),
-
- # glibc linux static lib and pkg-config file
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'centos6',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka-static.a',
- 'librdkafka_glibc_linux_amd64.a'),
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'centos6',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/pkgconfig/rdkafka-static.pc',
- 'librdkafka_glibc_linux_amd64.pc'),
-
- # glibc linux arm64 static lib and pkg-config file
- Mapping({'arch': 'arm64',
- 'plat': 'linux',
- 'dist': 'centos7',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka-static.a',
- 'librdkafka_glibc_linux_arm64.a'),
- Mapping({'arch': 'arm64',
- 'plat': 'linux',
- 'dist': 'centos7',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/pkgconfig/rdkafka-static.pc',
- 'librdkafka_glibc_linux_arm64.pc'),
-
- # musl linux static lib and pkg-config file
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'alpine',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka-static.a',
- 'librdkafka_musl_linux_amd64.a'),
- Mapping({'arch': 'x64',
- 'plat': 'linux',
- 'dist': 'alpine',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/pkgconfig/rdkafka-static.pc',
- 'librdkafka_musl_linux_amd64.pc'),
-
- # musl linux arm64 static lib and pkg-config file
- Mapping({'arch': 'arm64',
- 'plat': 'linux',
- 'dist': 'alpine',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka-static.a',
- 'librdkafka_musl_linux_arm64.a'),
- Mapping({'arch': 'arm64',
- 'plat': 'linux',
- 'dist': 'alpine',
- 'lnk': 'all',
- '!extra': 'gssapi'},
- 'librdkafka.tgz',
- './usr/local/lib/pkgconfig/rdkafka-static.pc',
- 'librdkafka_musl_linux_arm64.pc'),
-
- # osx x64 static lib and pkg-config file
- Mapping({'arch': 'x64',
- 'plat': 'osx',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka-static.a',
- 'librdkafka_darwin_amd64.a'),
- Mapping({'arch': 'x64',
- 'plat': 'osx',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/lib/pkgconfig/rdkafka-static.pc',
- 'librdkafka_darwin_amd64.pc'),
-
- # osx arm64 static lib and pkg-config file
- Mapping({'arch': 'arm64',
- 'plat': 'osx',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/lib/librdkafka-static.a',
- 'librdkafka_darwin_arm64.a'),
- Mapping({'arch': 'arm64',
- 'plat': 'osx',
- 'lnk': 'all'},
- 'librdkafka.tgz',
- './usr/local/lib/pkgconfig/rdkafka-static.pc',
- 'librdkafka_darwin_arm64.pc'),
-
- # win static lib and pkg-config file (mingw)
- Mapping({'arch': 'x64',
- 'plat': 'win',
- 'dist': 'mingw',
- 'lnk': 'static'},
- 'librdkafka.tgz',
- './lib/librdkafka-static.a', 'librdkafka_windows.a'),
- Mapping({'arch': 'x64',
- 'plat': 'win',
- 'dist': 'mingw',
- 'lnk': 'static'},
- 'librdkafka.tgz',
- './lib/pkgconfig/rdkafka-static.pc',
- 'librdkafka_windows.pc'),
- ]
-
- def __init__(self, version, arts):
- super(StaticPackage, self).__init__(version, arts)
-
- def cleanup(self):
- if os.path.isdir(self.stpath):
- shutil.rmtree(self.stpath)
-
- def build(self, buildtype):
- """ Build single package for all artifacts. """
-
- self.stpath = tempfile.mkdtemp(prefix="out-", dir=".")
-
- self.apply_mappings()
-
- print('Tree extracted to %s' % self.stpath)
-
- # After creating a bare-bone layout, create a tarball.
- outname = "librdkafka-static-bundle-%s.tgz" % self.version
- print('Writing to %s in %s' % (outname, self.stpath))
- subprocess.check_call("(cd %s && tar cvzf ../%s .)" %
- (self.stpath, outname),
- shell=True)
-
- return outname
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.nuspec b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.nuspec
deleted file mode 100644
index dbfd7b1aa..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.nuspec
+++ /dev/null
@@ -1,21 +0,0 @@
-<?xml version="1.0" encoding="utf-8"?>
-<package xmlns="http://schemas.microsoft.com/packaging/2011/10/nuspec.xsd">
- <metadata>
- <id>librdkafka.redist</id>
- <version>${version}</version>
- <title>librdkafka - redistributable</title>
- <authors>Magnus Edenhill, edenhill</authors>
- <owners>Confluent Inc.</owners>
- <requireLicenseAcceptance>false</requireLicenseAcceptance>
- <licenseUrl>https://github.com/confluentinc/librdkafka/blob/master/LICENSES.txt</licenseUrl>
- <projectUrl>https://github.com/confluentinc/librdkafka</projectUrl>
- <description>The Apache Kafka C/C++ client library - redistributable</description>
- <summary>The Apache Kafka C/C++ client library</summary>
- <releaseNotes>Release of librdkafka</releaseNotes>
- <copyright>Copyright 2012-2023</copyright>
- <tags>native apache kafka librdkafka C C++ nativepackage</tags>
- </metadata>
- <files>
- <file src="**" />
- </files>
-</package>
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.props b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.props
deleted file mode 100644
index c1615c61c..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.props
+++ /dev/null
@@ -1,18 +0,0 @@
-<?xml version="1.0" encoding="utf-8" ?>
-<Project ToolsVersion="12.0" xmlns="http://schemas.microsoft.com/developer/msbuild/2003">
- <ItemGroup>
- <Content Include="$(MSBuildThisFileDirectory)..\runtimes\win-x86\native\*">
- <Link>librdkafka\x86\%(Filename)%(Extension)</Link>
- <CopyToOutputDirectory>PreserveNewest</CopyToOutputDirectory>
- </Content>
- <Content Include="$(MSBuildThisFileDirectory)..\runtimes\win-x64\native\*">
- <Link>librdkafka\x64\%(Filename)%(Extension)</Link>
- <CopyToOutputDirectory>PreserveNewest</CopyToOutputDirectory>
- </Content>
- </ItemGroup>
- <ItemDefinitionGroup>
- <ClCompile>
- <AdditionalIncludeDirectories>$(MSBuildThisFileDirectory)include;%(AdditionalIncludeDirectories)</AdditionalIncludeDirectories>
- </ClCompile>
- </ItemDefinitionGroup>
-</Project>
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.targets b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.targets
deleted file mode 100644
index d174cda11..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/templates/librdkafka.redist.targets
+++ /dev/null
@@ -1,19 +0,0 @@
-<Project ToolsVersion="4.0" xmlns="http://schemas.microsoft.com/developer/msbuild/2003">
- <ItemDefinitionGroup>
- <Link>
- <AdditionalDependencies Condition="'$(Platform)' == 'x64'">$(MSBuildThisFileDirectory)lib\win\x64\win-x64-Release\v142\librdkafka.lib;%(AdditionalDependencies)</AdditionalDependencies>
- <AdditionalDependencies Condition="'$(Platform)' != 'x64'">$(MSBuildThisFileDirectory)lib\win\x86\win-x86-Release\v142\librdkafka.lib;%(AdditionalDependencies)</AdditionalDependencies>
- <AdditionalLibraryDirectories Condition="'$(Platform)' == 'x64'">$(MSBuildThisFileDirectory)lib\win\x64\win-x64-Release\v142;%(AdditionalLibraryDirectories)</AdditionalLibraryDirectories>
- <AdditionalLibraryDirectories Condition="'$(Platform)' != 'x64'">$(MSBuildThisFileDirectory)lib\win\x86\win-x86-Release\v142;%(AdditionalLibraryDirectories)</AdditionalLibraryDirectories>
- </Link>
- <ClCompile>
- <AdditionalIncludeDirectories>$(MSBuildThisFileDirectory)include;%(AdditionalIncludeDirectories)</AdditionalIncludeDirectories>
- </ClCompile>
- </ItemDefinitionGroup>
- <ItemGroup Condition="'$(Platform)' == 'x64'">
- <ReferenceCopyLocalPaths Include="$(MSBuildThisFileDirectory)..\..\runtimes\win-x64\native\*.dll" />
- </ItemGroup>
- <ItemGroup Condition="'$(Platform)' != 'x64'">
- <ReferenceCopyLocalPaths Include="$(MSBuildThisFileDirectory)..\..\runtimes\win-x86\native\*.dll" />
- </ItemGroup>
-</Project>
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/__init__.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/__init__.py
deleted file mode 100644
index e69de29bb..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/__init__.py
+++ /dev/null
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/zfile.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/zfile.py
deleted file mode 100644
index 51f2df25f..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/nuget/zfile/zfile.py
+++ /dev/null
@@ -1,98 +0,0 @@
-#!/usr/bin/env python3
-
-import os
-import tarfile
-import zipfile
-import rpmfile
-
-
-class ZFile (object):
- def __init__(self, path, mode='r', ext=None):
- super(ZFile, self).__init__()
-
- if ext is not None:
- _ext = ext
- else:
- _ext = os.path.splitext(path)[-1]
- if _ext.startswith('.'):
- _ext = _ext[1:]
-
- if zipfile.is_zipfile(path) or _ext == 'zip':
- self.f = zipfile.ZipFile(path, mode)
- elif tarfile.is_tarfile(path) or _ext in ('tar', 'tgz', 'gz'):
- self.f = tarfile.open(path, mode)
- elif _ext == 'rpm':
- self.f = rpmfile.open(path, mode + 'b')
- else:
- raise ValueError('Unsupported file extension: %s' % path)
-
- def __enter__(self):
- return self
-
- def __exit__(self, *args):
- if callable(getattr(self.f, 'close', None)):
- self.f.close()
-
- def getnames(self):
- if isinstance(self.f, zipfile.ZipFile):
- return self.f.namelist()
- elif isinstance(self.f, tarfile.TarFile):
- return self.f.getnames()
- elif isinstance(self.f, rpmfile.RPMFile):
- return [x.name for x in self.f.getmembers()]
- else:
- raise NotImplementedError
-
- def headers(self):
- if isinstance(self.f, rpmfile.RPMFile):
- return self.f.headers
- else:
- return dict()
-
- def extract_to(self, member, path):
- """ Extract compress file's \\p member to \\p path
- If \\p path is a directory the member's basename will used as
- filename, otherwise path is considered the full file path name. """
-
- if not os.path.isdir(os.path.dirname(path)):
- os.makedirs(os.path.dirname(path))
-
- if os.path.isdir(path):
- path = os.path.join(path, os.path.basename(member))
-
- with open(path, 'wb') as of:
- if isinstance(self.f, zipfile.ZipFile):
- zf = self.f.open(member)
- else:
- zf = self.f.extractfile(member)
-
- while True:
- b = zf.read(1024 * 100)
- if b:
- of.write(b)
- else:
- break
-
- zf.close()
-
- @classmethod
- def extract(cls, zpath, member, outpath):
- """
- Extract file member (full internal path) to output from
- archive zpath.
- """
-
- with ZFile(zpath) as zf:
- zf.extract_to(member, outpath)
-
- @classmethod
- def compress(cls, zpath, paths, stripcnt=0, ext=None):
- """
- Create new compressed file \\p zpath containing files in \\p paths
- """
-
- with ZFile(zpath, 'w', ext=ext) as zf:
- for p in paths:
- outp = os.path.sep.join(p.split(os.path.sep)[stripcnt:])
- print('zip %s to %s (stripcnt %d)' % (p, outp, stripcnt))
- zf.f.write(p, outp)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/.gitignore b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/.gitignore
deleted file mode 100644
index 4bfdf21ed..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/.gitignore
+++ /dev/null
@@ -1,7 +0,0 @@
-*.log
-available_pkgs
-installed_pkgs
-pkgs-*
-arts-*
-cache
-output
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/Makefile b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/Makefile
deleted file mode 100644
index c5c8f8c10..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/Makefile
+++ /dev/null
@@ -1,92 +0,0 @@
-PACKAGE_NAME?= librdkafka
-VERSION?= $(shell ../get_version.py ../../src/rdkafka.h)
-
-# Jenkins CI integration
-BUILD_NUMBER?= 1
-
-MOCK_CONFIG?=default
-
-RESULT_DIR?=pkgs-$(VERSION)-$(BUILD_NUMBER)-$(MOCK_CONFIG)
-
-# Where built packages are copied with `make copy-artifacts`
-ARTIFACTS_DIR?=../../artifacts
-
-all: rpm
-
-
-SOURCES:
- mkdir -p SOURCES
-
-archive: SOURCES
- cd ../../ && \
- git archive --prefix=$(PACKAGE_NAME)-$(VERSION)/ \
- -o packaging/rpm/SOURCES/$(PACKAGE_NAME)-$(VERSION).tar.gz HEAD
-
-
-build_prepare: archive
- mkdir -p $(RESULT_DIR)
- rm -f $(RESULT_DIR)/$(PACKAGE_NAME)*.rpm
-
-
-srpm: build_prepare
- /usr/bin/mock \
- -r $(MOCK_CONFIG) \
- $(MOCK_OPTIONS) \
- --define "__version $(VERSION)" \
- --define "__release $(BUILD_NUMBER)" \
- --enable-network \
- --resultdir=$(RESULT_DIR) \
- --no-clean --no-cleanup-after \
- --install epel-release \
- --buildsrpm \
- --spec=librdkafka.spec \
- --sources=SOURCES || \
- (tail -n 100 pkgs-$(VERSION)*/*log ; false)
- @echo "======= Source RPM now available in $(RESULT_DIR) ======="
-
-rpm: srpm
- /usr/bin/mock \
- -r $(MOCK_CONFIG) \
- $(MOCK_OPTIONS) \
- --define "__version $(VERSION)"\
- --define "__release $(BUILD_NUMBER)"\
- --enable-network \
- --resultdir=$(RESULT_DIR) \
- --no-clean --no-cleanup-after \
- --rebuild $(RESULT_DIR)/$(PACKAGE_NAME)*.src.rpm || \
- (tail -n 100 pkgs-$(VERSION)*/*log ; false)
- @echo "======= Binary RPMs now available in $(RESULT_DIR) ======="
-
-copy-artifacts:
- cp $(RESULT_DIR)/*rpm $(ARTIFACTS_DIR)
-
-clean:
- rm -rf SOURCES
- /usr/bin/mock -r $(MOCK_CONFIG) --clean
-
-distclean: clean
- rm -f build.log root.log state.log available_pkgs installed_pkgs \
- *.rpm *.tar.gz
-
-# Prepare ubuntu 14.04 for building RPMs with mock.
-# - older versions of mock needs the config file to reside in /etc/mock,
-# so we copy it there.
-# - add a mock system group (if not already exists)
-# - add the current user to the mock group.
-# - prepare mock environment with some needed packages.
-# NOTE: This target should be run with sudo.
-prepare_ubuntu:
- apt-get -qq update
- apt-get install -y -qq mock make git python-lzma
- cp *.cfg /etc/mock/
- addgroup --system mock || true
- adduser $$(whoami) mock
- /usr/bin/mock -r $(MOCK_CONFIG) --init
- /usr/bin/mock -r $(MOCK_CONFIG) \
- --enable-network \
- --no-cleanup-after \
- --install epel-release shadow-utils
-
-prepare_centos:
- yum install -y -q mock make git
- cp *.cfg /etc/mock/
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/README.md b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/README.md
deleted file mode 100644
index 92a6eca95..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/README.md
+++ /dev/null
@@ -1,23 +0,0 @@
-# RPM packages for librdkafka
-
-On a system with RPM mock installed, simply run make to create RPM packages:
-
- $ make
-
-Additional mock options may be specified using MOCK_OPTIONS:
-
- $ make MOCK_OPTIONS='--bootstrap-chroot'
-
-
-## Build with Mock on docker
-
-From the librdkafka top-level directory:
-
- $ packaging/rpm/mock-on-docker.sh
-
-Wait for packages to build, they will be copied to top-level dir artifacts/
-
-Test the packages:
-
- $ packaging/rpm/tests/test-on-docker.sh
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/el7-x86_64.cfg b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/el7-x86_64.cfg
deleted file mode 100644
index 502282749..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/el7-x86_64.cfg
+++ /dev/null
@@ -1,40 +0,0 @@
-config_opts['root'] = 'el7-x86_64'
-config_opts['target_arch'] = 'x86_64'
-config_opts['legal_host_arches'] = ('x86_64',)
-config_opts['chroot_setup_cmd'] = 'install @buildsys-build'
-config_opts['dist'] = 'el7' # only useful for --resultdir variable subst
-config_opts['releasever'] = '7'
-config_opts['docker_unshare_warning'] = False
-config_opts['nosync'] = True
-
-config_opts['yum.conf'] = """
-[main]
-keepcache=1
-debuglevel=2
-reposdir=/dev/null
-logfile=/var/log/yum.log
-retries=15
-obsoletes=1
-gpgcheck=0
-assumeyes=1
-syslog_ident=mock
-syslog_device=
-mdpolicy=group:primary
-
-# repos
-[base]
-name=BaseOS
-mirrorlist=http://mirrorlist.centos.org/?release=7&arch=x86_64&repo=os
-failovermethod=priority
-
-[updates]
-name=updates
-enabled=1
-mirrorlist=http://mirrorlist.centos.org/?release=7&arch=x86_64&repo=updates
-failovermethod=priority
-
-[epel]
-name=epel
-mirrorlist=http://mirrors.fedoraproject.org/mirrorlist?repo=epel-7&arch=x86_64
-failovermethod=priority
-"""
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/librdkafka.spec b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/librdkafka.spec
deleted file mode 100644
index 4f9e8c0d0..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/librdkafka.spec
+++ /dev/null
@@ -1,118 +0,0 @@
-Name: librdkafka
-Version: %{__version}
-Release: %{__release}%{?dist}
-%define soname 1
-
-Summary: The Apache Kafka C library
-Group: Development/Libraries/C and C++
-License: BSD-2-Clause
-URL: https://github.com/edenhill/librdkafka
-Source: librdkafka-%{version}.tar.gz
-
-BuildRequires: zlib-devel libstdc++-devel gcc >= 4.1 gcc-c++ cyrus-sasl-devel
-BuildRoot: %(mktemp -ud %{_tmppath}/%{name}-%{version}-%{release}-XXXXXX)
-
-%define _source_payload w9.gzdio
-%define _binary_payload w9.gzdio
-
-%description
-librdkafka is the C/C++ client library implementation of the Apache Kafka protocol, containing both Producer and Consumer support.
-
-
-%package -n %{name}%{soname}
-Summary: The Apache Kafka C library
-Group: Development/Libraries/C and C++
-Requires: zlib libstdc++ cyrus-sasl
-# openssl libraries were extract to openssl-libs in RHEL7
-%if 0%{?rhel} >= 7
-Requires: openssl-libs >= 1.0.2
-BuildRequires: openssl-devel >= 1.0.2 python3
-%else
-Requires: openssl
-# python34 is provided from epel-release, but that package needs to be installed
-# prior to rpmbuild working out these dependencies (such as from mock).
-BuildRequires: openssl-devel python34
-%endif
-
-%description -n %{name}%{soname}
-librdkafka is the C/C++ client library implementation of the Apache Kafka protocol, containing both Producer and Consumer support.
-
-
-%package -n %{name}-devel
-Summary: The Apache Kafka C library (Development Environment)
-Group: Development/Libraries/C and C++
-Requires: %{name}%{soname} = %{version}
-
-%description -n %{name}-devel
-librdkafka is the C/C++ client library implementation of the Apache Kafka protocol, containing both Producer and Consumer support.
-
-This package contains headers and libraries required to build applications
-using librdkafka.
-
-
-%prep
-%setup -q -n %{name}-%{version}
-
-# --install-deps will install missing dependencies that are not available
-# through BuildRequires, such as libzstd, which will be linked statically.
-%configure --install-deps --disable-lz4-ext
-
-%build
-cat config.log
-make
-examples/rdkafka_example -X builtin.features
-
-%install
-rm -rf %{buildroot}
-DESTDIR=%{buildroot} make install
-
-%clean
-rm -rf %{buildroot}
-
-%post -n %{name}%{soname} -p /sbin/ldconfig
-%postun -n %{name}%{soname} -p /sbin/ldconfig
-
-%files -n %{name}%{soname}
-%defattr(444,root,root)
-%{_libdir}/librdkafka.so.%{soname}
-%{_libdir}/librdkafka++.so.%{soname}
-%defattr(-,root,root)
-%doc %{_docdir}/librdkafka/README.md
-%doc %{_docdir}/librdkafka/LICENSE
-%doc %{_docdir}/librdkafka/CONFIGURATION.md
-%doc %{_docdir}/librdkafka/INTRODUCTION.md
-%doc %{_docdir}/librdkafka/STATISTICS.md
-%doc %{_docdir}/librdkafka/CHANGELOG.md
-%doc %{_docdir}/librdkafka/LICENSES.txt
-
-%defattr(-,root,root)
-#%{_bindir}/rdkafka_example
-#%{_bindir}/rdkafka_performance
-
-
-%files -n %{name}-devel
-%defattr(-,root,root)
-%{_includedir}/librdkafka
-%defattr(444,root,root)
-%{_libdir}/librdkafka.a
-%{_libdir}/librdkafka-static.a
-%{_libdir}/librdkafka.so
-%{_libdir}/librdkafka++.a
-%{_libdir}/librdkafka++.so
-%{_libdir}/pkgconfig/rdkafka++.pc
-%{_libdir}/pkgconfig/rdkafka.pc
-%{_libdir}/pkgconfig/rdkafka-static.pc
-%{_libdir}/pkgconfig/rdkafka++-static.pc
-
-%changelog
-* Thu Apr 09 2015 Eduard Iskandarov <e.iskandarov@corp.mail.ru> 0.8.6-0
-- 0.8.6 simplify build process
-
-* Fri Oct 24 2014 Magnus Edenhill <rdkafka@edenhill.se> 0.8.5-0
-- 0.8.5 release
-
-* Mon Aug 18 2014 Magnus Edenhill <rdkafka@edenhill.se> 0.8.4-0
-- 0.8.4 release
-
-* Mon Mar 17 2014 Magnus Edenhill <vk@edenhill.se> 0.8.3-0
-- Initial RPM package
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/mock-on-docker.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/mock-on-docker.sh
deleted file mode 100755
index eec3d54a7..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/mock-on-docker.sh
+++ /dev/null
@@ -1,97 +0,0 @@
-#!/bin/bash
-#
-#
-#
-# Run mock in docker to create RPM packages of librdkafka.
-#
-# Usage:
-# packaging/rpm/mock-on-docker.sh [<mock configs ..>]
-#
-
-set -ex
-
-_DOCKER_IMAGE=fedora:35
-_MOCK_CONFIGS="centos+epel-7-x86_64 centos-stream+epel-8-x86_64"
-
-if [[ $1 == "--build" ]]; then
- on_builder=1
- shift
-else
- on_builder=0
-fi
-
-
-if [[ -n $* ]]; then
- _MOCK_CONFIGS="$*"
-fi
-
-
-if [[ $on_builder == 0 ]]; then
- #
- # Running on host, fire up a docker container and run the latter
- # part of this script in docker.
- #
-
- if [[ ! -f configure.self ]]; then
- echo "$0 must be run from librdkafka top directory"
- exit 1
- fi
-
- mkdir -p ${PWD}/packaging/rpm/cache/mock
-
- docker run \
- --privileged \
- -t \
- -v ${PWD}/packaging/rpm/cache/mock:/var/cache/mock \
- -v ${PWD}:/io \
- $_DOCKER_IMAGE \
- /io/packaging/rpm/mock-on-docker.sh --build $_MOCK_CONFIGS
-
- mkdir -p artifacts
- for MOCK_CONFIG in $_MOCK_CONFIGS ; do
- cp -vr --no-preserve=ownership packaging/rpm/arts-${MOCK_CONFIG}/*rpm artifacts/
- done
-
- echo "All Done"
-
-else
- #
- # Running in docker container.
- #
-
- dnf install -y -q mock mock-core-configs make git
-
- echo "%_netsharedpath /sys:/proc" >> /etc/rpm/macros.netshared
-
- pushd /io/packaging/rpm
-
- for MOCK_CONFIG in $_MOCK_CONFIGS ; do
- cfg_file=/etc/mock/${MOCK_CONFIG}.cfg
- if [[ ! -f $cfg_file ]]; then
- echo "Error: Mock config $cfg_file does not exist"
- exit 1
- fi
-
- echo "config_opts['plugin_conf']['bind_mount_enable'] = False" >> $cfg_file
- echo "config_opts['docker_unshare_warning'] = False" >> $cfg_file
- echo "Building $MOCK_CONFIG in $PWD"
- cat $cfg_file
-
- echo "Setting git safe.directory"
- git config --global --add safe.directory /io
-
- export MOCK_CONFIG=$MOCK_CONFIG
- make all
-
- echo "Done building $MOCK_CONFIG: copying artifacts"
- artdir="arts-$MOCK_CONFIG"
- mkdir -p "$artdir"
- make ARTIFACTS_DIR="$artdir" copy-artifacts
-
- done
-
- popd
- echo "Done"
-fi
-
-exit 0
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/.gitignore b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/.gitignore
deleted file mode 100644
index 333a2b7ac..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/.gitignore
+++ /dev/null
@@ -1,2 +0,0 @@
-test
-testcpp
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/Makefile b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/Makefile
deleted file mode 100644
index edd457997..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/Makefile
+++ /dev/null
@@ -1,25 +0,0 @@
-
-PROGS?=test test-static testcpp testcpp-static
-
-all: $(PROGS)
-
-test: test.c
- $(CC) -O2 -Werror -Wall $^ -o $@ $$(pkg-config --libs rdkafka)
-
-test-static: test.c
- $(CC) -O2 -Werror -Wall $^ -o $@ $$(pkg-config --libs --static rdkafka-static)
-
-testcpp: test.cpp
- $(CXX) -O2 -Werror -Wall $^ -o $@ $$(pkg-config --libs rdkafka++)
-
-testcpp-static: test.cpp
- $(CXX) -O2 -Werror -Wall $^ -o $@ $$(pkg-config --libs rdkafka++-static)
-
-run:
- @(for p in $(PROGS); do \
- echo "# Running $$p" ; \
- ./$$p || (echo $$p failed ; exit 1) ; \
- done)
-
-clean:
- rm -f $(PROGS)
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/README.md b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/README.md
deleted file mode 100644
index 8d1107b66..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/README.md
+++ /dev/null
@@ -1,8 +0,0 @@
-# Test librdkafka RPMs using docker
-
-After building the RPMs (see README.md in parent directory) test
-the RPMs on the supported CentOS/RHEL versions using:
-
- $ packaging/rpm/tests/test-on-docker.sh
-
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/run-test.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/run-test.sh
deleted file mode 100755
index c1234a945..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/run-test.sh
+++ /dev/null
@@ -1,49 +0,0 @@
-#!/bin/bash
-#
-# This script runs in the docker container, performing:
-# * install build toolchain
-# * install librdkafka rpms
-# * builds test apps
-# * runs test apps
-#
-# Usage: $0 <docker-image-name>
-
-set -ex
-
-pushd /v
-
-_IMG=$1
-
-echo "Testing on $_IMG"
-
-if [[ $_IMG == "centos:6" ]]; then
- _EL=6
- _INST="yum install -y -q"
-elif [[ $_IMG == "centos:7" ]]; then
- _EL=7
- _INST="yum install -y -q"
- # centos:7 ships with openssl-libs 1.0.1 which is outdated and not
- # ABI-compatible with 1.0.2 (which we build with).
- # Upgrade openssl-libs, as users would, to prevent missing symbols.
- _UPG="yum upgrade -y openssl-libs"
-else
- _EL=8
- _INST="dnf install -y -q"
-fi
-
-$_INST gcc gcc-c++ make pkg-config
-
-if [[ -n $_UPG ]]; then
- $_UPG
-fi
-
-$_INST /rpms/librdkafka1-*el${_EL}.x86_64.rpm /rpms/librdkafka-devel-*el${_EL}.x86_64.rpm
-
-make clean all
-
-make run
-
-make clean
-
-echo "$_IMG is all good!"
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test-on-docker.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test-on-docker.sh
deleted file mode 100755
index 2c12ff792..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test-on-docker.sh
+++ /dev/null
@@ -1,56 +0,0 @@
-#!/bin/bash
-#
-#
-# Test librdkafka packages in <rpmdirectory> using docker.
-# Must be executed from the librdkafka top-level directory.
-#
-# Usage:
-# packaging/rpm/test-on-docker.sh [<rpm-dir>]
-
-set -ex
-
-if [[ ! -f configure.self ]]; then
- echo "Must be executed from the librdkafka top-level directory"
- exit 1
-fi
-
-_DOCKER_IMAGES="centos:7 redhat/ubi8:8.5-226"
-_RPMDIR=artifacts
-
-if [[ -n $1 ]]; then
- _RPMDIR="$1"
-fi
-
-_RPMDIR=$(readlink -f $_RPMDIR)
-
-if [[ ! -d $_RPMDIR ]]; then
- echo "$_RPMDIR does not exist"
- exit 1
-fi
-
-
-fails=""
-for _IMG in $_DOCKER_IMAGES ; do
- if ! docker run \
- -t \
- -v $_RPMDIR:/rpms \
- -v $(readlink -f packaging/rpm/tests):/v \
- $_IMG \
- /v/run-test.sh $_IMG ; then
- echo "ERROR: $_IMG FAILED"
- fails="${fails}$_IMG "
- fi
-done
-
-if [[ -n $fails ]]; then
- echo "##################################################"
- echo "# Package verification failed for:"
- echo "# $fails"
- echo "# See previous errors"
- echo "##################################################"
- exit 1
-fi
-
-exit 0
-
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.c b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.c
deleted file mode 100644
index cf39b6bcd..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.c
+++ /dev/null
@@ -1,77 +0,0 @@
-#include <stdio.h>
-#include <string.h>
-#include <librdkafka/rdkafka.h>
-
-int main(int argc, char **argv) {
- rd_kafka_conf_t *conf;
- rd_kafka_t *rk;
- char features[256];
- size_t fsize = sizeof(features);
- char errstr[512];
- const char *exp_features[] = {
- "gzip", "snappy", "ssl", "sasl", "regex",
- "lz4", "sasl_gssapi", "sasl_plain", "sasl_scram", "plugins",
- "zstd", "sasl_oauthbearer", NULL,
- };
- const char **exp;
- int missing = 0;
-
-
- printf("librdkafka %s\n", rd_kafka_version_str());
-
- conf = rd_kafka_conf_new();
- if (rd_kafka_conf_get(conf, "builtin.features", features, &fsize) !=
- RD_KAFKA_CONF_OK) {
- fprintf(stderr, "conf_get failed\n");
- return 1;
- }
-
- printf("builtin.features %s\n", features);
-
- /* Verify that expected features are enabled. */
- for (exp = exp_features; *exp; exp++) {
- const char *t = features;
- size_t elen = strlen(*exp);
- int match = 0;
-
- while ((t = strstr(t, *exp))) {
- if (t[elen] == ',' || t[elen] == '\0') {
- match = 1;
- break;
- }
- t += elen;
- }
-
- if (match)
- continue;
-
- fprintf(stderr, "ERROR: feature %s not found\n", *exp);
- missing++;
- }
-
- if (rd_kafka_conf_set(conf, "security.protocol", "SASL_SSL", errstr,
- sizeof(errstr)) ||
- rd_kafka_conf_set(conf, "sasl.mechanism", "PLAIN", errstr,
- sizeof(errstr)) ||
- rd_kafka_conf_set(conf, "sasl.username", "username", errstr,
- sizeof(errstr)) ||
- rd_kafka_conf_set(conf, "sasl.password", "password", errstr,
- sizeof(errstr)) ||
- rd_kafka_conf_set(conf, "debug", "security", errstr,
- sizeof(errstr))) {
- fprintf(stderr, "conf_set failed: %s\n", errstr);
- return 1;
- }
-
- rk = rd_kafka_new(RD_KAFKA_PRODUCER, conf, errstr, sizeof(errstr));
- if (!rk) {
- fprintf(stderr, "rd_kafka_new failed: %s\n", errstr);
- return 1;
- }
-
- printf("client name %s\n", rd_kafka_name(rk));
-
- rd_kafka_destroy(rk);
-
- return missing ? 1 : 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.cpp b/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.cpp
deleted file mode 100644
index d78a76710..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/rpm/tests/test.cpp
+++ /dev/null
@@ -1,34 +0,0 @@
-#include <iostream>
-#include <librdkafka/rdkafkacpp.h>
-
-
-int main() {
- std::cout << "librdkafka++ " << RdKafka::version_str() << std::endl;
-
- RdKafka::Conf *conf = RdKafka::Conf::create(RdKafka::Conf::CONF_GLOBAL);
-
- std::string features;
-
- if (conf->get("builtin.features", features) != RdKafka::Conf::CONF_OK) {
- std::cerr << "conf_get failed" << std::endl;
- return 1;
- }
-
- std::cout << "builtin.features " << features << std::endl;
-
- std::string errstr;
- RdKafka::Producer *producer = RdKafka::Producer::create(conf, errstr);
- if (!producer) {
- std::cerr << "Producer::create failed: " << errstr << std::endl;
- return 1;
- }
-
- delete conf;
-
- std::cout << "client name " << producer->name() << std::endl;
-
-
- delete producer;
-
- return 0;
-}
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-deb-package.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-deb-package.sh
deleted file mode 100755
index d9cad6d25..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-deb-package.sh
+++ /dev/null
@@ -1,64 +0,0 @@
-#!/bin/bash
-#
-# Build librdkafka Debian package on a bare-bone Debian host, such as ubuntu:16.04 (docker).
-#
-# Usage (from top-level librdkafka dir):
-# docker run -it -v $PWD:/v ubuntu:16.04 /v/packaging/tools/build-deb-package.sh 1.0.0 master
-#
-
-set -exu
-
-if [[ $# -ne 2 ]]; then
- echo "Usage: $0 <package-version> <librdkafka-branch-or-tag>"
- exit 1
-fi
-
-export VERSION=$1
-LRK_BRANCH=$2
-
-apt-get update
-
-# Install debian packaging tools and librdkafka build dependencies
-apt-get install -y git-buildpackage debhelper \
- zlib1g-dev libssl-dev libsasl2-dev liblz4-dev
-
-
-# Clone the librdkafka git repo to a new location to avoid messing
-# up the librdkafka working directory.
-
-
-BUILD_DIR=$(mktemp -d)
-
-pushd $BUILD_DIR
-
-git clone /v librdkafka
-
-pushd librdkafka
-
-export DEBEMAIL="librdkafka packaging <rdkafka@edenhill.se>"
-git config user.email "rdkafka@edenhill.se"
-git config user.name "librdkafka packaging"
-
-DEB_BRANCH=origin/confluent-debian
-TMP_BRANCH=tmp-debian
-git checkout -b $TMP_BRANCH $LRK_BRANCH
-git merge --no-edit $DEB_BRANCH
-
-dch --newversion ${VERSION/-/\~}-1 "Release version $VERSION" --urgency low && dch --release --distribution unstable ""
-
-git commit -a -m "Tag Debian release $VERSION."
-
-make archive
-mkdir -p ../tarballs || true
-mv librdkafka-${VERSION}.tar.gz ../tarballs/librdkafka_${VERSION}.orig.tar.gz
-
-gbp buildpackage -us -uc --git-debian-branch=$TMP_BRANCH \
- --git-upstream-tree=$LRK_BRANCH \
- --git-verbose \
- --git-builder="debuild --set-envvar=VERSION=$VERSION --set-envvar=SKIP_TESTS=y -i -I"
-
-
-popd # librdkafka
-
-popd # $BUILD_DIR
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-debian.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-debian.sh
deleted file mode 100755
index e62ee5f67..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-debian.sh
+++ /dev/null
@@ -1,65 +0,0 @@
-#!/bin/bash
-#
-# Build librdkafka on a bare-bone Debian host, such as the
-# mcr.microsoft.com/dotnet/sdk Docker image.
-#
-# Statically linked
-# WITH openssl 1.0, zlib
-# WITHOUT libsasl2, lz4(ext, using builtin instead)
-#
-# Usage (from top-level librdkafka dir):
-# docker run -it -v $PWD:/v mcr.microsoft.com/dotnet/sdk /v/packaging/tools/build-debian.sh /v /v/librdkafka-debian9.tgz
-#
-
-
-set -ex
-
-LRK_DIR=$1
-shift
-OUT_TGZ=$1
-shift
-CONFIG_ARGS=$*
-
-if [[ ! -f $LRK_DIR/configure.self || -z $OUT_TGZ ]]; then
- echo "Usage: $0 <librdkafka-root-direcotry> <output-tgz> [<configure-args..>]"
- exit 1
-fi
-
-set -u
-
-apt-get update
-apt-get install -y gcc g++ zlib1g-dev python3 git-core make patch
-
-
-# Copy the librdkafka git archive to a new location to avoid messing
-# up the librdkafka working directory.
-
-BUILD_DIR=$(mktemp -d)
-
-pushd $BUILD_DIR
-
-DEST_DIR=$PWD/dest
-mkdir -p $DEST_DIR
-
-# Workaround for newer Git not allowing clone directory to be owned by
-# another user (which is a questionable limitation for the read-only archive
-# command..)
-git config --global --add safe.directory /v
-
-(cd $LRK_DIR ; git archive --format tar HEAD) | tar xf -
-
-./configure --install-deps --disable-gssapi --disable-lz4-ext --enable-static --prefix=$DEST_DIR $CONFIG_ARGS
-make -j
-examples/rdkafka_example -X builtin.features
-CI=true make -C tests run_local_quick
-make install
-
-# Tar up the output directory
-pushd $DEST_DIR
-ldd lib/*.so.1
-tar cvzf $OUT_TGZ .
-popd # $DEST_DIR
-
-popd # $BUILD_DIR
-
-rm -rf "$BUILD_DIR"
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-manylinux.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-manylinux.sh
deleted file mode 100755
index 4aeaa9622..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-manylinux.sh
+++ /dev/null
@@ -1,68 +0,0 @@
-#!/bin/bash
-#
-# Build on a manylinux (https://github.com/pypa/manylinux) docker container.
-#
-# This will provide a self-contained librdkafka shared library that works
-# on most glibc-based Linuxes.
-#
-# Statically linked
-# WITH openssl 1.1.1, zlib, lz4(bundled)
-# WITHOUT libsasl2
-#
-#
-# Run:
-# docker run -t -v "$PWD:/v quay.io/pypa/manylinux2010_x86_64 /v/packaging/tools/build-manylinux.sh /v /v/artifacts/librdkafka-manylinux2010_x86_64.tgz $config_args"
-
-set -ex
-
-LRK_DIR=$1
-shift
-OUT_TGZ=$1
-shift
-CONFIG_ARGS=$*
-
-if [[ ! -f $LRK_DIR/configure.self || -z $OUT_TGZ ]]; then
- echo "Usage: $0 <librdkafka-root-direcotry> <output-tgz> [<configure-args..>]"
- exit 1
-fi
-
-set -u
-
-yum install -y libstdc++-devel gcc gcc-c++ python34
-
-# Copy the librdkafka git archive to a new location to avoid messing
-# up the librdkafka working directory.
-
-BUILD_DIR=$(mktemp -d)
-
-pushd $BUILD_DIR
-
-DEST_DIR=$PWD/dest
-mkdir -p $DEST_DIR
-
-# Workaround for newer Git not allowing clone directory to be owned by
-# another user (which is a questionable limitation for the read-only archive
-# command..)
-git config --global --add safe.directory /v
-
-(cd $LRK_DIR ; git archive --format tar HEAD) | tar xf -
-
-./configure --install-deps --source-deps-only --disable-gssapi --disable-lz4-ext --enable-static --prefix=$DEST_DIR $CONFIG_ARGS
-
-make -j
-
-examples/rdkafka_example -X builtin.features
-
-CI=true make -C tests run_local_quick
-
-make install
-
-# Tar up the output directory
-pushd $DEST_DIR
-ldd lib/*.so.1
-tar cvzf $OUT_TGZ .
-popd # $DEST_DIR
-
-popd # $BUILD_DIR
-
-rm -rf "$BUILD_DIR"
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-release-artifacts.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-release-artifacts.sh
deleted file mode 100755
index ea09aaf96..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/build-release-artifacts.sh
+++ /dev/null
@@ -1,138 +0,0 @@
-#!/bin/sh
-#
-# ^ NOTE: This needs to be sh, not bash, for alpine compatibility.
-#
-#
-# Build dynamic and statically linked librdkafka libraries useful for
-# release artifacts in high-level clients.
-#
-# Requires docker.
-# Supported docker images:
-# alpine:3.16
-# quay.io/pypa/manylinux2014_aarch64 (centos7)
-# quay.io/pypa/manylinux2014_x86_64 (centos7)
-# quay.io/pypa/manylinux2010_x86_64 (centos6)
-#
-# Usage:
-# packaging/tools/build-release-artifacts.sh [--disable-gssapi] <docker-image> <relative-output-tarball-path.tgz>
-#
-# The output path must be a relative path and inside the librdkafka directory
-# structure.
-#
-
-set -e
-
-docker_image=""
-extra_pkgs_rpm=""
-extra_pkgs_apk=""
-extra_config_args=""
-expected_features="gzip snappy ssl sasl regex lz4 sasl_plain sasl_scram plugins zstd sasl_oauthbearer http oidc"
-
-# Since cyrus-sasl is the only non-statically-linkable dependency,
-# we provide a --disable-gssapi option so that two different libraries
-# can be built: one with GSSAPI/Kerberos support, and one without, depending
-# on this option.
-if [ "$1" = "--disable-gssapi" ]; then
- extra_config_args="${extra_config_args} --disable-gssapi"
- disable_gssapi="$1"
- shift
-else
- extra_pkgs_rpm="${extra_pkgs_rpm} cyrus-sasl cyrus-sasl-devel"
- extra_pkgs_apk="${extra_pkgs_apk} cyrus-sasl cyrus-sasl-dev"
- expected_features="${expected_features} sasl_gssapi"
- disable_gssapi=""
-fi
-
-# Check if we're running on the host or the (docker) build target.
-if [ "$1" = "--in-docker" -a $# -eq 2 ]; then
- output="$2"
-elif [ $# -eq 2 ]; then
- docker_image="$1"
- output="$2"
-else
- echo "Usage: $0 [--disable-gssapi] <manylinux-docker-image> <output-path.tgz>"
- exit 1
-fi
-
-if [ -n "$docker_image" ]; then
- # Running on the host, spin up the docker builder.
- exec docker run -v "$PWD:/v" $docker_image /v/packaging/tools/build-release-artifacts.sh $disable_gssapi --in-docker "/v/$output"
- # Only reached on exec error
- exit $?
-fi
-
-
-########################################################################
-# Running in the docker instance, this is where we perform the build. #
-########################################################################
-
-
-# Packages required for building librdkafka (perl is for openssl).
-
-if grep -q alpine /etc/os-release 2>/dev/null ; then
- # Alpine
- apk add \
- bash curl gcc g++ make musl-dev linux-headers bsd-compat-headers git \
- python3 perl patch $extra_pkgs_apk
-
-else
- # CentOS
- yum install -y libstdc++-devel gcc gcc-c++ python3 git perl-IPC-Cmd $extra_pkgs_rpm
-fi
-
-
-# Clone the repo so other builds are unaffected of what we're doing
-# and we get a pristine build tree.
-git clone /v /librdkafka
-
-cd /librdkafka
-
-# Build librdkafka
-./configure \
- --install-deps --source-deps-only --disable-lz4-ext \
- --enable-static --enable-strip $extra_config_args
-
-make -j
-
-# Show library linkage (for troubleshooting) and checksums (for verification)
-for lib in src/librdkafka.so.1 src-cpp/librdkafka++.so.1; do
- echo "$0: LINKAGE ${lib}:"
- ldd src/librdkafka.so.1
- echo "$0: SHA256 ${lib}:"
- sha256sum "$lib"
-done
-
-# Verify that expected features are indeed built.
-features=$(examples/rdkafka_example -X builtin.features)
-echo "$0: FEATURES: $features"
-
-missing=""
-for f in $expected_features; do
- if ! echo "$features" | grep -q "$f" ; then
- echo "$0: BUILD IS MISSING FEATURE $f"
- missing="${missing} $f"
- fi
-done
-
-if [ -n "$missing" ]; then
- exit 1
-fi
-
-
-# Run quick test suite, mark it as CI to avoid time/resource sensitive
-# tests to fail in case the worker is under-powered.
-CI=true make -C tests run_local_quick
-
-
-# Install librdkafka and then make a tar ball of the installed files.
-mkdir -p /destdir
-
-DESTDIR=/destdir make install
-
-cd /destdir
-tar cvzf "$output" .
-
-# Emit output hash so that build logs can be used to verify artifacts later.
-echo "$0: SHA256 $output:"
-sha256sum "$output"
-
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/distro-build.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/distro-build.sh
deleted file mode 100755
index a4b5bfa61..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/distro-build.sh
+++ /dev/null
@@ -1,38 +0,0 @@
-#!/bin/bash
-#
-# Build librdkafka for different distros to produce distro-specific artifacts.
-# Requires docker.
-#
-
-set -e
-
-distro=$1
-shift
-config_args=$*
-
-case $distro in
- manylinux*)
- # Any pypa/manylinux docker image build.
- docker run -t -v "$PWD:/v" quay.io/pypa/$distro /v/packaging/tools/build-manylinux.sh /v /v/artifacts/librdkafka-${distro}.tgz $config_args
- ;;
- centos)
- if [[ -n $config_args ]]; then
- echo "Warning: configure arguments ignored for centos RPM build"
- fi
- packaging/rpm/mock-on-docker.sh
- packaging/rpm/tests/test-on-docker.sh
- ;;
- debian)
- docker run -it -v "$PWD:/v" mcr.microsoft.com/dotnet/sdk:3.1 /v/packaging/tools/build-debian.sh /v /v/artifacts/librdkafka-debian9.tgz $config_args
- ;;
- alpine)
- packaging/alpine/build-alpine.sh $config_args
- ;;
- alpine-static)
- packaging/alpine/build-alpine.sh --enable-static --source-deps-only $config_args
- ;;
- *)
- echo "Usage: $0 <centos|debian|alpine|alpine-static>"
- exit 1
- ;;
-esac
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/gh-release-checksums.py b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/gh-release-checksums.py
deleted file mode 100755
index e7259dc20..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/gh-release-checksums.py
+++ /dev/null
@@ -1,39 +0,0 @@
-#!/usr/bin/env python3
-#
-# Calculate checksums for GitHub release artifacts/assets.
-#
-# Use the direct links rather than getting the tarball URLs from
-# the GitHub API since the latter uses the git-sha1 rather than the tag
-# in its zipped up content, causing checksum mismatches.
-#
-
-import sys
-import requests
-import hashlib
-
-
-if __name__ == '__main__':
-
- if len(sys.argv) != 2:
- print("Usage: {} <tag>".format(sys.argv[0]))
- sys.exit(1)
-
- tag = sys.argv[1]
-
- print("## Checksums")
- print("Release asset checksums:")
-
- for ftype in ["zip", "tar.gz"]:
- url = "https://github.com/edenhill/librdkafka/archive/{}.{}".format(
- tag, ftype)
-
- h = hashlib.sha256()
-
- r = requests.get(url, stream=True)
- while True:
- buf = r.raw.read(100 * 1000)
- if len(buf) == 0:
- break
- h.update(buf)
-
- print(" * {}.{} SHA256 `{}`".format(tag, ftype, h.hexdigest()))
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/rdutcoverage.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/rdutcoverage.sh
deleted file mode 100755
index e99c51bdc..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/rdutcoverage.sh
+++ /dev/null
@@ -1,25 +0,0 @@
-#!/bin/bash
-#
-# Verify that code coverage numbers are not reused in multiple places.
-#
-
-set -e
-
-echo "Checking for duplicate coverage numbers:"
-cnt=0
-for d in $(egrep -Rsoh 'RD_UT_COVERAGE\([[:digit:]]+\)' src \
- | sort | uniq -c | \
- egrep -v '^[[:space:]]*1 ' | awk '{print $2}'); do
- grep -RsnF "$d" src
- cnt=$(expr $cnt + 1)
-done
-
-echo ""
-
-if [[ $cnt -gt 0 ]]; then
- echo "$cnt duplicates found: please use unique numbers"
- exit 1
-else
- echo "No duplicate(s) found"
- exit 0
-fi
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/requirements.txt b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/requirements.txt
deleted file mode 100644
index 43603098a..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/requirements.txt
+++ /dev/null
@@ -1,2 +0,0 @@
-flake8
-autopep8
diff --git a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/style-format.sh b/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/style-format.sh
deleted file mode 100755
index c59ecbe6a..000000000
--- a/fluent-bit/lib/librdkafka-2.1.0/packaging/tools/style-format.sh
+++ /dev/null
@@ -1,148 +0,0 @@
-#!/bin/bash
-#
-# Check or apply/fix the project coding style to all files passed as arguments.
-# Uses clang-format for C/C++ and flake8 for Python.
-#
-# Requires clang-format version 10 (apt install clang-format-10).
-#
-
-
-CLANG_FORMAT=${CLANG_FORMAT:-clang-format}
-
-set -e
-
-ret=0
-
-if [[ -z $1 ]]; then
- echo "Usage: $0 [--fix] srcfile1.c srcfile2.h srcfile3.c ..."
- echo ""
- exit 0
-fi
-
-if [[ $1 == "--fix" ]]; then
- fix=1
- shift
-else
- fix=0
-fi
-
-clang_format_version=$(${CLANG_FORMAT} --version | sed -Ee 's/.*version ([[:digit:]]+)\.[[:digit:]]+\.[[:digit:]]+.*/\1/')
-if [[ $clang_format_version != "10" ]] ; then
- echo "$0: clang-format version 10, '$clang_format_version' detected"
- exit 1
-fi
-
-# Get list of files from .formatignore to ignore formatting for.
-ignore_files=( $(grep '^[^#]..' .formatignore) )
-
-function ignore {
- local file=$1
-
- local f
- for f in "${ignore_files[@]}" ; do
- [[ $file == $f ]] && return 0
- done
-
- return 1
-}
-
-# Read the C++ style from src-cpp/.clang-format and store it
-# in a json-like string which is passed to --style.
-# (It would be great if clang-format could take a file path for the
-# format file..).
-cpp_style="{ $(grep -v '^...$' .clang-format-cpp | grep -v '^$' | tr '\n' ',' | sed -e 's/,$//') }"
-if [[ -z $cpp_style ]]; then
- echo "$0: Unable to read .clang-format-cpp"
- exit 1
-fi
-
-extra_info=""
-
-for f in $*; do
-
- if ignore $f ; then
- echo "$f is ignored by .formatignore" 1>&2
- continue
- fi
-
- lang="c"
- if [[ $f == *.cpp ]]; then
- style="$cpp_style"
- stylename="C++"
- elif [[ $f == *.h && $(basename $f) == *cpp* ]]; then
- style="$cpp_style"
- stylename="C++ (header)"
- elif [[ $f == *.py ]]; then
- lang="py"
- style="pep8"
- stylename="pep8"
- else
- style="file" # Use .clang-format
- stylename="C"
- fi
-
- check=0
-
- if [[ $fix == 1 ]]; then
- # Convert tabs to 8 spaces first.
- if grep -ql $'\t' "$f"; then
- sed -i -e 's/\t/ /g' "$f"
- echo "$f: tabs converted to spaces"
- fi
-
- if [[ $lang == c ]]; then
- # Run clang-format to reformat the file
- ${CLANG_FORMAT} --style="$style" "$f" > _styletmp
-
- else
- # Run autopep8 to reformat the file.
- python3 -m autopep8 -a "$f" > _styletmp
- # autopep8 can't fix all errors, so we also perform a flake8 check.
- check=1
- fi
-
- if ! cmp -s "$f" _styletmp; then
- echo "$f: style fixed ($stylename)"
- # Use cp to preserve target file mode/attrs.
- cp _styletmp "$f"
- rm _styletmp
- fi
- fi
-
- if [[ $fix == 0 || $check == 1 ]]; then
- # Check for tabs
- if grep -q $'\t' "$f" ; then
- echo "$f: contains tabs: convert to 8 spaces instead"
- ret=1
- fi
-
- # Check style
- if [[ $lang == c ]]; then
- if ! ${CLANG_FORMAT} --style="$style" --Werror --dry-run "$f" ; then
- echo "$f: had style errors ($stylename): see clang-format output above"
- ret=1
- fi
- elif [[ $lang == py ]]; then
- if ! python3 -m flake8 "$f"; then
- echo "$f: had style errors ($stylename): see flake8 output above"
- if [[ $fix == 1 ]]; then
- # autopep8 couldn't fix all errors. Let the user know.
- extra_info="Error: autopep8 could not fix all errors, fix the flake8 errors manually and run again."
- fi
- ret=1
- fi
- fi
- fi
-
-done
-
-rm -f _styletmp
-
-if [[ $ret != 0 ]]; then
- echo ""
- echo "You can run the following command to automatically fix the style:"
- echo " $ make style-fix"
- [[ -n $extra_info ]] && echo "$extra_info"
-fi
-
-exit $ret