python-confluent-kafka
No description set
- Developed at devel:languages:python
- Sources inherited from project openSUSE:Factory
-
3
derived packages
- Download package
-
Checkout Package
osc -A https://api.opensuse.org checkout openSUSE:Factory:PowerPC/python-confluent-kafka && cd $_
- Create Badge
Refresh
Refresh
Source Files
Filename | Size | Changed |
---|---|---|
confluent-kafka-1.1.0.tar.gz | 0000073333 71.6 KB | |
no-license-as-datafile.patch | 0000000333 333 Bytes | |
python-confluent-kafka.changes | 0000001861 1.82 KB | |
python-confluent-kafka.spec | 0000002048 2 KB |
Revision 2 (latest revision is 8)
Dominique Leuenberger (dimstar_suse)
accepted
request 745369
from
Tomáš Chvátal (scarabeus_iv)
(revision 2)
- update to 1.1.0: * confluent-kafka-python is based on librdkafka v1.1.0, see the librdkafka v1.1.0 release notes for a complete list of changes, enhancements, fixes and upgrade considerations. * ssl.endpoint.identification.algorithm=https (off by default) to validate the broker hostname matches the certificate. Requires OpenSSL >= 1.0.2(included with Wheel installations)) * Improved GSSAPI/Kerberos ticket refresh * Confluent monitoring interceptor package bumped to v0.11.1 (#634) New configuration properties: * ssl.key.pem - client's private key as a string in PEM format * ssl.certificate.pem - client's public key as a string in PEM format * enable.ssl.certificate.verification - enable(default)/disable OpenSSL's builtin broker certificate verification. * enable.ssl.endpoint.identification.algorithm - to verify the broker's hostname with its certificate (disabled by default). * Add new rd_kafka_conf_set_ssl_cert() to pass PKCS#12, DER or PEM certs in (binary) memory form to the configuration object. * The private key data is now securely cleared from memory after last use. * SASL GSSAPI/Kerberos: Don't run kinit refresh for each broker, just per client instance. * SASL GSSAPI/Kerberos: Changed sasl.kerberos.kinit.cmd to first attempt ticket refresh, then acquire. * SASL: Proper locking on broker name acquisition. * Consumer: max.poll.interval.ms now correctly handles blocking poll calls, allowing a longer poll timeout than the max poll interval. * configure: Fix libzstd static lib detection
Comments 0