"great" perftools (formerly google-perftools)
The gperftools package contains some utilities to improve and analyze the performance of C++ programs. This includes an optimized thread-caching malloc() and cpu and heap profiling utilities.
Formerly the google-perftools package.
- Developed at devel:libraries:c_c++
- Sources inherited from project openSUSE:Factory
-
5
derived packages
- Download package
-
Checkout Package
osc -A https://api.opensuse.org checkout openSUSE:Factory:Rebuild/gperftools && cd $_
- Create Badge
Refresh
Refresh
Source Files
Filename | Size | Changed |
---|---|---|
gperftools-2.5.tar.gz | 0001389081 1.32 MB | |
gperftools.changes | 0000017440 17 KB | |
gperftools.spec | 0000003885 3.79 KB | |
gperftools_fix_unassigned_malloc_in_unittest.patch | 0000000518 518 Bytes | |
gperftools_gcc46.patch | 0000000353 353 Bytes |
Revision 16 (latest revision is 33)
Dominique Leuenberger (dimstar_suse)
accepted
request 378351
from
Ismail Dönmez (namtrac)
(revision 16)
- gperftools 2.5, available 12 March 2016. See https://github.com/gperftools/gperftools/releases * Bryan Chan has contributed s390x support * stacktrace capturing via libgcc's _Unwind_Backtrace was implemented (for architectures with missing or broken libunwind). * "emergency malloc" was implemented. Which unbreaks recursive calls to malloc/free from stacktrace capturing functions (such us glib'c backtrace() or libunwind on arm). It is enabled by --enable-emergency-malloc configure flag or by default on arm when --enable-stacktrace-via-backtrace is given. It is another fix for a number common issues people had on platforms with missing or broken libunwind. * C++14 sized-deallocation is now supported (on gcc 5 and recent clangs). It is off by default and can be enabled at configure time via --enable-sized-delete. On GNU/Linux it can also be enabled at run-time by either TCMALLOC_ENABLE_SIZED_DELETE environment variable or by defining tcmalloc_sized_delete_enabled function which should return 1 to enable it. * we've lowered default value of transfer batch size to 512. Previous value (bumped up in 2.1) was too high and caused performance regression for some users. 512 should still give us performance boost for workloads that need higher transfer batch size while not penalizing other workloads too much. * Brian Silverman's patch finally stopped arming profiling timer unless profiling is started. * Andrew Morrow has contributed support for obtaining cache size of the current thread and softer idling (for use in MongoDB). * we've implemented few minor performance improvements, particularly on malloc fast-path. * issue that caused spurious profiler_unittest.sh failures was fixed.
Comments 0