|
|
|
|
Changelog for python38-charset-normalizer-2.1.1-1.1.noarch.rpm :
* Sat Sep 17 2022 Dirk Müller - update to 2.1.1: * Function `normalize` scheduled for removal in 3.0 * Removed useless call to decode in fn is_unprintable (#206) * Thu Aug 18 2022 Ben Greiner - Clean requirements: We don\'t need anything * Tue Jul 19 2022 Dirk Müller - update to 2.1.0: * Output the Unicode table version when running the CLI with `--version` * Re-use decoded buffer for single byte character sets * Fixing some performance bottlenecks * Workaround potential bug in cpython with Zero Width No-Break Space located * in Arabic Presentation Forms-B, Unicode 1.1 not acknowledged as space * CLI default threshold aligned with the API threshold from * Support for Python 3.5 (PR #192) * Use of backport unicodedata from `unicodedata2` as Python is quickly catching up, scheduled for removal in 3.0 * Tue Feb 15 2022 Dirk Müller - update to 2.0.12: * ASCII miss-detection on rare cases (PR #170) * Explicit support for Python 3.11 (PR #164) * The logging behavior have been completely reviewed, now using only TRACE and DEBUG levels * Mon Jan 10 2022 Dirk Müller - update to 2.0.10: * Fallback match entries might lead to UnicodeDecodeError for large bytes sequence * Skipping the language-detection (CD) on ASCII * Mon Dec 06 2021 Dirk Müller - update to 2.0.9: * Moderating the logging impact (since 2.0.8) for specific environments * Wrong logging level applied when setting kwarg `explain` to True * Mon Nov 29 2021 Dirk Müller - update to 2.0.8: * Improvement over Vietnamese detection * MD improvement on trailing data and long foreign (non-pure latin) * Efficiency improvements in cd/alphabet_languages * call sum() without an intermediary list following PEP 289 recommendations * Code style as refactored by Sourcery-AI * Minor adjustment on the MD around european words * Remove and replace SRTs from assets / tests * Initialize the library logger with a `NullHandler` by default * Setting kwarg `explain` to True will add provisionally * Fix large (misleading) sequence giving UnicodeDecodeError * Avoid using too insignificant chunk * Add and expose function `set_logging_handler` to configure a specific StreamHandler * Fri Nov 26 2021 Dirk Müller - require lower-case name instead of breaking build * Thu Nov 25 2021 Matej Cepl - Use lower-case name of prettytable package * Sun Oct 17 2021 Martin Hauke - Update to version 2.0.7 * Addition: bento Add support for Kazakh (Cyrillic) language detection * Improvement: sparkle Further improve inferring the language from a given code page (single-byte). * Removed: fire Remove redundant logging entry about detected language(s). * Improvement: zap Refactoring for potential performance improvements in loops. * Improvement: sparkles Various detection improvement (MD+CD). * Bugfix: bug Fix a minor inconsistency between Python 3.5 and other versions regarding language detection.- Update to version 2.0.6 * Bugfix: bug Unforeseen regression with the loss of the backward-compatibility with some older minor of Python 3.5.x. * Bugfix: bug Fix CLI crash when using --minimal output in certain cases. * Improvement: sparkles Minor improvement to the detection efficiency (less than 1%).- Update to version 2.0.5 * Improvement: sparkles The BC-support with v1.x was improved, the old staticmethods are restored. * Remove: fire The project no longer raise warning on tiny content given for detection, will be simply logged as warning instead. * Improvement: sparkles The Unicode detection is slightly improved, see #93 * Bugfix: bug In some rare case, the chunks extractor could cut in the middle of a multi-byte character and could mislead the mess detection. * Bugfix: bug Some rare \'space\' characters could trip up the UnprintablePlugin/Mess detection. * Improvement: art Add syntax sugar __bool__ for results CharsetMatches list-container.- Update to version 2.0.4 * Improvement: sparkle Adjust the MD to lower the sensitivity, thus improving the global detection reliability. * Improvement: sparkle Allow fallback on specified encoding if any. * Bugfix: bug The CLI no longer raise an unexpected exception when no encoding has been found. * Bugfix: bug Fix accessing the \'alphabets\' property when the payload contains surrogate characters. * Bugfix: bug pencil2 The logger could mislead (explain=True) on detected languages and the impact of one MBCS match (in #72) * Bugfix: bug Submatch factoring could be wrong in rare edge cases (in #72) * Bugfix: bug Multiple files given to the CLI were ignored when publishing results to STDOUT. (After the first path) (in #72) * Internal: art Fix line endings from CRLF to LF for certain files.- Update to version 2.0.3 * Improvement: sparkles Part of the detection mechanism has been improved to be less sensitive, resulting in more accurate detection results. Especially ASCII. #63 Fix #62 * Improvement: sparklesAccording to the community wishes, the detection will fall back on ASCII or UTF-8 in a last-resort case.- Update to version 2.0.2 * Bugfix: bug Empty/Too small JSON payload miss-detection fixed. * Improvement: sparkler Don\'t inject unicodedata2 into sys.modules- Update to version 2.0.1 * Bugfix: bug Make it work where there isn\'t a filesystem available, dropping assets frequencies.json. * Improvement: sparkles You may now use aliases in cp_isolation and cp_exclusion arguments. * Bugfix: bug Using explain=False permanently disable the verbose output in the current runtime #47 * Bugfix: bug One log entry (language target preemptive) was not show in logs when using explain=True #47 * Bugfix: bug Fix undesired exception (ValueError) on getitem of instance CharsetMatches #52 * Improvement: wrench Public function normalize default args values were not aligned with from_bytes #53- Update to version 2.0.0 * Performance: zap 4x to 5 times faster than the previous 1.4.0 release. * Performance: zap At least 2x faster than Chardet. * Performance: zap Accent has been made on UTF-8 detection, should perform rather instantaneous. * Improvement: back The backward compatibility with Chardet has been greatly improved. The legacy detect function returns an identical charset name whenever possible. * Improvement: sparkle The detection mechanism has been slightly improved, now Turkish content is detected correctly (most of the time) * Code: art The program has been rewritten to ease the readability and maintainability. (+Using static typing) * Tests: heavy_check_mark New workflows are now in place to verify the following aspects: Performance, Backward- Compatibility with Chardet, and Detection Coverage in addition# to currents tests. (+CodeQL) * Dependency: heavy_minus_sign This package no longer require anything when used with Python 3.5 (Dropped cached_property) * Docs: pencil2 Performance claims have been updated, the guide to contributing, and the issue template. * Improvement: sparkle Add --version argument to CLI * Bugfix: bug The CLI output used the relative path of the file(s). Should be absolute. * Deprecation: red_circle Methods coherence_non_latin, w_counter, chaos_secondary_pass of the class CharsetMatch are now deprecated and scheduled for removal in v3.0 * Improvement: sparkle If no language was detected in content, trying to infer it using the encoding name/alphabets used. * Removal: fire Removed support for these languages: Catalan, Esperanto, Kazakh, Baque, Volapük, Azeri, Galician, Nynorsk, Macedonian, and Serbocroatian. * Improvement: sparkle utf_7 detection has been reinstated. * Removal: fire The exception hook on UnicodeDecodeError has been removed.- Update to version 1.4.1 * Improvement: art Logger configuration/usage no longer conflict with others #44- Update to version 1.4.0 * Dependency: heavy_minus_sign Using standard logging instead of using the package loguru. * Dependency: heavy_minus_sign Dropping nose test framework in favor of the maintained pytest. * Dependency: heavy_minus_sign Choose to not use dragonmapper package to help with gibberish Chinese/CJK text. * Dependency: wrench heavy_minus_sign Require cached_property only for Python 3.5 due to constraint. Dropping for every other interpreter version. * Bugfix: bug BOM marker in a CharsetNormalizerMatch instance could be False in rare cases even if obviously present. Due to the sub-match factoring process. * Improvement: sparkler Return ASCII if given sequences fit. * Performance: zap Huge improvement over the larges payload. * Change: fire Stop support for UTF-7 that does not contain a SIG. (Contributions are welcome to improve that point) * Feature: sparkler CLI now produces JSON consumable output. * Dependency: Dropping PrettyTable, replaced with pure JSON output. * Bugfix: bug Not searching properly for the BOM when trying utf32/16 parent codec. * Other: zap Improving the package final size by compressing frequencies.json. * Thu May 20 2021 pgajdosAATTsuse.com- version update to 1.3.9 * Bugfix: bug In some very rare cases, you may end up getting encode/decode errors due to a bad bytes payload #40 * Bugfix: bug Empty given payload for detection may cause an exception if trying to access the alphabets property. #39 * Bugfix: bug The legacy detect function should return UTF-8-SIG if sig is present in the payload. #38 * Tue Feb 09 2021 John Vandenberg - Switch to PyPI source- Add Suggests: python-unicodedata2- Remove executable bit from charset_normalizer/assets/frequencies.json- Update to v1.3.6 * Allow prettytable 2.0- from v1.3.5 * Dependencies refactor and add support for py 3.9 and 3.10 * Fix version parsing * Mon May 25 2020 Petr Gajdos - %python3_only -> %python_alternative * Mon Jan 27 2020 Marketa Calabkova - Update to 1.3.4 * Improvement/Bugfix : False positive when searching for successive upper, lower char. (ProbeChaos) * Improvement : Noticeable better detection for jp * Bugfix : Passing zero-length bytes to from_bytes * Improvement : Expose version in package * Bugfix : Division by zero * Improvement : Prefers unicode (utf-8) when detected * Apparently dropped Python2 silently * Fri Oct 04 2019 Marketa Calabkova - Update to 1.3.0 * Backport unicodedata for v12 impl into python if available * Add aliases to CharsetNormalizerMatches class * Add feature preemptive behaviour, looking for encoding declaration * Add method to determine if specific encoding is multi byte * Add has_submatch property on a match * Add percent_chaos and percent_coherence * Coherence ratio based on mean instead of sum of best results * Using loguru for trace/debug <3 * from_byte method improved * Thu Sep 26 2019 Tomáš Chvátal - Update to 1.1.1: * from_bytes parameters steps and chunk_size were not adapted to sequence len if provided values were not fitted to content * Sequence having lenght bellow 10 chars was not checked * Legacy detect method inspired by chardet was not returning * Various more test updates * Fri Sep 13 2019 Tomáš Chvátal - Update to 0.3: * Improvement on detection * Performance loss to expect * Added --threshold option to CLI * Bugfix on UTF 7 support * Legacy detect(byte_str) method * BOM support (Unicode mostly) * Chaos prober improved on small text * Language detection has been reviewed to give better result * Bugfix on jp detection, every jp text was considered chaotic * Fri Aug 30 2019 Tomáš Chvátal - Fix the tarball to really be the one published by upstream * Wed Aug 28 2019 John Vandenberg - Initial spec for v0.1.8
|
|
|