id
int64
959M
2.55B
title
stringlengths
3
133
body
stringlengths
1
65.5k
state
stringclasses
2 values
created_at
stringlengths
20
20
updated_at
stringlengths
20
20
closed_at
stringlengths
20
20
user
stringclasses
174 values
2,550,841,506
Update datasets to 3.0.1
Update datasets to 3.0.1. Fix #3068.
open
2024-09-26T14:53:51Z
2024-09-26T16:21:52Z
null
albertvillanova
2,550,678,625
Fix CI by deleting test_polars_struct_thread_panic_error
Delete `test_polars_struct_thread_panic_error`. Note that the previous CI-Hub dataset (with the specific Parquet file) has been deleted. Fix #3069.
closed
2024-09-26T13:59:44Z
2024-09-26T15:23:33Z
2024-09-26T15:23:31Z
albertvillanova
2,550,490,787
CI worker test_polars_struct_thread_panic_error is broken
NOW: CI worker test `test_polars_struct_thread_panic_error` raises `RepositoryNotFoundError`: https://github.com/huggingface/dataset-viewer/actions/runs/11053126786/job/30706767804?pr=3037 ``` ERROR tests/job_runners/split/test_descriptive_statistics.py::test_polars_struct_thread_panic_error - huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66f560ea-0d3bd15934a45d7838ac6bb3;f1b1e4ce-2768-4f41-a5d8-f1c1c129addc) Repository Not Found for url: https://hub-ci.huggingface.co/datasets/__DUMMY_TRANSFORMERS_USER__/test_polars_panic_error/resolve/main/test_polars_panic_error.parquet. Please make sure you specified the correct `repo_id` and `repo_type`. If you are trying to access a private or gated repo, make sure you are authenticated. ``` Direct cause is that the dataset has been deleted from the CI Hub: https://hub-ci.huggingface.co/datasets/__DUMMY_TRANSFORMERS_USER__/test_polars_panic_error BEFORE: ~~CI worker test `test_polars_struct_thread_panic_error` raises `LocalEntryNotFoundError`: https://github.com/huggingface/dataset-viewer/actions/runs/11052280214/job/30704011957~~ ``` ERROR tests/job_runners/split/test_descriptive_statistics.py::test_polars_struct_thread_panic_error - huggingface_hub.utils._errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on. ``` ~~Direct cause is a 500 server error:~~ ``` E huggingface_hub.utils._errors.HfHubHTTPError: 500 Server Error: Internal Server Error for url: https://hub-ci.huggingface.co/datasets/__DUMMY_TRANSFORMERS_USER__/test_polars_panic_error/resolve/main/test_polars_panic_error.parquet (Request ID: Root=1-66f555d0-2480d6c65b27b8e7047e6f9c;093b5eeb-e881-4c93-bafd-12965f704b66) E E Cannot read properties of undefined (reading 'cloudfront') (internal error hidden in production) ```
closed
2024-09-26T12:50:09Z
2024-09-26T15:23:32Z
2024-09-26T15:23:32Z
albertvillanova
2,549,939,957
Update datasets to 3.0.1
Release: https://github.com/huggingface/datasets/releases/tag/3.0.1 This should fix issues with missing fields in JSON-lines datasets.
open
2024-09-26T08:50:12Z
2024-09-26T08:50:12Z
null
albertvillanova
2,543,318,612
fine grained token for admin
require fine grained token with repo.write permission to access the admin endpoints
closed
2024-09-23T18:09:11Z
2024-09-24T09:46:56Z
2024-09-24T09:46:54Z
lhoestq
2,541,131,118
`sabareesh88/FNSPID_external` and `sabareesh88/FNSPID_nasdaq` datasets not converting to duckdb
[User is reporting a few datasets ](https://discord.com/channels/879548962464493619/1286797561986023464/1286797561986023464)where the Search / SQL Console isn't showing (_because of job failing with duckdb branch_) - https://huggingface.co/datasets/sabareesh88/FNSPID_nasdaq - https://huggingface.co/datasets/sabareesh88/FNSPID_external I tried refreshing the job and the dataset, but still getting `Job manager crashed while running this job (missing heartbeats).` for `split-duckdb-index`.
closed
2024-09-22T15:29:18Z
2024-09-22T15:31:08Z
2024-09-22T15:31:08Z
cfahlgren1
2,538,934,712
Update cryptography >=43.0.1
Fix for https://github.com/huggingface/dataset-viewer/security/dependabot/610
closed
2024-09-20T14:11:20Z
2024-09-20T17:59:13Z
2024-09-20T17:59:11Z
AndreaFrancis
2,533,155,513
Upgrade huggingface_hub to 0.25.0
https://github.com/huggingface/huggingface_hub/releases/tag/v0.25.0
open
2024-09-18T09:09:12Z
2024-09-18T09:09:58Z
null
severo
2,533,153,387
Simplify test code where a dataset is set as gated
[huggingface_hub@0.25.0](https://github.com/huggingface/huggingface_hub/releases/tag/v0.25.0) provides an API to set a repository as gated. We had included a custom version of `update_repo_settings` because it lacked a `gated` parameter. Now we can switch back to the `huggingface_hub` method https://github.com/huggingface/dataset-viewer/blob/4859100ef282dcf73257dfb60e6b5a20d5955c68/jobs/cache_maintenance/tests/utils.py#L41 https://github.com/huggingface/dataset-viewer/blob/4859100ef282dcf73257dfb60e6b5a20d5955c68/services/admin/tests/fixtures/hub.py#L24 https://github.com/huggingface/dataset-viewer/blob/4859100ef282dcf73257dfb60e6b5a20d5955c68/services/worker/tests/fixtures/hub.py#L35
open
2024-09-18T09:08:14Z
2024-09-18T09:09:31Z
null
severo
2,531,125,441
update developer guide
null
closed
2024-09-17T13:09:25Z
2024-09-20T13:40:54Z
2024-09-20T13:40:52Z
severo
2,525,474,056
fix(chart): block admin metrics exposition
null
closed
2024-09-13T18:58:34Z
2024-09-18T05:02:10Z
2024-09-18T05:02:08Z
rtrompier
2,521,081,448
add documentation page for api playground
add api playground to docs _the playground doesn't support auth yet, but plan to add soon_ _**preview of docs page below**_ # Dataset Viewer API Playground The [API Playground](https://huggingface.co/spaces/cfahlgren1/datasets-api-playground) is a space that allows you to make requests to the Dataset Viewer API and visualize the results in real-time. It's a great way to explore the capabilities of the API and test different queries. <div class="flex justify-center"> <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/datasets-server/datasets-api-playground.png" alt="Dataset API Playground" /> </a> </div>
open
2024-09-12T00:52:05Z
2024-09-12T14:48:17Z
null
cfahlgren1
2,513,554,933
Add relaion2b dataset opt-out results
previously we had the laion2b results, but laion re-released the dataset with an improved filtering based on safety
closed
2024-09-09T10:23:54Z
2024-09-09T10:37:33Z
2024-09-09T10:37:31Z
lhoestq
2,507,235,555
Include splits in HF Croissant definitions.
null
open
2024-09-05T09:16:08Z
2024-09-26T08:23:15Z
null
ccl-core
2,504,025,093
Bump cryptography from 42.0.4 to 43.0.1 in /libs/libcommon in the pip group across 1 directory
[//]: # (dependabot-start) ⚠️ **Dependabot is rebasing this PR** ⚠️ Rebasing might not happen immediately, so don't worry if this takes some time. Note: if you make any changes to this PR yourself, they will take precedence over the rebase. --- [//]: # (dependabot-end) Bumps the pip group with 1 update in the /libs/libcommon directory: [cryptography](https://github.com/pyca/cryptography). Updates `cryptography` from 42.0.4 to 43.0.1 <details> <summary>Changelog</summary> <p><em>Sourced from <a href="https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst">cryptography's changelog</a>.</em></p> <blockquote> <p>43.0.1 - 2024-09-03</p> <pre><code> * Updated Windows, macOS, and Linux wheels to be compiled with OpenSSL 3.3.2. <p>.. _v43-0-0:</p> <p>43.0.0 - 2024-07-20<br /> </code></pre></p> <ul> <li><strong>BACKWARDS INCOMPATIBLE:</strong> Support for OpenSSL less than 1.1.1e has been removed. Users on older version of OpenSSL will need to upgrade.</li> <li><strong>BACKWARDS INCOMPATIBLE:</strong> Dropped support for LibreSSL &lt; 3.8.</li> <li>Updated Windows, macOS, and Linux wheels to be compiled with OpenSSL 3.3.1.</li> <li>Updated the minimum supported Rust version (MSRV) to 1.65.0, from 1.63.0.</li> <li>:func:<code>~cryptography.hazmat.primitives.asymmetric.rsa.generate_private_key</code> now enforces a minimum RSA key size of 1024-bit. Note that 1024-bit is still considered insecure, users should generally use a key size of 2048-bits.</li> <li>:func:<code>~cryptography.hazmat.primitives.serialization.pkcs7.serialize_certificates</code> now emits ASN.1 that more closely follows the recommendations in :rfc:<code>2315</code>.</li> <li>Added new :doc:<code>/hazmat/decrepit/index</code> module which contains outdated and insecure cryptographic primitives. :class:<code>~cryptography.hazmat.primitives.ciphers.algorithms.CAST5</code>, :class:<code>~cryptography.hazmat.primitives.ciphers.algorithms.SEED</code>, :class:<code>~cryptography.hazmat.primitives.ciphers.algorithms.IDEA</code>, and :class:<code>~cryptography.hazmat.primitives.ciphers.algorithms.Blowfish</code>, which were deprecated in 37.0.0, have been added to this module. They will be removed from the <code>cipher</code> module in 45.0.0.</li> <li>Moved :class:<code>~cryptography.hazmat.primitives.ciphers.algorithms.TripleDES</code> and :class:<code>~cryptography.hazmat.primitives.ciphers.algorithms.ARC4</code> into :doc:<code>/hazmat/decrepit/index</code> and deprecated them in the <code>cipher</code> module. They will be removed from the <code>cipher</code> module in 48.0.0.</li> <li>Added support for deterministic :class:<code>~cryptography.hazmat.primitives.asymmetric.ec.ECDSA</code> (:rfc:<code>6979</code>)</li> <li>Added support for client certificate verification to the :mod:<code>X.509 path validation &lt;cryptography.x509.verification&gt;</code> APIs in the form of :class:<code>~cryptography.x509.verification.ClientVerifier</code>, :class:<code>~cryptography.x509.verification.VerifiedClient</code>, and <code>PolicyBuilder</code> :meth:<code>~cryptography.x509.verification.PolicyBuilder.build_client_verifier</code>.</li> <li>Added Certificate :attr:<code>~cryptography.x509.Certificate.public_key_algorithm_oid</code> and Certificate Signing Request :attr:<code>~cryptography.x509.CertificateSigningRequest.public_key_algorithm_oid</code> to determine the :class:<code>~cryptography.hazmat._oid.PublicKeyAlgorithmOID</code> Object Identifier of the public key found inside the certificate.</li> <li>Added :attr:<code>~cryptography.x509.InvalidityDate.invalidity_date_utc</code>, a timezone-aware alternative to the naïve <code>datetime</code> attribute :attr:<code>~cryptography.x509.InvalidityDate.invalidity_date</code>.</li> <li>Added support for parsing empty DN string in</li> </ul> <!-- raw HTML omitted --> </blockquote> <p>... (truncated)</p> </details> <details> <summary>Commits</summary> <ul> <li><a href="https://github.com/pyca/cryptography/commit/a7733878281ca261c4ada04022fc706ba5de9d8b"><code>a773387</code></a> bump for 43.0.1 (<a href="https://redirect.github.com/pyca/cryptography/issues/11533">#11533</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/0393fef5758e55e3c7b3a3e6e5b77821c594a87f"><code>0393fef</code></a> Backport setuptools version ban (<a href="https://redirect.github.com/pyca/cryptography/issues/11526">#11526</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/6687bab97aef31d6ee6cc94ecc87a972137b5d4a"><code>6687bab</code></a> Bump openssl from 0.10.65 to 0.10.66 in /src/rust (<a href="https://redirect.github.com/pyca/cryptography/issues/11320">#11320</a>) (<a href="https://redirect.github.com/pyca/cryptography/issues/11324">#11324</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/ebf14f2edc8536f36797979cb0e075e766d978c5"><code>ebf14f2</code></a> bump for 43.0.0 and update changelog (<a href="https://redirect.github.com/pyca/cryptography/issues/11311">#11311</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/42788a0353e0ca0d922b6b8b9bde77cbb1c65984"><code>42788a0</code></a> Fix exchange with keys that had Q automatically computed (<a href="https://redirect.github.com/pyca/cryptography/issues/11309">#11309</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/2dbdfb8f3913cb9cef08218fcd48a9b4eaa8b57d"><code>2dbdfb8</code></a> don't assign unused name (<a href="https://redirect.github.com/pyca/cryptography/issues/11310">#11310</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/ccc66e6cdf92f4c29012f86f44ad183161eccaad"><code>ccc66e6</code></a> Bump openssl from 0.10.64 to 0.10.65 in /src/rust (<a href="https://redirect.github.com/pyca/cryptography/issues/11308">#11308</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/4310c8727b50fa5f713a0e863ee3defc0c831921"><code>4310c87</code></a> Bump sphinxcontrib-qthelp from 1.0.7 to 1.0.8 (<a href="https://redirect.github.com/pyca/cryptography/issues/11307">#11307</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/f66a9c4b4fe9b87825872fef7a36c319b823f322"><code>f66a9c4</code></a> Bump sphinxcontrib-htmlhelp from 2.0.5 to 2.0.6 (<a href="https://redirect.github.com/pyca/cryptography/issues/11306">#11306</a>)</li> <li><a href="https://github.com/pyca/cryptography/commit/a8fcf18ee0bb0570bd4c9041cf387dc7a9c1968a"><code>a8fcf18</code></a> Bump openssl-sys from 0.9.102 to 0.9.103 in /src/rust (<a href="https://redirect.github.com/pyca/cryptography/issues/11305">#11305</a>)</li> <li>Additional commits viewable in <a href="https://github.com/pyca/cryptography/compare/42.0.4...43.0.1">compare view</a></li> </ul> </details> <br /> [![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=cryptography&package-manager=pip&previous-version=42.0.4&new-version=43.0.1)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) --- <details> <summary>Dependabot commands and options</summary> <br /> You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show <dependency name> ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore <dependency name> major version` will close this group update PR and stop Dependabot creating any more for the specific dependency's major version (unless you unignore this specific dependency's major version or upgrade to it yourself) - `@dependabot ignore <dependency name> minor version` will close this group update PR and stop Dependabot creating any more for the specific dependency's minor version (unless you unignore this specific dependency's minor version or upgrade to it yourself) - `@dependabot ignore <dependency name>` will close this group update PR and stop Dependabot creating any more for the specific dependency (unless you unignore this specific dependency or upgrade to it yourself) - `@dependabot unignore <dependency name>` will remove all of the ignore conditions of the specified dependency - `@dependabot unignore <dependency name> <ignore condition>` will remove the ignore condition of the specified dependency and ignore conditions You can disable automated security fix PRs for this repo from the [Security Alerts page](https://github.com/huggingface/dataset-viewer/network/alerts). </details>
closed
2024-09-04T00:08:11Z
2024-09-20T17:59:55Z
2024-09-20T17:59:47Z
dependabot[bot]
2,502,373,594
The stale bot is broken (403 forbidden from GitHub)
See https://github.com/huggingface/dataset-viewer/actions/runs/10646207687 > github.GithubException.GithubException: 403 {"message": "Resource not accessible by integration", "documentation_url": "https://docs.github.com/rest/issues/issues#update-an-issue", "status": "403"} Some options: - fix the issue (seems to be an issue with a token) - delete the stale bot - replace this custom stale bot with something more standard
open
2024-09-03T09:27:57Z
2024-09-03T09:28:12Z
null
severo
2,499,053,306
Improve estimated row count
Currently estimated row count assumes file sizes are consistent across the entire set, from what I've seen this results in wildly inaccurate estimates for WebDataset. Typically WebDatasets are created with a set number of samples per file, therefore a simpler more accurate estimate can be calculated from the row count of one shard multiplied by the total number of shards.
open
2024-08-31T21:28:52Z
2024-09-04T09:45:10Z
null
hlky
2,495,811,839
Image URL detection
[`is_image_url`](https://github.com/huggingface/dataset-viewer/blob/946b0788fa426007161f2077a70b5ae64b211cf8/libs/libcommon/src/libcommon/utils.py#L131-L134) relies on a filename and extension being present, however, in some cases an image URL does not contain a filename. Example [dataset](https://huggingface.co/datasets/bigdata-pw/SteamScreenshots) and example [URL](https://steamuserimages-a.akamaihd.net/ugc/910172100453203507/062F4787060B2E4E93EFC4631E96183B027A860B/). This could be improved by checking the `content-type` header of the response or checking for strings like "image" in the URL.
open
2024-08-29T23:17:55Z
2024-09-03T09:17:27Z
null
hlky
2,492,135,837
Investigate the most common errors
All the entries in the cache that have a cause exception, sorted by number of occurrences: | error_code | cause_exception | count | | --- | --- | --- | | EmptyDatasetError | EmptyDatasetError | 17886 | | DataFilesNotFoundError | DataFilesNotFoundError | 9233 | | ComputationError | StatisticsComputationError | 6534 | | ComputationError | TypeError | 3063 | | DatasetGenerationCastError | DatasetGenerationCastError | 2403 | | PreviousStepStillProcessingError | CachedArtifactNotFoundError | 2160 | | DatasetGenerationError | ArrowInvalid | 2127 | | FeaturesError | ArrowInvalid | 1543 | | InfoError | HfHubHTTPError | 1354 | | DatasetGenerationError | UnicodeDecodeError | 1211 | | DatasetGenerationError | TypeError | 1176 | | UnexpectedError | TypeError | 1033 | | FeaturesError | ValueError | 969 | | DatasetGenerationError | ArrowNotImplementedError | 943 | | UnexpectedError | HfHubHTTPError | 896 | | FeaturesError | UnicodeDecodeError | 868 | | DatasetGenerationError | ValueError | 867 | | SplitNamesFromStreamingError | SplitsNotFoundError | 844 | | UnexpectedError | ValueError | 822 | | InfoError | BrokenPipeError | 817 | | InfoError | SplitsNotFoundError | 708 | | ConfigNamesError | ImportError | 651 | | UnexpectedError | ParserException | 621 | | UnexpectedError | ReadTimeout | 463 | | UnexpectedError | BinderException | 462 | | DatasetGenerationError | ParserError | 445 | | FeaturesError | ParserError | 436 | | ComputationError | ZeroDivisionError | 421 | | DatasetGenerationError | SchemaInferenceError | 387 | | UnexpectedError | FileNotFoundError | 343 | | ConfigNamesError | ValueError | 312 | | PolarsParquetReadError | FileNotFoundError | 273 | | FeaturesError | ZstdError | 266 | | StreamingRowsError | ValueError | 266 | | FileFormatMismatchBetweenSplitsError | ValueError | 221 | | UnexpectedError | RuntimeError | 207 | | UnexpectedError | PermissionError | 202 | | UnexpectedError | UnidentifiedImageError | 195 | | UnexpectedError | BadZipFile | 192 | | PolarsParquetReadError | ComputeError | 180 | | DatasetGenerationError | FileNotFoundError | 174 | | RowsPostProcessingError | ValueError | 168 | | ComputationError | UnidentifiedImageError | 166 | | UnexpectedError | UnicodeDecodeError | 163 | | UnexpectedError | EntryNotFoundError | 162 | | FeaturesError | ArrowTypeError | 134 | | UnexpectedError | ConnectionError | 133 | | StreamingRowsError | CastError | 132 | | DatasetGenerationError | KeyError | 131 | | ComputationError | SchemaError | 124 | | ConfigNamesError | BadZipFile | 111 | | DatasetGenerationError | ArrowTypeError | 108 | | UnexpectedError | ArrowInvalid | 104 | | DatasetGenerationError | CastError | 104 | | RowsPostProcessingError | KeyError | 103 | | StreamingRowsError | OSError | 94 | | UnexpectedError | ColumnNotFoundError | 93 | | StreamingRowsError | RuntimeError | 89 | | RowsPostProcessingError | UnidentifiedImageError | 86 | | UnexpectedError | SchemaError | 83 | | ConfigNamesError | FileNotFoundError | 82 | | StreamingRowsError | ArrowInvalid | 80 | | UnexpectedError | ReadError | 79 | | StreamingRowsError | KeyError | 78 | | UnexpectedError | ClientResponseError | 77 | | UnexpectedError | ComputeError | 74 | | InfoError | DatasetWithScriptNotSupportedError | 70 | | RowsPostProcessingError | TypeError | 68 | | InfoError | DatasetNotFoundError | 67 | | CreateCommitError | RepositoryNotFoundError | 65 | | FeaturesError | EmptyDataError | 65 | | UnexpectedError | InvalidInputException | 58 | | UnexpectedError | ServerDisconnectedError | 54 | | ComputationError | ValueError | 54 | | UnexpectedError | IOException | 52 | | UnexpectedError | ConversionException | 50 | | ConfigNamesError | TypeError | 47 | | InfoError | ReadTimeout | 42 | | DatasetGenerationError | EmptyDataError | 42 | | StreamingRowsError | TypeError | 40 | | UnexpectedError | KeyError | 38 | | PolarsParquetReadError | ColumnNotFoundError | 38 | | DatasetGenerationError | ConnectionError | 38 | | DatasetGenerationError | BadZipFile | 37 | | StreamingRowsError | UnicodeDecodeError | 36 | | UnexpectedError | NonMatchingSplitsSizesError | 36 | | UnexpectedError | OSError | 34 | | DatasetGenerationError | ArrowCapacityError | 33 | | UnexpectedError | ArrowTypeError | 33 | | SplitNamesFromStreamingError | FileNotFoundError | 33 | | DatasetGenerationError | GatedRepoError | 33 | | InfoError | ConnectionError | 31 | | RowsPostProcessingError | CouldntDecodeError | 30 | | DatasetGenerationError | OverflowError | 26 | | StreamingRowsError | UnidentifiedImageError | 25 | | FeaturesError | OverflowError | 22 | | StreamingRowsError | LibsndfileError | 21 | | UnexpectedError | ArrowCapacityError | 20 | | UnexpectedError | DecompressionBombError | 20 | | StreamingRowsError | FileNotFoundError | 19 | | UnexpectedError | NotImplementedError | 18 | | RowsPostProcessingError | OSError | 18 | | UnexpectedError | ZeroDivisionError | 18 | | ComputationError | DecompressionBombError | 18 | | NormalRowsError | DatasetGenerationError | 18 | | DatasetGenerationError | ReadError | 17 | | UnexpectedError | Error | 17 | | FeaturesError | RuntimeError | 16 | | UnexpectedError | DatasetGenerationError | 15 | | UnexpectedError | IndexError | 14 | | DatasetGenerationError | RuntimeError | 13 | | UnexpectedError | JSONDecodeError | 13 | | UnexpectedError | ParserError | 12 | | DatasetGenerationError | NotImplementedError | 12 | | ComputationError | ArrowInvalid | 12 | | StreamingRowsError | NotImplementedError | 11 | | StreamingRowsError | ParserError | 11 | | ConfigNamesError | InvalidConfigName | 11 | | DatasetGenerationError | ArrowIndexError | 11 | | ComputationError | DuplicateError | 11 | | StreamingRowsError | ArrowNotImplementedError | 10 | | FeaturesError | HfHubHTTPError | 10 | | FeaturesError | AttributeError | 9 | | CreateCommitError | BadRequestError | 8 | | StreamingRowsError | HfHubHTTPError | 8 | | NormalRowsError | FileNotFoundError | 8 | | FeaturesError | UnsupportedOperation | 8 | | UnexpectedError | FSTimeoutError | 8 | | ConfigNamesError | AttributeError | 8 | | UnexpectedError | AttributeError | 8 | | DatasetGenerationError | OSError | 7 | | FeaturesError | ArrowCapacityError | 7 | | DatasetGenerationError | AttributeError | 7 | | ConfigNamesError | ReadTimeout | 7 | | CreateCommitError | EntryNotFoundError | 6 | | FeaturesError | BadGzipFile | 6 | | FeaturesError | BadZipFile | 6 | | NormalRowsError | DatasetGenerationCastError | 6 | | ComputationError | InvalidOperationError | 6 | | ComputationError | OverflowError | 5 | | UnexpectedError | HTTPError | 5 | | InfoError | ValueError | 5 | | DatasetGenerationError | EOFError | 4 | | InfoError | FileNotFoundError | 4 | | FeaturesError | HTTPError | 4 | | UnexpectedError | TypeMismatchException | 4 | | NormalRowsError | HfHubHTTPError | 4 | | UnexpectedError | DatasetGenerationCastError | 4 | | UnexpectedError | InvalidOperationError | 4 | | UnexpectedError | ClientConnectorError | 4 | | StreamingRowsError | ReadError | 3 | | ConfigNamesError | UnicodeDecodeError | 3 | | FeaturesError | NotImplementedError | 3 | | PolarsParquetReadError | error | 3 | | NormalRowsError | OSError | 3 | | UnexpectedError | ExpectedMoreSplits | 3 | | FeaturesError | FileNotFoundError | 3 | | DatasetGenerationError | HfHubHTTPError | 3 | | UnexpectedError | error | 3 | | UnexpectedError | UnpicklingError | 3 | | StreamingRowsError | AssertionError | 3 | | StreamingRowsError | EmptyDataError | 3 | | StreamingRowsError | EntryNotFoundError | 2 | | DatasetGenerationError | UnsupportedOperation | 2 | | UnexpectedError | InternalException | 2 | | DatasetGenerationError | error | 2 | | ConfigNamesError | ScannerError | 2 | | StreamingRowsError | ArrowCapacityError | 2 | | RetryableConfigNamesError | HfHubHTTPError | 2 | | StreamingRowsError | DecompressionBombError | 2 | | SplitNamesFromStreamingError | HfHubHTTPError | 2 | | ConfigNamesError | JSONDecodeError | 2 | | ConfigNamesError | KeyError | 2 | | InfoError | DataFilesNotFoundError | 2 | | DatasetGenerationError | EntryNotFoundError | 2 | | InfoError | BadZipFile | 2 | | UnexpectedError | IsADirectoryError | 2 | | UnexpectedError | DuplicateError | 2 | | FeaturesError | KeyError | 1 | | RowsPostProcessingError | SyntaxError | 1 | | CreateCommitError | HfHubHTTPError | 1 | | FeaturesError | ConnectionError | 1 | | StreamingRowsError | error | 1 | | StreamingRowsError | OverflowError | 1 | | StreamingRowsError | HTTPError | 1 | | StreamingRowsError | ArrowTypeError | 1 | | StreamingRowsError | AttributeError | 1 | | UnexpectedError | ChunkedEncodingError | 1 | | UnexpectedError | EmptyDatasetError | 1 | | RowsPostProcessingError | LibsndfileError | 1 | | UnexpectedError | TransactionException | 1 | | UnexpectedError | EOFError | 1 | | SplitNamesFromStreamingError | ConnectionError | 1 | | InfoError | JSONDecodeError | 1 | | UnexpectedError | ClientPayloadError | 1 | | FeaturesError | ReadTimeout | 1 | | UnexpectedError | EmptyDataError | 1 | | ConfigNamesError | IsADirectoryError | 1 | | DatasetGenerationError | AssertionError | 1 | ```js db.cachedResponsesBlue.aggregate([ { $match: { "details.copied_from_artifact": {"$exists": false}, "details.cause_exception": {"$exists": true}, }, }, { $group: { _id: { error_code: "$error_code", cause_exception: "$details.cause_exception", }, count: { $sum: 1 }, }, }, { $sort: { count: -1 } } ]); ```
open
2024-08-28T13:48:06Z
2024-08-28T13:54:09Z
null
severo
2,491,704,042
Don't try to include binary cells in the /rows responses
https://huggingface.co/datasets/frutiemax/themoviedb_posters/discussions/2 > I have some columns that the viewer do not need to load i.e. T5 prompt embeds and VAE features. Currently, the viewer freezes because the dataset is too big to load. Thanks. <img width="1019" alt="Capture d’écran 2024-08-28 à 12 31 40" src="https://github.com/user-attachments/assets/62684493-2980-4658-aef7-33ca6139af70"> https://huggingface.co/datasets/frutiemax/themoviedb_posters/viewer/default/train lags forever to render the following: <img width="912" alt="Capture d’écran 2024-08-28 à 12 32 50" src="https://github.com/user-attachments/assets/00abc6a0-300b-4305-b803-242fc59bd42f"> We should ignore/hide the binary cells (`bytes (5KB)` for example)
open
2024-08-28T10:33:37Z
2024-08-28T10:34:22Z
null
severo
2,490,707,558
Fix typo "build > built"
null
closed
2024-08-28T01:55:19Z
2024-08-28T09:33:50Z
2024-08-28T09:33:50Z
david4096
2,489,615,941
remove obsolete retryable error codes
we don't have these error codes in the database anymore
closed
2024-08-27T14:50:53Z
2024-08-27T14:51:06Z
2024-08-27T14:51:04Z
severo
2,486,482,261
no need to keep the trace of the first exception
fixes #3048
closed
2024-08-26T10:08:39Z
2024-08-26T10:22:50Z
2024-08-26T10:22:49Z
severo
2,486,453,418
Hide `CachedArtifactError` to users
For example, currently on https://huggingface.co/datasets/ProGamerGov/synthetic-dataset-1m-dalle3-high-quality-captions ``` Cannot load the dataset split (in streaming mode) to extract the first rows. ``` ``` Error code: StreamingRowsError Exception: KeyError Message: 'jpeg' Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 322, in compute compute_first_rows_from_parquet_response( File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 88, in compute_first_rows_from_parquet_response rows_index = indexer.get_rows_index( File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 640, in get_rows_index return RowsIndex( File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 521, in __init__ self.parquet_index = self._init_parquet_index( File "/src/libs/libcommon/src/libcommon/parquet_utils.py", line 538, in _init_parquet_index response = get_previous_step_or_raise( File "/src/libs/libcommon/src/libcommon/simple_cache.py", line 591, in get_previous_step_or_raise raise CachedArtifactError( libcommon.simple_cache.CachedArtifactError: The previous step failed. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/src/worker/utils.py", line 96, in get_rows_or_raise return get_rows( File "/src/libs/libcommon/src/libcommon/utils.py", line 197, in decorator return func(*args, **kwargs) File "/src/services/worker/src/worker/utils.py", line 73, in get_rows rows_plus_one = list(itertools.islice(ds, rows_max_number + 1)) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 1816, in __iter__ for key, example in ex_iterable: File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/iterable_dataset.py", line 238, in __iter__ for key_example in islice(self.generate_examples_fn(**gen_kwags), shard_example_idx_start, None): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/webdataset/webdataset.py", line 115, in _generate_examples example[field_name] = {"path": example["__key__"] + "." + field_name, "bytes": example[field_name]} KeyError: 'jpeg' ```
closed
2024-08-26T09:55:24Z
2024-08-26T12:21:04Z
2024-08-26T10:22:49Z
severo
2,483,908,267
call pre_compute() and post_compute() in tests
see https://github.com/huggingface/dataset-viewer/pull/3046#issuecomment-2307793018
closed
2024-08-23T21:15:20Z
2024-08-23T21:25:08Z
2024-08-23T21:25:07Z
severo
2,483,501,949
remove support for script-based datasets
fixes #3004
closed
2024-08-23T16:44:06Z
2024-08-26T08:10:06Z
2024-08-26T08:10:04Z
severo
2,483,347,395
mnist and zalando are now data-only
see #3004
closed
2024-08-23T15:06:54Z
2024-08-23T15:07:24Z
2024-08-23T15:07:23Z
severo
2,482,760,646
Remove temporary retryable error codes
null
closed
2024-08-23T09:48:03Z
2024-08-23T09:48:12Z
2024-08-23T09:48:11Z
severo
2,482,714,622
Update values.yaml
remove obsolete datasets glob
closed
2024-08-23T09:24:20Z
2024-08-23T09:24:38Z
2024-08-23T09:24:37Z
severo
2,480,869,744
show the traceback of the cause, it's clearer
Current: https://huggingface.co/datasets/Recag/Rp_CommonC_53 ``` The dataset generation failed Error code: DatasetGenerationError Exception: DatasetGenerationError Message: An error occurred while generating the dataset Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 145, in _generate_tables dataset = json.load(f) File "/usr/local/lib/python3.9/json/__init__.py", line 293, in load return loads(fp.read(), File "/usr/local/lib/python3.9/json/__init__.py", line 346, in loads return _default_decoder.decode(s) File "/usr/local/lib/python3.9/json/decoder.py", line 340, in decode raise JSONDecodeError("Extra data", s, end) json.decoder.JSONDecodeError: Extra data: line 2 column 1 (char 8779) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1995, in _prepare_split_single for _, table in generator: File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 148, in _generate_tables raise e File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 122, in _generate_tables pa_table = paj.read_json( File "pyarrow/_json.pyx", line 308, in pyarrow._json.read_json File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: JSON parse error: Missing a closing quotation mark in string. in row 128 The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1321, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 935, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1027, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1122, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1882, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2038, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset ``` With this PR: ``` The dataset generation failed Error code: DatasetGenerationError Exception: pyarrow.lib.ArrowInvalid Message: JSON parse error: Missing a closing quotation mark in string. in row 128 Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 145, in _generate_tables dataset = json.load(f) File "/usr/local/lib/python3.9/json/__init__.py", line 293, in load return loads(fp.read(), File "/usr/local/lib/python3.9/json/__init__.py", line 346, in loads return _default_decoder.decode(s) File "/usr/local/lib/python3.9/json/decoder.py", line 340, in decode raise JSONDecodeError("Extra data", s, end) json.decoder.JSONDecodeError: Extra data: line 2 column 1 (char 8779) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1995, in _prepare_split_single for _, table in generator: File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 148, in _generate_tables raise e File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/json/json.py", line 122, in _generate_tables pa_table = paj.read_json( File "pyarrow/_json.pyx", line 308, in pyarrow._json.read_json File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowInvalid: JSON parse error: Missing a closing quotation mark in string. in row 128 ```
closed
2024-08-22T13:59:04Z
2024-08-23T09:47:14Z
2024-08-22T14:38:12Z
severo
2,480,735,834
Raise specific errors instead of ConfigNamesError when appropriate
It will help show them better in moonlanding see #3010
closed
2024-08-22T13:01:17Z
2024-08-22T14:34:52Z
2024-08-22T14:34:50Z
severo
2,480,397,730
Update datasets to 2.20.0
Update datasets to 2.20.0. This PR is intended to address the CI errors raised by this update, as a first step before: - #3037 Fixes after the update of `datasets`: - Pass `trust_remote_code=True` for script dataset: 227f2c4 - Use JSON-Lines (instead of JSON) dataset in `test_statistics_endpoint` to avoid pandas bug that downcasts float to int column: a86d040
closed
2024-08-22T10:16:02Z
2024-08-23T05:05:45Z
2024-08-23T05:05:43Z
albertvillanova
2,480,259,294
fix the lock release when finishing a job
fixes #1888
closed
2024-08-22T09:16:32Z
2024-08-22T09:48:28Z
2024-08-22T09:48:27Z
severo
2,480,119,650
Update huggingface-hub to 0.24.6
Update huggingface-hub to 0.24.6. Note that the fix: - #2781 - https://github.com/huggingface/huggingface_hub/pull/2271 was released in : https://github.com/huggingface/huggingface_hub/releases/tag/v0.23.1 Related to: - #3037
closed
2024-08-22T08:07:47Z
2024-08-22T08:28:47Z
2024-08-22T08:28:45Z
albertvillanova
2,479,891,224
Update datasets to 2.21.0
Update datasets to 2.21.0. Fix #3024.
closed
2024-08-22T05:45:39Z
2024-09-26T14:44:57Z
2024-09-26T14:44:54Z
albertvillanova
2,479,535,819
More explicit test and comments about offset-naive datetimes read from mongo
I think it's enough to close #862
closed
2024-08-22T01:31:14Z
2024-08-22T09:49:11Z
2024-08-22T09:49:10Z
severo
2,479,450,462
return 404 when accessing a renamed dataset, instead of unexpected error
fixes bug at https://github.com/huggingface/dataset-viewer/issues/2688#issuecomment-2303305529
closed
2024-08-22T00:07:40Z
2024-08-22T00:11:13Z
2024-08-22T00:11:11Z
severo
2,478,234,702
import from the correct file
see https://github.com/huggingface/huggingface_hub/pull/2474#issuecomment-2302132315
closed
2024-08-21T14:48:30Z
2024-08-21T15:22:08Z
2024-08-21T15:22:06Z
severo
2,478,105,982
Fix canonical dataset names
I replaced all the canonical datasets I could find in openapi.json and in the presidio list. Possibly we have more in the code (tests? docs?). For a next pass
closed
2024-08-21T13:52:34Z
2024-08-21T15:21:45Z
2024-08-21T15:21:43Z
severo
2,477,650,983
mediatype of opus is audio/ogg
I had set it to `audio/opus` which does not exist. https://datatracker.ietf.org/doc/html/rfc7845.html#section-9 > An "Ogg Opus file" consists of one or more sequentially multiplexed segments, each containing exactly one Ogg Opus stream. The RECOMMENDED mime-type for Ogg Opus files is "audio/ogg". > The RECOMMENDED filename extension for Ogg Opus files is '.opus'. > When Opus is concurrently multiplexed with other streams in an Ogg container, one SHOULD use one of the "audio/ogg", "video/ogg", or "application/ogg" mime-types, as defined in [[RFC5334](https://datatracker.ietf.org/doc/html/rfc5334)]. https://github.com/user-attachments/assets/caed4ee5-975d-4233-9cf8-31fd755b2429
closed
2024-08-21T10:17:43Z
2024-08-21T13:18:01Z
2024-08-21T10:39:11Z
severo
2,477,650,534
Update to new dataset ID in example
At the moment, the examples fail because rotten_tomatoes moved to https://huggingface.co/datasets/cornell-movie-review-data/rotten_tomatoes
closed
2024-08-21T10:17:31Z
2024-08-21T13:31:54Z
2024-08-21T13:31:34Z
davanstrien
2,476,445,158
[nit] admin and sse-api get requests from ALB on /healthcheck
<img width="680" alt="Capture d’écran 2024-08-20 à 22 01 39" src="https://github.com/user-attachments/assets/0a63090e-4fe0-4a26-9052-20cc07a844d3"> <img width="673" alt="Capture d’écran 2024-08-20 à 22 01 26" src="https://github.com/user-attachments/assets/aa4d2ad6-6721-435f-878c-a7d2abb08e7b">
closed
2024-08-20T20:00:48Z
2024-08-20T20:15:49Z
2024-08-20T20:11:03Z
severo
2,476,112,311
add ingress for /sse
https://datasets-server.us.dev.moon.huggingface.tech/sse/healthcheck or https://datasets-server.us.dev.moon.huggingface.tech/sse/hub-cache are currently returning 404 not found
closed
2024-08-20T16:52:19Z
2024-08-20T16:53:00Z
2024-08-20T16:52:59Z
severo
2,475,138,672
Use huggingface_hub to access /auth-check
See https://github.com/huggingface/huggingface_hub/issues/2466 Occurrences: https://github.com/search?q=repo%3Ahuggingface%2Fdataset-viewer%20auth-check&type=code
open
2024-08-20T09:17:33Z
2024-09-18T09:09:36Z
null
severo
2,473,536,178
Imagefolder: UnexpectedError with root cause: "[Errno 13] Permission denied: '/tmp/hf-datasets-cache/medium/datasets/....incomplete'"
We have some "UnexpectedError" with this kind of root cause: https://huggingface.co/datasets/abhi1505/Drone_Data/discussions/9 ``` [Errno 13] Permission denied: '/tmp/hf-datasets-cache/medium/datasets/59146511988270-config-parquet-and-info-abhi1505-Drone_Data-6928f977/downloads/a3d6d38d9159fb9d6429f28224c80ecc7a281b095c77183e4381751ea5f3ae72.incomplete' ``` https://huggingface.co/datasets/griffinbholt/augmented_waste_classification/discussions/1 ``` [Errno 13] Permission denied: '/tmp/hf-datasets-cache/medium/datasets/26961854788134-config-parquet-and-info-griffinbholt-augmented_wa-56266047/downloads/28fcc0b914dfa1c8da84fbbad2f870ba1dcffb9ca308dab7c4347a9a2a930a5f.incomplete ```
open
2024-08-19T14:41:36Z
2024-08-22T14:05:17Z
null
severo
2,473,134,535
Fix CI worker tests for gated datasets
Fix CI worker tests for gated datasets. Fix #3025.
closed
2024-08-19T11:31:18Z
2024-08-20T05:57:03Z
2024-08-20T05:57:02Z
albertvillanova
2,473,124,702
CI worker tests are broken for gated datasets: ConnectionError
CI worker tests are broken for gated datasets: https://github.com/huggingface/dataset-viewer/actions/runs/10319430999/job/28567795990 ``` FAILED tests/job_runners/config/test_split_names.py::test_compute_split_names_from_streaming_response[gated-False-SplitNamesFromStreamingError-DatasetNotFoundError] - AssertionError: assert 'ConnectionError' == 'DatasetNotFoundError' - DatasetNotFoundError + ConnectionError FAILED tests/job_runners/dataset/test_config_names.py::test_compute_splits_response[gated-False-ConfigNamesError-DatasetNotFoundError] - AssertionError: assert 'ConnectionError' == 'DatasetNotFoundError' - DatasetNotFoundError + ConnectionError ```
closed
2024-08-19T11:25:45Z
2024-08-20T05:57:03Z
2024-08-20T05:57:03Z
albertvillanova
2,471,073,549
Upgrade to datasets@2.21.0
https://github.com/huggingface/datasets/releases/tag/2.21.0 When done, we should refresh some datasets, like https://huggingface.co/datasets/ProGamerGov/synthetic-dataset-1m-dalle3-high-quality-captions/discussions/1#66bcd7e2f1685a3ade2e55f5
closed
2024-08-16T21:34:16Z
2024-09-26T14:44:56Z
2024-09-26T14:44:55Z
severo
2,470,439,282
fix anchor
found using https://github.com/raviqqe/muffet while checking for broken links ``` muffet --rate-limit 5 -i "https://huggingface.co/docs/dataset-viewer/*" https://huggingface.co/docs/dataset-viewer ```
closed
2024-08-16T14:56:43Z
2024-08-16T14:57:59Z
2024-08-16T14:57:45Z
severo
2,469,571,525
Fix CI e2e admin test_metrics with missing dataset_status
Fix CI e2e admin test_metrics by adding missing `dataset_status="normal"` label. This CI failure was introduced by: - #3008 Fix #3021.
closed
2024-08-16T06:15:00Z
2024-08-16T13:59:08Z
2024-08-16T08:32:31Z
albertvillanova
2,469,561,514
CI e2e admin test_metrics is broken
CI e2e test_metrics is broken: https://github.com/huggingface/dataset-viewer/actions/runs/10319430994/job/28567795517 ``` FAILED tests/test_31_admin_metrics.py::test_metrics - AssertionError: queue_jobs_total - queue=dataset-config-names found in {'starlette_requests_total{method="GET",path_template="/admin"}': 3.0, 'starlette_responses_total{method="GET",path_template="/admin",status_code="200"}': 2.0, 'starlette_requests_processing_time_seconds_sum{method="GET",path_template="/admin"}': 0.021941476999927545, 'starlette_requests_processing_time_seconds_bucket{le="0.005",method="GET",path_template="/admin"}': 1.0, 'starlette_requests_processing_time_seconds_bucket{le="0.01",method="GET",path_template="/admin"}': 1.0, 'starlette_requests_processing_time_seconds_bucket{le="0.025",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="0.05",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="0.075",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="0.1",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="0.25",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="0.5",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="0.75",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="1.0",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="2.5",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="5.0",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="7.5",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="10.0",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_bucket{le="+Inf",method="GET",path_template="/admin"}': 2.0, 'starlette_requests_processing_time_seconds_count{method="GET",path_template="/admin"}': 2.0, 'starlette_requests_in_progress{method="GET",path_template="/admin",pid="145"}': 1.0, 'starlette_requests_in_progress{method="GET",path_template="/admin",pid="23"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-config-names",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-filetypes",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-config-names",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-parquet-and-info",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-parquet",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-info",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-split-names",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-size",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-split-names",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-is-valid",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-opt-in-out-urls-count",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-duckdb-index-size",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-filetypes",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-modalities",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-duckdb-index-size",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-info",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-compatible-libraries",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-croissant-crumbs",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-is-valid",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-hub-cache",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-opt-in-out-urls-count",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-parquet",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-size",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-split-names",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-presidio-entities-count",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-split-names",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-first-rows",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-is-valid",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-opt-in-out-urls-count",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-duckdb-index-size",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-parquet-and-info",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-parquet",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-info",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-size",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-modalities",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-compatible-libraries",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-croissant-crumbs",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-hub-cache",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="dataset-presidio-entities-count",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-duckdb-index-size",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-is-valid",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-opt-in-out-urls-count",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-first-rows",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-is-valid",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-image-url-columns",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-info",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-parquet",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-parquet-metadata",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-size",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-is-valid",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-image-url-columns",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-opt-in-out-urls-scan",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-parquet-metadata",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-descriptive-statistics",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-presidio-scan",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-duckdb-index",status="waiting"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-opt-in-out-urls-scan",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-opt-in-out-urls-count",status="waiting"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-descriptive-statistics",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-duckdb-index",status="started"}': 1.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-presidio-scan",status="started"}': 0.0, 'queue_jobs_total{dataset_status="normal",pid="145",queue="split-opt-in-out-urls-count",status="started"}': 0.0, 'worker_size_jobs_count{pid="145",worker_size="medium"}': 1.0, 'worker_size_jobs_count{pid="145",worker_size="light"}': 8.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-config-names",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-filetypes",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-duckdb-index-size",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-info",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-is-valid",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-opt-in-out-urls-count",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-parquet",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-size",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-split-names",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-split-names",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-parquet-and-info",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-modalities",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-compatible-libraries",pid="145"}': 3.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-croissant-crumbs",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-hub-cache",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="dataset-presidio-entities-count",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-duckdb-index-size",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-is-valid",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-opt-in-out-urls-count",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="split-first-rows",pid="145"}': 3.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-info",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-parquet",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-size",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="split-is-valid",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="split-image-url-columns",pid="145"}': 3.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="config-parquet-metadata",pid="145"}': 5.0, 'responses_in_cache_total{error_code="MissingSpawningTokenError",http_status="500",kind="split-opt-in-out-urls-scan",pid="145"}': 7.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="split-descriptive-statistics",pid="145"}': 3.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="split-duckdb-index",pid="145"}': 2.0, 'responses_in_cache_total{error_code="None",http_status="200",kind="split-presidio-scan",pid="145"}': 2.0, 'responses_in_cache_total{error_code="MissingSpawningTokenError",http_status="500",kind="split-opt-in-out-urls-count",pid="145"}': 6.0, 'responses_in_cache_total{error_code="FileSystemError",http_status="500",kind="config-parquet-metadata",pid="145"}': 2.0, 'responses_in_cache_total{error_code="UnexpectedError",http_status="500",kind="dataset-compatible-libraries",pid="145"}': 4.0, 'responses_in_cache_total{error_code="InfoError",http_status="500",kind="split-first-rows",pid="145"}': 2.0, 'responses_in_cache_total{error_code="FileSystemError",http_status="500",kind="split-descriptive-statistics",pid="145"}': 2.0, 'responses_in_cache_total{error_code="FileSystemError",http_status="500",kind="split-duckdb-index",pid="145"}': 2.0, 'responses_in_cache_total{error_code="PresidioScanNotEnabledForThisDataset",http_status="501",kind="split-presidio-scan",pid="145"}': 3.0, 'responses_in_cache_total{error_code="InfoError",http_status="500",kind="split-image-url-columns",pid="145"}': 2.0, 'responses_in_cache_total{error_code="UnexpectedError",http_status="500",kind="split-first-rows",pid="145"}': 2.0, 'responses_in_cache_total{error_code="UnexpectedError",http_status="500",kind="split-descriptive-statistics",pid="145"}': 2.0, 'responses_in_cache_total{error_code="UnexpectedError",http_status="500",kind="split-duckdb-index",pid="145"}': 2.0, 'responses_in_cache_total{error_code="UnexpectedError",http_status="500",kind="split-image-url-columns",pid="145"}': 2.0, 'responses_in_cache_total{error_code="StreamingRowsError",http_status="500",kind="split-first-rows",pid="145"}': 0.0, 'responses_in_cache_total{error_code="StreamingRowsError",http_status="500",kind="split-image-url-columns",pid="145"}': 0.0, 'responses_in_cache_total{error_code="NormalRowsError",http_status="500",kind="split-presidio-scan",pid="145"}': 1.0, 'parquet_metadata_disk_usage{pid="145",type="total"}': 77851254784.0, 'parquet_metadata_disk_usage{pid="145",type="used"}': 76050628608.0, 'parquet_metadata_disk_usage{pid="145",type="free"}': 1783848960.0, 'parquet_metadata_disk_usage{pid="145",type="percent"}': 97.7} assert False + where False = has_metric(name='queue_jobs_total', labels={'pid': '[0-9]*', 'queue': 'dataset-config-names', 'status': 'started'}, metric_names={'parquet_metadata_disk_usage{pid="145",type="free"}', 'parquet_metadata_disk_usage{pid="145",type="percent"}', 'parquet_metadata_disk_usage{pid="145",type="total"}', 'parquet_metadata_disk_usage{pid="145",type="used"}', 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-duckdb-index-size",status="started"}', 'queue_jobs_total{dataset_status="normal",pid="145",queue="config-duckdb-index-size",status="waiting"}', ...}) ``` I think this failure was introduced after merging: https://github.com/huggingface/dataset-viewer/actions/runs/10182877247/job/28845665481 - #3008
closed
2024-08-16T06:07:49Z
2024-08-16T08:32:32Z
2024-08-16T08:32:32Z
albertvillanova
2,467,918,813
doc: Read parquet files with PySpark
Adding primary doc for reading parquet files with PySpark
closed
2024-08-15T12:00:40Z
2024-08-19T12:05:18Z
2024-08-19T12:05:17Z
AndreaFrancis
2,460,769,263
Fix CI test StorageClient.url_preparator
Fix CI test StorageClient.url_preparator. This fail in CI tests was introduced after: - #2966 See: https://github.com/huggingface/dataset-viewer/actions/runs/10350752694/job/28648303009?pr=3018 ``` FAILED tests/test_response.py::test_create_response_with_image - AssertionError: assert [{'row': {'im...d_cells': []}] == [{'row': {'im...d_cells': []}] At index 0 diff: {'row_idx': 0, 'row': {'image': {'src': 'http://localhost/cached-assets/ds_image/--/{dataset_git_revision}/--/default/train/0/image/image.jpg', 'height': 480, 'width': 640}}, 'truncated_cells': []} != {'row_idx': 0, 'row': {'image': {'src': 'http://localhost/cached-assets/ds_image/--/revision/--/default/train/0/image/image.jpg', 'height': 480, 'width': 640}}, 'truncated_cells': []} Full diff: [ { 'row': { 'image': { 'height': 480, - 'src': 'http://localhost/cached-assets/ds_image/--/revision/--/default/train/0/image/image.jpg', + 'src': 'http://localhost/cached-assets/ds_image/--/{dataset_git_revision}/--/default/train/0/image/image.jpg', ? +++++++++++++ + 'width': 640, }, }, 'row_idx': 0, 'truncated_cells': [], }, ] ```
closed
2024-08-12T11:43:04Z
2024-08-16T09:13:11Z
2024-08-16T09:13:09Z
albertvillanova
2,460,700,710
Update aiohttp 3.10.2 min version to fix vulnerability
Update aiohttp 3.10.2 min version to fix vulnerability. Fix 12 dependabot alerts.
closed
2024-08-12T11:10:10Z
2024-08-20T06:19:05Z
2024-08-20T06:19:04Z
albertvillanova
2,459,789,625
Re-enable Polars as a supported library for jsonl with glob paths
Ref https://github.com/huggingface/dataset-viewer/pull/3006 Issue fixed by https://github.com/pola-rs/polars/pull/17958
open
2024-08-11T22:32:25Z
2024-08-15T01:06:50Z
null
nameexhaustion
2,457,838,489
fix docstring
null
closed
2024-08-09T12:43:19Z
2024-08-09T12:43:28Z
2024-08-09T12:43:26Z
severo
2,457,830,514
obsolete
null
closed
2024-08-09T12:38:53Z
2024-08-09T12:39:05Z
2024-08-09T12:39:04Z
severo
2,444,676,341
Remove extra `label` column
In example dataset https://huggingface.co/datasets/datasets-examples/doc-audio-4, we have an "unexpected" label column with only `null` values. <img width="676" alt="Capture d’écran 2024-08-02 à 12 33 10" src="https://github.com/user-attachments/assets/e8e8a4b9-4681-4bb0-b8b4-31e4b0373a0d"> I think it's due to a "collision" between the heuristics that define splits and/or classes based on the directories. There is a `drop_labels=True` option in the datasets library, if it helps. Ideally, in this case, we should have two splits (train and test), and no additional `label` column. I think the issue also exists with image datasets.
open
2024-08-02T10:37:48Z
2024-08-02T10:38:19Z
null
severo
2,442,778,523
Add links to the Hub docs
fixes #2845
closed
2024-08-01T15:35:32Z
2024-08-01T21:50:02Z
2024-08-01T21:50:00Z
severo
2,442,267,210
Update README.md
fixes #2668
closed
2024-08-01T11:58:35Z
2024-08-01T15:11:02Z
2024-08-01T15:11:00Z
severo
2,442,175,079
Replace `DatasetGenerationError` with the underlying error?
50K cache entries (0.5%) have the `DatasetGenerationError`. It would be better to show the underlying error, and help the user debug their data files.
closed
2024-08-01T11:10:29Z
2024-08-22T14:38:13Z
2024-08-22T14:38:13Z
severo
2,442,173,496
Replace ConfigNamesError with the underlying error
100K cache entries (1%) have the `ConfigNamesError`. It would be better to show the underlying error, and help the user debug their data files.
closed
2024-08-01T11:09:46Z
2024-08-23T09:38:17Z
2024-08-22T14:34:51Z
severo
2,441,056,538
add a field to the index + it should fix the deployment issue
deployment (migration script) error: ``` INFO: 2024-07-31 22:33:14,880 - root - Start migrations INFO: 2024-07-31 22:33:14,926 - root - 72 migrations have already been applied. They will be skipped. INFO: 2024-07-31 22:33:14,927 - root - Migrate 20240731143600: add to the migrations collection INFO: 2024-07-31 22:33:14,931 - root - Migrate 20240731143600: apply INFO: 2024-07-31 22:33:14,931 - root - If missing, add the 'dataset_status' field with the default value 'normal' to the jobs metrics INFO: 2024-07-31 22:33:14,941 - root - Migrate 20240731143600: validate INFO: 2024-07-31 22:33:14,941 - root - Ensure that a random selection of jobs metrics have the 'dataset_status' field ERROR: 2024-07-31 22:33:14,943 - root - Migration failed: An existing index has the same name as the requested index. When index names are not specified, they are auto generated and can cause conflicts. Please refer to our documentation. Requested index: { v: 2, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false }, existing index: { v: 2, unique: true, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false, sparse: false }, full error: {'ok': 0.0, 'errmsg': 'An existing index has the same name as the requested index. When index names are not specified, they are auto generated and can cause conflicts. Please refer to our documentation. Requested index: { v: 2, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false }, existing index: { v: 2, unique: true, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false, sparse: false }', 'code': 86, 'codeName': 'IndexKeySpecsConflict', '$clusterTime': {'clusterTime': Timestamp(1722465194, 566), 'signature': {'hash': b'\xa0C\xe2\xd5;Z\xfd:8\x8d\xfe\x0fX\x1a\xbd\x87\x94D\xe3\xd4', 'keyId': 7345684769667022921}}, 'operationTime': Timestamp(1722465194, 566)} INFO: 2024-07-31 22:33:14,943 - root - Start rollback INFO: 2024-07-31 22:33:14,943 - root - Rollback 20240731143600: roll back INFO: 2024-07-31 22:33:14,943 - root - Remove the 'dataset_status' field from all the jobs metrics INFO: 2024-07-31 22:33:14,952 - root - Rollback 20240731143600: removed from the migrations collection INFO: 2024-07-31 22:33:14,956 - root - Rollback 20240731143600: done INFO: 2024-07-31 22:33:14,956 - root - All executed migrations have been rolled back ERROR: 2024-07-31 22:33:14,959 - root - An existing index has the same name as the requested index. When index names are not specified, they are auto generated and can cause conflicts. Please refer to our documentation. Requested index: { v: 2, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false }, existing index: { v: 2, unique: true, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false, sparse: false }, full error: {'ok': 0.0, 'errmsg': 'An existing index has the same name as the requested index. When index names are not specified, they are auto generated and can cause conflicts. Please refer to our documentation. Requested index: { v: 2, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false }, existing index: { v: 2, unique: true, key: { job_type: 1, status: 1 }, name: "job_type_1_status_1", background: false, sparse: false }', 'code': 86, 'codeName': 'IndexKeySpecsConflict', '$clusterTime': {'clusterTime': Timestamp(1722465194, 566), 'signature': {'hash': b'\xa0C\xe2\xd5;Z\xfd:8\x8d\xfe\x0fX\x1a\xbd\x87\x94D\xe3\xd4', 'keyId': 7345684769667022921}}, 'operationTime': Timestamp(1722465194, 566)} ```
closed
2024-07-31T22:36:47Z
2024-07-31T22:36:59Z
2024-07-31T22:36:57Z
severo
2,439,978,856
add dataset_status='normal|blocked' to job metrics
It will help show more precise graphs in Grafana, separating the jobs for blocked datasets from the rest of the jobs. Two notes: - the lookup for jobs in blocked datasets is maybe not optimal - we only take this into account in the cron job that recomputes the metrics, not in every job creation/update/deletion (ie: every 10 minutes). I think it's good enough
closed
2024-07-31T12:48:01Z
2024-07-31T15:12:10Z
2024-07-31T15:12:08Z
severo
2,439,957,953
Stats for datetimes
null
open
2024-07-31T12:38:13Z
2024-08-12T11:26:03Z
null
polinaeterna
2,439,654,893
fix 'FileNotFoundError' in polars detection function
We simply don't put Polars as a supported library when we're not able to detect if it's JSON or JSONL. TODO: transform `**/*` to a list of files? In the case of `datasets/tcor0005/langchain-docs-400-chunksize`, it's a single file --- The error we fix: ``` tests/job_runners/dataset/test_compatible_libraries.py:564: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ src/worker/job_runners/dataset/compatible_libraries.py:700: in get_polars_compatible_library is_json_lines = ".jsonl" in first_file or HfFileSystem(token=hf_token).open(first_file, "r").read(1) != "[" .venv/lib/python3.9/site-packages/fsspec/spec.py:1281: in open self.open( .venv/lib/python3.9/site-packages/fsspec/spec.py:1293: in open f = self._open( .venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py:236: in _open return HfFileSystemFile(self, path, mode=mode, revision=revision, block_size=block_size, **kwargs) .venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py:687: in __init__ self.details = fs.info(self.resolved_path.unresolve(), expand_info=False) .venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py:520: in info self.ls(parent_path, expand_info=False) .venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py:294: in ls _raise_file_not_found(path, None) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ path = 'datasets/tcor0005/langchain-docs-400-chunksize/**', err = None def _raise_file_not_found(path: str, err: Optional[Exception]) -> NoReturn: msg = path if isinstance(err, RepositoryNotFoundError): msg = f"{path} (repository not found)" elif isinstance(err, RevisionNotFoundError): msg = f"{path} (revision not found)" elif isinstance(err, HFValidationError): msg = f"{path} (invalid repository id)" > raise FileNotFoundError(msg) from err E FileNotFoundError: datasets/tcor0005/langchain-docs-400-chunksize/** .venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py:868: FileNotFoundError ```
closed
2024-07-31T10:05:52Z
2024-07-31T11:44:47Z
2024-07-31T10:40:12Z
severo
2,438,220,970
remove code for 'manual download' script datasets
fixes #2478
closed
2024-07-30T16:23:17Z
2024-07-31T08:23:01Z
2024-07-31T08:22:59Z
severo
2,438,216,759
Remove allow list for script-based datasets
https://github.com/huggingface/dataset-viewer/blob/5c5be128c205ae5c1f55440178497db564d81868/chart/values.yaml#L83-L86 Then remove the code that allow script-based datasets in this repo + exception `DatasetModuleNotInstalledError`. Related to #2478.
closed
2024-07-30T16:20:49Z
2024-08-26T08:10:05Z
2024-08-26T08:10:05Z
severo
2,437,703,019
Fix json compatible libraries
json datasets currently don't have a "Use this dataset" button because the compatible-libraries job fails note that this doesn't impact datasets with .jsonl files
closed
2024-07-30T12:19:46Z
2024-07-30T12:28:51Z
2024-07-30T12:28:50Z
lhoestq
2,432,448,352
increment version to recompute dataset-compatible-libraries
null
closed
2024-07-26T15:12:47Z
2024-07-26T15:13:01Z
2024-07-26T15:12:59Z
severo
2,430,004,954
otherwise we have dataset_name: 'parquet'
Not sure what to with the existing entries (we don't want to recompute all the entries). Alternative: remove the field?
closed
2024-07-25T13:39:49Z
2024-07-25T16:01:38Z
2024-07-25T16:01:37Z
severo
2,429,632,916
publish dataset-hub-cache at /hub-cache?dataset=...
useful for local Hub development
closed
2024-07-25T10:40:36Z
2024-07-25T10:43:41Z
2024-07-25T10:43:39Z
severo
2,429,466,797
Catch another UnexpectedException
https://discuss.huggingface.co/t/strange-problems-with-datasets-server/43871/8 https://huggingface.co/datasets/rippleripple/ProbMed The error was in step `config-parquet-and-info`: ``` { "error": "(ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed connection without response')), '(Request ID: 2529cb6c-46ac-4c17-8e7a-44488f47d04e)')", "cause_exception": "ConnectionError", "cause_message": "(ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed connection without response')), '(Request ID: 2529cb6c-46ac-4c17-8e7a-44488f47d04e)')", "cause_traceback": [ "Traceback (most recent call last):\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/connectionpool.py\", line 715, in urlopen\n httplib_response = self._make_request(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/connectionpool.py\", line 467, in _make_request\n six.raise_from(e, None)\n", " File \"<string>\", line 3, in raise_from\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/connectionpool.py\", line 462, in _make_request\n httplib_response = conn.getresponse()\n", " File \"/usr/local/lib/python3.9/http/client.py\", line 1377, in getresponse\n response.begin()\n", " File \"/usr/local/lib/python3.9/http/client.py\", line 320, in begin\n version, status, reason = self._read_status()\n", " File \"/usr/local/lib/python3.9/http/client.py\", line 289, in _read_status\n raise RemoteDisconnected(\"Remote end closed connection without\"\n", "http.client.RemoteDisconnected: Remote end closed connection without response\n", "\nDuring handling of the above exception, another exception occurred:\n\n", "Traceback (most recent call last):\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/requests/adapters.py\", line 589, in send\n resp = conn.urlopen(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/connectionpool.py\", line 801, in urlopen\n retries = retries.increment(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/util/retry.py\", line 552, in increment\n raise six.reraise(type(error), error, _stacktrace)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/packages/six.py\", line 769, in reraise\n raise value.with_traceback(tb)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/connectionpool.py\", line 715, in urlopen\n httplib_response = self._make_request(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/connectionpool.py\", line 467, in _make_request\n six.raise_from(e, None)\n", " File \"<string>\", line 3, in raise_from\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/urllib3/connectionpool.py\", line 462, in _make_request\n httplib_response = conn.getresponse()\n", " File \"/usr/local/lib/python3.9/http/client.py\", line 1377, in getresponse\n response.begin()\n", " File \"/usr/local/lib/python3.9/http/client.py\", line 320, in begin\n version, status, reason = self._read_status()\n", " File \"/usr/local/lib/python3.9/http/client.py\", line 289, in _read_status\n raise RemoteDisconnected(\"Remote end closed connection without\"\n", "urllib3.exceptions.ProtocolError: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response'))\n", "\nDuring handling of the above exception, another exception occurred:\n\n", "Traceback (most recent call last):\n", " File \"/src/services/worker/src/worker/job_manager.py\", line 127, in process\n job_result = self.job_runner.compute()\n", " File \"/src/services/worker/src/worker/job_runners/config/parquet_and_info.py\", line 1671, in compute\n compute_config_parquet_and_info_response(\n", " File \"/src/services/worker/src/worker/job_runners/config/parquet_and_info.py\", line 1577, in compute_config_parquet_and_info_response\n parquet_operations = convert_to_parquet(builder)\n", " File \"/src/services/worker/src/worker/job_runners/config/parquet_and_info.py\", line 1191, in convert_to_parquet\n builder.download_and_prepare(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py\", line 1027, in download_and_prepare\n self._download_and_prepare(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py\", line 1789, in _download_and_prepare\n super()._download_and_prepare(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py\", line 1100, in _download_and_prepare\n split_generators = self._split_generators(dl_manager, **split_generators_kwargs)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/folder_based_builder/folder_based_builder.py\", line 114, in _split_generators\n downloaded_files = dl_manager.download(files)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/download_manager.py\", line 257, in download\n downloaded_path_or_paths = map_nested(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py\", line 511, in map_nested\n mapped = [\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py\", line 512, in <listcomp>\n _single_map_nested((function, obj, batched, batch_size, types, None, True, None))\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py\", line 380, in _single_map_nested\n return [mapped_item for batch in iter_batched(data_struct, batch_size) for mapped_item in function(batch)]\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/py_utils.py\", line 380, in <listcomp>\n return [mapped_item for batch in iter_batched(data_struct, batch_size) for mapped_item in function(batch)]\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/download_manager.py\", line 300, in _download_batched\n return thread_map(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/tqdm/contrib/concurrent.py\", line 69, in thread_map\n return _executor_map(ThreadPoolExecutor, fn, *iterables, **tqdm_kwargs)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/tqdm/contrib/concurrent.py\", line 51, in _executor_map\n return list(tqdm_class(ex.map(fn, *iterables, chunksize=chunksize), **kwargs))\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/tqdm/std.py\", line 1181, in __iter__\n for obj in iterable:\n", " File \"/usr/local/lib/python3.9/concurrent/futures/_base.py\", line 609, in result_iterator\n yield fs.pop().result()\n", " File \"/usr/local/lib/python3.9/concurrent/futures/_base.py\", line 446, in result\n return self.__get_result()\n", " File \"/usr/local/lib/python3.9/concurrent/futures/_base.py\", line 391, in __get_result\n raise self._exception\n", " File \"/usr/local/lib/python3.9/concurrent/futures/thread.py\", line 58, in run\n result = self.fn(*self.args, **self.kwargs)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/download/download_manager.py\", line 323, in _download_single\n out = cached_path(url_or_filename, download_config=download_config)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/file_utils.py\", line 201, in cached_path\n output_path = get_from_cache(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/file_utils.py\", line 676, in get_from_cache\n fsspec_get(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/file_utils.py\", line 385, in fsspec_get\n fs.get_file(path, temp_file.name, callback=callback)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/hf_file_system.py\", line 636, in get_file\n http_get(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/file_download.py\", line 456, in http_get\n r = _request_wrapper(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/file_download.py\", line 392, in _request_wrapper\n response = get_session().request(method=method, url=url, **params)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py\", line 589, in request\n resp = self.send(prep, **send_kwargs)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py\", line 724, in send\n history = [resp for resp in gen]\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py\", line 724, in <listcomp>\n history = [resp for resp in gen]\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py\", line 265, in resolve_redirects\n resp = self.send(\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/requests/sessions.py\", line 703, in send\n r = adapter.send(request, **kwargs)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/huggingface_hub/utils/_http.py\", line 66, in send\n return super().send(request, *args, **kwargs)\n", " File \"/src/services/worker/.venv/lib/python3.9/site-packages/requests/adapters.py\", line 604, in send\n raise ConnectionError(err, request=request)\n", "requests.exceptions.ConnectionError: (ProtocolError('Connection aborted.', RemoteDisconnected('Remote end closed connection without response')), '(Request ID: 2529cb6c-46ac-4c17-8e7a-44488f47d04e)')\n" ] } ```
open
2024-07-25T09:20:17Z
2024-07-25T09:23:17Z
null
severo
2,428,053,115
bump croissant job version
recompute croissant files after https://github.com/huggingface/dataset-viewer/pull/2943 cc @marcenacp
closed
2024-07-24T16:51:21Z
2024-07-25T09:27:12Z
2024-07-25T09:27:11Z
lhoestq
2,427,760,387
Add Polars loading code
Hello from the Polars team! We've recently added support for scanning Hugging Face datasets, and to make it easier for users we're hoping that Polars code snippets could be added under the "Use this dataset" section on the dataset viewer webpage, next to the other libraries. Specifically, here's where we'd like to add Polars as an option: <img width="247" alt="image" src="https://github.com/user-attachments/assets/53868c96-7e70-4489-a785-2e4287a69f52">
closed
2024-07-24T14:34:14Z
2024-07-26T14:07:14Z
2024-07-26T13:42:24Z
nameexhaustion
2,425,473,680
Modalities not detected for some datasets using the Webdatasets format
I have found 2 examples of the modality detection code failing to recognize modalities in text and image datasets using the Webdataset format: * https://huggingface.co/datasets/ProGamerGov/synthetic-dataset-1m-dalle3-high-quality-captions * https://huggingface.co/datasets/CaptionEmporium/midjourney-niji-1m-llavanext I'm not sure where in the modality detection code that things are failing: https://github.com/huggingface/dataset-viewer/blob/main/services/worker/src/worker/job_runners/dataset/modalities.py
open
2024-07-23T15:15:38Z
2024-08-23T15:04:07Z
null
ProGamerGov
2,425,350,169
move the docs page
This change should move https://huggingface.co/docs/datasets-server to https://huggingface.co/docs/dataset-viewer. cc @mishig25: we also need to redirect https://huggingface.co/docs/datasets-server to https://huggingface.co/docs/dataset-viewer. Where do we do it?
closed
2024-07-23T14:22:25Z
2024-07-23T15:49:18Z
2024-07-23T15:49:16Z
severo
2,425,160,398
Compute leaks between splits?
See https://huggingface.co/blog/lbourdois/lle Also: should we find the duplicate rows?
open
2024-07-23T13:00:39Z
2024-07-23T15:19:13Z
null
severo
2,422,729,998
replace configuration with subset where appropriate
see also https://github.com/huggingface/hub-docs/pull/1347.
closed
2024-07-22T12:16:42Z
2024-07-22T14:11:36Z
2024-07-22T14:11:34Z
severo
2,422,651,235
Unescaped config names with special characters in the URL
When playing with mlcroissant, we observed the following issue: [bigcode/commitpackft](https://huggingface.co/datasets/bigcode/commitpackft) has both the configs `c` and `c#`. When going to https://huggingface.co/api/datasets/bigcode/commitpackft/parquet/c#/train/0.parquet, it lists https://huggingface.co/api/datasets/bigcode/commitpackft/parquet/c/train/0.parquet (instead of https://huggingface.co/api/datasets/bigcode/commitpackft/parquet/c%23/train/0.parquet). Should dataset names / config names be escaped in the URLs? cc @severo @lhoestq
open
2024-07-22T11:37:09Z
2024-07-29T08:30:36Z
null
marcenacp
2,422,430,142
Create a special column type when it contains PDF bytes or PDF URL
In that case, we would generate an image (thumbnail of the first page), stored as an asset, to populate /first-rows and /rows and display in the dataset viewer. asked internally on Slack: https://huggingface.slack.com/archives/C064HCHEJ2H/p1721215883166569 cc @Pleias
open
2024-07-22T09:45:00Z
2024-07-22T10:49:42Z
null
severo
2,418,023,119
Inquiry about the Frontend Technologies Used in Dataset Viewer
Hi team, I'm currently researching the Dataset Viewer project and would like to understand more about the frontend technologies used. Specifically, I'm interested in knowing: 1. Which frontend framework is being utilized (e.g., React, Vue, etc.)? 2. Are there any specific libraries or components being used for UI (e.g., Material-UI, Ant Design)? 3. Any other notable frontend tools or technologies that are part of this project? Your assistance in providing these details would be greatly appreciated. Thank you for your time and effort! Best regards
closed
2024-07-19T06:06:20Z
2024-08-19T13:42:45Z
2024-08-19T13:42:44Z
jacob-rodgers-max
2,410,308,970
Update setuptools to 70.3.0 to fix vulnerability
Update setuptools to 70.3.0 to fix vulnerability. It will close 1 Dependabot alert.
closed
2024-07-16T06:20:29Z
2024-07-17T09:18:47Z
2024-07-17T09:18:45Z
albertvillanova
2,405,926,121
Count image urls as image modality
close https://github.com/huggingface/dataset-viewer/issues/2970 for images (audio can be taken care of later)
closed
2024-07-12T16:07:30Z
2024-07-15T16:48:12Z
2024-07-15T16:48:10Z
lhoestq
2,405,821,839
fix text and arrow format
fix "0 dataset in text format" reported at https://x.com/jedmaczan/status/1809280782588158452 (+ detect arrow format as well) ![image](https://github.com/user-attachments/assets/fa328156-7650-4aa2-a828-41d5ab13c37c) I'll re-launch the job for text datasets
closed
2024-07-12T15:05:54Z
2024-07-16T09:32:17Z
2024-07-12T15:06:45Z
lhoestq
2,405,408,924
Include code snippets for other libraries?
For example, in https://github.com/huggingface/huggingface.js/pull/797, we add `distilabel`, `fiftyone` and `argilla` to the list of libraries the Hub knows. However, the aim is only to handle the user-defined tags better, not to show code snippets. In this issue, I propose to discuss if we should expand the list of dataset libraries for which we show code snippets. For now, we support pandas, HF datasets, webdatasets, mlcroissant and dask. We already mentioned polars as a potential new lib, I think. Maybe duckdb too?
open
2024-07-12T11:57:43Z
2024-07-12T14:39:59Z
null
severo
2,403,541,322
Skip smart update when language tag is updated
null
closed
2024-07-11T15:56:59Z
2024-07-11T16:11:39Z
2024-07-11T16:11:37Z
AndreaFrancis
2,399,927,837
Update zipp to 3.19.2 to fix vulnerability
Update zipp to 3.19.2 to fix vulnerability. It will close 1 Dependabot alert.
closed
2024-07-10T07:04:00Z
2024-07-10T07:06:44Z
2024-07-10T07:06:42Z
albertvillanova
2,398,580,183
Ignore transformed data in split-duckdb-index when stats throws error
Part of https://github.com/huggingface/dataset-viewer/issues/1443: Currently, we have 3054 cache entries with UnexpectedError because stats fail for some columns like: ``` details: { error: 'expected str, bytes or os.PathLike object, not BytesIO', cause_exception: 'TypeError', cause_message: 'expected str, bytes or os.PathLike object, not BytesIO', cause_traceback: [ 'Traceback (most recent call last):\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/librosa/core/audio.py", line 803, in get_duration\n' + ' return sf.info(path).duration # type: ignore\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/soundfile.py", line 467, in info\n' + ' return _SoundFileInfo(file, verbose)\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/soundfile.py", line 412, in __init__\n' + ' with SoundFile(file) as f:\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/soundfile.py", line 658, in __init__\n' + ' self._file = self._open(file, mode_int, closefd)\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/soundfile.py", line 1216, in _open\n' + ' raise LibsndfileError(err, prefix="Error opening {0!r}: ".format(self.name))\n', 'soundfile.LibsndfileError: Error opening <_io.BytesIO object at 0x7f6f99e39db0>: Format not recognised.\n', '\n' + 'During handling of the above exception, another exception occurred:\n' + '\n', 'Traceback (most recent call last):\n', ' File "/src/services/worker/src/worker/job_manager.py", line 125, in process\n' + ' job_result = self.job_runner.compute()\n', ' File "/src/services/worker/src/worker/job_runners/split/duckdb_index.py", line 448, in compute\n' + ' compute_split_duckdb_index_response(\n', ' File "/src/services/worker/src/worker/job_runners/split/duckdb_index.py", line 278, in compute_split_duckdb_index_response\n' + ' transformed_df = compute_transformed_data(split_parquet_directory, features)\n', ' File "/src/services/worker/src/worker/job_runners/split/duckdb_index.py", line 161, in compute_transformed_data\n' + ' transformed_df = compute_audio_duration_column(parquet_directory, feature_name, transformed_df)\n', ' File "/src/services/worker/src/worker/job_runners/split/duckdb_index.py", line 124, in compute_audio_duration_column\n' + ' durations = AudioColumn.compute_transformed_data(parquet_directory, column_name, AudioColumn.get_duration)\n', ' File "/src/services/worker/src/worker/statistics_utils.py", line 606, in compute_transformed_data\n' + ' shard_transformed_values = thread_map(\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/tqdm/contrib/concurrent.py", line 69, in thread_map\n' + ' return _executor_map(ThreadPoolExecutor, fn, *iterables, **tqdm_kwargs)\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/tqdm/contrib/concurrent.py", line 51, in _executor_map\n' + ' return list(tqdm_class(ex.map(fn, *iterables, chunksize=chunksize), **kwargs))\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/tqdm/std.py", line 1181, in __iter__\n' + ' for obj in iterable:\n', ' File "/usr/local/lib/python3.9/concurrent/futures/_base.py", line 609, in result_iterator\n' + ' yield fs.pop().result()\n', ' File "/usr/local/lib/python3.9/concurrent/futures/_base.py", line 446, in result\n' + ' return self.__get_result()\n', ' File "/usr/local/lib/python3.9/concurrent/futures/_base.py", line 391, in __get_result\n' + ' raise self._exception\n', ' File "/usr/local/lib/python3.9/concurrent/futures/thread.py", line 58, in run\n' + ' result = self.fn(*self.args, **self.kwargs)\n', ' File "/src/services/worker/src/worker/statistics_utils.py", line 672, in get_duration\n' + ' return librosa.get_duration(path=f) # type: ignore # expects PathLike but BytesIO also works\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/librosa/core/audio.py", line 812, in get_duration\n' + ' with audioread.audio_open(path) as fdesc:\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/audioread/__init__.py", line 127, in audio_open\n' + ' return BackendClass(path)\n', ' File "/src/services/worker/.venv/lib/python3.9/site-packages/audioread/rawread.py", line 59, in __init__\n' + " self._fh = open(filename, 'rb')\n", 'TypeError: expected str, bytes or os.PathLike object, not BytesIO\n' ] }, ``` Query to identify the affected records: ``` Atlas atlas-x5jgb3-shard-0 [primary] datasets_server_cache> db.cachedResponsesBlue.aggregate([{$match: {error_code: "UnexpectedError", "details.cause_exception":"TypeError", kind:"split-duckdb-index", "details.copied_from_artifact":{$exists:false}}},{$group: {_id: {error: "$details.error"}, count: {$sum: 1}}},{$sort: {count: -1}}]) [ { _id: { error: 'expected str, bytes or os.PathLike object, not BytesIO' }, count: 3054 }, ... ] ``` If transformed values fail, indexing should continue without fail. It will just not add the transformed columns, but that is not a problem since they won't be visible in the UI either because split-descriptive-statistics fails.
closed
2024-07-09T15:49:57Z
2024-07-11T16:11:56Z
2024-07-11T16:11:54Z
AndreaFrancis
2,398,326,202
Docs: mention partial parquet conversion for big row groups
asked at https://huggingface.slack.com/archives/C0709MK0PUN/p1720433431019569?thread_ts=1720210655.175099&cid=C0709MK0PUN
closed
2024-07-09T14:04:46Z
2024-07-12T09:43:23Z
2024-07-09T16:59:14Z
lhoestq
2,398,269,890
run presidio on UpVoteWeb
asked at https://huggingface.slack.com/archives/C01GSG1QFPF/p1720525385703189?thread_ts=1720079065.204769&cid=C01GSG1QFPF
closed
2024-07-09T13:42:18Z
2024-07-29T10:26:21Z
2024-07-25T12:02:13Z
lhoestq
2,397,691,922
remove unused actions
null
closed
2024-07-09T09:45:05Z
2024-07-09T12:40:14Z
2024-07-09T12:40:12Z
severo
2,397,687,415
Revert "the repository will be moved (#2977)"
This reverts commit 858ec8c147959cd426c2c662a2ff067dcc7149c3.
closed
2024-07-09T09:43:02Z
2024-07-09T09:43:15Z
2024-07-09T09:43:13Z
severo
2,397,673,196
remove the 'tags' field from the Hub cache response
Note that I don't migrate to remove the field from the cache (we could do it, but not sure it's worth the pain) Better wait for https://github.com/huggingface-internal/moon-landing/pull/10591 (internal) to be deployed on the Hub before merging.
closed
2024-07-09T09:36:33Z
2024-07-10T08:50:53Z
2024-07-10T08:50:51Z
severo
2,397,652,674
the repository will be moved
null
closed
2024-07-09T09:28:13Z
2024-07-09T09:37:00Z
2024-07-09T09:36:57Z
severo
2,395,588,139
Update statistics.md: Add the length calculation statement.
specify that the "string_text " length is calculated by its character number.
closed
2024-07-08T13:03:13Z
2024-07-25T07:56:43Z
2024-07-25T07:56:43Z
Siki-cloud
2,395,165,781
Upgrade to mongo v7?
null
open
2024-07-08T09:47:17Z
2024-07-08T09:47:31Z
null
severo
2,394,700,278
Update certifi to 2024.7.4 to fix vulnerability
Update certifi to 2024.7.4 to fix vulnerability. It will close 14 Dependabot alerts.
closed
2024-07-08T05:53:11Z
2024-07-08T07:00:12Z
2024-07-08T07:00:10Z
albertvillanova
2,393,271,806
Bump certifi from 2023.7.22 to 2024.7.4 in /docs in the pip group across 1 directory
Bumps the pip group with 1 update in the /docs directory: [certifi](https://github.com/certifi/python-certifi). Updates `certifi` from 2023.7.22 to 2024.7.4 <details> <summary>Commits</summary> <ul> <li><a href="https://github.com/certifi/python-certifi/commit/bd8153872e9c6fc98f4023df9c2deaffea2fa463"><code>bd81538</code></a> 2024.07.04 (<a href="https://redirect.github.com/certifi/python-certifi/issues/295">#295</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/06a2cbf21f345563dde6c28b60e29d57e9b210b3"><code>06a2cbf</code></a> Bump peter-evans/create-pull-request from 6.0.5 to 6.1.0 (<a href="https://redirect.github.com/certifi/python-certifi/issues/294">#294</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/13bba02b72bac97c432c277158bc04b4d2a6bc23"><code>13bba02</code></a> Bump actions/checkout from 4.1.6 to 4.1.7 (<a href="https://redirect.github.com/certifi/python-certifi/issues/293">#293</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/e8abcd0e62b334c164b95d49fcabdc9ecbca0554"><code>e8abcd0</code></a> Bump pypa/gh-action-pypi-publish from 1.8.14 to 1.9.0 (<a href="https://redirect.github.com/certifi/python-certifi/issues/292">#292</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/124f4adf171e15cd9a91a8b6e0325ecc97be8fe1"><code>124f4ad</code></a> 2024.06.02 (<a href="https://redirect.github.com/certifi/python-certifi/issues/291">#291</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/c2196ce5d6ee675b27755a19948480a7823e2c6a"><code>c2196ce</code></a> --- (<a href="https://redirect.github.com/certifi/python-certifi/issues/290">#290</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/fefdeec7588ff1c05214b85a552afcad5fdb51b2"><code>fefdeec</code></a> Bump actions/checkout from 4.1.4 to 4.1.5 (<a href="https://redirect.github.com/certifi/python-certifi/issues/289">#289</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/3c5fb1560b826a7f83f1f9750173ff766492c9cf"><code>3c5fb15</code></a> Bump actions/download-artifact from 4.1.6 to 4.1.7 (<a href="https://redirect.github.com/certifi/python-certifi/issues/286">#286</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/4a9569a3eb58db8548536fc16c5c5c7af946a5b1"><code>4a9569a</code></a> Bump actions/checkout from 4.1.2 to 4.1.4 (<a href="https://redirect.github.com/certifi/python-certifi/issues/287">#287</a>)</li> <li><a href="https://github.com/certifi/python-certifi/commit/1fc808626a895a916b1e4c2b63abae6c5eafdbe3"><code>1fc8086</code></a> Bump peter-evans/create-pull-request from 6.0.4 to 6.0.5 (<a href="https://redirect.github.com/certifi/python-certifi/issues/288">#288</a>)</li> <li>Additional commits viewable in <a href="https://github.com/certifi/python-certifi/compare/2023.07.22...2024.07.04">compare view</a></li> </ul> </details> <br /> [![Dependabot compatibility score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=certifi&package-manager=pip&previous-version=2023.7.22&new-version=2024.7.4)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores) You can trigger a rebase of this PR by commenting `@dependabot rebase`. [//]: # (dependabot-automerge-start) [//]: # (dependabot-automerge-end) --- <details> <summary>Dependabot commands and options</summary> <br /> You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show <dependency name> ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore <dependency name> major version` will close this group update PR and stop Dependabot creating any more for the specific dependency's major version (unless you unignore this specific dependency's major version or upgrade to it yourself) - `@dependabot ignore <dependency name> minor version` will close this group update PR and stop Dependabot creating any more for the specific dependency's minor version (unless you unignore this specific dependency's minor version or upgrade to it yourself) - `@dependabot ignore <dependency name>` will close this group update PR and stop Dependabot creating any more for the specific dependency (unless you unignore this specific dependency or upgrade to it yourself) - `@dependabot unignore <dependency name>` will remove all of the ignore conditions of the specified dependency - `@dependabot unignore <dependency name> <ignore condition>` will remove the ignore condition of the specified dependency and ignore conditions You can disable automated security fix PRs for this repo from the [Security Alerts page](https://github.com/huggingface/dataset-viewer/network/alerts). </details> > **Note** > Automatic rebases have been disabled on this pull request as it has been open for over 30 days.
closed
2024-07-06T00:37:38Z
2024-08-26T08:27:44Z
2024-08-26T08:27:42Z
dependabot[bot]
2,392,849,689
Decrease worker size metrics when dataset is blocked
Part of https://github.com/huggingface/dataset-viewer/issues/2945 Context: https://github.com/huggingface/dataset-viewer/issues/2945#issuecomment-2192170023 - Decrease jobs when a dataset is blocked.
closed
2024-07-05T16:00:41Z
2024-07-08T15:26:33Z
2024-07-08T15:26:31Z
AndreaFrancis

This dataset contains the GitHub issues for the Dataset Viewer repository.

image/png

https://github.com/huggingface/dataset-viewer

How to get the data:

import requests
import pandas as pd
from datasets import Dataset

def fetch_github_issues(repo_owner, repo_name, token=None):
    url = f"https://api.github.com/repos/{repo_owner}/{repo_name}/issues"
    headers = {'Authorization': f'token {token}'} if token else {}
    issues = []
    page = 1
    
    while True:
        response = requests.get(url, headers=headers, params={'state': 'all', 'page': page, 'per_page': 100})
        if response.status_code != 200:
            raise Exception(f"Failed to fetch issues: {response.status_code}")
        
        page_issues = response.json()
        if not page_issues:
            break
            
        issues.extend(page_issues)
        page += 1

    issues_df = pd.DataFrame(issues)
    
    return issues_df

def prepare_hf_dataset(issues_df):
    issues_df = issues_df[['id', 'title', 'body', 'state', 'created_at', 'updated_at', 'closed_at', 'user']]

    issues_df['user'] = issues_df['user'].apply(lambda x: x['login'] if isinstance(x, dict) else None)

    hf_dataset = Dataset.from_pandas(issues_df)
    
    return hf_dataset

def main():
    # Replace with your repository details
    repo_owner = "huggingface"
    repo_name = "dataset-viewer"
    github_token = None

    issues_df = fetch_github_issues(repo_owner, repo_name, token=github_token)
    print("Fetched issues:", len(issues_df))

    hf_dataset = prepare_hf_dataset(issues_df)
    print("Hugging Face Dataset:", hf_dataset)

    # hf_dataset.push_to_hub("your_dataset_name")

if __name__ == "__main__":
    main()
Downloads last month
1
Edit dataset card