initial
Browse files- common_crawl.py +3 -3
common_crawl.py
CHANGED
@@ -5,7 +5,7 @@ import requests
|
|
5 |
import urllib.request
|
6 |
from typing import Optional
|
7 |
from warcio.archiveiterator import ArchiveIterator
|
8 |
-
|
9 |
from datasets import (
|
10 |
BuilderConfig, Version, GeneratorBasedBuilder, DownloadManager, SplitGenerator, Split, Features, Value, DatasetInfo
|
11 |
)
|
@@ -40,7 +40,7 @@ def _decode_text(_bytes):
|
|
40 |
|
41 |
class CommonCrawlConfig(BuilderConfig):
|
42 |
def __init__(self, features, data_url, citation, url, version, label_classes: Optional[tuple] = None, **kwargs):
|
43 |
-
super().__init__(version=version, **kwargs)
|
44 |
self.features = features
|
45 |
self.label_classes = label_classes
|
46 |
self.data_url = data_url
|
@@ -56,7 +56,7 @@ class CommonCrawl(GeneratorBasedBuilder):
|
|
56 |
]
|
57 |
|
58 |
def _info(self):
|
59 |
-
return DatasetInfo(
|
60 |
description="",
|
61 |
features=Features(
|
62 |
{
|
|
|
5 |
import urllib.request
|
6 |
from typing import Optional
|
7 |
from warcio.archiveiterator import ArchiveIterator
|
8 |
+
import datasets
|
9 |
from datasets import (
|
10 |
BuilderConfig, Version, GeneratorBasedBuilder, DownloadManager, SplitGenerator, Split, Features, Value, DatasetInfo
|
11 |
)
|
|
|
40 |
|
41 |
class CommonCrawlConfig(BuilderConfig):
|
42 |
def __init__(self, features, data_url, citation, url, version, label_classes: Optional[tuple] = None, **kwargs):
|
43 |
+
super(CommonCrawlConfig, self).__init__(version=version, **kwargs)
|
44 |
self.features = features
|
45 |
self.label_classes = label_classes
|
46 |
self.data_url = data_url
|
|
|
56 |
]
|
57 |
|
58 |
def _info(self):
|
59 |
+
return datasets.DatasetInfo(
|
60 |
description="",
|
61 |
features=Features(
|
62 |
{
|