feat: Choose dataset with options
This commit is contained in:
parent
20bdd4f566
commit
81c767371e
5 changed files with 67 additions and 60 deletions
|
|
@ -10,11 +10,14 @@ Author: Tibo De Peuter
|
|||
class Dataset(TorchDataset, ABC):
|
||||
"""Abstract base class for datasets."""
|
||||
@abstractmethod
|
||||
def __init__(self, root: str, transform: Callable = None):
|
||||
def __init__(self, name: str, root: str | None, transform: Callable = None):
|
||||
"""
|
||||
:param root: Relative path to the dataset root directory
|
||||
"""
|
||||
self._root: str = join(curdir, 'data', root)
|
||||
if root is None:
|
||||
root = join(curdir, 'data')
|
||||
|
||||
self._root = join(root, name)
|
||||
self.transform = transform
|
||||
self.dataset = None
|
||||
|
||||
|
|
|
|||
|
|
@ -1,18 +1,20 @@
|
|||
from datasets import load_dataset
|
||||
from torch.utils.data import Dataset
|
||||
import torch
|
||||
from os.path import curdir, join
|
||||
from typing import Callable
|
||||
|
||||
import torch
|
||||
from datasets import load_dataset
|
||||
|
||||
from .Dataset import Dataset
|
||||
|
||||
|
||||
class EnWik9DataSet(Dataset):
|
||||
def __init__(self, root: str = "data", transform: Callable | None = None):
|
||||
super().__init__()
|
||||
self.transform = transform
|
||||
"""
|
||||
Hugging Face: https://huggingface.co/datasets/haukur/enwik9
|
||||
"""
|
||||
def __init__(self, root: str | None = None, transform: Callable | None = None):
|
||||
super().__init__('enwik9', root, transform)
|
||||
|
||||
# HuggingFace dataset: string text
|
||||
path = join(curdir, root)
|
||||
data = load_dataset("haukur/enwik9", cache_dir=path, split="train")
|
||||
data = load_dataset("haukur/enwik9", cache_dir=self.root, split="train")
|
||||
|
||||
# Extract raw text
|
||||
text = data["text"]
|
||||
|
|
@ -31,7 +33,7 @@ class EnWik9DataSet(Dataset):
|
|||
|
||||
def __getitem__(self, idx):
|
||||
# context window
|
||||
x = self.data[idx : idx + self.context_length]
|
||||
x = self.data[idx: idx + self.context_length]
|
||||
|
||||
# next byte target
|
||||
y = self.data[idx + self.context_length]
|
||||
|
|
@ -40,4 +42,3 @@ class EnWik9DataSet(Dataset):
|
|||
x = self.transform(x)
|
||||
|
||||
return x, y
|
||||
|
||||
|
|
|
|||
|
|
@ -1,21 +1,19 @@
|
|||
from typing import Callable
|
||||
|
||||
import torch
|
||||
from os.path import curdir, join
|
||||
from lorem.text import TextLorem
|
||||
|
||||
from .Dataset import Dataset
|
||||
|
||||
|
||||
class LoremIpsumDataset(Dataset):
|
||||
def __init__(self, root: str = "data", transform: Callable = None):
|
||||
super().__init__(root, transform)
|
||||
def __init__(self, root: str | None = None, transform: Callable = None, size: int = 512):
|
||||
super().__init__('lorem_ipsum', root, transform)
|
||||
|
||||
# Generate text and convert to bytes
|
||||
_lorem = TextLorem()
|
||||
_text = ' '.join(_lorem._word() for _ in range(512))
|
||||
_text = ' '.join(_lorem._word() for _ in range(size))
|
||||
|
||||
path = join(curdir, "data")
|
||||
self._root = path
|
||||
# Convert text to bytes (UTF-8 encoded)
|
||||
self.dataset = torch.tensor([ord(c) % 256 for c in list(_text)], dtype=torch.long)
|
||||
self.context_length = 128
|
||||
|
|
|
|||
|
|
@ -1,3 +1,8 @@
|
|||
from .Dataset import Dataset
|
||||
from .EnWik9 import EnWik9DataSet
|
||||
from .LoremIpsumDataset import LoremIpsumDataset
|
||||
from .Dataset import Dataset
|
||||
|
||||
dataset_called: dict[str, type[Dataset]] = {
|
||||
'enwik9': EnWik9DataSet,
|
||||
'lorem_ipsum': LoremIpsumDataset
|
||||
}
|
||||
|
|
|
|||
Reference in a new issue