# Dataset Structure The organization of this dataset was inspired by the WILDS benchmark and torchgeo python package. There are three overarching datasets: 'Points', 'Polygons' and 'Boxes' based on the annotation geometry. ## Data download ```python dataset = TreePointsDataset(download=True, root_dir=) ``` ## Dataset Access Methods ### DataFrame Interface Each dataset maintains a pandas DataFrame containing all annotations and metadata, accessible via the `df` attribute: ```python dataset = TreePointsDataset() dataset.df # Access full DataFrame with annotations and metadata ``` The DataFrame contains: - `filename`: Image filename - `x`, `y`: Point coordinates (TreePoints) - `xmin`, `ymin`, `xmax`, `ymax`: Box coordinates (TreeBoxes) - `source`: Original data source - `split`: Train/test/validation split - `source_id`: Numeric ID for each source - `filename_id`: Numeric ID for each image ### Lookup Dictionaries Helpful mappings between IDs and names: ```python # Map numeric source IDs to source names dataset._source_id_to_code # {0: 'source1', 1: 'source2', ...} # Map numeric filename IDs to actual filenames dataset._filename_id_to_code # {0: 'image1.jpg', 1: 'image2.jpg', ...} # Map filenames to annotation indices dataset._input_lookup # {'image1.jpg': array([0,1,2]), ...} ``` For example, if you want to get the annotations for a specific image, you can use the lookup dictionary: ``` from milliontrees import get_dataset dataset = get_dataset("TreePoints") indices = dataset._input_lookup["IMG_904.jpg"] coordinates = dataset._y_array[indices] ``` ## Dataloaders Part of the inspiration of this package is to keep most users from needing to interact with the filesystem. The dataloaders are built in, and for many applications, the user will never need to mess around with csv files or image paths. All annotations are pytorch dataloaders and can be iterated over. ```python for image, label, metadata in dataset: assert image.shape == (3, 100, 100) assert label.shape == (2,) assert len(metadata) == 2 ``` Users can select a subset of the dataset and optionally supply a torchvision transform: ```python transform = transforms.Compose([ transforms.Resize((448, 448)), transforms.RandomHorizontalFlip(p=0.5), transforms.ToTensor() ]) train_dataset = dataset.get_subset("train", transform=transform) for image, label, metadata in train_dataset: assert image.shape == (3, 448, 448) assert label.shape == (4,) assert len(metadata) == 2 ``` ## Split Schemes One of the great things about supplying data as dataloaders is easy access to different ways to combine datasets. The MillionTrees benchmark has multiple tasks, and each of these is a 'split_scheme', following the terminology from the WILDS benchmark. ```python dataset = TreePointsDataset(download=True, root_dir=, split_scheme="official") ``` This looks at the file official.csv and gets the 'split' column that designates which images are in train/test/val depending on the task. ## Dataset Splits The MillionTrees benchmark supports multiple dataset split schemes to accommodate various tasks: - **Official**: For each source, 80% of the data is used for training, and 20% is reserved for testing. - **Crossgeometry**: Combines Boxes and Points annotations to predict Polygons. - **Zeroshot**: Entire sources are held out for testing, simulating a zero-shot learning scenario. ## Annotation Geometry ### Boxes Boxes annotations are given as xmin, ymin, xmax, ymax coordinates relative to the image origin (top-left). ### Points Points annotations are given as x,y coordinate relative to the image origin. ### Polygons Polygon annotations are given as well-known text coordinates, e.g. "POLYGON((x1 y2, x2 y2, x3, y3 ...))" relative to the image origin.