You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -48,7 +48,7 @@ Additional dependencies for working with specific cloud services can be installe
48
48
-`roboflow`: Dependencies for downloading datasets from Roboflow
49
49
-`mlflow`: Dependencies for working with MLFlow
50
50
51
-
> \[!NOTE\]
51
+
> [!NOTE]
52
52
> If some of the additional dependencies are required but not installed (_e.g._ attempting to use Google Cloud Storage without installing the `gcs` extra), then the missing dependencies will be installed automatically.
Copy file name to clipboardExpand all lines: luxonis_ml/data/README.md
+14-14Lines changed: 14 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@
4
4
5
5
LuxonisML Data is a library for creating and interacting with datasets in the LuxonisDataFormat (LDF).
6
6
7
-
> \[!NOTE\]
7
+
> [!NOTE]
8
8
> For hands-on examples of how to prepare and iteract with `LuxonisML` datasets, check out [this guide](https://github.com/luxonis/ai-tutorials/tree/main/training#%EF%B8%8F-prepare-data-using-luxonis-ml).
9
9
10
10
The lifecycle of an LDF dataset is as follows:
@@ -77,7 +77,7 @@ You can create as many datasets as you want, each with a unique name.
77
77
78
78
Datasets can be stored locally or in one of the supported cloud storage providers.
79
79
80
-
> \[!NOTE\]
80
+
> [!NOTE]
81
81
> 📚 For a complete list of all parameters and methods of the `LuxonisDataset` class, see the [datasets README.md](datasets/README.md).
82
82
83
83
### Dataset Creation
@@ -92,10 +92,10 @@ dataset_name = "parking_lot"
92
92
dataset = LuxonisDataset(dataset_name)
93
93
```
94
94
95
-
> \[!NOTE\]
95
+
> [!NOTE]
96
96
> By default, the dataset will be created locally. For more information on creating a remote dataset, see [this section](datasets/README.md#creating-a-dataset-remotely).
97
97
98
-
> \[!NOTE\]
98
+
> [!NOTE]
99
99
> If there already is a dataset with the same name, it will be loaded instead of creating a new one.
100
100
> If you want to always create a new dataset, you can pass `delete_local=True` to the `LuxonisDataset` constructor.\
101
101
> For detailed information about how the luxonis-ml dataset is stored in both local and remote storage, please check the [datasets README.md](datasets/README.md#in-depth-explanation-of-luxonis-ml-dataset-storage)
@@ -254,7 +254,7 @@ Once you've defined your data source, pass it to the dataset's add method:
254
254
dataset.add(generator())
255
255
```
256
256
257
-
> \[!NOTE\]
257
+
> [!NOTE]
258
258
> The `add` method accepts any iterable, not only generators.
259
259
260
260
### Defining Splits
@@ -291,7 +291,7 @@ Calling `make_splits` with no arguments will default to an 80/10/10 split.
291
291
In order for splits to be created, there must be some new data in the dataset. If no new data were added, calling `make_splits` will raise an error.
292
292
If you wish to delete old splits and create new ones using all the data, pass `redefine_splits=True` to the method call.
293
293
294
-
> \[!NOTE\]
294
+
> [!NOTE]
295
295
> There are no restrictions on the split names,
296
296
> however for most cases one should stick to `"train"`, `"val"`, and `"test"`.
297
297
@@ -338,8 +338,8 @@ The available commands are:
338
338
-`luxonis_ml data ls` - lists all datasets
339
339
-`luxonis_ml data info <dataset_name>` - prints information about the dataset
340
340
-`luxonis_ml data inspect <dataset_name>` - renders the data in the dataset on screen using `cv2`
341
-
-`luxonis_ml data health <dataset_name>` - checks the health of the dataset and logs and renders dataset statistics
342
-
-`luxonis_ml data sanitize <dataset_name>` - removes duplicate files and duplicate annotations from the dataset
341
+
-`luxonis_ml data health <dataset_name>` - checks the health of the dataset and logs and renders dataset statistics
342
+
-`luxonis_ml data sanitize <dataset_name>` - removes duplicate files and duplicate annotations from the dataset
343
343
-`luxonis_ml data delete <dataset_name>` - deletes the dataset
344
344
-`luxonis_ml data export <dataset_name>` - exports the dataset to a chosen format and directory
345
345
-`luxonis_ml data push <dataset_name>` - pushes local dataset to remote storage
@@ -357,7 +357,7 @@ This guide covers the loading of datasets using the `LuxonisLoader` class.
357
357
358
358
The `LuxonisLoader` class can also take care of data augmentation, for more info see [Augmentation](#augmentation).
359
359
360
-
> \[!NOTE\]
360
+
> [!NOTE]
361
361
> 📚 For a complete list of all parameters of the `LuxonisLoader` class, see the [loaders README.md](loaders/README.md).
362
362
363
363
### Dataset Loading
@@ -609,7 +609,7 @@ The directory can also be a zip file containing the dataset.
609
609
The `task_name` argument can be specified as a single string or as a dictionary. If a string is provided, it will be used as the task name for all records.
610
610
Alternatively, you can provide a dictionary that maps class names to task names for better dataset organization. See the example below.
611
611
612
-
> \[!NOTE\]
612
+
> [!NOTE]
613
613
> 📚 For a complete list of all parameters of the `LuxonisParser` class, see the [parsers README.md](parsers/README.md).
614
614
615
615
```python
@@ -664,7 +664,7 @@ A single class label for the entire image.
664
664
}
665
665
```
666
666
667
-
>\[!NOTE\]
667
+
> [!NOTE]
668
668
> The `classification` task is always added to the dataset.
669
669
670
670
### Bounding Box
@@ -794,10 +794,10 @@ The `counts` field contains either a **compressed byte string** or an **uncompre
794
794
795
795
```
796
796
797
-
>\[!NOTE\]
797
+
> [!NOTE]
798
798
> The RLEformatisnot intended for regular use andis provided mainly to support datasets that may already be in this format.
799
799
800
-
>\[!NOTE\]
800
+
> [!NOTE]
801
801
> Masks provided as numpy arrays are converted to RLEformat internally.
802
802
803
803
### Array
@@ -993,7 +993,7 @@ The following example demonstrates a simple augmentation pipeline:
993
993
994
994
```
995
995
996
-
>\[!NOTE\]
996
+
> [!NOTE]
997
997
> The augmentations are **not** applied in order. Instead, an optimal order is determined based on the type of the augmentations to minimize the computational cost.
[A remote dataset functions similarly to a local dataset](#in-depth-explanation-of-luxonis-ml-dataset-storage). When a remote dataset is created, the same folder structure appears locally, and the equivalent structure appears in the cloud. The media folder is empty locally but is filled with images on the remote storage, where filenames become UUIDs with the appropriate suffix.
232
232
233
-
> \[!NOTE\]
233
+
> [!NOTE]
234
234
> **IMPORTANT:** Be careful when creating a remote dataset with the same name as an already existing local dataset, because corruption of datasets may occur if not handled properly.
235
235
>
236
236
> Use `delete_local=True` and `delete_remote=True` to create a new dataset (deleting both local and remote storage) before calling `dataset.add()`, or use `dataset.push_to_cloud()` to push an existing local dataset to the cloud. To append data to an existing dataset using `dataset.add()`, keep `delete_local=False` and `delete_remote=False`. In that case, ensure both local and remote datasets are healthy. If the local dataset might be corrupted but the remote version is healthy, use `delete_local=True` and `delete_remote=False` so that the local dataset is deleted, while the remote stays intact.
0 commit comments