There are two ways to upload local files into Avala: the Add Dataset wizard in the web app (drag-and-drop) and theDocumentation Index
Fetch the complete documentation index at: https://avala.ai/docs/llms.txt
Use this file to discover all available pages before exploring further.
avala datasets upload CLI command. Both write to the same managed storage, both enforce the same per-user 10 GB cap, and both create a normal Avala dataset that you can label, share, and export.
Use the wizard for ad-hoc uploads of a few hundred files. Use the CLI when you have many files, large files, or want to script the upload.
Storage cap
Local uploads count against a 10 GB per-user quota. Each individual file is capped at 2 GB. Files larger than 2 GB are not currently supported via either the wizard or the CLI — split them, or use one of the cloud-storage data sources instead. You can see your current usage on the local-upload step of the Add Dataset wizard, or fetch it from the API:Web wizard (drag-and-drop)
- Open Mission Control and click Add Dataset.
- Step 1 — Common data: name the dataset, pick the data type (image, video, lidar, splat, MCAP, etc.), and set labels.
- Step 2 — Source: pick Local files.
- Step 3 — Upload: drag files into the drop zone or click to browse. The picker accepts only file types that match the data type from step 1 — for example, a Lidar dataset accepts
.pcd,.bin,.las,.laz,.plyand rejects everything else with an inline reason. Per-file progress shows on each row. - Click Submit once every selected file has finished uploading. (The wizard keeps the button disabled while any file is still pending, uploading, or failed — partial batches would create a dataset with only some of the files you picked.) The dataset is created in your account immediately and is ready to label.
CLI (avala datasets upload)
Install the SDK:
Common flags
| Flag | What it does |
|---|---|
--source <path> | Local file or directory containing files to upload. Required. |
--name <name> | Display name for the new dataset. Required. |
--slug <slug> | URL-friendly identifier for the dataset. Required. |
--data-type <type> | One of image, video, lidar, splat, mcap. Required. |
--visibility <vis> | private (default) or public. |
--dry-run | List files that would upload, with their sizes, and exit. |
--workers N | Parallel uploads (default 8). |
Quota errors
If your upload would push you over 10 GB, the API returns HTTP 413 and the CLI prints:After the upload
Either path produces a normal dataset. From there you can:- Define a project over it and assign annotation tasks.
- Trigger an export once labelling is done.
- Browse items via
client.datasets.get(<uid>).items()in the Python SDK.
provider_config handles access, so you don’t need separate AWS credentials to view them.