clientMetadata
) to specify key frames, custom metadata, and custom embeddings. For more information go here or here for information on using the SDK.skip_duplicate_urls
is set to true
, all object URLs that exactly match existing images/videos in the dataset are skipped.Audio files
audiometadata
flag. When the audiometadata
flag is present in the JSON file, we directly use the supplied metadata without performing any additional validation, and do not store the file on our servers. To guarantee accurate labels, it is crucial that the metadata you provide is accurate.Text Files
Single images
Image groups
skip_duplicate_urls
is set to true
, all URLs exactly matching existing image groups in the dataset are skipped.objectUrl_{position_number}
).Image sequences
image_groups
array with the createVideo
flag set to true
represents a single image sequence.skip_duplicate_urls
is set to true
, all URLs exactly matching existing image sequences in the dataset are skipped.createVideo
flag to be set to true
. Both use the key image_groups
.objectUrl_{position_number}
).DICOM
dicom_series
element can contain one or more DICOM series.skip_duplicate_urls
is set to true
, all object URLs exactly matching existing DICOM files in the dataset will be skipped..dcm
file and does not have to be specific during the upload to Encord. DicomSeries
file is created. This file represents the series file-set. Only DicomSeries
are displayed in the Encord application.Multiple file types
ObjectUrl
column is interpreted as a request for video upload. If your objects are of a different type (for example, images), this error displays: “Expected a video, got a file of type XXX”.Videos
objectUrl
. This field is mandatory for each file, as it specifies the full URL of the video resource.
video_title
. If left blank, the original file name is used.
ObjectUrl | Video title |
---|---|
path/to/storage-location/frame1.mp4 | Video 1 |
path/to/storage-location/frame2.mp4 | Video 2 |
path/to/storage-location/frame3.mp4 | |
path/to/storage-location/frame4.mp4 | Video 3 |
Single images
objectUrl
. This field is mandatory for each file, as it specifies the full URL of the image resource.
image_title
. If left blank, the original file name is used.
ObjectUrl | Image title |
---|---|
path/to/storage-location/frame1.jpg | Image 1 |
path/to/storage-location/frame2.jpg | Image 2 |
path/to/storage-location/frame3.jpg | |
path/to/storage-location/frame4.jpg | Image 3 |
Image groups
objectUrl
. This field is mandatory for each file, as it specifies the full URL of the resource.
image_group_title
. This field is mandatory, as it determines which image group a file will be assigned to.
ObjectUrl | Image group title | Create video |
---|---|---|
path/to/storage-location/frame1.jpg | Group 1 | false |
path/to/storage-location/frame2.jpg | Group 1 | false |
path/to/storage-location/frame3.jpg | Group 2 | false |
path/to/storage-location/frame4.jpg | Group 2 | false |
Image sequences
objectUrl
. This field is mandatory for each file, as it specifies the full URL of the resource.
image_group_title
. This field is mandatory, as it determines which image sequence a file will be assigned to. The dimensions of the image sequence are determined by the first file in the sequence.
ObjectUrl | Image group title | Create video |
---|---|---|
path/to/storage-location/frame1.jpg | Sequence 1 | true |
path/to/storage-location/frame2.jpg | Sequence 1 | true |
path/to/storage-location/frame3.jpg | Sequence 2 | true |
path/to/storage-location/frame4.jpg | Sequence 2 | true |
DICOM
objectUrl
. This field is mandatory for each file, as it specifies the full URL of the resource.
dicom_title
. When two files are given the same title they are grouped into the same DICOM series. If left blank, the original file name is used.
ObjectUrl | Series title |
---|---|
path/to/storage-location/frame1.dcm | dicom series 1 |
path/to/storage-location/frame2.dcm | dicom series 1 |
path/to/storage-location/frame3.dcm | dicom series 2 |
path/to/storage-location/frame4.dcm | dicom series 2 |
path/to/storage-location/frame5.dcm | dicom series 3 |
Multiple file types
true
all files that are not image sequences must contain the value false
ObjectUrl | Image group title | Create video |
---|---|---|
path/to/storage-location/frame1.jpg | Sequence 1 | true |
path/to/storage-location/frame2.jpg | Sequence 1 | true |
path/to/storage-location/frame3.jpg | Sequence 2 | true |
path/to/storage-location/frame4.jpg | Sequence 2 | true |
path/to/storage-location/frame5.jpg | Group 1 | false |
path/to/storage-location/frame6.jpg | Group 1 | false |
ObjectUrl | Image title | Create video |
path/to/storage-location/frame1.jpg | Image 1 | false |
ObjectUrl | Image title | Create video |
full/storage/path/video.mp4 | Video 1 | false |
Upload is still in progress, try again later!
is returned, use the
script to check the upload status to see whether the upload has finished.<private_key_path>
with the path to your private key.<integration_title>
with the title of the integration you want to use.<folder_name>
with the folder name. The scripts assume that the specified folder name is unique.path/to/json/file.json
with the path to a JSON file specifying which cloud storage files should be uploaded.A folder to store my files
with a meaningful description for your folder."my": "folder_metadata"
with any metadata you want to add to the folder."Upload is still in progress, try again later!"
, run the following code to query the Encord server again. Ensure that you replace <upload_job_id>
with the output by the previous code. In the example above upload_job_id=c4026edb-4fw2-40a0-8f05-a1af7f465727
.
The script has several possible outputs:
timeout_seconds
argument from the
add_private_data_to_dataset_get_result() method performs status checks until the status upload has finished.<private_key_path>
with the file path for your private key.<private_key_path>
with the path to your private key.<folder_name>
with the name you want to give your Storage folder.<dataset_hash>
with the hash of the Dataset you want to add the data units to.<private_key_path>
with the path to your private key.<dataset_hash>
with the hash of your Dataset.