Log in with your account
Click 'Datasets' > 'Create'
Provide a name and description
Click 'Select Files' to choose which files to add
Click 'Upload' to save selected files to dataset
Click 'View Dataset' to confirm. You can add more content with 'Add Files'.
Add metadata, terms of use, etc.
Some metadata may automatically be generated depending on the types of files uploaded. Metadata can be manually added to files or datasets at any time.
Clowder also includes a RESTful API that allows programmatic interactions such as creating new datasets and downloading files. For example, one can request a list of datasets using: GET _clowder home URL_/api/datasets. The current API schema for a Clowder instance can be accessed by selecting API from the ? Help menu in the upper-right corner of the application.
For typical workflows, the following steps are sufficient to push data into Clowder in an organized fashion:
Create a collection to hold relevant datasets (optional) POST /api/collections
provide a name; returns collection ID
Create a dataset to hold relevant files and add it to the collection POST /api/datasets/createempty
provide a name; returns dataset ID POST /api/collections/<collection id>/datasets/<dataset id>
Upload files and metadata to dataset POST /api/datasets/uploadToDataset/<dataset id>
provide file(s) and metadata
An extensive API reference can be found here.
Some files, e.g. those transferred via Globus, will be moved to the server without triggering Clowder's normal upload paths. These must be transmitted in a certain way to ensure proper handling.
Log into Globus and click 'Transfer Files'.
Select your source endpoint, and Terraref as the destination. You need to contact NCSA to ensure you have the necessary credentials and folder space to utilize Globus - unrecognized Globus accounts will not be trusted.
Transfer your files. You will receive a Task ID when the transfer starts.
Send this Task ID and requisite information about the transfer to the TERRAREF Globus Monitor API as a JSON object:
In addition to username and Task ID, you must also send a "contents" object containing each dataset that should be created in Clowder, and the files that belong to that dataset. This allows Clowder to verify it has handled every file in the Globus task.
The JSON object is sent to the API via an HTTP request: POST 141.142.168.72:5454/tasks
For example, with cURL this would be done with: curl -X POST -u <globus_username>:<globus_password> -d <json_object> 141.142.168.72:5454/tasks
In this way Clowder indexes a pointer to the file on disk rather than making a new copy of the file; thus the file will still be accessible via Globus, FTP, or other methods directed at the filesystem.