Skip to content

Commit

Permalink
hot-fix: processor full name (#101)
Browse files Browse the repository at this point in the history
  • Loading branch information
jtyoung84 authored May 7, 2024
1 parent 8e753ab commit e82e0e3
Show file tree
Hide file tree
Showing 3 changed files with 2 additions and 7 deletions.
3 changes: 0 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,6 @@ You can go to http://aind-data-transfer-service to submit a `.csv` or `.xlsx` fi

What each column means in the job submission template:

- **processor_full_name**: Name of the person submitting the upload job
- **project_name**: Project name. A full list can be downloaded at [Project Names](http://aind-metadata-service/project_names)
- **process_capsule_id**: Optional Code Ocean capsule or pipeline to run when data is uploaded
- **platform**: For a list of platforms click [here](https://github.com/AllenNeuralDynamics/aind-data-schema/blob/main/src/aind_data_schema/models/platforms.py).
Expand Down Expand Up @@ -73,11 +72,9 @@ platform = Platform.BEHAVIOR
behavior_config = ModalityConfigs(modality=Modality.BEHAVIOR, source=(source_dir / "Behavior"))
behavior_videos_config = ModalityConfigs(modality=Modality.BEHAVIOR_VIDEOS, source=(source_dir / "Behavior videos"))
metadata_dir = source_dir / "Config" # This is an optional folder of pre-compiled metadata json files
processor_full_name="Anna Apple"
project_name="Ephys Platform"

upload_job_configs = BasicUploadJobConfigs(
processor_full_name=processor_full_name,
project_name=project_name,
s3_bucket = s3_bucket,
platform = platform,
Expand Down
5 changes: 2 additions & 3 deletions src/aind_data_transfer_service/templates/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -145,16 +145,15 @@ <h2>Submit Jobs</h2>
let jobsLength = jobs.length;
var table = document.createElement('table'), tr, td, row;
addTableRow(
[ "processor_full_name", "project_name", "process_capsule_id", "s3_bucket", "platform", "subject_id", "acq_datetime", "metadata_dir", "modality", "modality.source" ],
[ "project_name", "process_capsule_id", "s3_bucket", "platform", "subject_id", "acq_datetime", "metadata_dir", "modality", "modality.source" ],
table, tr, td, true
);
for (row = 0; row < jobsLength; row++) {
let job = JSON.parse(jobs[row]);
let modalities = job.modalities;
let modalitiesLength = modalities.length;
addTableRow(
[ { value: job.processor_full_name, rowspan: modalitiesLength },
{ value: job.project_name, rowspan: modalitiesLength },
[ { value: job.project_name, rowspan: modalitiesLength },
{ value: job.process_capsule_id ?? "", rowspan: modalitiesLength },
{ value: job.s3_bucket, rowspan: modalitiesLength },
{ value: job.platform.abbreviation, rowspan: modalitiesLength },
Expand Down
1 change: 0 additions & 1 deletion tests/test_hpc_models.py
Original file line number Diff line number Diff line change
Expand Up @@ -133,7 +133,6 @@ class TestHpcJobSubmitSettings(unittest.TestCase):

example_config = BasicUploadJobConfigs(
aws_param_store_name="/some/param/store",
processor_full_name="John Smith",
project_name="Behavior Platform",
s3_bucket="some_bucket",
platform=Platform.ECEPHYS,
Expand Down

0 comments on commit e82e0e3

Please sign in to comment.