At the moment, we sometimes see two datasets kicked off at the same time, probably because of lag in the dataset scan. We should: - [ ] Add some logging to the scanner so we can see everything being run - This should be set up independently of the (dataset-specific) pipeline logging - [ ] Decouple the scan from the pipeline processing so there is always a single scanning process - This would mean the core pipeline is run with something like `analysis_driver --sample <sample_id>` - The status manipulation (reset, resume, etc.) could be moved somewhere else
At the moment, we sometimes see two datasets kicked off at the same time, probably because of lag in the dataset scan. We should:
analysis_driver --sample <sample_id>