At the moment, we sometimes see two datasets kicked off at the same time, probably because of lag in the dataset scan. We should: - [ ] Add some logging to the scanner so we can see everything being run - This should be set up independently of the (dataset-specific) pipeline logging - [ ] Decouple the scan from the pipeline processing so there is always a single scanning process - This would mean the core pipeline is run with something like `analysis_driver --sample <sample_id>` - The status manipulation (reset, resume, etc.) could be moved somewhere else