You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I ran vcf_to_bq without --run_annotation_pipeline - it ran fine and BigQuery tables were created.
When I added the '--run_annotation_pipeline true' parameter - 8570 output files were generated, but none had the **_vep_output.vcf extension. The output file structure was 'annotation/shards/LONG_UUID' with a single file in each called 'count_20000'.
Traceback (most recent call last):
File "/usr/lib/python2.7/runpy.py", line 174, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "/usr/lib/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
File "/opt/gcp_variant_transforms/src/gcp_variant_transforms/vcf_to_bq.py", line 643, in <module>
raise e
IOError: No files found based on the file pattern gs://my_output_bucket/annotation/**_vep_output.vcf
The text was updated successfully, but these errors were encountered:
I think what is needed for VEP is the --vcf flag.
I'm running VEP manually and the command looks like: vep -i NA12878.g.vcf.gz --fork=4 --vcf --allele_number --assembly=GRCh38 --cache --offline --force_overwrite
Which outputs a VCF formatted file.
Also, I tried adding '--shard_variants false' to the parameter list - no change - still failed with the same error.
I believe this is due to the vep_runner using the v2alpha Genomics API. We will look at updating this. In the meantime you can enable the Genomics API here to resolve the issue.
The input xxxx.g.vcf.gz file was generated using the BAM to VCF Cromwell pipeline: https://github.com/broadinstitute/wdl-runner
When I ran vcf_to_bq without --run_annotation_pipeline - it ran fine and BigQuery tables were created.
When I added the '--run_annotation_pipeline true' parameter - 8570 output files were generated, but none had the **_vep_output.vcf extension. The output file structure was 'annotation/shards/LONG_UUID' with a single file in each called 'count_20000'.
The command I ran was:
The output error was:
The text was updated successfully, but these errors were encountered: