Lift over of GWAS summary stat file from Hg38 to Hg19
1
1
Entering edit mode
3.4 years ago
AVA ▴ 40

Hello,

I am using the UCSC lift over tool and the associated chain to lift over the results of my GWAS summary statistic file (a tab separated file) from build 38 to build 37. The GWAS summary stat file looks like:

1 chr1_17626_G_A 17626 A G 0.016 -0.0332 0.0237 0.161
1 chr_20184_G_A  20184 A G 0.113 -0.185  0.023 0.419

Follwing is the UCSC tool with the associated chain I am using:

I want to create a file in bed format from GWAS summary stat fle that is the required input by the tool, where I would like the first three columns to be tab separated and rest of the columns to be merged in a single column and separated by a non tab separator such as "." so as to preserve them while running the lift over. The first three columns of the input bed file would be:

awk '{print chr$1, $3-1, $3}' GWAS summary stat file > ucsc.input.file

#$1 = chrx - where x is chromosome number 
#$2  position -1  for SNPs
#$3  bp position hg38 for SNPs

The above three are the required columns for the tool.

My questions are:

  1. How can I use a non tab separator say ":" to merge rest of the columns of the GWAS summary stat file in one column?
  2. After running the liftover, how can I unpack the columns separated by :?
Hg38 linux Liftover GWAS Hg19 • 2.6k views
ADD COMMENT
0
Entering edit mode
22 months ago
arturtjaro ▴ 40

There's a simpler and cleaner solution.

  1. Make the bed file as normal using awk, and carry over all the columns ($0 here):
    awk 'BEGIN {OFS="\t"} {print chr$1, $3-1, $3, $0}' GWAS_summary_stat_file > ucsc.input.bed
    
  2. Lift over using the -bedPlus=3 option. This does the liftover based on the first three columns, and carries over all the remaining columns for the ride:
    liftOver -bedPlus=3 -tab ucsc.input.bed hg38ToHg19.over.chain ucsc.output.bed ucsc.unmapped.bed
    
ADD COMMENT

Login before adding your answer.

Traffic: 2364 users visited in the last hour
Help About
FAQ
Access RSS
API
Stats

Use of this site constitutes acceptance of our User Agreement and Privacy Policy.

Powered by the version 2.3.6