My new catch-phrase every time Pip comments on my complete lack of ability to explain things with pith and piquancy: Haud yer wheesht, laddie! Dae ye nae ken that genetics is gey complex? (Though, Pip: trust but verify. All I truly know about DNA is that I know less now than I thought I did a few years ago.)
Julie: The short answer is: Not that I know of; and probably not very soon, especially since the sales growth for genetic genealogy tests have been in the tank since mid-2018.
Unfortunately--you guessed it!--it'll take a bit to answer the questions with any meaningful detail.
Whole genome sequencing doesn't result in just one raw data file like our inexpensive microarray chip tests. With my 30X sequencing last year I got--lemme check--12 distinct files, and missing at least two I sorely wish had been provided.
It's worth noting that the files' contents all are dependent upon which version of the human genome reference map was in use by the testing company at the time the results were provided. For example, none of the providers of WGS tests that I'm aware of still use data mapped against GRCh37/hg19, which was released February 2009 and has since been found to have thousands of discrepancies and inaccuracies. GRCh38/hg38 was released December 2013. But we're still using GRCh37 for genealogy because that's what we started with and companies wouldn't be able to cross-compare unless they all switched to a newer release simultaneously...and they'd have to either convert your previous raw data to the new reference map for you, or tell you to go fish; figure it out yourself. That might effectively kill what market remains for genealogy because kits between old and new reference mappings really couldn't be directly compared. And my prediction is if we do up and move to a new reference, it won't be GRCh38...which really is at the end of its own lifecycle; it will the be next major release after that.
What GEDmatch used to accept as an upload but stopped is what's called a VCF file; stands for variant call format: https://en.wikipedia.org/wiki/Variant_Call_Format.
It's a plain text file in a standardized format that's extremely useful as a standalone for medical and population genetics purposes...but not for genealogy. The issue there is that the file shows only those alleles where your genome differs from the reference genome. Sounds okay at first glance, but the problem is that for genealogy we need to be able to verify that a shared segment is really a segment. The VCF only deals in where you're different, not where you're also the same. So you and your cousin might share an interesting indel (insertion/deletion) at a particular location on a chromosome, but it could involve only several or even a single nucleic acid. No segment to compare.
That may well have been why GEDmatch axed the uploads. VCFs take up a lot more room than our microarray test results; about 20MB compared to 1GB. If a VCF is uploaded in combination with, say, three sets of results from different microarray chips and versions, the combined superkit could be very useful because the microarray results give us a comparison baseline to go with all the differences. But just the boatload of differences by themselves aren't terribly helpful unless everyone has uploaded such a file.
One of the types of files I wish I'd received, but didn't, is called a gVCF, "genomic variant call format." This is an extended VCF that includes certain blocks of allele information that do match the reference genome map. GEDmatch never accepted these, though. And it's still not your complete genome: those reference matching blocks are selected because they provide qualifying information about the differences, and not all of them will be genealogically useful.
As I mentioned upstream, the real raw data from a WGS comes in the form of FASTQ or BAM files. These are huge. I think 30X coverage for WGS is and will prove to be the functional minimum for accuracy, but we may well see 60X become price-point feasible for general consumer purchases in the near future. My two FASTQ files, one forward and one reverse read, total right at 270 gigabytes. A 60X test won't double that, but it won't be remarkably smaller than double. There's just no networking infrastructure on the horizon that will let us routinely upload 400-gigabyte files that cloud server resources on the other end will then manipulate and store and subsequently run multiple comparisons at a time against.
I can think of at least two scenarios where WGS for genealogy would be feasible, and I'll bet someone is working on one, both, or something I haven't thought of. But to my knowledge, there's no packaged solution yet. With some lengthy and tedious work, I could take my GRCh38 FASTQ data and eventually extract GRCh37-mapped data--based on Illumina and Thermo-Fisher microarray specs plus info about rsIDs tested by the major companies plus my variants as identified in my VCF...plus maybe whatever else I could research about SNPs seen as indicative by population geneticists--and build a honkin' big data file formatted just like, say, the 23andMe standard, and then begin trying to upload it to GEDmatch to see how large a kit I can get accepted for tokenizing.
But, yeah. Get back to me on that. By the time I can find enough hours to get something like that done, someone will already have a commercial solution for us.