Nanopore DNA Sequencing 51
mindpixel writes: "Harvard scientists have concieved a revolutionary technology for probing, and eventually sequencing, individual DNA molecules using single-channel recording techniques. The technique essentially pulls a single strand of DNA through a nanopore, reading off the individual bases electrically. The technique could allow for decoding of a person's genome in hours instead of years." While the sequencing in hours instead of years is something that's pretty darn cool, our holdup in using this data is actually now what the genes are, and how they interact. That will still take years for us to figure out.
still vaporware & hype at this point (Score:1)
There are several companies rushing to commercialize this general way of sequencing and haplotyping. The most interesting one I've seen so far is US Genomics [usgenomics.com].
The problem is still at the stage of "translating our pretty drawings and ideas into working hardware". There are tremendous engineering challenges both with the hardware (detection devices and materials handling) and software (algorithms etc.)
The science is solid and the ideas are clever. Its just going to take a while to build a system that can reliably resolve single molecules as they speed down a tiny channel.
just my $.02
Re:Protein Applications? (Score:1)
A genome a day? (Score:1)
To answer an earlier question, sequencing is reading the DNA and recording the base pairs in some fashon. Often this information ends up as an asci file. Because of the mechanical limits of the current sequencing technologies, chemicals called restriction enzymes are used to snip long strands of DNA at particular patterns of sequence. A sample of DNA is amplified using a technique called polymerase chain reaction ( PCR ) which gives many many copies of the precisely same sample. It is then split into multiple samples each "cut" with a different restriction enzyme each with a different "trigger sequence" so the different samples end up being cut in different places which means that the cut pieces of different samples overlap the cuts in other samples. Then all the samples are seperatly sequenced into files. The files can then be compared for overlaps and "assembled" into longer sequences until the sequence of the original sample ( which was too big for the machine ) is known. After the sequence is known, it can be converted from nucleic acid ( G A T C ) code to amino acid code ( google can probably find you the genetic code ) which is composed of 3 nucleic bases per amino acid. Because we really do not know where a given amino acid starts within the nucleic file, we often do a sliding frame translation, decoding the entire sequence for each possible starting point of amino acid translation. The proper translation is often indicated by having a minimal number of null translation frames. The decoded sequence can be compared to other previously decoded sequence which have been analyzed for function - these blocks end up being called genes. A gene from a mous which makes a particular protein is the same as that gene from a human, so by starting with simple organisims, many genes for specific proteins have been found which help decode the more complex organisims....
hmmmmm sorry for the drif folks...
Z
Erm.. (Score:1)
Re:Lame Posts (Score:1)
Re:Not Hours ... buy Days Maybe (Score:1)
Anyway, I see two issues. One, what size DNA strand can be sequenced using a nanopore without breaking? Two, how many nanopores can be made to operate simultaneously? With 100 nanopores operating in parallel on 10-kilobase DNA fragments at 10 seconds per nanopore per fragment, you're getting 10 kilobases per second, or a billion bases every three hours. At these speeds, the real limiting factor is probably going to be something other than raw sequencing speed. I think it's a very exciting technology.
makes me dream... (Score:1)
Huge! (Score:1)
But it's very important empirically to have more than one dataset. Who's GNOME did we sequence anyway? And comparing the sequence of many different instances of genes from different people is not time consuming. You're talking about figuring out how the Human GNOME works. That's a completely different dicipline.
Being able to decode strands of Nucleic Acids (and the technique might be applicatble to Poly Peptides(Protiens) as well) in a matter of minuites without the mess of Gel Electrophoresis would be HUGE.
Not Hours ... buy Days Maybe (Score:1)
Because the pore in the channel is large enough to admit only a single strand of DNA, the time it takes for the DNA to be drawn through the channel (enlarged view) effectively measures the length of the DNA molecule (here, 1,300 ms corresponding to a 1,060 nt polymer).
So if there are 3 billion bases in the Human GNOME,
Of course if the apparatus is easy to use I supose you could have several running at the same time provided you could prepare the material fast enough.Re:i don't get it? (Score:1)
Decoding: finding out what the sequence leads to as far as expression of traits goes. Like why you have an aversion to bio-tech speak (grin).
i don't get it? (Score:1)
I guess I'm a little naive on the genome thing... but the way quantum computing is going, I better brush up on it, eh?
Re:i don't get it? (Score:1)
At the moment DNA gets broken up into small segments and fired at a device that can decode each of thease segments,
The problem it you know what the segments are but not what order there in, like a big jigsaw.
Sequencing is putting thease pieces back in the correct order.
This new process decodes the DNA in sequence so there hard and sloq sequencing step is no longer needed
Re:DNA Fingerprinting (Score:1)
Such a system could be used to trace all kinds of "things" and "substances"
Re:A "Watson, come quick!" moment (Score:1)
So when can I back up my hard drive (Score:1)
darn straight (Score:1)
Old News (Score:1)
Re:DNA Code (Score:1)
So?
"I have no child."
"There's no child unto me."
You can understand that, although we wouldn't say it that way. So, too, for other meanings of the verb "to have":
"You have my stuff."
"My stuff is with you."
"I have to go."
"I need to go."
"I've never gone."
"I never went."
There is no particularly compelling reason that a language should use "to have". In fact, if English
~
Re:What will it be used for? (Score:1)
Gattaca (Score:1)
Mysterious Mobious (Score:1)
P.S. if you're reading this Dan, do you need an electrophysiologist? G. (LOL)
few minutes (Score:1)
A "Watson, come quick!" moment (Score:1)
"Professor, the computer is decoding the first base pairs right now!"
"Record them for posterity, this is history in the making"
"Ok... reading now... G... A... T... T... A... C... A..."
Re:Sounds like good news for systematic zoology! (Score:1)
DNA testing kit! (Score:1)
Re:Protein Folding (Score:1)
Here are the links along with key excerpts:
"...It has been put forward that many of these chromosomal changes are caused by the ability of certain specific types of DNA sequences to fold into unusual structures which interfere with the faithful copying of chromosomes."
www.cpa.ed.ac.uk/news/research/07/item3.html
"...The knots and kinks in the DNA provide crucial topological stop-and-go signals for the enzymes."
www.khouse.org/articles/technical/19971201143.h
Protein Folding (Score:1)
Re:Lame Posts (Score:1)
The implications of this discovery are incredible (Score:1)
Re:i don't get it? (Score:2)
Re:i don't get it? (Score:2)
--
DNA Fingerprinting (Score:2)
Get arrested, give a blood sample. It'll only take a few hours to verify who you are. None of this "probably" stuff, they'll have YOUR sequence on file, and there won't be any doubt (unless you have an identical twin).
Re:Who told you it takes years? (Score:2)
The faster better technique is electrostatically driven capliary tubes. They suck the sample through a microcapliary and shine a laser on it as it passes by. Generally the same 72 channels in parallel as the gel method - something about legacy analysis software as well as the plates that hold the samples and the robots which manipulate them. Kind of like the gauge of the railroad tracks used to carry space shuttle boosters being determined by the wheel spacing of the wheels on a roman chariot which was determined by the space required by two side by side horses.
we have the most advanced transportation system in the world having a major design parameter determined by the width of a roman horses ass.
Z
Storage Network Needed (Score:2)
The article mentions 3 billion bases in less than 2 hours. That comes to
(3/2) x 10e9 / (3.6 x 10e3) bases/sec
= 416667 bases/sec.
So you would need a sustained writing speed of about 400 kilobytes/sec, or if you compress it into 4 bases per byte, say 100 kilobytes/sec. to write to about 3 GB (or 730MB compressed) of disk space.
You could fit it onto an IBM Microdrive attached to your Palm!
Re:Protein Folding (Score:2)
You are both right and wrong with your post, and your enthusiasm for learning is encouraging.
You're correct: The protein folding ploblem is is hard, in the worst (best?) sense of the word. That is to say, NO, there have been no interesting advances in the field; All that has been done so far, more or less, is to throw some processor cycles at the problem. There have some good results from this aproach, but no new insights or understanding of the problem.
This would be a good project for an interested and motivated student looking for a trip to Stockholm.
With an understanding of protein folding, we could rapidly discard the current: "piss on the porch, see if the cat licks it up," era of biotech, and actually do some engineering. Proteins are, to put it mildly, amazing: they are nanomachines that work, sometines with such elegance that when you see what is really going on it is enough to take your breath away as you try to yell Holly Sh*t and wonder if you are living your life right.
Designing proteins to functional specifications would be highly interesting, i.e., one of the simplest scenarios: "Uh, how about we have it arange the glucose carbons in a cubic crystal latice and let the O and H diffuse out as water?"
Today, we can pretty much string together any arbitrarilly long sequence of amino acids we want, but they are just so much fried egg until the folding problem is solved. When, and if, that day comes, our world is going to change in surprising ways. The implications reach far beyond medicine and agriculture.
You're a bit wrong: DNA folding is not, by itself, a barrier to our understanding of the genome. Actually, we have a pretty good understanding of how the whole transcription translation thing works, with a pretty solid understanding of splicing, targeting, etc., even some understanding of promoters, enhancers, and so on, we just don't know, for the most part, what the hell the products of this process (proteins) are doing.
Re:i don't get it? (Score:2)
"Decoding" is a useful but too-broad term; it covers mapping and everything that comes afterward, such as figuring which of the mapped genes will express when, and to what degree, and how the proteins thus expressed will fold, and how they'll interact with each other. Once you've got all that, you have Mastered The Secrets Of Life Itself.
There is, to put it mildly, a lot of work to be done here. Job security for bioinformaticians.
Re:What will it be used for? (Score:2)
Immortality, my friend. Immortality.
The honolulu(sp?) technique used in cloning (done with mice - an incredibly difficult subject for cloning because the eggs cycle so quickly), unlike the faulty technique used with Dolly (which involved starving the nucleus-donor cell, and then using an electric shock to cause it to merge with the nucleus-free egg), has fertility rates that are rapidly approaching those of normal mice, and little to no genetic damage per breed. Yet, at around 5 generations of clones (cloning a clone of a clone of a clone of a clone), we start to see premature aging. Why? Because, as DNA lives, it slowly mutates. As must mutations are bad, it steadily poisons the mice's genetics.
The key is in a DNA backup.
Digital data doesn't corrode. It can be verified, backed up, copied, you name it. With the recent production of a completely DNA-synthesized fruit fly, the possibility approaches that we can completely re-create an individual's DNA strand. Then, when cloning organs for the individual, we use the backup DNA, not DNA from the person themself, as that DNA has become slightly corrupted overtime.
Digitizing DNA strands is a key to immortality.
That is why this is important.
-= rei =-
P.S. - I vote for genetically altered population. I certainly hope we see that day soon when people can make choices on whether or not they want their children to have to suffer.
Protein Applications? (Score:2)
I was thinking that this method could also be used to sequence proteins - a process which is now done using an automated process which can only produce (correct me if I'm wrong, and if I am I'll eat a bug) maybe 30 "letters" of sequence. Compare this to many hundreds at a run from DNA sequencing. If proteins could be sequenced hundreds of amino acids at a time, you could sequence a whole protein in one run. This would be better than the current method, where fragments are sequenced, and then the overlaps are compared to piece together the whole sequence.
Re:A genome a day.... (Score:2)
And will they be able to tell A from G (both purines) and C from T (both pyrimidines)? I don't have the charges handy but I think C and T are pretty close.
Also, DNA breaks very easily. No way are you going to be able to pull a whole chromosome through at once. If they get just 100 bases at a time, will that be useful?
What will it be used for? (Score:2)
Every drug company has a Genomics division these days, to analize the existing data from the Human Genome Project [doe.gov]. Now that new data can be gathered at such increadible speeds, are we any closer to improving the quality of life based on this work. Probably not, and the cause is a double edged sword.
The problem is the restrictions through international treadies and government regulation, on gentic engineering of humans. Don't get me wrong, I'm not in favor of such modification of the human genome, howeer, this leaves only one recourse. They can create medications that the sufferer of a genetic defect can take every day for their entire lives to prevent the ocurrance of an illness that they are genetically predisposed to. This is a boon for drug companies. If they can generate long term revenue streams by creating medications which reduce the chances of developing illnesses to which certain people are genetically predisposed to, and clain that they are doing this, instead of developing ways to repair a gene at birth - not because it's more proficable to do it this way but - because this is the only avenue they're allowed to pursue due to federal and international regulations against messing with the human genome; then who are the regulations truly serving? the population, or the drug companies?
Along the same lines, there will always be countries which are not signatories to the afore mentioned international regulations - in which drug companies can deelop the gene theropies which could truly benefit sufferers of gentic diseases and defects. That said, there will always be a black market for these theropies, once deeloped.
The question becomes which is a better world to live in: one where we have a drug dependant population, or one where we have a genetically altered population.
At this point I'll conclude my analysis because any further speculation will lead to the realm of Gattica style science-fiction. There is, however a great deal to consider...
--CTH
--
initial feasibility studies show good results (Score:2)
That was my first take as well, but then when I looked through the references, I found that many feasibility questions seem to be resolved already. For instance, I read the main page [harvard.edu] and thought, "Sure, but how do you transport the strand through the nanopore?" Then I checked the first reference listed [pnas.org], and what do you know: "We show that an electric field can drive single-stranded RNA and DNA molecules through a 2.6-nm diameter ion channel in a lipid bilayer membrane."
The final system may still be largely conceptual, but it's by no means blue sky. I tend to be a techno-skeptic but this work impresses me.
The page sounds to me like a breathless plea for lots of venture capital funding.
This is grossly unfair. The language and style are well within the normal bounds for scientific papers. The word "revolutionary" is appropriate for a technology that would do years of work in hours. And in case you didn't notice, it's not private research -- it's being done at The Department of Molecular and Cellular Biology, The Biological Laboratories, Harvard University [harvard.edu]. What interest would a university laboratory have in "venture capital"? If they later spin it off into private industry for product development, then they might go for venture funding, but it simply makes no sense to do so now. There's a big difference between research sponsorship and venture funding.
Tim
After ten years sequencencing the human genome (Score:2)
It must be a bit like climbing mount Everest the hard way and while you are sitting at the top eating your Kendal Mint Cake, someone rides up the access road on a bicycle.
So what was the point of spending several hundred million doing the job the hard way? Oh they filled a gazillion patents on the sequences the read out. And there I thought you had to invent something to get a patent.
Re:A genome a day.... (Score:2)
Just to correct a few inaccuracies. Firstly it hasn't been completed yet. Celera Genomics decided to claim they'd finished it, which spurred the public project to also claim the same. In reality both had only 90 - 95% of the consensus sequence and only then at "draft" quality.
Secondly, I feel it's wrong to claim that Celera were the people to "complete" it. Celera used their own data in conjunction with the public data, and yet they have (more or less) comparable results in terms of coverage, number of contigs, quality and so on. Personally I feel that this is like admitting that their own work doesn't add anything new to the public effort - ie they failed.
The bottleneck with sequencing at the moment is in the "finishing" process - tidying up the results to produce highly accurate answers. This is largely caused due to the randomness of the shotgun sequencing approach. As it's effectively solving a jigsaw puzzle from lots of randomly cut pieces of DNA, in some places you'll get lots stacking up and in others you'll find none. It is unrealistic (not to mention expensive) to keep increasing the coverage so that everywhere gets covered by the random shotgun process.
Instead a fixed depth (only 3 or 4 fold in the draft sequences) is used followed by directed sequencing where the user, or an automatic program, analyses the data set and chooses experiments to perform (primer walking typically). The graphs I've seen of draft vs finished data show quite well how the finishing is lagging behind.
However this new strategy, not being random, will greatly reduce the amount of finishing needed. However note that at present they are using it for probes rather than full scale sequencing. It has great potential, but looks to be years away from replacing the current work.
STM and AFM sequencing (Score:2)
AMAZING new technology! (Score:2)
This sounds exactly like late-night infomercials that invariably say things like "our scientists [actors in white lab coats conspicuously walking around behind the one being interviewed] have devised a revolutionary new formula that will make you lose weight without dieting or exercising!"
That is, if the people selling something describe it as "revolutionary" themselves, it isn't. If it really is revolutionary, we'll hear about it in other places. The HP-35 from the slide rule article -- that's revolutionary.
So while this may be a significant improvement, I'd change the prose if I was them.
Re:After ten years sequencencing the human genome (Score:3)
My point is that all this pain grief and expense was absolutely required to get where we are today and in most fields this will continue to be the way things are as we have no way of knowing what is hard and how it can be easy if what we are doing is new and innovative.
A genome a day.... (Score:3)
It will not replace conventional sequencing technology, unless it can beat the now pretty cheap cost. Conventional sequencing is based on labelling the individual DNA bases with a different flourescent dye, and running the DNA through a gel which seperates the DNA according to size: As each base runs through the gel, it goes past a laser/detector which can detect the specific DNA base (A,T,C or G) at that position. Due to gradual impovements to this technique over the last 20 or so years (originally it employed radiation, rather then flourescence) the speed, sensitivity and cost has decreased dramatically. For example, the human genome project started in ernest about 10 yrs ago. Celera Genomics, using modern technology (and alot of financial backing, and the fact they are a subsidiary of the people who make sequencing machines, [appliedbiosystems.com]) competed the genome in a matter of months. The increase in DNA sequencing capacity puts moore's law to shame.
For example, our lab could process around 100kb (thats KiloBases guys!) of data a day, but we never even touch this with our machine [appliedbiosystems.com]. No need, and the same stands for many small-medium research labs. Alot of people like us will probably stick with conventional sequencing technology for a long time (it works well, is high enough throuput, cheap & easy).
However, the are some exciting applications with single strand sequencing. For example forensics. Also, it allows the oppotunirty of sequencing RNA (this is the "messenger" which passes the "important" part of the DNA message to the ribosomes, which then "compile" a protein - the stuff which actually does things, like an enzyme or structural component). Sequencing RNA is exciting, as currently you have to convert the RNA back to DNA (which can cause problems) and then sequence that.
Another obvious application for this would be very high throuput sequencing which would be employed by the major sequencing centres. Yes, i know we already have the Human Genome, but a fashionable idea at the moment is comparative genomics. This is very much taking biology back to its roots (i.e. like Darwin and Wallace comparing the morphological characteristics of certain species and infering adaption), but at a molecular level. This will yield amazing insights with discoveries having important implications from medicine to evolution. In fact I think the general public & media will soon be bored of this. Each week it will be a new genome being announced; mouse, chicken, rat, pufferfish, rice, corn, dog, cat, cow, chimp......
Re:Huge! (Score:3)
Hey, dont knock electrophoresis mate. It is the basis of PCR resolution, Southerns, Northens, SSCP, conventional sequencing methods plus a multitude of other applications. Furthermore, refinements to this approach (read: capillary electrophoresis) have supplied one of the major advances to sequencing methodology in recent years, unlike the technology we are (were?) discussing.
Who's GNOME did we sequence anyway
Good Question. Apparantly, they took DNA from around 100 (i forget the exact figures here) US citizens of various sex/ethnicity, picked 7 out of the hat, and sequenced portions of each (most from a single, unidentified individual - although if nanopore technology comes to fruition, i reckons we can track him) down
p.s. thats one of my favourite all-time techno-typo's: the Human Gnome Project. Almost as good as sequencing my ARS (a yeast thang....)
DNA Code (Score:4)
But it is a place to start.
Side note:
while looking up the Finnish Language pages for this comment, I came across this tidbit: That Finnish has "no equivalent of the verb to have [finland.fi]". This has interesting philosophic implications in the history of open source, etc.
Check out the Vinny the Vampire [eplugz.com] comic strip
Re:A genome a day.... (Score:4)
Yeah. Sure, i agree with you. Once youve taken DNA out of the nucleus and stripped out the protein complement of chromatin, it is darn hard to get high-quality genomic DNA (for PFGE or whatever). Furthermore, yanking the DNA through that little hole is gonna probably cause problems.
However, you dont need unfragmented DNA. For example, you could fragment genomic DNA, and pieces of this would randomly pass through the pore - essentially a shotgun approach. This wouldn't be a bad way of doing it. This would also get round problems with the detection mechanism. For example, if they could only disringuish between C/T with 80% accuracy, multiple reads of the same sequence could clear this up.
The technology could also overcome problems such as cloning bias, problems with sequencing microsats (e.g. (AT)n), GC rich regions etc. Which the HGMP is still having problems with.
Sounds like good news for systematic zoology! (Score:4)
The human genomes was sequenced by taking lots of DNA, cutting it up randomly sequencing the random pieces of cut up DNA.
In my field, we work with much smaller amounts of DNA. Sometimes I only have a single specimen of a tiny insect, or unique material (from rare or extinct species) to try and get some DNA out of. In older material, DNA is usually degraded and many times we end up with nothing but a destroyed or damaged specimen.
With small amounts of DNA to begin with, we have to amplify (PCR) single genes or regions by using general primers, which means that they don't only fit on the insect DNA, but fungi and human DNA too, making contamination of your material very real risk.
If this technology turns out to work on a larger scale, it's amazing news for me and my collegues.
The nanopore technolgy sequences single moleculer, which means the PCR step becomes unneccesary! This means that we can get sequences from specimens with severely degrades DNA, and we don't have to be as afraid of grinding up rare material in hope of getting sequences.