Wednesday, 2 September 2015

New Ion S5 sequencher announced

It's been a while since Life Technologies, now part of Thermo Fisher, have announced major improvements on its semiconductor sequencing platforms.

Ion S5 and S5 XL NGS Sequencing Systems

Now the company announced this new entry it its NGS sequenchers: the Ion S5.
The Ion S5 looks similar to the Ion Proton and place as compromise solution between the Proton and the PGM machines. The new machine will be available in two version, standard and XL, with the XL version having more computational power.
The Ion S5 works with three different sequencing chips. As reported in the official brochure, the 3 new chips are named Ion 520, resembling the PGM 318 chip with up to 1 Gb throughput; Ion 530, up to 4 Gb; and Ion 540, similar to the Proton PI chip with 15 Gb.

The idea is clearly to offer a machine that combines the use of fast and low cost gene panels with the ability to deliver human exomes and RNA-Seq.
In this way the machine became an interesting choice to perform both cost-effective diagnostic panels based on the AmpliSeq technology, but maintaining also a research flavor with exome sequencing and counting applications.
The sequencing kits have been reconfigured and are now sold as prepacked cartridges and bottles that combines with a barcode scan system to make the use simpler and easier. Ideally the new system have to be integrated with the Ion Chef system for automated template preparation and chip loading. Moreover, thank to a recent upgrade to eh software, this system will also prepare AmpliSeq libraries. In this way all your diagnostic (or research) AmpliSeq panels could be prepared in automated fashion from DNA to ready-to-sequence chip with minimal hand-on time, increasing standardization and performances.

It has been clear in the past few years that where Ion technology deliver its best, and has more commercial opprtunities, is on fast and low cost targeted sequence with an expanding range of solution available in terms of diagnostic gene panels and targeted panel for cancer research. Also this new machine go in this direction, as it seems clearly aimed at clinical applications.

With a price tag of $65K ($150K for the XL model) the Ion S5 sequencher is a new interesting competitor in the field!

Tuesday, 3 March 2015

Main new technologies from AGBT15

Ion Torrent
  • The new avalanche chemistry.
Risultati immagini per ion torrentThis new chemistry based on isothermal amplification will provide longer reads up to 600bp on PGM an 400bp on Proton PI with lower costs for template preparation. 
  • New protocols for Ampliseq library preparation on Ion Chef system. 
The company aims to transfer all the processing steps on the new Ion Chef that will beacm the all-in-one solution from library prep to chip loading, minimizing hand on time. Take a look to the video on you tube.
  • News on the PII chip.
They finally reported about a full working version of the PII chip producing up to 300M reads with 100bp length. However an official release date was not announced


  • NeoPrep, the new automated system for library preparation. 
Risultati immagini per neoprepHaving already presented some new sequencers few months ago, Illumina come to the stage with this new equipment that will provide fast and accurate library preparation and can prepare 16 libraries at with as little as 30 minutes of hands-on time. NeoPrep is based on the electrowetting microfluidic technology and use A cartridge with reagents and your sheared DNA. The instrument generate and quantify each library, ready for pooling. Roughly is required per run. All this stuff at the price of $49K ($39K introductory pricing for the first 6 months). More details at the Illumina page or on Omics Omics blog.

10X Genomics

Product Shot
10X Genomics is the real innovation this year. They revealed their new GemCode technology to reconstruct long reads and haplotypes from standard short reads sequencing. The new instrument will integrate in the standard Illumina based library preparation, thus configuring as an add-on for already equipped laboratories. Using dedicated pen source software and genome browsers from 10X Genomics one can then generate standard file formats (such as phased VCF or BAM with phase tags) and visualize the reconstructed haplotypes. 
Take a look at this nice video or read more details from Omics Omics

Wednesday, 25 February 2015

AGBT 2015 is here!!

Here it is another edition of the exciting conference AGBT (advances in genome biology an technology).
As usual there are a lot of new technologies announced and hundreds of interesting talks touching every aspects of genomics and presenting the latest technologies and methods!

In this first day one there is a lot of talking about the new platform from 10X Genomics that promise to use massive fragment labeling to assemble long reads from standard short reads produce by Illumina technology. The new instrument will thus be an add-on for seqeucning labs instead of a completely new sequencing technologies with the ability to deliver long haplotypes.
Let see what they will reveal at the official presentations

Don't forget to follow the conference in real-time on twitter #AGBT15
If you want an overview of what is going on take a look to the official agenda and to the full list of posters and software presented on site.

Thursday, 30 October 2014

Exome Aggregation Consortium release its data on 63,000 exomes!

On October 29th, the Exome Aggregation Consoritum as released its browser based on the impressive number of 63,000 human exomes.
This database is the larger collection of human exome data so far and provide both a web base interface to retrieve variants in your gene of interest or the download of a VCF file containing the list of all the annotated variants.

The final dataset is based on sequences from several consortia working on complex disorders and also includes 1000G and ESP6500 data.

The first aim of the Consortium is to study the distribution of "human knockout", that is people having both copies of a given gene inactivated by severe mutations. The analysis of associated phenotype data promise to reveal lot of interesting information of the actual role of single human genes. Moreover, the study of subjects carrying inactivating mutations on known disease genes but not showing the expected phenotype could lead to identification of new therapeutic targets.

See more information on Nature News and Genome Web!

Tuesday, 2 September 2014

PubMed Highlight: New relese of ENCODE and modENCODE

Five papers that summarize the latest data from ENCODE and modENCODE consortia have recently been published on Nature. Together, the publications add more than 1,600 new data sets, bringing the total number of data sets from ENCODE and modENCODE to around 3,300.

The growth of ENCODE and modENCODE data sets.

The authors analyze RNA-Seq data produced in the three species and an extensive effort was conducted in Drosophila to investigate genes expressed only in specific tissue, developmental stages or only after specific perturbations.  The analysis also identified many new candidate long non-coding RNAs, including ones that overlap with previously defined mutations that have been associated with developmental defects.
Other data sets derive from chromatin binding assays focused on transcription-regulatory factors in human cell lines, Drosophila and C. elegans; and on study of DNA accessibility and certain modifications to histone proteins. These new chromatin data sets led to identification of several features common to the three species, such as shared histone-modification patterns around genes and regulatory regions.
The new transcriptome data sets will result in more precise gene annotations in all three species, which should be released soon. The access to the data on chromatin features, regulatory-factor binding sites, and the regulatory-element predictions seem more difficult. We have to wait for them to be integrated in user-friendly portals for data visualization and flexible analyses. The UCSC Genome Browser, Ensembl, ENCODE consortium are all working to provide the solution.

Meanwhile take a look to the papers:
Diversity and dynamics of the Drosophila transcriptome Regulatory analysis of the C. elegans genome with spatiotemporal resolution Comparative analysis of metazoan chromatin organization

Monday, 14 July 2014

New challenges in NGS

After about a decade from the first appearance if NGS sequencing we have seen incredible improvements in throughput, accuracy and analysis methods and sequencing is now more diffused and easy to achieve also for small labs. Researchers have produced tons of sequencing data and the new technology allowed us to investigate DNA and human genomic variations at unprecedent scale and precision.
However, beside the milestones achieved, we have now to deal with new challenges that were largely underestimated in the early days of NGS.

MassGenomics has a nice blog post underlining the main ones, that I reported here:

Data Storage. 
Where do we put all those data from large genomic sequencing projects? Can we afford the cost of store everything or we have to be more selectively on what to keep in our hard drives?

Statistical significance.
GWAS studies have showed us that large numbers, in the order of 10 thousands of samples, are needed to achieve statistical significance for association studies, particularly for common diseases. Even when you consider the present low price of 1,000$ / genome it will require around 10 millions $ for such a sequencing project. So we can reduce our sample size (and thus significance) or create mega consortium with all the managing issues.

Samples became precious resources.
In the present scenario sequencing power is not longer a limitation. The real matter is find enough well-characterized samples to sequence!

Functional validation.
Whole genome and whole exome approaches let researchers to rapidly identify new variants potentially related to phenotypes. But which of them are truly relevant? Our present knowledge do not allow for a confident prediction of functional impact of genetic variation and thus functional studies are often needed to assess the actual role of each variants. These studies, based on cellular models or animal models, could be expensive and complicated.

With large and increasing amount of genomic data available to the community and studies showing that people ancestry and living location could be traced using them (at least in a proportion of cases), there are concerns about how "anonymous" these kind of data could really be. This is going to became a real problem has more and more genomes are sequenced.

Friday, 4 July 2014

PubMed highlight: Literome help you find relevant papers in the "genomic" literature

This tool mines the "genomic" literature for your gene of interest and reports a list of interactions with other genes, specifying also the kind of the relation (inhibit, activate, regulate...). It can also search for a SNP and find phenotypes associated to it by GWAS. You can then filter the results and also report if the listed interactions are actually real or not.

Good stuff to quickly identify relevant papers in the large amount of genomic researches!

Literome: PubMed-scale genomic knowledge base in the cloud

Hoifung Poon, Chris Quirk, Charlie DeZiel and David Heckerman

Motivation: Advances in sequencing technology have led to an exponential growth of genomics data, yet it remains a formidable challenge to interpret such data for identifying disease genes and drug targets. There has been increasing interest in adopting a systems approach that incorporates prior knowledge such as gene networks and genotype–phenotype associations. The majority of such knowledge resides in text such as journal publications, which has been undergoing its own exponential growth. It has thus become a significant bottleneck to identify relevant knowledge for genomic interpretation as well as to keep up with new genomics findings.
Results: In the Literome project, we have developed an automatic curation system to extract genomic knowledge from PubMed articles and made this knowledge available in the cloud with a Web site to facilitate browsing, searching and reasoning. Currently, Literome focuses on two types of knowledge most pertinent to genomic medicine: directed genic interactions such as pathways and genotype–phenotype associations. Users can search for interacting genes and the nature of the interactions, as well as diseases and drugs associated with a single nucleotide polymorphism or gene. Users can also search for indirect connections between two entities, e.g. a gene and a disease might be linked because an interacting gene is associated with a related disease.

Availability and implementation: Literome is freely available at Download for non-commercial use is available via Web services.

Wednesday, 25 June 2014

National Children's Study stopped, waiting for revisions

One of the most ambitious project and one of the few attempts to really perform "personal genomics", is (or I may say was) the National Children's Study (NCS) sustained by NIH and the US government.

The project try to investigate the relation between genomics and environmental factors to define their impact on human life and define which advantages this kind of genomic screening could provide for the human health. The massive longitudinal project that would sequence the genomes of 100,000 US babies and collect loads of environmental, lifestyle, and medical data on them until the age of 21.
However the NIH director, Francis Collins, has recently announced that the project will be stopped waiting for a detailed review on the methodologies applied and the opportunity to complete it in its present form. Few key questions has to be addressed: Is the study actually feasible, particularly in light of budget constraints? If so, what changes need to be made? If not, are there other methods for answering the key research questions the study was designed to address?

As GenomeWeb reports, National Academy of Sciences (NAS) released a report saying the NCS needs some major changes to its design, management, and oversight.  The NAS recommendations include making some changes to the core hypotheses behind the study, beefing up scientific input and oversight, and enrolling the subjects during pregnancy, instead of at birth, as is the current plan.

Monday, 23 June 2014

One banana a day will keep the doctor away

According to GenomeWeb and The Guardian, researchers from Australia are tweaking the genome of the banana in order to get it to deliver higher levels of vitamin A. The study is aimed to supplement vitamin A in Uganda and other similar population, where banana is one of the main food sources and deficiency in vitamin A cause blindness and death in children.

The group of professor James Dale, from the Queensland University of Technology, received a $10 million grant from the Bill and Melinda Gates Foundation to support this 9 year project.

Dale said that by 2020 vitamin A-enriched banana varieties would be grown by farmers in Uganda, where about 70% of the population survive on the fruit.

The genome of a baby sequenced before birth raises questions on the opportunity and pitfalls of genome screening

Khan, graduate student at the University of California, Davis, and blogger at The Unz Review, decided that he wanted detailed genetic information on his child as soon as he knew that his wife was pregnant. After a genetic test for chromosomal abnormalities he asked to have the DNA sample back and managed to have the baby's genome sequenced on one of the University NGS instruments.

MIT technology review reports about the whole story and Khan tells on the many difficulties he faced to have the genome sequencing done. Most of the medical staff tried to discourage him from performing this kind of test, afraid that the couple could take irrevocable decisions, such as pregnancy termination, based on the presence of putative deleterious mutations in the baby's genome. This case raised again the question of how much information could be extracted from a single genome, which part of this information is really useful on a medical care basis and which part is actionable in nowadays.

It seems to me that by now our ability to robustly correlate genotypes to phenotypes is still scarce. This is due to incomplete knowledge about causative and risk associated mutations as well as on the molecular and genetic mechanisms that lead from genetic variants to phenotypes. Studies in the last years have demonstrated that this path is not straightforward and the actual phenotypes often depend on the interaction of several genetic components and regulatory mechanisms, living aside the environmental factors.
Several disease mutations show incomplete penetrance and many example exist of variants linked to phenotypes only in specific populations, so a reliable interpretation of genomic data seems far away by now.
However, many decision can be made knowing your DNA sequence and this information will become even more interesting as researchers continue to find new associations and elucidate genotype-phenotype correlation mechanisms.
Moreover, if the health public service continues to stand against whole genome screening, people will soon turn to private companies, that can already provide this kind of services. This policy will thus increase the risk of incomplete or misleading interpretations without any kind of support from medical stuff.
A lot has to be discussed on the practical and ethical point of view, but we have to face the reality that since these kind of tests are going to became easily accessible in the near future, we have also to find a way to provide the correct information to the subject analyzed.
The topic of genomic risk assessment in healthy people has been recently discussed also on the New England Jornal of Medicine, that published a review on clinical whole exome and whole genome sequencing. The journal also presented the hypothetical scenario of a subject which discovers some cancer affected relatives and wants to undergo genetic testing. They propose 2 strategies, gene panel or whole exome/genome sequencing and the case is open for readers to comment with even a pool to vote for your preferred solution.