Removing unwanted variation in machine learning for personalized medicine

author: Terry Speed, Walter and Eliza Hall Institute of Medical Research
published: July 18, 2016,   recorded: May 2016,   views: 1189
Categories

Slides

Related content

Report a problem or upload files

If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Lecture popularity: You need to login to cast your vote.
  Delicious Bibliography

Description

Machine Learning for Personalized Medicine will inevitably build on large omics datasets. These are often collected over months or years, and sometimes involve multiple labs. Unwanted variation (UV) can arise from technical elements such as batches, different platforms or laboratories, or from biological signals such as heterogeneity in age, ethnicity or cellular composition, which are unrelated to the factor of interest in the study. Similar issues arise when the goal is to combine several smaller studies. A very important task is to remove these UV factors without losing the factors of interest. Some years ago we proposed a general framework (called RUV) for removing UV in microarray data using negative control genes. It showed very good behavior for differential expression analysis (i.e., with a known factor of interest) when applied to several datasets. Our objective in this talk is to describe our recent results doing similar things in a machine learning context, specifically when carrying out classification.

Link this page

Would you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !

Write your own review or comment:

make sure you have javascript enabled or clear this field: