<span class="live-editor-title live-editor-title-145" data-post-id="145" data-post-date="2019-04-07 06:40:50">Roche study solidifies AI’s immense value in diagnostics</span>

Roche study solidifies AI’s immense value in diagnostics

 

I have said this before, whenever speaking about AI in the context of healthcare, and here we are, with Roche, once again proving that AI has a huge potential to transform diagnostics. Now, I also strongly believe that AI (and in the current context we are talking about Narrow, Weak AI) has the potential to transform therapy and the fused emerging field, theragnostics. However, diagnostics is a low-hanging fruit, and will be first conquered by AI.

The current Roche study

Roche scientists deserve kudos for fighting genericide, having picked Diabetes Macular Edema (DME), a specific condition affecting specific cohorts within those living with Diabetes. The biggest effect of diabetes is in the secondary effects that cause more harm to patients many a time. Without appropriate diagnosis, Macular Degeneration can lead to permanent damage and consequent degeneration of sight.

The choice of a specific eye condition is not a weakness, but a very good way to conduct research, and make a comparison between AI, which uses the two-dmensional Color Fundus Photos (CFP) versus the three-dimensional Optical Coherence Tomography (OCT), which is the gold standard, and requires human expertise for successful diagnosis. Plus, as Susan Shepard lays out in her review (link below), the former technique is available through telemedicine centers, while the latter is not, as it is more expensive as well. This also has implications for underserved populations in the developing world, as diabetes is a global health condition.

AI and Imaging based Diagnostics

Another aspect that makes this particular study powerful, and also demonstrates the power of narrow AI, specifically Deep Learning is the choice of diagnostics that relies on imaging. Over the past few decades, quite a bit of research and development has focused on making Deep Learning powerful, especially when it comes to imaging viz-a-viz applications in facial recognition and more broadly, image recognition. This has resulted in knowledge in the form of algorithms and techniques that can now be successfully extended to AI.

 

 

Explainability

The authors of the study also allude to the fact that they have been able to get the algorithm to spit out its secrets as to how it is arriving at the diagnoses. This is a fundamental requirement of AI. The more an algorithm is understood in terms of how it recognizes patterns and arrives at conclusions, the more value it has in aiding diagnoses.

Explainability is critical, not just for healthcare applications, but in many fields where AI will be put to use. It has also been an ongoing challenge, and will be a significant hurdle when it comes to application. It is of such high importance that DARPA is continuing to make investments to have explainability or XAI be a functional component of AI going forward.

 

Where we go from here

AI can make diagnosis, likeĀ  many other activities and tasks faster, cheaper and more accurate, supplementing and complementing humans. As I mentioned before, this can have an effect and applications of immense value with very high Returns on Investment (ROI) for governments, private organizations, patients and stakeholders sundry. There are also several social applications where it will greatly benefit humanity.

There will be progressive churn, and the competition to make AI a market reality, already very intense is only going to grow stronger over time.

Yet, I do not expect the transition to be smooth, and there will definitely bumps in the road ahead, in terms of technological, regulatory and other challenges.

 

References:

  1. The Qmed Article by Susan Shepard: https://www.mddionline.com/study-shows-using-ai-eye-screenings-could-improve-vision-outcomes?ADTRK=UBM&elq_mid=8095&elq_cid=74447
  2. Close-up Image of the Eye: https://www.pexels.com/photo/eye-iris-anatomy-biology-8588/
  3. Image of person checking blood sugar: https://pixabay.com/photos/diabetes-blood-finger-glucose-777002/
  4. Image of X-ray: https://unsplash.com/photos/ouyjDk-KdfY
  5. Image of Doctors Collaborating: https://stocksnap.io/photo/GSBJXWRSDV
  6. Image of Subway Stairs: https://burst.shopify.com/photos/up-subway-stairs?q=stairs

The importance of providing Labels and Structure to your Healthcare Data towards AI Success

The worn cliche: Data is the new oil!

Your organization may be big. It might be small. It might be a mature business, with an approved device pipeline, or, it might be an early stage one. Your organization might have any number of adjectives describing its nature, size and characteristics. The commonality, is that, if it is a healthcare organization (or any type of organization, though we are focused on healthcare here), it has data, that can be classified and categorized in several ways, and can indicate to a point in the immediate or distant future, where it could have significant impact and ROI depending on how it is analyzed and applied in alignment with your organizational goals.

Thus, whether the cliche seems tired to you or not, there is a lot of truth to the statement that your data can be useful in several ways in the future – it could represent inherent value for investors, it could help with forensics and root cause analysis, in new designs, in optimizing performance, manufacturing, service, competitiveness and possibly in many different ways you haven’t yet imagined! This then indicates you should have a sound data strategy.

Storage is Tactical, Not Strategic

If you think, “well, our strategy is to acquire the data, store it in the cloud and back up for when we are ready for AI”, you are confusing tact with strategy. Some might argue that the act of storing data is even more banal than being tact. There might be truth to that, but let us make sure that you do have sound plans to acquire and store data with back up plans that are better than what MySpace appears to have been employing (for the Rip Van Winkles out there, I will provide a link of the recent disaster that MySpace was engaged in – essentially losing approximately a decade’s worth of its users’ music)!

So, your strategy for data needs to be a lot more than strategy.

However, do not take storage lightly! In my career, I have seen organizations with weak leadership that engaged in such stupid ways (and this is putting it kindly) of collecting and archiving data, that they lost critical and crucial, historic information relating to verification and validation!!

So, while it might seem simple, make sure your organization has well tested and trusted methods of collecting and archiving data, backed up appropriately and retrieved securely, with reasonable ease. At a minimum, you should consult with Data Scientists, if your size and budget does not permit developing a Data Science team in-house. It might be appropriate for the Quality and/or Regulatory functions within a small organization to bear the responsibility for this. For device and drug companies, regulations already require much of this, and yet you would be surprised how easy it is for organizations to miss the mark on this.

Labeling

There are several definitions for data labeling and definition of structure, as it pertains to data. I will leave this as a basic exercise for you, the reader, but I am sure, you have a commonsense understanding of what the terms allude to, and so let us use that as a yardstick to get to the thrust of this blog, rather than reducing it to pedantic and semantic exercises. Data labeling alludes to the fact that you mark data in terms of body temperature, instrument temperature, body weight, age, sex, heart rate and on and on. For anyone who has done even the simple exercise of bench top testing, it will be obvious that you already have to take a lot of care in labeling data, or the losses can be quite catastrophic.

Labeling tends to be the most expensive and time consuming exercise in preparing your data for Machine Learning, as anyone who has attended a talk by me on this topic has been clearly warned before!

Labeling should be understood in broad terms that go beyond merely the titles of columns and rows on a spreadsheet. For instance, being able to pull several spreadsheets, and readily understand which data came from acute or chronic studies, bench-top or human trials, real vs. simulated data must all be considered. Confuse these, and your machine learning might be trained with the wrong data and the results can be quite undesirable.

Data Labeling should also be very clear at every level and this cannot be stressed enough! If you have data collected by 4 thermocouples, two years from now, when someone pulls up the data for instance, they should be able to identify exactly where the data came!

In effect, you should establish a discipline of rigorous labeling for data, from the macro levels, going into the individual levels, through your SOPs, Work Instructions, Forms and make this a part of organizational culture.

Structuring Data

Data can be classified into Structured Data and Unstructured Data. Structured Data is reflected by the type of data discussed above – data obtained manually, through machines automatically, by programming, through logs, etc. (There is a note about data logs I will make in a moment). Unstructured Data on the other hand, could be data coming from sources such as online customer reviews (if you think this is not for you, just search for reviews on IUDs or Heart Burn Medications on the internet and you are in for a surprise), physician or nurse feedback, complaint information, procedure videos and other sources. Depending on the device, drug, biologic, app or the combinations thereof, your unstructured data can take many forms, and you should develop a very deep understanding of these sources, and how to collect, and structure this data as much as possible.

While, we all assume that some day, in the near or far future, depending on the various predictions, Artificial General Intelligence will allow us to completely skip data labeling (Neural Networks can do this to an extent currently), and analyze data, providing insights. However, there are limitations with current Machine Learning and Deep Learning Approaches, and you are better off providing as much structure to your data as possible.

For instance, if you are capturing procedure videos, whether on models or on humans, you might benefit from using fiducial markers for instance. Such markers are common in imaging systems already. Your device could also have identification markers on it, that show position, location etc., making human and machine understanding fast and easy. This can be very helpful when dealing with explainability and machine training. This could essentially convert the status of your data from ‘unstructured‘ to ‘semi-structured‘ or, if you can further template and refine it, ‘structured‘. Of course, you might hit a point where you get diminishing returns, or, as is the case, the machine can handle semi-structured data well enough.

Identifying the sources of your unstructured data, and striving to provide structure to your data, early, and strategically, will ensure that Data Analysis and Machine Learning Projects will be less expensive and better facilitated.

A note on Machine Logging

I was once involved in a project where the capital equipment logged critical data such as currents, voltages and other data associated with procedures. Various regulations require this, and for Failure Analysis, Adverse Event Investigation, etc. this is generally very essential. The fundamental problem with this (and others I have seen) particular data logging was the lack of foresight and poor implementation. Starting from the very labeling of the log files, to the time stamps, to the data collection frequency and various other features, it was a complete mess. The data had to be pulled from the logs into a spreadsheet program, and a number of routines had to be run on it, to make it passably useful. I used to wonder how a diligent auditor would take to the cumbersome and questionable logging, if he or she were to stumble on it!

Worse, the project leader prided himself in thinking that he had created an innovation in translating the gibberish into a-little-less-gibberish through the scripts! That sort of paucity in planning and respect for data will not help you, if you are to succeed.

The Bigger Picture

I want to make it clear. This is not just exhortation to make machine logs readily readable. And yes, your software team MUST do this. The point here is awareness, education, discipline and diligence towards data. Your organization needs to strive towards all of these aspects – and two more, which follow in the next section. It is not just leadership, but it is your entire organization that needs to focus on this for success.

Patient Privacy and Data Security

Increasingly, ranging from Government Organizations to Data Silo Owners and even other third parties, several organizations are admitting to failures in securing data, exposing data belonging to stakeholders. As you can imagine, exposing your patients’ data, erodes trust and is unforgivable, if the root cause for this is carelessness and sloppy, outdated practices.

Even when patient privacy is not involved, losing data of any other kind, such as test data, designs, manufacturing data etc. can be catastrophic to your competitiveness, and can make or break your organization.

Efforts to secure your data and protect your patients can never cease, and there can never be complacency in any of these areas!

Conclusion

The larger conclusion I want you to draw from my post here is this: you need a Data Strategy. Within that strategy lie elements such as labeling, structuring, security and privacy, none of which can be ignored. Your organizational members need to be educated of the importance of your strategy and they must have a healthy respect to help formulate, improve and maintain the strategy and discipline it will take you to succeed.

References:

  1. Tactic Vs. Strategy:Ā https://en.wikipedia.org/wiki/Tactic_(method)
  2. Image of Tags: https://www.pexels.com/photo/several-assorted-color-tags-697059/
  3. Image of Book Stacks: https://www.pexels.com/photo/pile-of-books-on-gray-metal-rack-1853836/
  4. Image of Escalators: https://www.pexels.com/photo/building-escalator-1769356/
  5. Image of Labeled Beans: https://www.pexels.com/photo/assorted-beans-in-white-sacks-1024005/
  6. Image of Laptop: https://www.pexels.com/photo/low-angle-view-of-human-representation-of-grass-296085/
  7. Data Loss at MySpace (An NY Times Subscription may be required): https://www.nytimes.com/2019/03/19/business/myspace-user-data.html

Interesting AI & Immune System Initiative by Microsoft and Adaptive Biotechnologies that leaves behind some questions

It is no secret that the larger technology companies including Google, Microsoft, nVidia, IBM, Apple and others want to dominate AI, as well as healthcare, in an ever expanding competitive landscape. While it is anyone’s guess if they will succeed, or get upstaged by smaller, nimbler firms in either arena, the moves they make are definitely interesting to watch. A lot of the moves appear benign, but could lead to cannibalization, such as the “AI Contests” some of the organizations put up (more posts to come on this).

Partnerships can go both ways I suppose, and are probably a strategic way to externalize any risk of failure. In that sense, in the current example, both Microsoft and Adaptive Biotechnologies appear to want to play up their strengths. The premise of what they want to do with the partnership is also quite intriguing. You can read it from the horses’ mouths in the links provided below. I will summarize them and lay out a couple of thoughts that come to my mind.

Essentially, the project would turn the body’s immune system itself into the data source for diagnosis. Because every time the immune system responds to a disease, T-Cell receptor (TCR) proteins are expressed to combat antigens. Mapping the TCRs, through a simple blood test, as Microsoft and Adaptive postulate, can go a long way in early diagnoses of an array of diseases. To say the least, the project is ambitious, and here are a few thoughts:

  1. Accurate diagnosis and personalized therapy require knowledge of the state of the human body and its disease. Simply mapping the genetics of a person, and considering their epigenetics and their lifestyles, etc. is complex enough, but it still might not be enough! Thus, TCRs could be mapped and allow for quicker diagnoses, if the theory pans out on a large-scale. It is yet unclear to me that mapped TCRs can actually yield the necessary diagnostic clues, machine learning or not, for a larger variety of diseases. However, it might supplement diagnostic efforts alongside genetics, epigenetics and other health data sources.
  2. From a business angle, I also find this to be intriguingly different from the general bedlam of text processing a la Watson, and all the algorithms rushing to read and reinterpret imaging as with nVidia and others. Microsoft has appeared to have looked for and found a partner with a unique approach to the application of machine learning in healthcare.
  3. Any large set of unknown targets, powered by data might appear to be a classic problem for machine learning to solve. However, Microsoft and Adaptive (Microsoft has invested in them now, apparently) might have joined a Kool-Aid club that bridles the horses behind carts. What I mean when I array out those cliches is this: medicine already has a problem of knowledge paucity when it comes to diagnoses, until more clarity becomes available, by way of a progressed disease. This is fundamentally because disease precursors are poorly understood, from want of clinical research, not lack of intelligence. What is to say the TCRs won’t just set off an array of false and confounding alarms? Yes, with liquid biopsy and other such hyped up methods out there, the industry is in a rush for quick fixes. It might well turn out that this is much harder to resolve, with clinical studies and protocols that will require to demonstrate that TCR expression, their proportional presence, etc., do truly indicate the preliminary stages of a disease being present in the body. I am not convinced yet.
  4. There is also a maddening rush out the gate to define universal tests with pinpricks of blood. While I am not suggesting we practice good old, barbaric bloodletting (although there are all kinds of people out there “thirsting” for a comeback to this practice), I think this is an unnecessarily over-constrained problem definition, perhaps making titillatingĀ fodder for press releases and blog posts. There might also be an urge to combine these pinprick tests with diabetes monitoring and such. While it is tempting to fantasize about such possibilities, and at some point, these might come to fruition, there is no need to go to such extremes before solving fundamental problems in medicineaccurate diagnosis and targeted therapy. For example, when should a person’s blood be drawn? How frequently? Would the frequency vary when a certain set of TCRs are observed? There are so many things to worry about here. I would think companies would stop using overly broad terms such as “universal”.
  5. In my posts here, in my talks, daily discussions and so on, I always come back to a few bugaboos. Who will own the TCR mappings? Who owns the product of the machine learning algorithms? Will they be patented and bridled off? How will such diagnostic methods be regulated? Validated? Will Explainable AI, something I expect to be a fundamental principle that should be applied to healthcare be required (see explanation from DARPA linked below) and used judiciously? And on and on we go.
  6. Data has been walled off quite well in the healthcare industry up to this point. Yes, we got the human genome, but much much more sits behind curtains and masks and other cliches you can think off, that every new technology that promises to expose and dig through data always concerns me, surrounding ownership.
  7. The “Theranos” Effect:Ā If you are like me, and know about the story of Theranos, you are still sitting up at nights, jaws dropped, wondering how in the hell, the company is still in vogue (I have written about this on my medical devices blog, in fact, using the same Pixabay image! See link below). I have also linked, one of several dozen well-written write ups that offer us a tale of caution, and I plan to call this, as I have named it, the “Theranos” Effect. In summary, this company went into the “over promise, and extreme under-delivery” (or never delivery, to date..) business. They engaged in egregious and unethical business practices, fooling the industry, investors, partners and more, along the way. How do we make sure that with all the promise of AI, companies don’t make such ugly incidents repeat themselves? Mind you, this is not me pointing fingers at Microsoft. I think this is indeed a great effort. I am just offering this up as an important tale of caution, for people in healthcare, and in any industry for that matter. I understand, as much as anyone else that businesses need hype to push their products. However, it would behoove you to make sure you don’t push things off a cliff…

CONCLUSION

In mankind’s march towards the goal of a healthy future for all, we have many strides to make. Naturally, we want to be as accurate and as thorough, yet economic as we can. Therefore, we rely on technological breakthroughs on one end, such as anything ranging from improvements in basic science, to sensors and AI, and on the economics of lower thresholds for test materials consumed, time to diagnosis and other aspects on the other end. What Microsoft and Adaptive aim to do with their (investment based) symbiotic looking partnership is commendable. It may take us one step closer to our goal, but it may not be the one to take us there at all. Only time can tell, and in the meanwhile, I hope commonsense and ethics prevail over hype and fantastic marketing materials.

Should you have something to add, please leave a comment below.

Subscribe and Support, Please!Ā 

Did you enjoy this post? Please subscribe for more updates, using the sidebar. Have ideas or blog posts youā€™d like to see here? Contact me at yamanoor at gmail dot com.

References:

  1. The Microsoft Blog Post:Ā https://blogs.microsoft.com/blog/2018/01/04/microsoft-adaptive-biotechnologies-announce-partnership-using-ai-decode-immune-system-diagnose-treat-disease/?imm_mid=0fa701&cmp=em-data-na-na-newsltr_ai_20180115
  2. Adaptive’s Press Release:Ā https://www.businesswire.com/news/home/20180104005464/en/Adaptive-Biotechnologies-Announces-Partnership-Microsoft-Decode-Human
  3. DARPA, on Explainable AI:Ā https://www.darpa.mil/program/explainable-artificial-intelligence
  4. Vanity Fair on Theranos: https://www.vanityfair.com/news/2016/09/elizabeth-holmes-theranos-exclusive
  5. Myself, writing with incredulity on Theranos’s longevity:Ā http://chaaraka.blogspot.com/2017/12/theranos-lives-to-die-another-day.html
  6. Image, Courtesy, Pexels+Pixabay:Ā https://www.pexels.com/photo/white-and-clear-glass-syringe-161628/

The FDA Designates an Eye-Diagnostic System as a “Breakthrough Device”

 

Last year, in my talk at the San Jose BIOMEDevice Conference, I had postulated that one of the key factors in the rate and level of adoption of AI based devices and systems would depend on how the FDA would act on such submissions. It is good to see that one company seems have to have proceeded with taking the FDA to task. The result is positive, and good news for companies pushing AI based systems, but also carries a tale of caution, for the road ahead for everyone involved. I will lay out the case briefly in this blog, but this is not the last time you will hear me talk about the effect of regulatory agency response to such device submissions and their long-term implications.

I just reviewed news that IDx, a company that designed what is being described as “an AI based autonomous diabetic retinopathy” detection system has been given “breakthrough” status by the FDA for their system. This provides the system with an expedited review, and potentially quick approval. This is very encouraging, not only for the organization with the retinopathy system, but also for all companies vending AI systems.

Here are a couple of key thoughts I have on this:

  1. A very promising outlook:Ā I am not aware of how much Scott Gottlieb had, in terms of a personal hand in this decision, but I am sure his leadership had a measurable impact on the decision to expedite review. He has taken a stance of modernizing some of the approval processes, especially in the area of Digital Health. The FDA’s backing for AI based systems, is a very positive thing for companies with an AI focus on healthcare.
  2. Good for competition in the healthcare industry:Ā Once the FDA sets AI based apps and systems on the path to commercialization, there will be no excuse for organizations that do not adopt Machine Learning and Neural Networks. Of course watch for a good spike in M&A activity in this regard.
  3. The Hype Cycle is around the corner:Ā While companies with real ML/DL products and services will be out there, helping patients, there will also be the fakes and the wannabes that we will have to contend with. And it is quite possible that the rotten apples will ruin it for everyone. This is something to watch for.
  4. Unrealistic expectations and unintended consequences:Ā I know this is closely related to the point about the hype cycle, but I want to make a refined point here. Even the companies with good AI based tools might push the envelope on hyping up the utility and efficiency of their products. Over-promise and under-delivery lead to angry and disappointed customers and eventually will create problems for everyone in the industry.
  5. How will CMS and private insurance companies handle reimbursement? This is a big question and can be the topic (and probably will be) of several blog posts. However, I just wanted to highlight that while AI might improve efficiency and automate diagnoses and treatments, companies have to resolve how and where the reimbursement will come from.
  6. A word of caution on regulation:Ā While it is commendable that the FDA is moving forward with initiatives on things such as AI. However, I do not always agree they are doing a great job on things. I was irked by the craze with which the orphan applications were being reviewed by the FDA. The FDA, in the past, has gotten too cozy with the industry and Congress, with terrible consequences. Those who remember the Menaflex incident (reference below) remember how quickly the public’s trust in the FDA eroded, and how the organization took years to recover from that. Therefore, there is cause for concern here, and one hopes there will be a balancing act in making sure that the approval processes remain rigorous enough.

Conclusion:Ā The current “breakthrough” designation and fast track review show great promise, but one hopes the FDA will balance it out with an equal dose of caution.

Subscribe and Support, Please!Ā 

Did you enjoy this post? Please subscribe for more updates, using the sidebar. Have ideas or blog posts you’d like to see here? Contact me at yamanoor at gmail dot com.

References:

  1. The Press Release on IDx:Ā https://www.eyediagnosis.net/single-post/2018/02/05/FDA-to-expedite-review-of-IDx-DR-a-breakthrough-AI-diagnostic-system?lipi=urn%3Ali%3Apage%3Ad_flagship3_feed%3B6Bf9xSfCRQSJXa4nvXjScA%3D%3D
  2. On what happened with Menaflex:Ā https://www.fda.gov/downloads/NewsEvents/PublicHealthFocus/UCM183642.pdf
  3. Image, Courtesy, Pexels:Ā https://www.pexels.com/photo/people-face-child-eye-32267/

Personalized Medicine might be the true ticket for AI’s march on Healthcare

I came across this interesting piece on siliconANGLEĀ  focused on how big data is coming to cardiovascular medicine, via a precision medicine initiative. I found the read to be fascinating, and I recommend you read the article, provided in the references section below.

In summary, the American Heart Association (AHA) is collaborating with Amazon Web Services to build a data analytics platform. Details are thin, so I am unsure of how much is hype and how much is reality. I assume, as time progresses, it will indeed turn out to be very useful. The article describes the platform as a “marketplace” (yes, concerning, in more ways than one) for various forms of patient data:

  1. Personal Data
  2. Clinical Trial Data
  3. Pharmaceutical Data
  4. Hospital Data, and apparently, other data.

The goal here is to use this to personalize treatments for patients. Read more in the article. I believe that personalized medicine and AI share a deep connection. I want to share some of my own thoughts of how such efforts could give way to lasting influence by AI on healthcare:

  1. Personalized medicine needs knowledge, which comes through data. Similarly, learning requires data. Data helps you glean patterns, which is what learning is. Thus, AI and Personalized medicine are the two snakes of Caduceus, which in this case is the data itself. Thus, efforts to bring about personalized medicine should be expanded vastly.
  2. Knowledge is supreme. When it comes to personalized medicine and such, you hear all this hype of n=1. Unfortunately, for medicine to be effective, first you need large quantities of data, and with that the ability to glean actionable information from the data, to understand diseases, treatments, and their effectiveness. Whether you see this on a gross scale, or all the way down to a Single Nucleotide Polymorphism (SNP), this type of knowledge needs to be the horse in front of the cart. Without actionability, the patterns recognized through the data might not be useful. This knowledge doesn’t exist, so the hype cycle might cause enough damage and reduce funding levels later on. Realistic expectations ought to be set so that this doesn’t happen.
  3. There are two problems with healthcare data that come to the fore. First, we don’t really know that is being collected is appropriate. There is a lot of day dreaming about genetic data, racial data, physiological parameters (one company claims to be measuring 300 parameters!)Ā  and other measurands and the hope that these will somehow magically transform themselves into diagnoses and treatments. With enough research, they will, but this will take expensive effort and time. Low hanging fruits such as breast cancer diagnoses, etc., are pushing people to imagine that the same level of rapid success will be seen throughout the healthcare continuum. There is no way to know if this will be the case.
  4. The second problem with healthcare data is about ownership and walls around the data. For example, will AHA share the data with the general public? If not, why not? Terms like “marketplace” give me pause. They might very well be just terms used to explain things colloquially. On the other hand, it is important to think about the issue of data sitting behind paywalls, or blocked of with HIPAA as a convenient excuse (more posts to come on all this). Data Democracy is a critical need for the progress of medicine and for Artificial Intelligence, eventually.
  5. What if the data is not free? What about the insights? Do we suddenly now start paying for the insights? Or, would CMS start reimbursing Doctors for personalized medicine? For the use of AI tools? There are several questions here, and only time will answer them.

Conclusion:Ā 

Data will give us clinical insights. These insights will eventually fragment into meaningful personalized healthcare. The same data will lead to even more advances with AI. However, unanswered questions over data ownership and access remain.

 

References:

  1. https://siliconangle.com/blog/2018/01/22/can-precision-medicine-break-chokehold-on-healthcare-big-data-reinvent-womenintech/
  2. Image, Courtesy Pexels:Ā https://www.pexels.com/photo/interior-of-office-building-325229/

 

Machine Learning shows promise in Dementia prediction

Rubik's Cube

I was flipping through some archives, and found this Scope (a Stanford University, School of Medicine Publication) article that delineates a machine learning tool (link below, Scope calls it AI, the authors of the tool, in their paper, also linked below, rightfully categorize it as ML, a subset of AI as we generally describe it). I always love when you have access to the paper linked to a study. It always makes things easy.

So, the folks at McGill, trained an ML system using PET scans from people who demonstrate symptoms of mild cognitive impairment, to see who among them would develop Alzheimer’s, given that not all of them do. They taught the system to focus on the elevated protein expression in specific brain regions to train and make predictions.

Used on an independent set, the tool had an 84% prediction accuracy of dementia progression. Read more in the paper. I want to share a few thoughts below.

  1. I think tools like this will become the norm over time. However, right now, they lack the kind of standardization and maturity required for integration into clinical practice. I don’t mean to state that in a negative sense. Such efforts take time, effort and funding, of course.
  2. An 84% percent efficiency is not enough, not even for a supporting tool, not even when humans are completely in charge. This is also achieved through training with large data sets, the use of better algorithms and other improvement methodologies. This could also use some standardization, that can then be spread to all ML, DL and AI tools, which use imaging for diagnostics in healthcare.
  3. The future should consist of such tools passively (and when necessary, actively) siphoning your imaging and other data off your EHRs, and then parsing them to see if predictions can be made. This however, requires more groups such as the ADNI (Alzheimerā€™s Disease Neuroimaging Initiative), from whose participants the imaging and other data was used, collaborations from hospitals, insurance companies and governments.
  4. To improve diagnosis across ages, sexes, races and other discriminating factors, global co-operation would be required.
  5. Of course, we need to take various types of data, ranging from imaging to genetics, to epigenetics and other sources to make diagnosis quite efficient. Perhaps, this combination is one way to get around the 84% efficiency in this tool, till a time comes when imaging alone produces better results. At that point, say you make predictions based on imaging, genetics, lifestyle and other factors, and they all chime in. You can probably use whatever interventions are available (this is a key factor, missing in all the hype about machine learning. You learn something, yes, but what do you DO?) to delay, treat and cure patients.

When I find more such interesting studies, I will share similar and other thoughts on Machine Learning, Deep Learning and AI, and their impact on Healthcare.

Subscribe and Support, Please!Ā 

Did you enjoy this post? Please subscribe for more updates, using the sidebar. Have ideas or blog posts you’d like to see here? Contact me at yamanoor at gmail dot com.

References:Ā 

  1. The Stanford SCOPE Article:Ā http://scopeblog.stanford.edu/2017/08/29/artificial-intelligence-can-help-predict-who-will-develop-dementia-a-new-study-finds/?imm_mid=0f5d75&cmp=em-data-na-na-newsltr_ai_20170904
  2. The McGill Paper:Ā http://www.neurobiologyofaging.org/article/S0197-4580(17)30229-4/fulltext
  3. Image courtesy, Pexels:Ā https://www.pexels.com/photo/brain-color-colorful-cube-19677/

Interesting AI+Health Imaging Investment and a few thoughts on where the field might be headed

 

This morning, a LinkedIn contact of mine shared an interesting MassDevice article about “Bay Labs”, a company focused on bringing Deep Learning to Cardiovascular Health Imaging (though they define themselves in broader terms as an algorithm company). It appears that Khosla Ventures (after the same much-hated Khosla from California), an existing investor (which must mean seed funding) has backed the organization with $6mn in series A funding.

What is interesting

  1. So far, what we have seen is, the really big companies, such as nVidia and G.E. working on backing up imaging with AI. They even recently announced a collaboration. This would have been a scary scenario for start-ups trying to break into the space. Even now, one should not take too much comfort. Bay Labs apparently won in an AI company contest conducted by nVidia. It clearly shows that big players are keeping an eye on the start up space, potentially looking to scoop up promising organizations.
  2. Though Bay Labs likes to talk about themselves in somewhat general terms; they state this on their website, “In order to serve the largest number of people, we aim to amplify the benefits of deep learning by providing high-performance algorithmic capabilities to assist with healthcare challenges on a global scale.”, they are at least currently focused on Deep Learning, as it relates to Imaging for Cardiovascular Health. Don’t get me wrong, I like both ends of the spectrum displayed here. They appear to have a broad vision of where they want to end up, and yet, as is important for a startup, they have a somewhat keen sense of focus.
    1. Given that, even with all the healthcare and lifestyle influences urged on by healthcare systems, governments, non-profits and others, cardiovascular diseases remain the cause behind the most deaths in the US and around the world. Thus, even within just this field, in imaging alone, the opportunities abound.
    2. Plus, cardiovascular health also happens to be one of the most expensive, and is a low hanging-fruit for startups, expecting to break into a new field and work towards financial success. The other field is probably oncology, which however is crowded with Watson and others. Still, I expect at least a handful of startups focused on various subsets of oncological diagnosis and treatment to emerge out of the woods as time progresses.
    3. It is also one of the fields where practitioners are prone to adopt cutting edge solutions.
  3. Bubbles:Ā Bay Labs appears to be an organization founded on some sound principles. The next few should probably follow along the same path. One has to wonder though, at what point, in the not too far futureĀ  will we see organizations touting AI-this and AI-that, without any sound basis? This should be interesting to track.
  4. Algorithm Patents: The debate on whether software and algorithms can be patented, and under what conditions was never properly settled (what, ever is?). A new wave of companies touting algorithms will now come along, and generally speaking, take each other to court, giving headaches to Luddite Justices, such as in the US Supreme Court who are still hemming and hawing about allowing basic AV equipment within the courts (ironically AI has been trained to predict how individual US Supreme Court Judges would rule, with fairly good efficiency). This whole space is something to watch closely.
  5. Assets:Ā Besides patents, there are other things to be concerned about here. Who ultimately owns the data? The results from what the Machine Learning and Deep Learning algorithms find? Are the findings marketable? If not, who pays to get this done? How do governments, healthcare systems and private organizations, all of which form the backbone of medicine share the costs and the benefits?
  6. Mergers & Acquisitions: Once the hype cycle is ridden and who owns what is also a bit more clear, comes the M&A song and dance. This also should be interesting to watch.
  7. Interoperability:Ā Data interoperability in heatlhcare, is already a major headache. This is currently one of my core research areas. I am deeply interested in unraveling the challenges of how data should be cleaned and labeled to be used in AI applications. On the surface, it appears that all you have to do is collect the data, just transform it for various purposes – to upload to EHRs, to use in Machine Learning and so on and then get on with it. The underlying issues are much more complicated. Who collects the data? How is it blinded? Who pays for all this? How does the learning get applied? How do you glean precision medicine uses out of all this? And, many, many more…
  8. Regulatory Dilemmas:Ā I have a series of blog posts planned on this topic, but let’s take a quick dive here. How do you get approvals for such AI systems? How do you fix problems? Should improvements, fed by data volume and algorithm changes require regulatory updates? What happens when an algorithm produces defective results? These and many others problems will need some type of resolution, and will create new problems and opportunities of their own.

Summary:Ā In summary, I’d say, news of companies such as Bay Labs is good. However, headwinds lie ahead for them, and everyone in industry. We are looking at huge paradigm shifts and thus, not much is predictable. New issues will unfold while ones that seem dire right now, might become trivial. We are in for the ride of our lifetimes!

References:

  1. The MassDevice Article:Ā http://www.massdevice.com/cardiovascular-ai-imaging-dev-bay-labs-raises-6m-series/?lipi=urn%3Ali%3Apage%3Ad_flagship3_feed%3BMBJbTpHMSb6f%2B2WzYDP%2BUA%3D%3D
  2. Bay Labs:Ā https://baylabs.io/
  3. Image, Courtesy Pexels:Ā https://www.pexels.com/photo/low-angle-view-of-spiral-staircase-against-black-background-247676/

Slides – How Artificial Intelligence is Changing Medical Devices

The main goal of this site is to air my thoughts on how new technologies and paradigms impact the practice of medicine. In relation to this, I made a presentation, just yesterday, December 6, 2017, on the topic of Artificial Intelligence, and the growing influence of this discipline on the field of medical devices.

The slides will be available to download this through the site.

I do have to caution you. I do not present a lot of text in the slides. I have written some notes on the slides, but together, they still will not be enough to understand all of what I spoke about.

However, do not worry. I will blog about each slide, concept and use-case presented on this presentation over the next few weeks and months. Subscribe for updates and stay tuned! In addition, if you feel you have a set of burning questions and need them answered as soon as possible, do feel free to contact me and I will answer them as early as possible.

The download link:Ā https://drive.google.com/file/d/12o1DHzJ2f_jNJi8MzZFPqhOwydVoDU52/view?usp=sharing