MEDICAL IP, AI body composition analysis solution 'DeepCatch' medical device class 2 certification
MEDICAL IP, AI body composition analysis solution ‘DeepCatch’ certified as medical device 2nd grade
▶ Obtained MFDS certification for ‘DeepCatch’, a medical image analysis software that automatically calculates body composition analysis results
▶ DeepCatch is being used in a wide range of researches such as cancer, liver disease, obesity, sarcopenia, and osteoporosis at medical institutions around the world
<2020-12-22> MEDICALIP acquired medical device grade 2 certification for ‘DeepCath’ from the Ministry of Food and Drug Safety.
DeepCatch is a medical image analysis software that uses AI deep learning technology to automatically analyze whole body components, such as bone, muscle, fat, etc., from CT images and calculate the volume and area of each component.
When the whole body CT image is uploaded, the body composition is divided into 7 structures (skin, bone, muscle, visceral fat, subcutaneous fat, organs, and cerebrospinal cord). In addition, automatic analysis reports are provided for muscle and fat volume and area, abdominal circumference, and body fat percentage.
According to a company official, changes in body composition are directly related to metabolic diseases and sarcopenia, which are closely related to the prognosis and mortality of patients. As research on this is being actively conducted in many countries around the world, inquiries for DeepCatch Demo are continuing from domestic and foreign medical institutions, as well as global medical equipment companies and clinical trial institutions.
LATEST NEWS
Oct 25 / 2024
MEDICAL IP participates in HLTH 2024, showcasing X-ray based AI solutions
Oct 23 / 2024
MEDICAL IP and kt wiz sign partnership to enhance player health management
Oct 4 / 2024
MEDICAL IP unveils innovative AI solutions at K-hospital, marking visible achievements
Sep 30 / 2024
MEDICAL IP Showcases New Medical Education Possibilities with AI & Metaverse at the ‘Medical Metaverse Conference’