AI-Powered Medical Imaging Solutions Advance Healthcare

The use of edge AI in medical imaging offers the possibility of enormous benefits to stakeholders throughout the healthcare sector.

On the provider side, edge AI imaging can improve diagnostic accuracy, boost physician efficiency, speed case processing timelines, and reduce the burden on overstretched medical personnel. Patients benefit from shorter wait times for their diagnostic test results and a better overall quality of care.

But it can be challenging to develop AI-powered solutions needed to make this promise a reality. The computing requirements to implement edge AI in medicine are high, which has historically made it both difficult and expensive to obtain adequate computing resources. It can also be hard to customize the underlying hardware components well enough to suit medical imaging use cases.

It’s a frustrating situation for anyone wanting to offer innovative AI-enabled imaging solutions to the medical sector—because while the market demand certainly exists, it’s not easy to build products that are effective, efficient, and profitable all at the same time.

But now independent software vendors (ISVs), original equipment manufacturers (OEMs), and system integrators (SIs) are better positioned to innovate edge AI-enabled medical imaging solutions. The prevalence of rich edge-capable hardware options and the increasing availability of flexible AI solution reference designs make this possible.

AI Bone Density Detection: A Case Study

The AI Reasoning Solution from HY Medical, a developer of computer vision medical imaging systems is a case in point. The company wanted to offer clinicians an AI-enabled tool to proactively screen for possible bone density problems in patients so that timely preventive steps could be taken.

They needed an edge AI deployment that would put the computational work of AI inferencing closer to the imaging devices, thereby reducing network latency and bandwidth usage while ensuring better patient data privacy and system security. But there were challenges.

The edge computing power requirements for a medical imaging application are high due to the complexity of the AI models, need for fast processing times, and sheer amount of visual data to be processed.

In addition, special challenges involved developing an AI solution for use in medical settings: an unusually high demand for stability, the need for waterproof and antimicrobial design elements, and the requirement that medical professionals approve the solution before use.

The solution can automatically measure and analyze a patient’s bone density and tissue composition based on the #CT scan data, making it a valuable screening tool for #physicians. HY Medical (Huiyihuiying) via @insightdottech

HY Medical leveraged Intel’s medical imaging AI reference design and Intel® Arc graphics cards to develop a solution that takes image data from CT scans and then processes it using computer vision algorithms. The solution can automatically measure and analyze a patient’s bone density and tissue composition based on the CT scan data, making it a valuable screening tool for physicians.

The solution also meets the stringent performance requirements of the medical sector. In testing, HY Medical found that their system had an average AI inference calculation time of under 10 seconds.

Intel processors offer a powerful platform for medical edge computing, which allows the company to meet its performance goals with ease. Intel technology also provides tremendous flexibility and stability, enabling the wide-scale application of this technology in bone density screening scenarios.

Reference Designs Speed AI Solution Development

HY Medical’s experience with developing their bone density screening solution is a promising story—and one that will likely become more common thanks to the availability of AI reference designs. These reference architectures make it possible for ISVs, OEMs, and SIs to develop medical imaging solutions for a hungry market both quickly and efficiently.

Intel’s edge AI inferencing reference design for medical imaging applications supports this goal in several ways:

Tight integration with high-performance edge hardware: Ensures that solutions built with the reference design will be optimized for computer vision workloads at the edge. The result is improved real-world performance, better AI model optimization for the underlying hardware, and increased energy efficiency.

Flexible approach to AI algorithms: Because different software developers work with different tools, multiple AI model frameworks are supported. Models written in PyTorch, TensorFlow, ONNX, PaddlePaddle, and other frameworks can all be used without sacrificing compatibility or performance.

AI inferencing optimization: The Intel® OpenVINO toolkit makes it possible to optimize edge AI models for faster and more efficient inferencing performance.

Customized hardware support: The reference design also factors in the special needs of the medical sector that require customized hardware configurations—for example, heat-dissipating architectures, low-noise hardware, and rich I/O ports to enable connection with other devices in clinical settings.

The result of reference architectures such as this one is that they shorten time-to-market and reduce the inherent risks of the product development phase, giving innovators a clear path to rapid, performant, and profitable solution development. That’s a win for everyone involved—from solutions developers and hospital administrators to frontline medical professionals and their patients.

The Future of AI in Medical Imaging

The ability to develop innovative, tailored solutions quickly and cost-effectively makes it likely that far more AI-enabled medical imaging solutions will emerge in the coming years. The potential impact is huge, because medical imaging covers a lot of territory—from routine screenings, preventive care, and diagnosis to support for physicians treating diseases or involved in medical research.

Hospitals will be able to use this technology to improve their medical image reading capabilities significantly while reducing the burden on doctors and other medical staff. The application of edge AI to medical imaging represents a major step forward for the digital transformation of healthcare.

 

Edited by Georganne Benesch, Editorial Director for insight.tech.

Patient-Centered AI Redefines Continuum of Care

Healthcare professionals have a singular mission: provide the best possible care for patients. But from admittance to discharge and everything in between, they face countless challenges.

Persistent staff shortages, constrained resources, and tight budgets are just a few. The greatest challenge is access to essential information about a patient’s condition throughout their hospital journey, specifically the second-by-second time series waveform data generated from the biomedical devices monitoring a patient. When seconds matter, how do hospitals harness this data and make it easily accessible to their healthcare teams?

Why Time Series Data Matters

The answer to this challenge is a single, open platform that continuously collects, processes, and unifies disparate data and presents it to clinicians in real time. Take, for example, an eight-hospital system in Houston that was confronting staffing issues and limited provider coverage—especially overnight. This forced difficult decisions, like hiring more travel nurses and physicians or turning patients away. All that changed when the organization implemented the Sickbay® Clinical platform, a vendor-neutral, software-based monitoring and analytics solution, from Medical Informatics Corp. (MIC).

The platform enables flexible care models and the #development and deployment of patient-centered #AI at scale on a single, interconnected architecture. @sickbayMIC via @insightdottech

Sickbay is an FDA-cleared, software-based clinical platform that can help hospitals standardize patient monitoring. The platform enables flexible care models and the development and deployment of patient-centered AI at scale on a single, interconnected architecture. Sickbay redefines the traditional approach of storage and access to static data contained in EMR systems and PACS imaging. The web-based architecture brings near real-time streaming and standardized retrospective data to care teams wherever they are to support a variety of workflows with the same integration. This includes embedded EMR reporting and monitoring data on PCs and mobile devices.

“Out of about 800,000 data points generated each hour for a single patient from bedside monitoring equipment, only about two dozen data points are available for clinical use,” says Craig Rusin, Chief Product & Innovation Officer and cofounder at MIC. It’s not widely known that alarms from non-networked devices such as ventilators outside of a patient’s room are difficult for staff to hear or view remotely. Similarly, current patient monitoring doesn’t use AI tools with the existing data to inform patient care.

Measuring the Impact

Hospitals and healthcare systems using Sickbay have redefined patient monitoring and have created a new standard of flexible, data-driven care by demonstrating the ability to:

  • Rapidly add bed and staff capacity while creating flexible virtual care models that go beyond traditional tele-sitting, admit, and discharge.
  • Provide more near real-time and retrospective data to staff already on unit, on service, or on call to improve their workflows and delivery of care.
  • Create virtual nursing stations where one nurse can monitor 50+ patients on a single user interface across units and/or facilities.
  • Leverage the same infrastructure to create virtual command centers that monitor patients across the continuum of care.

No matter the method of deployment, Sickbay gives control back to healthcare teams and provides direct benefit back to the hospital. Benefits reported include reduced labor, capital, and annual maintenance costs as well as improved staff, patient, and family satisfaction. Most important, clients using Sickbay see direct impact on improvements in quality of care and outcomes, including reductions in length of stay, code blue events, ICU transfers, time on vent, time for dual sign-off, and time to treat.

Results such as these provide the pathway for other hospitals to rethink patient monitoring and realize the vision of near real-time, patient-centered AI. Healthcare leaders have proven that going back to team-based nursing by adding virtual staff can help reverse the staffing crisis. “This isn’t about taking nurses away from patients. This is about taking some of the tasks and centralizing them,” says Rusin. “There will never be enough nurses, physicians, and respiratory therapists to cover all of the demand required for the foreseeable future. We need to get bedside teams back to bedside care. Flexible, virtual care support makes that a reality.”

Changing the Economics of Care

Sickbay provides the ability to change the economics of monitoring patients and directly impact improvement in quality and outcomes.

The ability to integrate with different devices, regardless of function or brand, is the key. “We have created an environment that allows our healers to get access to data they have never had before and build content on top of that, in an economically viable way that has never been achieved,” Rusin says.

For healthcare providers, having the data available is game-changing, says MIC EVP of Strategic Market Engagement, Heather Hitchcock. As one doctor commented: “In a single minute, I have to process 300 data points. No machine is ever going to make a decision for me, but Sickbay helps me process that data faster so I can make the right decision and save more lives.”

From Scalable Patient Monitoring to Predictive Analytics

Sickbay’s value extends beyond near real-time patient monitoring and virtual care to long-term treatment improvements. Sickbay supports the ability to leverage the same data to develop and deploy predictive analytics to help get ahead of deterioration and risk.

Clients currently and continuously develop analytics on Sickbay. For example, one client integrates 32 near real-time, multimodal risk scores into its virtual care workflow. Another client created a Sickbay algorithm that analyzes data generated by two separate monitoring devices to determine ideal blood pressure levels in patients. “The particular analytic requires the blood pressure waveform from a bedside monitor and a measure of cerebral blood density from a different monitor,” says Rusin.

Saving Lives with Data

Treatment of patients across the care continuum today will lead to improved care tomorrow. To do that, reliable, specific data is the very starting block. Without it, clinicians are left to their best guesses to solve the body’s most urgent care needs without the data-driven decision-making support they desire. That’s slow, costly, unfair to caregivers, and ultimately not providing the best benefit for the patient.

To truly realize a future where treatment is as specific and individual as the person it serves, healthcare must harness patient data in a way that is most impactful—specific, accurate, near real-time, vendor-agnostic, transformable, and instantly accessible. Leveraging the power of time series data empowers healthcare providers to help more people than it ever has before, and more effectively. After all, saving lives is healthcare’s primary mission.

 

Edited by Georganne Benesch, Editorial Director for insight.tech.

AI Workloads Scale with Next-Gen Processors and COM Express

X-rays, ultrasounds, and endoscopy machines generate massive volumes of data—sometimes too much to make sense of. In response, medical device OEMs integrate AI directly into medical imaging and diagnostics machines to make screening procedures more efficient, effective, and accessible for clinicians and patients alike.

Supporting AI-enabled medical imaging and diagnostics requires high-end hardware with the graphics and compute performance to execute intelligent imaging workloads in real time. Until recently, the easiest way to enable these capabilities was through discrete accelerators—an approach that can be expensive and inefficient in terms of upfront hardware costs and power consumption.

But by far the most costly design decision is the wrong system architecture. AI is evolving rapidly, so without flexible, adaptable, and upgradable system hardware, equipment can become obsolete before it is adequately broken in.

“AI workloads are advancing so quickly, it’s sort of dangerous when you start talking about hardware at all,” says Zeljko Loncaric, Market Segment Manager for Infrastructure at congatec AG, a global leader in embedded solutions. “That’s one of the most significant challenges facing medical device designers. They also face hurdles in implementing newer functionality in long-lifecycle systems.”

COM Express modules based on Intel® Core Ultra mobile processors address these challenges. They offer superior performance and efficiency in AI workload processing thanks to integrated GPUs and NPUs. And their inherent modularity streamlines the initial design process while enabling easy upgrades, processor generation over processor generation.

#AI #technology represents a meaningful advancement for #medical imaging, with the potential to significantly improve diagnostic efficiency and accuracy. @congatecAG via @insightdottech

Balancing Edge AI Longevity and Innovation in Embedded Computing

Because medical imaging devices must undergo a comprehensive certification process before they can be used, their lifecycles tend to average a decade or more. Meanwhile, AI technology represents a meaningful advancement for medical imaging, with the potential to significantly improve diagnostic efficiency and accuracy in ultrasounds, mobile ultrasounds, endoscopy machines, X-rays, and more.

But faced with the time and expense of redesigning and recertifying a medical device, OEMs hesitate to transition to next-generation platforms that support AI without an extremely compelling business case. And without being able to answer how long a system design will remain relevant, that business case becomes less compelling.

Enter new Intel Core Ultra Mobile processors, the first x86 processors to integrate an NPU, and one of the most power-efficient SoC families on the market today. The integrated NPU enables support for advanced AI workloads without the added cost and complexity of a discrete accelerator. When paired with the SoC’s leading performance-per-watt, medical device designers can better manage power consumption and thermal efficiency in resource-constrained edge AI deployments.

“The processor’s per-watt performance is also highly interesting in the context of mobile ultrasound devices and other battery-powered systems,” notes Maximilian Gerstl, Product Line Manager at congatec. “What Intel did with the architecture is very impressive. The numbers look great in terms of performance—not only on the CPU side, but also in terms of graphics. The new processors also offer an unprecedented level of flexibility to customers, allowing them to upgrade their systems across multiple generations while retaining the same form factor.”

“If there’s not a great new technology coming up, organizations will stay on the same module for 10 years or more so that they don’t have to recertify,” he continues. “Intel Core Ultra Mobile processors are a big step up. Healthcare organizations will have to think about changing to it.”

Open-Standard Modules Fast-Track System Upgrades

The latest congatec conga-TC700 COM Express Compact module incorporates the processing performance and application-ready AI capabilities of Intel Core Ultra Mobile processors in a plug-and-play form factor. Medical device designers can leverage the module as a shortcut to building efficient edge AI systems while significantly improving time-to-market and reducing total cost of ownership (TCO). And since COM Express is an open hardware standard governed by the global technology consortium, PICMG, the TC700 provides a vendor-neutral path to system upgrade whereby a legacy module can simply be swapped out for a higher-performance one with the same interfaces.

“The ability to quickly swap hardware means an organization can have its applications running for a very long time,” Gerstl explains. “Though they have to recertify new hardware components, they can bring over a lot of their software and hardware designs from previous applications.”

Intelligent Healthcare, Enabled by Edge AI Solutions

The conga-TC700 is supported by congatec’s OEM solution-focused ecosystem, which features efficient active and passive thermal management solutions, long-term support, and ready-to-use evaluation carrier boards. The company is also exploring how the open-source Intel® OpenVINO toolkit can empower its customers in the development and deployment of AI vision systems. According to Gerstl, the company is working on early benchmarking with specific use cases to help customers get their applications up and running more quickly.

For congatec, the availability of Intel Core Ultra Mobile processors represents a considerable step forward in the price, performance, and power consumption of next-generation edge AI devices. For medical device OEMs, these processors provide a compelling path to new, AI-enabled imaging and diagnostics equipment.

“We will continue to enable AI acceleration, hardware, and software and bring it to our products,” Gerstl says. “We want to enable this new trend.”

 

This article was edited by Georganne Benesch, Editorial Director for insight.tech.

Harmonizing Innovation with Audio-Based Generative AI

Artificial intelligence is an umbrella term for many different technologies. Generative AI is one we hear a lot about—particularly ChatGPT. And ChatGPT gets a whole lot of press, but it’s not at all the only song in the generative AI playbook. And one tune that Ria Cheruvu, AI Software Architect and Generative AI Evangelist at Intel, has been excited about lately is generative AI for the audio space (Video 1). 

Video 1. Ria Cheruvu, Generative AI Evangelist for Intel, explores the business and development opportunities for audio-based generative AI. (Source: insight.tech)

But generative AI of any kind can be intimidating, and developers don’t always know exactly where to start or, once they get going, how to optimize their models. Partnering with Intel can really simplify the process. For example, beginning developers can leverage the Intel® OpenVINO notebooks to take advantage of tutorials and sample codes that will help them get started playing around with GenAI. And then, when they’re ready to take it to the next level or ready to scale, Intel will be right there with them.

Ria Cheruvu talks with us about the OpenVINO notebook repository, as well as the real-world applications suggested by generative AI for audio, and the differences between the aspect of it that works for call centers and the aspect that can actually work for musicians.

What are the different areas of generative AI?

This space is definitely developing in terms of the types of generative AI out there. ChatGPT is not the only example of it! Text generation is a very important form of generative AI, of course, but there is also image generation, for example, using models like Stable Diffusion to produce art and prototypes and different types of images. And there’s also the audio domain, where you can start to make music, or make audio for synthetic avatars, as well as many other types of use cases.

In the audio domain, a fast runtime is especially important, and that’s one of the common pain points. You want models that are super powerful and able to generate outputs with high quality really quickly, and that takes up a lot of compute. So I’d say that the tech stack around optimizing generative AI models is definitely crucial, and it’s something I investigate as part of my day-to-day role at Intel.

What are the specific business opportunities around generative AI for audio?

It’s really interesting to think about using voice AI or conversational AI for reading in and processing audio, which is what you do with a voice agent, like a voice assistant on your phone. Compare that to generative AI for audio, where you’re actually creating the content—being able to generate synthetic avatars or voices to call and talk to, for example. And definitely the first business applications you think about are call centers, or metaverse applications where there are simulated environments that use this created audio.

But there are also some nontraditional business uses cases in the creative domain, in content creation, and that’s where we start to see some of the applications related to generative AI for music. And to me this is incredibly exciting. Intel is starting to look at how generative AI can complement artists’ workflows: for example, in creating a composition and using generative AI to sample beats. There’s also a very interesting cultural element to how musicians and music producers can leverage generative AI as part of their content-creation workflows.

And so while it’s not a traditional business use case—like what you would see in call centers, or in interactive kiosks that use audio for retail—I do believe that generative AI for music has some great applications for content creation. Eventually it could also come into other types of domains where there is a need to generate sound bites, for example, creating synthetic data for AI system training.

“#GenerativeAI for music has some great applications for content creation. Eventually it could also come into other types of domains where there is a need to generate sound bites” – Ria Cheruvu, @intel via @insightdottech

What is the development process for generative AI for audio?

There are a couple of different ways that the generative AI domain is currently approaching this. One of them is definitely adapting the model architectures that are already out there for other types of generative AI models. For example, Riffusion is based on the architecture for Stable Diffusion, the image-generation model; it just generates waveforms instead of images.

I was speaking recently to someone who is doing research in the music domain, and one of the things we talked about was the diversity of input data that you can give these audio-domain models. It could be notes—maybe as part of a piano composition—all the way to just waveforms or specific types of input that are specialized for use cases like MIDI formats. There’s a lot of diversity there.

What technologies are required to train and deploy these models?

We’ve been investigating a lot of interesting generative AI workloads as part of the Intel OpenVINO toolkit and the OpenVINO Notebooks repository. We are incorporating a lot of key examples of audio generation as very useful use cases to prompt and test generative AI capabilities. We had a really fun time partnering with other teams across Intel to create Taylor Swift-type pop beats using the Riffusion model—all the way to more advanced models that generate audio to match something that someone is speaking.

And one of the things that I see with OpenVINO is being able to optimize all these models, especially when it comes to memory and model size, but also enabling flexibility between the edge and the cloud and the client.

OpenVINO really targets that optimization part. There’s a fundamental notion that generative AI models are big in terms of their size and their memory footprint; and the foundations for all of these models—be it audio, image, or text generation—certain elements of them just are very large. By halving the model footprint using compression and quantization-related techniques, we’re able to achieve a lot of reduction of the model size while still ensuring that performance is very similar.

And all of this is motivated by a very interesting notion of local development. Music creators or audio creators are looking to move toward their PCs when creating content—as well as being able to work on the cloud in terms of intensive work like gathering audio data, recording it, annotating it, and collaborating with different experts to create a data set. And then they would be able to do other workloads on a PC and say, “Okay, now let me generate some interesting pop beats locally on my system and then prototype that in a room.”

What are some examples of how developers can get started with generative AI?

One example that I really love to talk about is how exactly you take some of these OpenVINO tutorials and workloads that we’re showing in the notebooks repo and then turn them into reality. At Intel we partner with Audacity, a tool that essentially enables open-source audio-related editing creation. It’s really a one-stop, Photoshop kind of a tool for audio editing. And one of the things we’ve done is integrate OpenVINO with it through a plugin that we provide. Our engineering team took the code in the OpenVINO Notebooks repo from Python, converted it to C++, and then deployed it as part of Audacity.

It allows for more of that performance and memory improvement I mentioned before, but it’s also integrated directly into the same workflow that many different people who are editing and just playing around with audio are leveraging. You just highlight a sound bite and say “Generate,” and OpenVINO will generate the rest of it.

That’s an example of workflow integration that can be used for artist workflows; or to create synthetic audio for voice production for the movie industry; or for interactive kiosks in the retail industry; or for patient-practitioner conversations in healthcare. That seamless integration into workflows is the next step that Intel is very excited to drive and to help collaborate on.

What else is in store for generative AI—especially generative AI for audio?

When it comes to generative AI for audio, I think it’s “blink and you may miss it” for any particular moment in this space. It’s just amazing to see how many workloads have been added. But just looking into the near future—maybe end of year or next year—some of the developments I can start to see popping up are definitely around those workflows I mentioned before, and identifying where exactly you want to run them—is it on your local system, or is it on the cloud, or on some sort of mix of the two? That is definitely something that really interests me.

We are trying some things around audio generation on the AI PC with the Intel® Core Ultra and similar types of platforms, where—when you’re sitting in a room prototyping with a bunch of fellow musicians and just playing around—ideally you won’t have to access the cloud for that. Instead, you’ll be able to do it locally, export it to the cloud, and just move your workloads back and forth. And key to this is asking how we incorporate our stakeholders as part of that process—how do we exactly create generative AI solutions, instantiate them, and then maintain them over time?

Can you leave us with a final bit of generative AI evangelism?

Generative AI is kind of a flashy space right now, but almost everyone sees the value that can be extracted out of it if there is a future-proof strategy. The Intel value prop for the industry is really being able to hold the hands of developers, to show them what they can do with the technology, and also to help them every step of the way to achieve what they want.

Generative AI for audio—generative AI in general—is just moving so fast. So keep an eye on the workloads, evaluating, testing, and prototyping; they are definitely all key as we move forward into this new era of audio generation, synthetic generation, and so many more of these exciting domains.

Related Content

To learn more about generative AI, read Generative AI Solutions: From Hype to Reality and listen to Generative AI Composes New Opportunities in Audio Creation. For the latest innovations from Intel, follow them on X at @IntelAI and on LinkedIn.

 

This article was edited by Erin Noble, copy editor.

All-in-One Medical AI PCs Meet Healthcare Computing Needs

In the healthcare industry, there are companies that build medical equipment and providers that use these machines and devices. While the business models of these two groups are completely different, they are guided by the same challenges and opportunities. Both are eager to deploy the latest technologies but are faced with strict regulatory requirements and short product life cycles.

A traditional medical device is designed with data residing locally on the device. With the increasing demand of interoperability within healthcare facilities, healthcare professionals can improve efficiency effectively with AI-enabled medical PC solutions designed to sustain the mission-critical environment and process patient data throughout the treatment process.

Managing all these constraints is a tall order, but today’s medical computers are up to the task. Hygienic, compact, and portable, medical-grade AI computers can be used by practitioners throughout hospitals and clinics. And high-performance processors enable near-real-time AI analytics, helping doctors and nurses make faster, better-informed diagnostic and treatment decisions.

#EdgeAI and #ComputerVision have become increasingly important to today’s imaging and patient monitoring machines, which can swiftly analyze #data and support physicians with diagnoses. @OnyxHealthcare1 via @insightdottech

Keeping Up with AI Innovations in IoT Medical Devices

Edge AI and computer vision have become increasingly important to today’s imaging and patient monitoring machines, which can swiftly analyze data and support physicians with diagnoses. But for medical device development, incorporating these cutting-edge capabilities can be a struggle. Medical device development takes on average from eight to 24 months to implement hardware and software design changes in accordance with regulations, and another two to three years to obtain certification via clinical trial.

“They don’t have the luxury to continuously upgrade to the latest technology,” says John Chuang, President of Onyx Healthcare, Inc., a Taiwan-based global producer of medical PCs and hospital IT solutions.

And once those finished medical devices are released, they need to stay in service for a long time. Hospitals have a complex mix of technology, and don’t usually upgrade their equipment for 10 years or more—an eternity in the fast-moving world of medical AI and computer vision development.

To keep machines as up-to-date as possible, Onyx collaborates with medical device companies, hospitals, and Intel, which supplies the processors for the all-in-one (AIO) medical computer the company produces for hospitals and clinics. Intel high-performance processing power is the key that enables software to run edge AI analytics.

Working closely with Intel, Onyx can provide a scalable custom design that allows medical device companies to incorporate the latest processors into its medical-grade computing technology. “By providing the latest technology to medical OEMs and ODMs, we help them keep a step ahead, so they don’t have to worry their technology is outdated by the time their devices are launched,” Chuang says.

Delivering Machine Information Where It’s Needed

In hospitals, medical devices are part of an elaborate symphony that requires precise timing and coordination. Doctors rely on information from many sources to diagnose and treat patients, including medical records and lab results, blood pressure and oxygen monitors, and images from X-ray, CT, and ultrasound scanners. But since these machines are made by different manufacturers and use different software protocols, they typically don’t connect with one another—or with hospital IT systems. As a result, doctors often must examine disjointed patient data.

A system like the Onyx AIO medical AI computer serves as a symphony conductor, integrating data from all sources—including patient records and off-site machines. It enables the transmission of high-resolution images and the performance of AI analytics, giving doctors a comprehensive, near-real-time view of a patient’s condition.

“The data transmitted is informative enough for physicians to make sound, timely treatment decisions. That’s especially crucial for patients in critical care, and in situations where the doctor needs to determine whether surgery is required,” Chuang says.

The ONYX AIO AI computer is also designed to meet hospitals’ rigorous sanitary requirements. For example, instead of using a fan for cooling, it uses an onboard heat sink, creating a closed system that won’t transfer germ-carrying air into hospital corridors or patient rooms. “We are able to use a fanless design because of the efficiency of low-wattage Intel processor technology,” Chuang says.

Medical IoT in Action: Mobile Nursing and Telehealth Solutions

Connecting patient information via medical computers can help hospitals and clinics achieve greater interoperability. That’s an important goal for the CAIH, a French government alliance formed to consolidate technology requirements across the country’s hospital networks. Onyx developed two solutions to help the organization achieve its objectives.

The first is mobile nursing stations—carts containing an AIO AI computer that nurses can bring on their rounds. The medical computer enables them to keep an eye on every patient under their care as they go from room to room. In addition to keeping nurses apprised of patients’ vital signs, the AIO helps monitor equipment, letting nurses know, for example, if an IV is running low on fluid.

AI monitoring helps short-staffed hospitals better attend to patients’ needs, Chuang says. It also helps them deal with the fast-growing use of telehealth. In a second solution it developed for the CAIH, Onyx enables AIO computers to connect doctors with patients, caregivers, and medical equipment at remote facilities—including skilled-nursing homes, where a physician may not be present.

Doctors can view patients from their own AIO computer and guide nurses in using medical instruments, such as portable ultrasound machines or scopes for examining the ear, nose, throat, or skin. Devices are equipped with high-definition cameras that relay medical-grade images to the doctor.

“With this information, physicians can do some diagnostics and quickly determine whether a patient needs to come to the hospital right away,” Chuang says. Otherwise, many would have no choice but to be transported there—often a challenge for those in a skilled-care institution.

Onyx AIO computers are also enabled for 5G communications, allowing remote facilities with a 5G network to relay alerts for patient vital data or slip-and-fall accidents directly to doctors or nurses, instead of waiting for the information to be processed in the cloud.

Building Future-Ready Technology

As AI capabilities expand, medical computers are assuming a greater role in patient care. But to stay useful, they must evolve along with the machines they connect with, Chuang says.

“Medical computers need to become more like medical devices themselves. We’re seeing greater demand for them to interface with specialized machines, and demand for data processing speed is also increasing. By building the latest Intel technology into our computers, we are able to satisfy those needs,” Chuang adds.

 

This article was edited by Georganne Benesch, Editorial Director for insight.tech.

AI and Computer Vision Boost Biomedical Research

Breakthroughs in biomedical research often come from understanding correlations and causality—the what and how of the body’s physiological processes.

Scientists make observations such as a higher rate of cancer or better reaction to vaccines by correlating data sets. They then research underlying reasons for these correlations. Systematically plodding through these cycles of biomedical research is tedious but rewarding work.

Now AI-driven computer vision applied to medical imaging accelerates discovery of data correlations. It finds problem points worth exploring much more quickly. In doing so, AI helps humans zero in on problems faster. And that can help human scientists arrive at life-changing medical solutions much more quickly.

AI Models in Medical Imaging

One use case is how Mayo Clinic uses AI and machine learning to profoundly enhance the capabilities of ultrasound imaging. As a starting point, the medical institution uses the latest technologies, tools, and products from Intel and Dell—the Intel® Geti platform and the Intel® OpenVINO toolkit running on Dell edge systems—to find kidney stones from endoscopy videos of the organ, and to assess Point of Care (POC) Ultrasound images for cardiac function.

Mayo Clinic’s work in the use of AI ultrasound imaging is a particularly useful case of the technology, says Alex Long, Global Head of Life Sciences Strategy at Dell Technologies, a solutions provider that offers an extensive product portfolio and comprehensive services. For too long, interpretation of ultrasounds has been subjective, prone to errors, and requires specialized training.

Visual #AI models, trained on banks of #data, can help providers offer more personalized care at the bedside. Augmenting care with AI can find anomalies faster, more accurately, and with minimal training. @DellTech via @insightdottech

But visual AI models, trained on banks of data, can help providers offer more personalized care at the bedside. Augmenting care with AI can find anomalies faster, more accurately, and with minimal training. Modern approaches which leverage pre-trained models and active learning enable the rapid development and deployment of these models. “Our care providers understand the benefit of using AI to aid in patient care, but in cases like the POC ultrasound, there wasn’t a viable AI model available,” says Dr. David Holmes of Mayo Clinic. His team of engineers leveraged interactive AI modeling tools to rapidly develop an AI solution that assess the quality of the images at the bedside in order to ensure the best images are used in the patient care.

The use of AI in medical imaging is about more than its capacity as a diagnostic tool. “It’s about leveraging visual AI to interpret imaging data and to accurately augment the capabilities of the human,” Long says. Diagnosticians trained to sift through files to find problems—evaluating mammograms to find early signs of breast cancer is a good example—can also benefit from AI guiding them to more places to evaluate. The advantage of AI is that it finds patterns that the human eye, due to confirmation bias, might miss.

A variety of additional scenarios in biomedical research can benefit from AI, especially if they involve imaging data. “It turns out there’s a lot of other medical systems that are visual in nature,” Long says. And they could all benefit from using AI as a tool to augment human abilities.

Collaboration Propels Innovation

A partnership between Intel and Dell Technologies enables these AI-driven breakthroughs. “The definition of community is a group of people with like-minded aspirations who are trying to achieve a goal together,” Long says. “We’re seeing a healthcare life sciences community being born between Dell and Intel.”

Collaboration between the two companies has evolved organically over many years, and Dr. Holmes’ work is one example of how the two bring their strengths to the table. The companies’ healthcare solution teams and their technology and product platforms enable collaborations with leading biomedical researchers and providers.

“The depth of our portfolio, the depth of our partnership, and the expertise in IT and infrastructure required to deliver” are what Dell brings to the table, Long says. In addition, Dell keeps in mind that the healthcare industry places heavy emphasis on privacy and protection of sensitive patient health information. “It’s not just about technology adoption to mitigate costs,” Long says, “it’s about technology to advance the human initiative of improved health. We’re passionate about really advancing the care of human beings.”

The Future of AI in Healthcare

The Mayo Clinic use case offers a glimpse of what is possible with AI models in biomedical research. We are just beginning to explore ways that AI can find correlations on visual imaging data, directing humans to new avenues for further exploration.

Researchers almost always try to find correlative data to drive conclusions, and “if you want something to identify a correlation, there’s nothing better than AI,” Long says. “I’m very excited about AI’s potential in accelerating diagnostics, improving patient care, and rapidly getting to understand the next wave of heuristics and treatments.”

When it comes to the human body, there’s a lot left to discover. It’s an exciting time to work at the intersection of technology and medicine because the volume of discoveries that AI can facilitate is simply mind-boggling. AI can train its eyes on years of data. The results are likely to be nothing short of revolutionary.

 

Edited by Georganne Benesch, Editorial Director for insight.tech.

Machine Builders Gain an Edge with Next-Gen Products

Edge AI and computer vision technologies are finding new use cases in nearly every industry. In the factory, applications like automated optical inspection (AOI) and industrial robotics improve operational efficiency. In healthcare, these technologies augment medical imaging and diagnostics. And they enable smarter traffic management in our cities, and enhanced security in our offices and public spaces.

This adoption of AI in so many different sectors also changes the outlook of business leaders. AI is no longer seen as some promising technology on the far horizon. AI already is beginning to deliver positive outcomes for organizations of all types. “The implementation of AI in real-world scenarios is currently happening,” says Christine Liu, Product Manager at Advantech, an edge AI platform and AIoT solutions provider. “Decision-makers today view AI as a ‘must-have’ in order to remain competitive.”

It’s a time of great opportunity for AI solutions developers, but they face challenges that need to be overcome, such as adopting AI computing solutions, integrating software SDKs, AI model training, and so on.

The good news is that embedded hardware partnerships enable powerful, development-ready AI computing with products like Advantech’s GPU Card EAI-3101, designed with the Intel® Arc A380 GPU. GPUs primarily offer visual image optimization and are currently one of the primary AI accelerators used to enhance AI computing power. 

The Latest Embedded GPU Card Supports Multiple AI Use Cases

The EAI series product line offers comprehensive AI acceleration and graphics solutions, including several PCIe and MXM GPU cards with Intel Arc Graphics. With the coming launch of Intel Arc A380E, Advantech offers EAI-3101, a new embedded PCIe GPU card powered by Intel Arc A380E with five-year longevity. Featuring 128 Intel Xe matrix AI engines, this GPU card delivers 5.018 TFLOPS AI computing power. With optimized thermal solutions and auto smart fans, these GPU cards can fulfill different use cases, such as gaming, medical analysis, and more. The designs are proven to outperform the competition in AI inference capability and graphics computing.

The diversity of choices means that OEMs, ODMs, and machine builders are more likely to find a computing platform to suit their needs, regardless of intended use case. Machine builders for the industrial sector, for example, would most likely select one of the commonly used PCIe configuration cards—while the smaller form factor and shock and vibration resistance of the MXM card might appeal to manufacturers of medical devices.

“Intel® Dynamic Power Share Arc GPUs and Intel CPUs can automatically and dynamically (re)distribute power between processing engines to boost performance depending on the use case—providing stable, high-performance computing for all kinds of edge workloads,” says Liu. “And the Intel® OpenVINO toolkit helps us accelerate AI inference times, reduce the AI model footprint, and optimize hardware usage.”

Advantech’s #development partnership with @Intel enables the company to bring the latest Intel products to market faster since it has early access to Intel’s latest-generation #processors. @Advantech_USA via @insightdottech

Advantech’s development partnership with Intel enables the company to bring the latest Intel products to market faster since it has early access to Intel’s latest-generation processors. This benefits Advantech customers even when they already have existing solutions in full deployment. For example, ISSD Electronics, a maker of intelligent traffic management solutions, deployed a smart traffic management system in Turkey and recently upgraded the solution to incorporate Advantech’s EAI-3100 series. As a result, the company has already improved its system’s accuracy, reduced AI inferencing latency, and cut construction costs by 33%, says Liu.

Advantech is also announcing new models in its AIR series of edge AI inferencing appliances:

  • AIR-150: compact, fanless edge AI inferencing system based on 13th Gen Intel® Core processors
  • AIR-310: edge AI box with MXM-GPU card supported by 14th Gen Intel® Core processors
  • AIR-510: edge AI workstation based on 14th Gen Intel® Core processors with RTX 6000 Ada

These edge AI platforms adopting the latest Intel platform fit in many different scenarios. Businesses might opt for the relatively lightweight AIR-150 for their offices. To achieve factory AMR automation management, the AIR-310 provides industrial protocols and scalable GPU computing power needed. And for creating a computer vision-assisted medical imaging solution that would likely have heavier graphical computing requirements, the more robust AIR-510 is the right fit.

Leveling the Playing Field for AI Application Development

Alongside its hardware products, Advantech offers a cross-platform edge AI software development kit (SDK). The SDK provides benchmarking tools to evaluate an AI application’s hardware requirements early in the solution development process. This helps developers select the best hardware for their solution—and prevents them from overspending on excessive computing power. In addition, the SDK enables real-time monitoring and over-the-air (OTA) AI model updates post-deployment.

As part of the SDK, OpenVINO provides model optimization and hardware acceleration benefits. The open-source inferencing toolkit also helps AI developers simplify their model deployments and software development workflows by supporting multiple AI model frameworks, including PyTorch, TensorFlow, and PaddlePaddle.

The availability of open-source toolkits and SDKs, coupled with a mature edge AI product ecosystem, will help more machine builders, OEMs, and ODMs to compete more effectively with a stable, development-ready AI computing environment. They help shorten the overall solution development time and allow designers to get innovative products to market faster.

Advantech also offers Edge AI SDK, the AI toolkit, to build an friendly environment from evaluation, SDK adopting, to deployment on all EAI and AIR series products as mentioned above

In the coming years, then, expect to see a far more level playing field for AI application development—what some have called “the democratization of AI.”

In Liu’s view, this is the correct path forward for our increasingly AI-enabled world. “The power of AI shouldn’t be limited to just a few companies. Resources such as our edge computing platforms, our SDK, and OpenVINO are there to be leveraged by everyone,” she says. “AI will be everywhere in the future—which is why we need these open and powerful solutions.”

 

This article was edited by Georganne Benesch, Editorial Director for insight.tech.

Optimizing Surgical Teams: AI’s Role in the OR

When you or a loved one faces surgery, you naturally want to make sure that you have the most skilled surgeon. Not to mention that it is in the most up-to-date facility with the most sophisticated technology to ensure the best possible outcome. After all, the stakes can be incredibly high. So it’s disconcerting to think that, until recently, surgeons lacked decision-support resources in the OR.

While medical use of technology has evolved rapidly over the past few years, the surgical field has been a little slower to adopt these advances than some other industries. Surgeons are used to relying on the skill of their hands and the knowledge gained through experience—with good reason. But medical technology isn’t all about robot arms and AI-guided surgery; there’s a lot to be gained just by freeing healthcare data from its traditional silos and giving surgeons access to that information where and when they need it most—in the OR.

We talk more about this with Dennis Kogan, Founder and CEO of digital surgery platform provider Caresyntax, as well as the dynamic challenges of the operating room, the importance of a good partner ecosystem, and how AI-assisted surgery can improve patient outcomes (Video 1).

Video 1. Dennis Kogan, CEO of Caresyntax, discusses the integration of real-time AI-driven data into surgical procedures, emphasizing its critical timing and impact on surgeons. (Source: insight.tech)

How are technological advancements in the OR changing healthcare expectations?

My dad is a surgeon, and years ago when I was in college, I was talking to him about how much decision support athletes get around things like performance management, situational awareness, and analytics. And he told me, “We have nothing like this in surgery. We have very interesting and important medical devices, and we’re continuously getting clinical innovation into our hands, but there isn’t really a lot of data-usage and decision-making support.”

And up until a few years ago that hadn’t changed much. There was a ton of innovation around medical devices, but at the end of the day that was still helping only how surgeons operated with their hands. The advancements that we see now enable surgical teams to have better decision-support mechanisms as well.

I think there is more and more expectation that surgeons cannot just be thinking about the risks of the procedure by themselves. And they do want support; they do want additional information to stratify risks more. And doing it all in their heads is probably no longer acceptable anymore.

What are some of the challenges integrating new technologies into the OR ?

Relative to other types of therapies, patients are probably less aware of what’s happening in the OR. Naturally—they’re under anesthesia. What they want is to understand how likely they are to have a good outcome. And I think they would probably be surprised that not as much integrated decision-making support is available to their surgical teams as they would expect.

The challenge to innovating the surgical field with technology is that surgery is a real-time intervention, and you have to integrate the AI and the software so that it runs in that setting. There should be almost no lag time in the OR. And that by itself is a higher hurdle than for a lot of other information technology used in healthcare. Of course, there is also a pretty high threshold for quality and operational effectiveness.

The surgical environment is also extremely dynamic. So how does a surgeon adapt to a changing clinical picture during the procedure? And it’s not only quantifiable activities and techniques; there’s also communication and teamwork. Surgery is actually a team sport. Part of the outcome depends on how well a surgeon does a certain maneuver, but another part of it is how well they communicate with the nursing staff and anesthesiologist. It’s so complex that it’s almost impossible to foresee how it could be replaced by artificial intelligence in the foreseeable future.

But AI does have a lot to give in terms of bringing the right information and options to the fingertips of physicians, just because of that dynamism. In one day a surgical team may be operating on very different types of patients: a healthy 25-year-old female and then a very sick 85-year-old male. The team has to be able to adjust a lot of inputs and make a lot of decisions.

That cognitive overload can cause suboptimal decisions or mistakes. Probably one out of seven cases has some sort of significant complication—over 15%. And so what we’re talking about here is proactive risk management through situational awareness—through automation. It’s about reducing and removing unwarranted variability driven by cognitive overload and a changing clinical picture. The best use cases we see right now for AI are in showcasing specific information about a given patient and a given procedure to be able to guide the entire pathway for that procedure and have the outcome be better than it would have been without that support.

“The challenge to innovating the surgical field with #technology is that surgery is a real-time intervention, and you have to integrate the #AI and the #software so that it runs in that setting” – Dennis Kogan, @caresyntax via @insightdottech

What is the benefit of combining AI with patient data?

First and foremost, truly integrated surgical-decision support touches on all points of the peri-operative cycle. Because everything that happens before and after a surgery is also extremely important, the best-integrated platforms allow for connectivity between the operating room and the pre- and postoperative spaces, times, and activities.

There are decisions made right before the patient enters the operating room—preparing the right tools, the right medications, having the right people at the table. It also includes the electronic medical record, because that has a trove of data about the patient and his or her predispositions. Then there’s the situation inside the OR, where medical devices and video cameras can be connected. And then afterward: knowing what level of risk that patient is exiting the OR with may change the protocol of how they are going to be taken care of. Maybe they can go home; maybe they need to be in the ICU; maybe they need an extra dose of antibiotics.

So to get the best, smartest insights you have to have a full peri-operative clinical and operational record, but the crown jewel is the intra-operative space—because that of course is the most mission-critical piece, where things can really go wrong. And because of that, and because the OR is real time, it requires an additional level of sophistication. And it’s not, in technical terms, a cloud-friendly territory. It’s all on the edge, because you cannot rely on two-second upload and download from a cloud. So edge computing and the Internet of Things technology toolkit are extremely important here.

At the same time, this technology solution has to be very robust and attractive from the perspective of deployment and cost. Because at the end of the day, anything that is overly expensive or unwieldy—another huge machine being rolled into an already very packed operating room—is just not going to work.

It took us at Caresyntax—with the help of a few technology partners—years to develop this platform in a way that achieves all these parameters. But I do know that it’s possible. Things are still sort of at the beginning, but I think the next decade will probably see every OR being equipped with these kinds of systems. And in 10 years physicians will be wondering how they were working without it.

How can hospitals future-proof this kind of investment?

Every industry goes through a cycle of having a few vendors create kind of a walled garden at first, and then gradually users expect more and more flexibility to add value and to add new applications. I think surgery and healthcare will need to undergo the same change.

The medical-device world has a lot of proprietary intellectual property, for some good reasons. Historically that’s been a dominant mindset for physicians, too—thinking of the operating room through the prism of a device and a vendor, to a certain degree. So the first investment that needs to be made is in reinventing and recalibrating that mindset. The operating room should be seen not as an extension of a leading device platform but as belonging to that horizontal process of achieving the best outcome.

Do you have any use cases or customer examples you can share?

So we’ve been able to show that using these advanced platforms in the OR can lift performance level, and not only for surgeons but also for other physicians and clinical collaborators as well. For example, nurses. After the pandemic a lot of folks entered the nursing workforce without maybe as much training as they would have had before. And then there’s a lot of surgical volume right now because so many surgeries were bumped. So there are a lot of newer nurses who need to come up to speed very quickly. We’re increasingly deploying something like an interactive, step-by-step navigation guide in the OR. Getting step-by-step support in the right moment of the procedure can be extremely helpful to someone who may still be lacking confidence or experience in that setting.

How does Caresyntax work with partners to bring these platforms into ORs?

Being surgery specialists, we have a very good view of what the end applications and use cases should be, but we don’t have as much experience building the infrastructure. We don’t have the benchmarks and comparables from other use cases that may be similar in terms of the rigor and the actual architecture. And an integrated smart-surgery platform that is plug-and-play, that is very smart but not very heavy in terms of hardware content, something that is able to generate information but also has the capability and bandwidth to receive algorithm and produce AI and showcase it in real time—that’s a pretty sophisticated set of requirements.

Intel has been one of the partners that has really plugged in with us, almost inside our team, to make this happen. Designing the architecture, finding the right components, utilizing some of their components—such as OpenVINO that allows for this AI penetration and usage—all of these things are very important. Without a partner like Intel we would have been, at the very least, much slower, looking for every piece ourselves and probably making more mistakes.

Alongside Intel, of course, we also work with cloud-solution providers—AWS and Google Cloud. Because there has to be an edge-to-cloud transition. As I mentioned before, it’s a preoperative, intra-operative, and postoperative space, so you have to continuously go to the edge and back to the cloud and make the information interchangeable. And actually they all collaborate in between themselves—Intel and Google, Intel and AWS—which has been very rewarding as well.

Of course, the pandemic was an impediment to innovation, but that has subsided lately. I think everybody’s really looking at surgery and saying, “It’s not as safe as flying; it’s not as safe as even some other medical procedures. It’s time to improve it.” And it takes an ecosystem of players to achieve that.

What’s your most important takeaway about the use of AI in surgery?

I very often see that folks think of surgery as something that’s been figured out, something that’s reached maturity and doesn’t require innovation. It doesn’t give me any pleasure to say that this is not the case. But there is the opportunity to get surgery to the same place as, say, aviation. I don’t think you and I would accept getting on a plane with a 15% chance of something going wrong in that flight.

It’s a huge problem that has not only clinical implications but cost implications. Next to pharmaceutical therapies, surgical therapies are the second-most-used way of correcting a disease. It’s probably 20%, 30% of all of healthcare spend in the US.

So if we’re going into a surgery, I think we should have the feeling that everything is going to be okay. And that should be backed by real statistics. We really can make surgery safer and smarter. It will have broad impact on patient health for millions of people, and a broad impact on cost as well. There’s ample room for improvement as long as the mindset for innovation is there.

Related Content

To learn more about AI-assisted surgical technology, listen to Staffing AI in the OR: With Caresyntax and follow Caresyntax at @caresyntax and on LinkedIn.

 

This article was edited by Erin Noble, copy editor.

Generative AI Composes New Opportunities in Audio Creation

Despite what many may think, generative AI extends beyond just generating text and voice responses. Among the growing fields is audio-based generative AI, which harnesses AI models to create and/or compose fresh and original audio content. This opens a world of new possibilities for developers and business solutions.

In this podcast, we discuss the opportunities presented by audio-based generative AI and provide insights into how developers can start building these types of applications. Additionally, we explore the various tools and technologies making audio-based generative AI applications possible.

Listen Here

[Podcast Player]

Apple Podcasts      Spotify      Google Podcasts      Amazon Music

Our Guest: Intel

Our guest this episode is Ria Cheruvu, AI Software Architect and Generative AI Evangelist for Intel. Ria has been with Intel for more than five years in various roles, including AI Ethics Lead Architect, AI Deep Learning Research, and AI Research Engineer.

Podcast Topics

Ria answers our questions about:

  • (1:52) Generative AI landscape overview
  • (4:01) Generative AI for audio business opportunities
  • (6:29) Developing generative AI audio applications
  • (8:24) Available generative AI technology stack
  • (11:45) Developer resources for generative AI development
  • (14:36) What else we can expect from this space

Related Content

To learn more about generative AI, read Generative AI Solutions: From Hype to Reality. For the latest innovations from Intel, follow them on X at @IntelAI and on LinkedIn.

Transcript

Christina Cardoza: Hello and welcome to the IoT Chat, where we explore not only the latest developments in the Internet of Things, but AI, computer vision, 5G, and more. Today we’re going to be talking about generative AI, but a very interesting area of generative AI, which is the audio space, with a familiar face and friend of the podcast, Ria Cheruvu from Intel. Thanks for joining us again, Ria.

Ria Cheruvu: Thank you, Christina, excited to be here.

Christina Cardoza: So, not only are you AI Software Evangelist for Intel, but you’re also a Generative AI Evangelist. So, what can you tell us about what that means and what you’re doing at Intel these days?

Ria Cheruvu: Definitely. Generative AI is one of those transformational spaces in the AI industry that’s impacting so many different sectors, from retail to healthcare, aerospace, and so many different areas. I’d say that as part of being an AI evangelist it’s our role to keep up to date and help educate and evangelize these materials regarding AI.

But with generative AI that’s especially the case. The field moves so rapidly, and it can be challenging to keep up to date with what’s going on. So that’s one of the things that really excites me about being an evangelist in the generative AI space around what are some of the newer domains and sectors that we can innovate in.

Christina Cardoza: Absolutely. And not only is there always so much going on, but take generative AI for example: there’s so many different areas of generative AI. I almost think of it—like, artificial intelligence is an umbrella term for so many of these different technologies—generative AI is also sort of an umbrella for so many different things that you can do. I think a lot of people consider ChatGPT as generative AI, and they don’t realize that it really goes beyond that.

So that’s sort of where I wanted to start the conversation today. If you could tell us a little bit more about the generative AI landscape: where things are moving towards, and the different areas of development that we have, such as the text based or audio areas.

Ria Cheruvu: Sure. I think the generative AI space is definitely developing in terms of the types of generative AI. And, exactly as you mentioned, ChatGPT is not the only type of generative AI out there, although it does represent a very important form of text generation. We also have image generation, where we’re able to generate cool art and prototypes and different types of images using models like Stable Diffusion. And then of course there’s the audio domain, which is bringing in some really unique use cases where we can start to generate music: we can start to generate audio for synthetic avatars, and so many other different types of use cases.

So I know that you mentioned the tech stack, and I think that that’s especially critical when it comes to being able to understand what are the technologies powering generative AI. So, especially with generative AI there’s a couple of common pain points. One of them is a fast run time. You want these models that are super powerful and taking up a lot of compute to be able to generate outputs really quickly and also with high quality. That pertains to both text, to image, and then all the way to audio too.

For the audio domain it’s especially important, because you have these synthetic audio elements that are being generated, or music being generated, and it’s one of those elements that we pay a lot of attention to, similar to images and text. So I’d say that the tech stack around optimizing generative AI models is definitely crucial and what I investigate as part of my day-to-day role.

Christina Cardoza: I’m looking forward to getting a little bit deeper into that tech stack that you just mentioned. I just want to call out generative AI for audio. You mentioned the music-generation portion of this, and I just want to call that out because we’ve had conversations around voice AI and conversational AI, and this is sort of separate from that area. It’s probably adjacent to it, but we’re not exactly talking about those AI avatars or chatbots that you’re communicating with and that you can have conversations with.

But, like you said, the music composition of this, the audio composition of this—so I’m curious, what are the business opportunities for generative AI for audio? Just so that we can get an understanding of the type of use cases that we’re looking at before we dive deeper a little bit into that tech stack and development.

Ria Cheruvu: Yeah, and I think that you brought up a great point in terms of conversational voice agents and how does this actually relate. And I think it’s really interesting to think about how we use AI for reading in and processing audio, which is what we do with a voice agent—like a voice assistant on our phones compared to generative AI for audio, where we’re actually creating this content.

And I know you mentioned, for example, being able to generate these synthetic avatars or this voice for being able to communicate and call and talk to. And I think, definitely, the business applications for those, the first ones that we think about are call centers, or, again, metaverse applications where we have simulated environments and we have parties or actors that are operating using this audio. There’s also additional use cases for interaction in those elements.

And then we go into the creative domain, and that’s where we start to see some of the generative AI for music-related applications. And this, to me, is incredibly exciting. Because we’re able to start to look at how generative AI can complement artists’ workflows, whether you’re creating a composition and using generative AI to figure out and sample some beats and tunes in a certain space, and also dig deeper into existing composition. So, to me, that’s also a very interesting cultural element of how musicians and music producers can connect and leverage generative AI as part of their content-creation workflows.

So, while that’s not a traditional business use case—like what we would see in call centers, interactive kiosks that can use audio for retail, and other use cases—I also believe that generative AI for music has some great applications in the content creation, artistic domain. And eventually, that could also come into other types of domains as well where we need to generate certain sound bites, for example, training synthetic data for AI systems to get even better at this.

Christina Cardoza: Yeah, it’s such an exciting space. I love how you mentioned the artistic side of this. Because we see generative AI with the image creation, like you mentioned, creating all of these different types of pictures for people and paintings—things like that. So it’s interesting to see this other form that people can take and express their artistic capabilities with generative AI.

Because we talked about how generative AI—you can use it for text or image generation—I’m curious what the development for generative AI for audio is. Are there similarities that developers can take from text or image generation? Or is this a standalone development process.

Ria Cheruvu: That’s a great question. I think that there’s a couple of different ways to approach it as it is currently in the generative AI domain. One of the approaches is definitely adapting the model architectures that are already out there when it comes to audio and music generation, and also leveraging the architectures for other types of generative AI models. So, for example, Riffusion, which is a really popular earlier model in the generative AI-for-audio space, although considerably it’s pretty new, but with the advancements in generative AI there’s just more and more models being created every day.

This particular Riffusion model is based on the architecture for Stable Diffusion, the image-generation model, in that sense that we’re actually being able to generate waveforms instead of images leveraging Riffusion model. And there are similar variants that are popping up, as well as newer ones that are saying, “How do we optimize the architecture that we’re leveraging for generative AI and structure it in a way that you can generate audio sound bites or audio sound tokens or things like this that are customized for the audio space?”

I was talking to someone who is doing research in the music domain, and one of the things that we were talking about is the diversity and the variety of input data that you can give these models as part of the audio domain—whether that’s notes, like as part of a piano composition, all the way to just waveforms, or specific types of input formats as well that are specialized for different use cases, like MIDI or MIDI formats. There’s a lot of different diversity and application of the types of input data and outputs that we’re expecting from these models.

Christina Cardoza: And I assume with these models, in order to optimize them and to get them to perform well and to deploy them, there is a lot of hardware and software that’s going to go into this. We mentioned a little bit of that tech stack in the beginning. So, what types of technologies make these happen, or train these models and deploy these models, especially in the Intel space? How can developers partner with Intel to start working towards some of these generative AI audio use cases and leverage the technologies that the company has available?

Ria Cheruvu: As part of the Intel® OpenVINO toolkit, we’ve been investigating a lot of interesting generative AI workloads, but audio is definitely something that is continuing to come back again and again as a very useful and interesting use case, in a way to prompt and test generative AI capabilities. I’d say that as part of the OpenVINO Notebooks repository we are incorporating a lot of key examples when it comes to audio generation—whether it’s the Riffusion model, which we had a really fun time partnering with other teams across Intel to generate pop beats, similar to something that Taylor Swift would make, to some more of these advanced models, like generating audio, again, for being able to match it to something that someone is speaking. So there’s a lot of different use cases and complexity.

With OpenVINO we are really targeting that optimization part, which is based on this fundamental notion that we are recognizing that generative AI models are big in terms of their size and their memory footprint. And naturally the foundations for all of these models—be it audio, image generation, text generation—there’s certain elements of it that are just very large and that can be optimized further. So by halving model footprint or the model size by using compression and quantization-related techniques, we’re able to achieve a lot of reduction in terms of the model size, while also ensuring that the performance is very similar.

Then all of this is motivated by a very interesting notion of local development, where you’re starting to see music creators or audio creators looking to move towards their PCs in terms of creating content, as well as working on the cloud. So with that flexibility you’re able to essentially do what you need to do on the cloud in terms of some of your intensive work—like annotating audio data, gathering it, recording it, collaborating with different experts to create a data set that you need. And then you’re able to do some of your workloads on your PC or on your system, where you’re saying, “Okay, now let me generate some interesting pop beats locally on my system and then prototype it in a room.” Right?

So there’s a lot of different use cases for local versus cloud computing. And one of the things that I see with OpenVINO is optimizing these architectures, especially the bigger elements when it comes to memory and model size, but also being able to enable that flexibility between traversing the edge and the cloud and the client.

Christina Cardoza: I always love hearing about these different tools and resources. Because generative AI—this space—it can be intimidating, and developers don’t know exactly where to start or how they can optimize their model. So I think it’s great that they can partner with Intel or use these technologies, and it really simplifies the process and makes it easy for them so they can focus on the use case, and they don’t have to worry about any of the other complications that they may come across.

And I love that you mentioned the OpenVINO Notebooks. We love the OpenVINO Notebook repository, because you guys just provided a wealth of different tutorials, sample codes, information for all of these different things we talk about in the podcast—how developers can get started, experiment with it, and then really create their own real-world business use cases. Where else do you think developers can learn about generative AI for audio, can learn how to develop it, build it?

Ria Cheruvu: Yeah. I think that—and definitely, Christina, I think we’re very excited about being able to advance a lot of the development, but also the short prototypes that you can do to actually take this forward and partner with developers in this space, and also be able to take it further with additional, deeper engagements and efforts such as that.

I think to answer your question about a deeper tech stack, one of the examples that I really love to talk about—and I was able to witness firsthand as part of working through and creating this—is how do you exactly tape some of the tutorials and the workloads that we’re showing in the OpenVINO Notebooks repo, and then turn it into a reality for your use cases?

So, at Intel we partner with Audacity, a tool that is essentially enabling open-source, audio-related editing creation and a couple of other different efforts. It’s really this one-stop Photoshop kind of tool for audio editing. And one of the things that we’ve done is integrated OpenVINO through a plugin that we provide with that platform. So, as part of that what our engineering team did is they took the code in the OpenVINO Notebooks repo from Python, converted it to C++, and then were able to deploy it as part of Audacity.

So now you’re getting even more of that performance and memory improvement, but you’re also having it integrated directly into the same workflow that many different people who are looking to edit and play around with audio are leveraging. So that means that you just highlight a sound bite and then you say “Generate” with OpenVINO, and then it’ll generate the rest of it, and you’re able to compare and contrast.

So, to me, that’s an example of workflow integration, which can eventually, again, be used for artist workflows, all the way to creating synthetic audio for voice production as part of the movie industry; or, again, interactive kiosks as part of the retail industry for being able to communicate back and forth; or patient-practitioner conversations as part of healthcare. So I’d say that that seamless integration into workflows is the next step that Intel is very excited to drive and help collaborate on.

Christina Cardoza: Yeah, that’s a great point. Because beginner developers can leverage some of these notebooks or at least samples to get started and start playing around with generative AI, especially generative AI for audio. But then when they’re ready to take it to the next level, ready to scale, Intel is still there with them, making sure that everything is running smoothly and as easy as possible. They can start continuing on their journey for generative AI.

I know in the beginning we mentioned how a lot of people consider, like, ChatGPT or text-based AI as generative AI, when it really is all of these other different forms associated with it also. So I think probably still early days in this space, and I’m looking forward to the additional opportunities that are going to come. I’m curious, from your perspective, where do you think this space is going in the next year or so? And what is the future of generative AI, especially generative AI for audio? And how do you envision Intel is going to play a role in that future?

Ria Cheruvu: Sure. And I completely agree. I think that it’s blink and you may miss it when it comes to generative AI for audio, even with the growth of the OpenVINO Notebooks repository. As an observer and a contributor, it’s just amazing to see how many workloads around audio have continued to be added in terms of generative AI workloads and some of the interesting elements and ways that we can implement and optimize that.

But I’d say, just looking into the near future, maybe end of year or maybe next year or so, some of the developments that we can start to see that are popping up are definitely those workflows that we think about. Now we have these models and these technologies, and we’re seeing a lot of companies in the industry creating platforms and toolboxes, as they call it, for audio editing and audio generation and some of these elements using generative AI. So I would say that identifying where exactly you want to run these workloads—is it on your local system, or is it on the cloud, or some sort of mix of it?—is definitely something that really interests me, as I mentioned earlier.

And with Intel, some of the things that we are trying are around audio generation on AI PC with the Intel® Core Ultra and similar types of platforms around what can you achieve locally when you’re sitting in a room, prototyping with a bunch of fellow artists for music, and you’re just playing around and trying to do some things. And ideally you’re not exactly having to access the cloud for that, but you’re actually able to do it locally, export it to the cloud, and move your workloads back and forth.

So I’d say that that really is the key of it, which is what exactly is going to happen with generative AI for audio. How do we get it to be incorporating our stakeholders as part of that process—whether we’re, again, generating audio for these avatars—how do we exactly create that, instantiate that, and then maintain it over time? I think that these are a lot of the questions that are going to be coming up in the next year. And I’m excited to be collaborating with our teams at Intel and across the industry to see what we’re going to achieve.

Christina Cardoza: Great. And I love that you mentioned maintaining it over time. Because we want to make sure that anything that we do today is still going to make sense tomorrow. How can we future-proof the developments that we’re doing? And Intel is always leading the way to make sure that developers can plug and play or add new capabilities, make their solutions more intelligent without having to rewrite their entire application. Intel has always been great at partnering with developers and letting them take advantage of the latest innovations and technologies. So I can’t wait to see where else the company takes this.

We are running a little bit out of time. So, before we go, I just want to ask you one last time, Ria, if there’s anything else about this space that we should know, or there’s any takeaways that you want to leave our listeners with today.

Ria Cheruvu: I think one takeaway is exactly rephrasing what you said in terms of there’s a lot of steps towards being able to enable and deploy generative AI. It’s kind of that flashy space right now, but almost everyone sees the value that we can extract out of this if we have that future-proof strategy and that mindset. Definitely couldn’t have phrased it better in terms of our value prop or value add that we want to provide to the industry—is really being able to hold the hands of developers, show you what you can do with the technology and the foundations, and also help you with every step of the way in order to achieve what you want.

But I’d say, based off of everything that we’ve gathered up until now, as I mentioned earlier, generative AI for audio and specifically generative AI in general is just moving so fast. So, keeping an eye on the workloads, evaluating, testing and prototyping is definitely key as we move forward into this new era of audio generation, synthetic generation, and so many more of these exciting domains.

Christina Cardoza: Of course we’ll also be keeping an eye on the work that you’re doing at Intel. I know you often write and publish a lot of blogs on the Intel or OpenVINO media channels and different areas. There’s different edge reference kits that are published online every day. So we’ll continue to keep an eye on this space and the work that you guys are doing.

So, just want to thank you again for joining us on the podcast and for the insightful conversation. And thank you to our listeners for tuning into this episode. Until next time, this has been the IoT Chat.

The preceding transcript is provided to ensure accessibility and is intended to accurately capture an informal conversation. The transcript may contain improper uses of trademarked terms and as such should not be used for any other purposes. For more information, please see the Intel® trademark information.

This transcript was edited by Erin Noble, copy editor.

Medical Panel PCs Fulfill Hospitals’ Rigorous Requirements

From MRIs and CT scanners to electrocardiograms, oximeters, and blood pressure monitors, today’s medical machines collect an enormous amount of data about patient health. But getting that information into the hands of hospital doctors and nurses who need it is often a struggle.

One problem is that medical machines and devices are made by a wide range of manufacturers. As a result, the information they produce is scattered in incompatible formats across multiple data bases and IT systems, making it hard for doctors and nurses to assemble a coherent picture of patient health.

Another issue is hospitals’ restrictive hardware requirements. Nursing stations, operating rooms, and intensive care units—as well as the labs and pharmacies they connect with—must use computers that meet rigorous hygienic standards. And to successfully transmit vital patient data, the computers must also be very fast, reliable, easy to use, and secure.

Modern medical Panel PCs are designed to meet these challenges. They support data integration from a diverse range of medical equipment via an all-in-one, compact computer that meets hospital sanitary, usability, security, and reliability requirements. And they have the computing power to run the sophisticated medical software that provides caregivers the data they need not only to respond to emergencies but to consistently measure patient progress.

Unifying Patient Monitoring Solutions

Because healthcare technology continuously advances, most hospitals contain a wide range of medical equipment makes and models.

“Hospitals modernize their technology in phases, so they have a very heterogeneous hardware and software environment. Communication protocols are a big pain point,” says Guenter Deisenhofer, Product Manager at Kontron AG, a manufacturer of IoT equipment.

Machines that don’t intercommunicate erode procedural efficiency and make treatment decisions difficult—a problem Deisenhofer recently experienced firsthand after taking his son to a local hospital. First, an intake worker measured the boy’s heart rate, blood pressure, and oxygen level, recording the information on a slip of paper. His son was later seen by a doctor, who took his vital measurements again, writing them on a different paper before sending him off for X-rays—where the process was repeated yet again.

“At the end of the day, there were probably five pieces of paper. They had not continuously monitored his condition and nobody had an overview of it,” Deisenhofer says.

Fortunately, the situation turned out not to be serious. But with vital information arriving from different devices at different times—whether it’s recorded on paper or encoded in incompatible software programs—doctors can never be sure of what they might miss.

Medical edge computers, such as Kontron’s MediClient Panel PC, close the information gap, using a standard protocol to integrate data from machines, wearables, and patient health records. The Panel PC satisfies hospitals’ strict sanitary regulations and is readily accessible to caregivers, even if they’re wearing gloves. It conveys information from patient monitoring machines through wired or wireless connections to hospital communication hubs, such as nursing stations. High-performance Intel® processors enable the monitoring machines’ software to run near-real time analytics on incoming results, helping doctors and nurses see patterns and spot anomalies that may suggest a diagnosis, or point to the need for specific tests.

“With continuous monitoring data available, doctors aren’t just reacting to an emergency. They can see, for example, if the heart rate is dropping and recovering over time. It helps them make better diagnostic and treatment decisions,” Deisenhofer says.

Monitoring can continue after patients are released from the hospital, with wearable devices seamlessly transmitting their information to the MediClient, where it can be integrated with patients’ previous records.

With hospitals increasingly relying on advanced #medical equipment, machine #manufacturers must be vigilant in keeping them updated, both to enable new #IoT functions and to guard against the latest cyberthreats. @Kontron via @insightdottech

Improving Machine Life Cycle Management with Medical Panel PCs

With hospitals increasingly relying on advanced medical equipment, machine manufacturers must be vigilant in keeping them updated, both to enable new IoT functions and to guard against the latest cyberthreats. Making these changes often involves not only software but firmware and sometimes hardware. That means devices like the MediClient PC must also be updated to continue providing hospitals the machines’ vital data.

As technology innovation accelerates, machines that were built to last 10 or 15 years often require several major updates. “Machine life cycle is getting more and more difficult to manage,” Deisenhofer says.

Kontron works closely with medical equipment manufacturers to keep up with planned changes and is often included in early product planning cycles. Because every hardware change requires extensive testing and recertification, close communication saves time. Manufacturers can get their products recertified in one go, instead of having to do so again after making modifications. Kontron also does third-party software installation and electrical testing of manufacturers’ equipment to help them resolve potential problems before release.

Collaboration with Kontron allows manufacturers to deliver upgrades to hospitals sooner—and hospitals can integrate the new capabilities into their medical systems as soon as they are available.

Bringing New Capabilities to the Edge

Working together, OEMs and Panel PC makers can extend the value of monitoring machines as technology advances. The more data the machines collect, the more system builders can improve their AI software, reducing false alarms and pinpointing problems. “For example, you could use AI to create warning scores for recognizing conditions well in advance of a critical situation,” Deisenhofer says.

And the processing speed of edge Panel PCs will improve, helping caregivers respond to identified health threats sooner.

Benefits like these suggest a bright future for patient monitoring machines and the medical Panel PCs that connect their data to doctors and nurses. As Deisenhofer says, “Our device cannot make decisions. But by bringing all the data together at the edge, it can help doctors make the right decisions.”

Edited by Georganne Benesch, Editorial Director for insight.tech.