Landing AI develops “visual prompting” technology for teaching edge AI

Landing AI develops “visual prompting” technology for teaching edge AI

During the Computex 2023 event, Landing AI, a company specializing in cloud-based computer vision solutions, created a visual promotion technology that leverages Nvidia’s Metropolis for Factories. This solution brings the concept of prompt-based interaction to the computer vision ecosystem, expanding on the framework of text prompting used in technologies like ChatGPT.

Just like in ChatGPT, where a text prompt is utilized to express desired text processing tasks, the visual prompting technology enables users to precisely indicate image processing tasks through visual prompts. Leveraging large pre-trained vision transformers, this technology streamlines the development process by requiring only a few straightforward prompts and examples.

Using this approach, users can deploy computer vision systems and make inferences, all within a significantly reduced development time. Incorporating visual prompts alongside the pre-trained vision transformers empowers users to interact efficiently with the system and rapidly achieve their desired outcomes in the field of computer vision.

“Access to Nvidia AI and Metropolis for Factories enables us to enhance and accelerate our Visual Prompting technology and apply it to make spaces and operations safer and more efficient,” says Carl Lewis, senior director of partnership at Landing AI. “Nvidia’s assistance has been invaluable for advancing our technology roadmap, and we are thrilled to accelerate our collaboration through Metropolis.”

What is visual prompting technology?

To streamline the labeling process of images in the training dataset, visual prompting technology leverages large pre-trained vision transformers. Unlike traditional methods that necessitate labeling every image in its entirety, this advanced approach enables users to label only a few specific areas within an image. This efficient labeling strategy accelerates the process, reducing the user’s workload while still achieving accurate and comprehensive training data.

Recognizing the inefficiency of manually labeling each defect in an image, Landing AI says it decided to mimic the way humans teach. Rather than laboriously labeling every single defect in an image, the user tags an area of the image, allowing the machine to autonomously find additional defects.

Landing AI has successfully implemented the visual prompting technology into the workflow of one of its life sciences clients, enabling them to create a deployable cell detection model. The customer faced challenges associated with labeling too many cells, which often led to the omission of some cells in the process. With visual prompting, the client achieved their desired results, reporting that the entire task was completed within 10 minutes.

Article Topics

 |   |   |   |   | 

Comments

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Sponsored Links

Avassa: Empowers companies to bridge the gap between modern containerized applications development and operations and distributed edge infrastructure. https://avassa.io/

DataBank: We believe there is a different edge to be served - the “middle edge" - that will become the first step for many in their journey to the edge. https://www.databank.com/

Latitude.sh: Where the power of bare metal meets the flexibility of the cloud. Deploy physical servers across 23 global locations in as little as 5 seconds. https://www.latitude.sh/

Zenlayer: A massively distributed edge cloud service provider operating over 270 PoPs around the world, with expertise in fast-growing emerging markets. https://www.zenlayer.com/

OnLogic: A global industrial PC manufacturer and solution provider focused on hardware for IoT and edge AI, OnLogic designs highly-configurable computers engineered for reliability. https://www.onlogic.com/

Featured Company

Latest News