Tuesday, March 28, 2023
  • Login
CEO North America
  • Home
  • Business
    • Entrepreneur
    • Industry
    • Innovation
    • Management & Leadership
  • CEO Interviews
  • CEO Life
    • Art & Culture
    • Food
    • Health
    • Travel
    • Environment
  • Opinion
  • News
  • Multimedia
No Result
View All Result
  • Home
  • Business
    • Entrepreneur
    • Industry
    • Innovation
    • Management & Leadership
  • CEO Interviews
  • CEO Life
    • Art & Culture
    • Food
    • Health
    • Travel
    • Environment
  • Opinion
  • News
  • Multimedia
No Result
View All Result
CEO North America
No Result
View All Result

A Simpler Path to Better Computer Vision

in Business, Innovation
A simpler path to better computer vision
Share on LinkedinShare on WhatsApp

New research reveals a scalable technique that uses synthetic data to improve the accuracy of AI models that recognize images.

Before a machine-learning model can complete a task, such as identifying cancer in medical images, the model must be trained. Training image classification models typically involves showing the model millions of example images gathered into a massive dataset.

However, using real image data can raise practical and ethical concerns: The images could run afoul of copyright laws, violate people’s privacy, or be biased against a certain racial or ethnic group. To avoid these pitfalls, researchers can use image generation programs to create synthetic data for model training. But these techniques are limited because expert knowledge is often needed to hand-design an image generation program that can create effective training data.

Researchers from MIT, the MIT-IBM Watson AI Lab, and elsewhere took a different approach. Instead of designing customized image generation programs for a particular training task, they gathered a dataset of 21,000 publicly available programs from the internet. Then they used this large collection of basic image generation programs to train a computer vision model.

These programs produce diverse images that display simple colors and textures. The researchers didn’t curate or alter the programs, which each comprised just a few lines of code.

The models they trained with this large dataset of programs classified images more accurately than other synthetically trained models. And, while their models underperformed those trained with real data, the researchers showed that increasing the number of image programs in the dataset also increased model performance, revealing a path to attaining higher accuracy.

“It turns out that using lots of programs that are uncurated is actually better than using a small set of programs that people need to manipulate. Data are important, but we have shown that you can go pretty far without real data,” says Manel Baradad, an electrical engineering and computer science (EECS) graduate student working in the Computer Science and Artificial Intelligence Laboratory (CSAIL) and lead author of the paper describing this technique.

Co-authors include Tongzhou Wang, an EECS grad student in CSAIL; Rogerio Feris, principal scientist and manager at the MIT-IBM Watson AI Lab; Antonio Torralba, the Delta Electronics Professor of Electrical Engineering and Computer Science and a member of CSAIL; and senior author Phillip Isola, an associate professor in EECS and CSAIL; along with others at JPMorgan Chase Bank and Xyla, Inc. The research will be presented at the Conference on Neural Information Processing Systems.

Rethinking pretraining

Machine-learning models are typically pretrained, which means they are trained on one dataset first to help them build parameters that can be used to tackle a different task. A model for classifying X-rays might be pretrained using a huge dataset of synthetically generated images before it is trained for its actual task using a much smaller dataset of real X-rays.

These researchers previously showed that they could use a handful of image generation programs to create synthetic data for model pretraining, but the programs needed to be carefully designed so the synthetic images matched up with certain properties of real images. This made the technique difficult to scale up.

In the new work, they used an enormous dataset of uncurated image generation programs instead.

They began by gathering a collection of 21,000 images generation programs from the internet. All the programs are written in a simple programming language and comprise just a few snippets of code, so they generate images rapidly.

“These programs have been designed by developers all over the world to produce images that have some of the properties we are interested in. They produce images that look kind of like abstract art,” Baradad explains.

These simple programs can run so quickly that the researchers didn’t need to produce images in advance to train the model. The researchers found they could generate images and train the model simultaneously, which streamlines the process.

They used their massive dataset of image generation programs to pretrain computer vision models for both supervised and unsupervised image classification tasks. In supervised learning, the image data are labeled, while in unsupervised learning the model learns to categorize images without labels.

Improving accuracy

When they compared their pretrained models to state-of-the-art computer vision models that had been pretrained using synthetic data, their models were more accurate, meaning they put images into the correct categories more often. While the accuracy levels were still less than models trained on real data, their technique narrowed the performance gap between models trained on real data and those trained on synthetic data by 38 percent.

“Importantly, we show that for the number of programs you collect, performance scales logarithmically. We do not saturate performance, so if we collect more programs, the model would perform even better. So, there is a way to extend our approach,” Manel says.

The researchers also used each individual image generation program for pretraining, in an effort to uncover factors that contribute to model accuracy. They found that when a program generates a more diverse set of images, the model performs better. They also found that colorful images with scenes that fill the entire canvas tend to improve model performance the most.

Now that they have demonstrated the success of this pretraining approach, the researchers want to extend their technique to other types of data, such as multimodal data that include text and images. They also want to continue exploring ways to improve image classification performance.

“There is still a gap to close with models trained on real data. This gives our research a direction that we hope others will follow,” he says.

Courtesy MIT. By Adam Zewe Article available here

Tags: artificial intelligenceComputer visionMachine learningMIT

Related Posts

When cutting costs, don’t lose sight of long-term organizational health
Business

When Cutting Costs, Don’t Lose Sight of Long-Term Organizational Health

U. S. Businesses need to be more prepared for physical risks
Business

U.S. Businesses Need to Be More Prepared for Physical Risks

Keeping track of communications when going remote
Business

Keeping Track Of Communications When Going Remote

Research: the complicated effects of pay transparency
Business

Research: The Complicated Effects of Pay Transparency

Esg outlook 2023: greater scrutiny on companies & investments
Business

ESG Outlook 2023: Greater Scrutiny on Companies & Investments

Upwork study finds 60 million americans freelancing in 2022
Business

Upwork Study Finds 60 Million Americans Freelancing in 2022

Global tech spend will slow to 4. 7% in 2023
Business

Global Tech Spend Will Slow To 4.7% In 2023

What’s hot this year in supply chain technology
Business

What’s hot this year in supply chain technology

The pros and cons of using ghost jobs in your recruiting strategy
Business

The Pros and Cons of Using Ghost Jobs in Your Recruiting Strategy

Remote work, intermittent leave make fmla compliance more challenging
Business

Remote work, intermittent leave make FMLA compliance more challenging

No Result
View All Result

Recent Posts

  • Symbotic’s CEO Rick Cohen shines a light on the consumer goods supply revolution
  • CEO John Wynne outlines how Fortis Solution Group is disrupting the packaging industry
  • CEO Keh-Shew Lu explains how Diodes Inc is pushing the frontiers of connectivity
  • CEO NA Andrea Pirondini explains why the Prysmian Group is best-in-class at cable solutions
  • Eric Clark on how ONE NTT is driving client-led innovation

Recent Comments

    Archives

    Categories

    • Art & Culture
    • Business
    • CEO Interviews
    • CEO Life
    • Editor´s Choice
    • Entrepreneur
    • Environment
    • Food
    • Health
    • Highlights
    • Industry
    • Innovation
    • Issues
    • Management & Leadership
    • Multimedia
    • News
    • Opinion
    • PrimeZone
    • Printed Version
    • Travel
    • Uncategorized

    Meta

    • Log in
    • Entries feed
    • Comments feed
    • WordPress.org

    • CONTACT
    • GENERAL ENQUIRIES
    • ADVERTISING
    • MEDIA KIT
    • DIRECTORY
    • TERMS AND CONDITIONS

    Editorials – stuart.james@ceo-na.com

    Advertising – media@ceo-na.com

    NEW YORK

    110 Wall St.,
    3rd Floor
    New York, NY.
    10005
    USA
    +1 212 432 5800

     

    MEXICO CITY

    Paseo de la Reforma 296,
    Floor 38
    Mexico City
    06600
    MEXICO

    • CONTACT
    • GENERAL ENQUIRIES
    • ADVERTISING
    • MEDIA KIT
    • DIRECTORY
    • TERMS AND CONDITIONS

    Editorials –
    stuart.james@ceo-na.com

    Editor-In-Chief –

    caroline.sposto@ceo-na.com

    Editorials – editorials@ceo-na.com

    Advertising –
    media@ceo-na.com

    NEW YORK

    110 Wall St.,
    3rd Floor
    New York, NY.
    10005
    USA
    +1 212 432 5800

    MEXICO CITY

    Paseo de la Reforma 296,
    Floor 38
    Mexico City
    06600
    MEXICO

    CEO North America © 2022 - Sitemap

    No Result
    View All Result
    • Home
    • Business
      • Entrepreneur
      • Industry
      • Innovation
      • Management & Leadership
    • CEO Interviews
    • CEO Life
      • Art & Culture
      • Food
      • Health
      • Travel
      • Environment
    • Opinion
    • News
    • Multimedia

    © 2023 JNews - Premium WordPress news & magazine theme by Jegtheme.

    Welcome Back!

    Login to your account below

    Forgotten Password?

    Retrieve your password

    Please enter your username or email address to reset your password.

    Log In

    Warning: array_sum() expects parameter 1 to be array, null given in /home/ceonacom/public_html/wp-content/plugins/jnews-social-share/class.jnews-social-background-process.php on line 111