With AI-powered object detection, you can now seamlessly integrate the latest advances in foundation models into your warehouse and construction site safety operations. As the demand for better safety monitoring continues to rise, it's essential for teams to maximize protective equipment use to mitigate potential hazards. Labelbox empowers the world’s largest organizations to leverage AI solutions tailored to their unique safety detection challenges.
However, teams can face multiple challenges when implementing AI for large-scale safety detection. This includes:
Labelbox is a data-centric AI platform that empowers businesses to transform their safety detection through advanced computer vision techniques. Instead of relying on time-consuming manual human review, companies can leverage Labelbox’s AI-assisted data enrichment and flexible training frameworks to quickly build task-specific models that uncover actionable insights for the prevention of supply chain mistakes.
In this guide, we’ll walk through an end-to-end workflow on how your team can leverage Labelbox’s platform to build a powerful task-specific model to improve safety detection using personal protective equipment as an example. Specifically, this guide will walk through how you can explore and better understand your visual assets to make more data-driven business decisions for worker safety.
The walkthrough below covers Labelbox’s platform across Catalog, Annotate, and Model. We recommend that you create a free Labelbox account to best follow along with this tutorial.
Part 1: Explore and enhance your data with Catalog and Foundry
Part 2: Create a model run and evaluate model performance
You can follow along with both parts of the tutorial below via:
Follow along with the tutorial and walkthrough in the Colab Notebook. If you are following along, please make a copy of the notebook.
Ingest data into Labelbox
For this tutorial, we’ll be working with a dataset that detects workers wearing safety equipment – with the goal of quickly curating data and finding protective equipment (e..g, helmets, goggles, reflective vests, gloves, masks, etc) from high-volumes of images.
The first step will be to gather data:
Please download the dataset and store it in an appropriate location on your environment. You'll also need to update the read/write file paths throughout the notebook to reflect relevant locations on your environment. You'll also need to update all references to API keys, and Labelbox ontology, project, and model run IDs
Once you’ve uploaded your dataset, you should see your image data rendered in Labelbox Catalog. You can browse through the dataset and visualize your data in a no-code interface to quickly pinpoint and curate data for model training.
You’ll now be able to see your dataset in Labelbox Catalog. With Catalog, you can contextualize your data with custom metadata and attachments to each asset for greater context.
In this demo, we'll be using Catalog to find relevant images of equipment for our dataset with the goal of annotating bounding boxes for the personal protective equipment using foundation models.
Leverage custom and out-of-the-box smart filters and embeddings to quickly explore product listings, surface similar data, and optimize data curation for ML. You can:
In this next step, we'll walk through how you can take a human-in-the-loop approach to iterate or modify pre-labels and speed up the annotation process.
Model Foundry enables teams to choose from a library of models and in this case, we'll be using GroundingDINO to generate previews and attach them as pre-labels.
With Model Foundry, you can automate data workflows, including data labeling with world-class foundation models. Leverage a variety of open source or third-party models to accelerate pre-labeling and cut labeling costs by up to 90%.
The next step is to take all of our labeled data and train our model on it. This allows us to make predictions on this data and for Labelbox to calculate evaluation metrics so that we can see where the model is going wrong and improve model performance.
As a last step, let's compare model inferences with ground-truth annotations to see where the model may be underperforming. A disagreement between model predictions and ground truth labels is typically due to either a model error (poor model prediction) or a labeling mistake (ground truth is wrong).
After running error analysis, you can make more informed decisions on how to iterate and improve your model’s performance with corrective action or targeted data selection and additional labeling.
By analyzing high-volumes of images and videos using foundation models and human alignment, Labelbox provides teams with the ability to inject valuable insights for delivering better protective equipment detection models for warehouses and construction sites that allow you to improve operational efficiency, compliance and overall worker safety.
Labelbox is a data-centric AI platform that empowers teams to iteratively build powerful task-specific models. To get started, sign up for a free Labelbox account or request a demo.