Get AI summaries of any video or article — Sign up free
Labeling (3)  - Data Management - Full Stack Deep Learning thumbnail

Labeling (3) - Data Management - Full Stack Deep Learning

The Full Stack·
5 min read

Based on The Full Stack's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Labeling tool features are standard across providers, but quality hinges on annotator instructions for subjective edge cases (e.g., occlusion boundaries).

Briefing

Data labeling hinges less on the annotation software’s feature list and more on the human decisions inside the labeling workflow—especially when “reasonable” people would draw different boundaries. Vision labeling tools typically offer the same core primitives—bounding boxes, segmentation polygons, keypoints, and cuboids for 3D or video—but quality depends on clear instructions, consistent interpretation, and active quality control. A concrete example is an object partially hidden behind another item: annotators must be told whether to label the entire imagined object or only the visible portion. Without that guidance, label noise rises from subjective judgment calls, and even well-intentioned annotators can drift.

That quality challenge shapes the three main ways teams source labeling labor: hire in-house annotators, crowdsource through marketplaces, or outsource to full-service labeling companies. Hiring directly offers tight control and faster iteration once trusted annotators are identified, including the ability to promote top performers into quality control roles and secure the work via contracts. The tradeoffs are cost, slower scaling (finding and managing real people takes time), and administrative overhead. Crowdsourcing—via services such as Amazon Mechanical Turk—can reduce cost and speed up throughput, but it adds complexity because annotators are transient and often poorly paid, which typically requires an additional quality-control layer.

Full-service data labeling companies aim to remove that operational burden by providing both the annotation interface and the quality assurance machinery. The workflow often looks like: spend a few days evaluating vendors, create a gold-standard labeled dataset internally to understand the true labeling complexity, send that same dataset to candidate providers, and compare their outputs against the internal standard. Pricing matters because vendors charge for recruiting labor, enforcing quality, and managing the process. Examples mentioned include Figure Eight (formerly CrowdFlower), which has labeled over 100 million images and accumulated over 1 billion human judgments by 2018, and Scale AI, which is positioned as an API-like service for labeling new image content. Other providers include Labelbox and Supervisely, while Prodigy is highlighted as an annotation tool that supports active learning and can reduce the amount of labeling needed by prioritizing uncertain samples.

When teams can’t afford heavy quality control, the transcript’s practical stance is blunt: robustness to bad samples can’t be assumed without seeing the relevant distribution. Models may generalize to unseen conditions, but there’s no reliable way to guarantee resilience to specific outliers unless those cases are labeled, tested, and then incorporated into training if needed.

Finally, the discussion turns to aggregation and “subjective” labels. If multiple annotators label the same task, teams can use majority vote or averaging for objective outputs like bounding boxes. For ranking or search relevance—where people disagree—there are multiple strategies: bin items into clearly high/low categories and treat the ambiguous region as noisy, or frame ranking as a regression problem by predicting average rank. Related questions about leveraging negative examples (e.g., images a user didn’t choose, or search links a user didn’t click) remain an active research area because “not chosen” doesn’t always mean “incorrect”—it may reflect randomness or user behavior rather than ground truth.

Cornell Notes

Labeling quality depends on human judgment, not just tool features. Even with standard primitives like bounding boxes and segmentation, teams must provide unambiguous instructions for edge cases (e.g., partially occluded objects) and enforce quality through training and review. Labor can be sourced by hiring in-house (high control but costly), crowdsource marketplaces like Amazon Mechanical Turk (cheaper but needs extra quality control), or outsource to full-service vendors such as Figure Eight or Scale AI (faster operationally but pricier). If quality control is limited, robustness to outliers can’t be guaranteed without labeling and testing those cases, then adding them to training. For subjective tasks like ranking, aggregation may use binning rules or regression-style targets like average rank.

Why do labeling instructions matter as much as the labeling tool’s features?

Because many labeling decisions are inherently subjective. The transcript’s example is a fox partially hidden behind a rock: annotators must be told whether to label the full imagined fox or only the visible portion. Without explicit tie-breaking rules, different annotators can produce reasonable but inconsistent labels, increasing noise even when everyone uses the same interface primitives (bounding boxes, polygons, keypoints, cuboids).

What are the tradeoffs between hiring annotators, crowdsource labor, and using full-service labeling companies?

Hiring annotators in-house offers control: teams can identify the best workers, train them, and promote top performers into quality control roles; contracts can secure the process. Downsides are expense, slower scaling, and admin overhead. Crowdsourcing (e.g., Amazon Mechanical Turk) can be faster and cheaper, but annotators are transient and often poorly paid, so it typically requires a separate quality-control layer. Full-service vendors package both the annotation interface and quality assurance; teams mainly need to train the vendor on their task and compare outputs against a gold-standard dataset, but they pay a premium for the operational overhead.

How should a team evaluate labeling vendors before scaling up?

The transcript recommends dedicating several days to vendor selection. Teams should label a gold-standard dataset themselves first to understand the labeling complexity and desired conventions. Then they send that same dataset to multiple contenders and compare each vendor’s results against the internal gold standard. Pricing also becomes a key comparison point because vendors charge for labor, quality management, and workflow polish.

What does the transcript say about making models robust when label quality control is limited?

It argues there’s no reliable way to claim robustness to specific outliers without having those cases in the labeled data distribution. A model might generalize to unseen conditions, but it may fail on particular parts of the distribution it hasn’t seen. The only practical test is to label and evaluate those cases; if performance breaks, those examples should be added to training.

How can multiple annotators’ outputs be combined when labels are subjective (like ranking)?

For more objective tasks (e.g., bounding boxes), teams can use majority rule or averaging across labelers. For ranking or search relevance—where annotators disagree on exact ordering—one approach is binning: identify items that are consistently top-ranked or bottom-ranked and treat the ambiguous middle as noisy. Another approach is to treat ranking as regression by predicting an average rank rather than an exact ordering.

Can “negative” user feedback (not choosing an item) be treated as ground truth?

The transcript frames it as complicated: in search or selection interfaces, a user clicking one link doesn’t prove the others are incorrect. The clicked result might be the best, or it might reflect randomness or user behavior. This uncertainty is why leveraging negatives from interaction data is an active area of research and development.

Review Questions

  1. What edge-case decision in labeling requires explicit instructions to reduce subjective disagreement, and why?
  2. Compare how quality control needs differ between in-house annotators, Amazon Mechanical Turk-style crowdsourcing, and full-service labeling vendors.
  3. For ranking tasks with annotator disagreement, what two aggregation strategies are suggested, and how do they differ conceptually?

Key Points

  1. 1

    Labeling tool features are standard across providers, but quality hinges on annotator instructions for subjective edge cases (e.g., occlusion boundaries).

  2. 2

    Quality assurance is essential because reasonable people can disagree; tie-breaking rules and monitoring reduce label noise.

  3. 3

    In-house annotators provide maximum control and allow building a trusted quality-control pipeline, but they cost more and scale more slowly.

  4. 4

    Crowdsourcing can be cheaper and faster, yet transient annotators typically require an additional quality-control layer.

  5. 5

    Full-service labeling companies shift operational burden to vendors; teams should validate them using a gold-standard dataset and output comparisons.

  6. 6

    Model robustness to outliers can’t be assumed without labeling and testing those cases; failures should feed back into training data.

  7. 7

    For subjective tasks like ranking, aggregation may use binning rules or regression targets such as average rank rather than forcing a single “true” ordering.

Highlights

The transcript’s core labeling lesson: standard annotation primitives don’t prevent disagreement—unclear conventions for edge cases do.
Vendor evaluation should start with an internal gold-standard dataset, then compare vendor outputs directly against it.
Robustness to bad samples isn’t guaranteed by hope; it requires labeling, testing, and retraining on the relevant distribution.
Ranking aggregation is harder than bounding boxes; binning or regression-style targets can better handle disagreement.
“Not clicked” or “not chosen” isn’t automatically “wrong,” making negative-example learning a research challenge.

Topics

  • Data Labeling
  • Annotation Quality
  • Labeling Labor
  • Vendor Evaluation
  • Subjective Aggregation

Mentioned