Components / LiDAR
Sensor 3 variants normalized

LiDAR

LiDAR appears across 17 tracked robots, concentrated in Humanoid, Cleaning, and Commercial. Use this page to understand why the signal matters, who relies on it most, and which live profiles deserve the first comparison click.

Tracked robots

17

Ready now

11

Manufacturers

16

Public prices

5

Why it matters

What it tends to unlock

Perception, mapping, detection, and safer motion decisions, cleaner autonomy loops when the robot needs environmental context, and higher-quality data for navigation, manipulation, or monitoring.

What to verify

Do not stop at the label

Coverage, placement, and how the sensor performs in messy conditions, what decisions actually rely on the sensor versus backup systems, and whether the label signals depth, proximity, or full-scene understanding.

Coverage

5 categories

The heaviest concentration is in Humanoid (10), Cleaning (2), and Commercial (2). Top manufacturers include Boston Dynamics (2), Agile Robots (1), and Agility (1).

Research brief

Research first. Sweep the roster second.

The useful questions here are how common LiDAR really is, which robot classes depend on it, and which live profiles are worth opening before you compare the whole stack.

Verified 30d

10

17 in the last 90 days

Top category

Humanoid

10 tracked robots

Paired most often with

Wi-Fi, IMU, and Ethernet

Sensor

Decision brief

What matters before you compare implementations

Where it helps most

  • perception, mapping, detection, and safer motion decisions
  • cleaner autonomy loops when the robot needs environmental context
  • higher-quality data for navigation, manipulation, or monitoring

What to validate

  • coverage, placement, and how the sensor performs in messy conditions
  • what decisions actually rely on the sensor versus backup systems
  • whether the label signals depth, proximity, or full-scene understanding

Evidence basis

What this route is grounded in

  • Aggregated from each robot's `specs.sensors` field in ui44 data.

Market snapshot

Use the structure first: which categories lean on LiDAR, which manufacturers repeat it, and what usually ships beside it.

Lead category

Humanoid

10 tracked robots currently anchor this label.

Most repeated manufacturer

Boston Dynamics

2 tracked robots make this the clearest manufacturer-level signal on the route.

Most common adjacent signal

Wi-Fi

12 shared robots pair this component with Wi-Fi.

Top categories

# Name Usage
1 Humanoid 10 robots
2 Cleaning 2 robots
3 Commercial 2 robots
4 Research 2 robots
5 Home Assistants 1 robot

Top manufacturers

# Name Usage
1 Boston Dynamics 2 robots
2 Agile Robots 1 robot
3 Agility 1 robot
4 Dyson 1 robot
5 Italian Institute of Technology 1 robot
6 KAIST 1 robot

Commonly paired with LiDAR

# Name Shared robots
1 Wi-Fi 12 robots
2 IMU 9 robots
3 Ethernet 5 robots
4 Bluetooth 4 robots
5 Cliff Sensors 3 robots
6 Force/Torque Sensors 3 robots

How to read the market

Structure first, prose second.

Category concentration tells you where the component is actually doing work, manufacturer repetition shows whether the signal is market-wide or vendor-specific, and pairings reveal which neighboring technologies usually ship alongside it.

At a glance

Kind Sensor
Tracked robots 17
Ready now 11
Public prices 5
Official sources 17
Variants normalized 3

Robot directory · LiDAR

The old card wall is replaced with a featured first-click strip and a dense inventory table so the route behaves like a serious directory.

Directory briefing

Featured first, dense sweep second.

Open the clearest profiles first, then sweep the full inventory in a denser table. Featured cards are selected by readiness, image quality, and official source availability, so the first click is usually the most informative one.

Ready now

11

Public price

5

Official links

17

Featured now

3

How to scan this directory

Use the shortest credible path through the roster.

  • Featured cards: start with the strongest documented profiles to understand real implementation quality fast.
  • Inventory table: sweep the whole market once you know which profiles deserve serious comparison.
  • Compare intent: use status, official links, and standout specs before treating the label itself as proof.

Best first clicks

Open these before sweeping the full inventory

These robots score highest on readiness, public detail quality, and image clarity, making them the fastest way to understand how LiDAR shows up in practice.

Atlas (Electric) by Boston Dynamics — Humanoid robot
Active Humanoid
Boston Dynamics Since 2024

Atlas (Electric)

Boston Dynamics' fully electric humanoid robot, unveiled at CES 2026, designed for a wide array of industrial tasks from material handling to order fulfillment. Production began immediately at Boston headquarters, with 2026 deployments fully committed — fleets shipping to Hyundai's Robotics Metaplant Application Center (RMAC) and Google DeepMind, with additional customers planned for early 2027. Atlas features 56 degrees of freedom, fully rotational joints, a 2.3m reach, and 50kg instant lift capacity. The robot autonomously swaps its own batteries in under 3 minutes for continuous 24/7 operation. Trained using AI foundation models including a partnership with Google DeepMind, with fleet-wide task replication — once one Atlas learns a task, it deploys across the entire fleet. IP67-rated for harsh environments, with fenceless human safety guarding. Offered to qualified enterprise prospects, not sold to normal consumers. Successor to the hydraulic Atlas research platform.

Public price

Price TBA

No official pricing published

Battery

~4 hours

Charge Not disclosed

Shortlist read

Active in the catalog with enough detail to review immediately.

Profile
BellaBot by Pudu Robotics — Commercial robot
Active Commercial
Pudu Robotics Since 2020

BellaBot

Pudu Robotics' premium food delivery robot, one of the most widely deployed commercial service robots in the world. BellaBot features an innovative bionic cat-face design with multimodal interaction (touch, light, voice), 3D omnidirectional obstacle avoidance with RGBD cameras and LiDAR, and a dual SLAM positioning system (LiDAR + Visual SLAM). The robot navigates autonomously through restaurants, hotels, and healthcare facilities, delivering food and items on up to four trays. It supports hot-swappable batteries for 24/7 operation. Deployed in over 60 countries across 600+ cities with tens of thousands of units in service. BellaBot responds to petting with cat-like animations and sounds, making it a crowd favorite in the hospitality industry.

Public price

Price TBA

Contact manufacturer for pricing…

Battery

13 hours (no load)

Charge 4.5 hours (or instant with battery swap)

Shortlist read

Active in the catalog with enough detail to review immediately.

Profile
Digit by Agility — Humanoid robot
Active Humanoid
Agility Since 2023

Digit

Purpose-built humanoid for logistics and warehouse operations. Commercially deployed at multiple Fortune 500 companies including Amazon, Toyota Motor Manufacturing Canada (RaaS deal, Feb 2026), Mercado Libre (Dec 2025), Schaeffler, and GXO. Produced at Agility's RoboFacility in Salem, Oregon. Enterprise RaaS (Robots-as-a-Service) model — no consumer pricing available.

Public price

Price TBA

Enterprise deployment via Agility sales…

Battery

~4 hours

Charge ~2 hours

Shortlist read

Active in the catalog with enough detail to review immediately.

Profile

Full inventory · 17 robots

Compact mobile scan: status, price, standout context, and links stay visible without sideways scrolling.

Quick answers

FAQ

The short version of what this label means in the ui44 catalog, where it matters, and how to compare it without over-reading the marketing copy.

Frequently Asked Questions

How common is LiDAR in the database?

LiDAR currently appears on 17 tracked robots across 16 manufacturers. That makes this route useful for both deep research and fast shortlist scanning, not just one-off editorial reading.

Which robot categories lean on LiDAR the most?

The strongest concentration is in Humanoid (10), Cleaning (2), and Commercial (2). Category mix is the fastest clue for whether this component behaves like baseline plumbing or a more selective differentiator.

Does LiDAR usually show up on ready-to-buy robots?

11 of the 17 tracked profiles are currently marked Available or Active. That means the label has live market relevance here, but you should still open the profiles with public pricing or official links first before treating it as a clean buyer signal.

What should I compare first on this page?

Start with readiness, official source quality, and the standout spec column in the inventory table. On component routes, those three signals usually remove weak profiles faster than reading every descriptive paragraph.

What usually ships alongside LiDAR?

The strongest shared-stack signals here are Wi-Fi (12), IMU (9), and Ethernet (5). Use those pairings to branch into adjacent component pages when one label is too narrow for the decision.

Are there enough public price points to benchmark this component?

5 matching robots currently expose public pricing. That is enough to create directional context, but not enough to treat one price bracket as the whole market. Use the directory to find the transparent profiles first, then widen the sweep.

Which manufacturers are worth opening first?

Start with Boston Dynamics (2), Agile Robots (1), and Agility (1). Repetition across manufacturers is often the clearest signal that the component is part of a stable market pattern rather than a one-off marketing callout.

Reference library

The original long-form component research is still here, but collapsed so the main route can prioritize hierarchy and scan speed.

Fundamentals

The baseline explanation of what LiDAR is, why it matters, and how to think about it before comparing implementations.

What Is LiDAR?

LiDAR is a sensor component found in 17 robots tracked in the ui44 Home Robot Database. As a sensor technology, LiDAR plays a specific role in enabling robot perception, interaction, or operation depending on its implementation in each platform.

At a Glance

Component Type

Sensor

Used By

17 robots

Manufacturers

Agile Robots, Boston Dynamics, Pudu Robotics +13 more

Categories

Humanoid, Commercial, Research +2 more

Price Range

$1.2k – $150k

Available Now

11 robots

Sensors are the perceptual backbone of any robot. They convert physical phenomena — light, sound, distance, motion, temperature — into digital signals that the robot's AI can process and act upon.

Key Points

  • Convert physical phenomena into digital signals
  • Enable obstacle detection, navigation, and object recognition
  • Without sensors, a robot cannot interact safely with its environment

In the ui44 database, LiDAR is categorized under Sensor components. For a comprehensive explanation of all component types, consult the components glossary.

Why LiDAR Matters in Robotics

The sensor suite is one of the most important differentiators between robots. Robots with richer sensor arrays can navigate more complex environments, avoid obstacles more reliably, and perform more nuanced tasks.

Directly impacts what a robot can actually do in practice — not just on paper

Richer sensor arrays enable more complex navigation and interaction

Determines obstacle avoidance reliability and object/person recognition

LiDAR Adoption

Used in 17 robots across 5 categories (Humanoid, Commercial, Research, Cleaning…), indicating broad applicability across the robotics industry.

How LiDAR Works

Modern robot sensors work by emitting or detecting various forms of energy. The robot's processor fuses data from multiple sensors simultaneously (sensor fusion) to build a coherent understanding of its surroundings.

1

Active sensors

LiDAR and ultrasonic emit signals and measure reflections to determine distance and shape

2

Passive sensors

Cameras and microphones detect ambient light and sound without emitting anything

3

Sensor fusion

The processor combines data from all sensors simultaneously for a coherent environmental picture

LiDAR Integration

Implementation varies by robot platform and manufacturer. Each robot integrates LiDAR differently depending on system architecture, use case, and target tasks. Integration with other onboard sensors and the main processing unit determines real-world performance.

Technical notes and use cases

Deeper technical framing, matched technology profiles, and the longer use-case treatment for LiDAR.

LiDAR: Detailed Technology Analysis

In-depth technical analysis of 1 technology domain relevant to this component

Technology Overview

While the sections above cover general sensor principles, this analysis focuses on the particular technology domains relevant to LiDAR based on its implementation characteristics.

LiDAR & Time-of-Flight Ranging

LiDAR (Light Detection and Ranging) and time-of-flight sensors measure distances by emitting light pulses and measuring the time they take to reflect back from surfaces. This principle enables precise, three-dimensional mapping of the robot's environment regardless of ambient lighting conditions — a significant advantage over camera-only systems that struggle in darkness or strong direct sunlight. In home robotics, LiDAR has become the gold standard for floor plan mapping and systematic navigation.

Read full technical analysis

Two main LiDAR architectures exist in consumer robotics. Mechanical spinning LiDAR uses a rotating mirror or emitter assembly to sweep a laser beam 360° around the robot, building a complete horizontal distance profile with each revolution. This technology is proven and reliable but involves moving parts that can wear over time. Solid-state LiDAR eliminates moving components by using arrays of emitters and detectors, or MEMS (micro-electromechanical) mirrors, to steer the beam electronically. Solid-state designs are more compact, potentially more durable, and increasingly cost-effective, though they may have slightly different field-of-view characteristics than spinning units.

Time-of-flight sensors used in robotics typically operate with infrared laser diodes at wavelengths around 850-940 nm, which are invisible to the human eye. Consumer robots universally use Class 1 eye-safe lasers, meaning the beam intensity is low enough to be safe even with direct eye exposure. The precision of these sensors — typically 1-3 cm at ranges up to 12 meters for consumer-grade units — enables robots to build room maps accurate enough for efficient navigation and furniture avoidance. More advanced implementations combine LiDAR distance data with camera imagery in a process called sensor fusion, creating rich 3D environmental models that combine the geometric precision of LiDAR with the semantic richness of visual data.

LiDAR: Technical Deep Dive

Beyond the high-level overview, understanding the technical foundations of sensor technologies like LiDAR helps buyers and researchers evaluate implementations more critically.

Engineering Principles

Every sensor converts a physical quantity into an electrical signal that can be digitized and processed. The raw analog output is conditioned through amplification, filtering, and A/D conversion before reaching the processor.

  • Optical sensors use photodiodes or CMOS arrays to detect photons
  • Acoustic sensors use piezoelectric elements to detect pressure waves
  • Inertial sensors use MEMS to detect acceleration and rotation
  • Range sensors use time-of-flight or structured light for distance measurement

Performance Characteristics

Sensor performance involves key metrics with inherent engineering trade-offs.

Accuracy How close the reading is to the true value
Precision Consistency across repeated measurements
Resolution Smallest detectable change in measurement
Sampling rate Reading frequency — critical for fast-moving robots
Field of view Spatial coverage area of the sensor

Technological Evolution

Sensor technology in robotics has evolved dramatically over the past decade.

Early home robots relied on simple bump sensors and infrared proximity detectors

Today's platforms incorporate multi-spectral cameras, solid-state LiDAR, and millimeter-wave radar

Miniaturization: sensors that filled circuit boards now fit into fingernail-sized packages

Next frontier: sensor fusion at the hardware level — multiple sensing modalities in single chip-scale packages

Known Limitations

No sensor is perfect in all conditions. Understanding limitations is critical for evaluating robots in specific environments.

  • Optical sensors struggle in direct sunlight or complete darkness
  • LiDAR can be confused by mirrors, glass, and highly reflective surfaces
  • Ultrasonic sensors may produce false readings in complex acoustic environments
  • Dust, fog, rain, and temperature extremes can degrade performance

Use Cases & Applications for LiDAR

Key application domains for sensor technologies like LiDAR.

Autonomous Navigation

Sensors enable robots to build maps of their environment, detect obstacles in real time, and plan collision-free paths. This is essential for both indoor robots (navigating furniture and doorways) and outdoor robots (handling terrain variations and weather conditions). The quality and coverage of the sensor array directly determines how reliably a robot can navigate without human intervention.

Object Recognition & Manipulation

Advanced sensors allow robots to identify objects by shape, color, and texture, enabling tasks like picking up items, sorting packages, or recognizing faces. Depth-sensing technologies are particularly important for calculating object distances and sizes, which is necessary for precise manipulation in both home and industrial settings.

Safety & Collision Avoidance

In environments shared with humans, sensors provide the critical safety layer that prevents robots from causing harm. Proximity sensors, bumper sensors, and vision systems work together to detect people and obstacles, triggering immediate stop or avoidance maneuvers. This is a fundamental requirement for any robot operating in homes, hospitals, or public spaces.

Environmental Monitoring

Sensors can measure temperature, humidity, air quality, and other environmental parameters. Robots equipped with these sensors can perform automated monitoring rounds in warehouses, data centers, or homes, alerting users to abnormal conditions like water leaks, temperature spikes, or poor air quality.

Human-Robot Interaction

Microphones, cameras, and touch sensors enable natural interaction between robots and humans. These sensors allow robots to recognize voice commands, detect gestures, respond to touch, and maintain appropriate social distances during conversations or collaborative tasks.

150 Capabilities Across 17 robots

Bipedal Walking Autonomous Navigation Dexterous Manipulation (21-joint hands) Force-Controlled Grasping Human-Robot Interaction Material Handling Machine Tending Tool Use Precision Assembly Heavy Lifting (50kg Instant, 30kg Sustained) Precise Manipulation Dynamic Recovery Industrial Tasks Self-Swappable Batteries (<3 min) Barcode/RFID Integration Fenceless Human Safety Guarding +134 more

Visit each robot's detail page to see which capabilities are available on specific models.

Market breakdown and adjacent routes

Manufacturer mix, specs context, price context, category overlap, and adjacent components worth branching into next.

LiDAR by Manufacturer

LiDAR is used by 16 manufacturers — showing how widely this technology is deployed across the industry.

Manufacturer Models
Boston Dynamics 2 robots
Agile Robots 1 robot
Pudu Robotics 1 robot
Agility 1 robot
KAIST 1 robot
Italian Institute of Technology 1 robot
XPENG Robotics 1 robot
Kawasaki Heavy Industries 1 robot
Leju Robotics 1 robot
LimX Dynamics 1 robot
Shark 1 robot
Pollen Robotics 1 robot
Oversonic Robotics 1 robot
Dyson 1 robot
Zeroth Robotics 1 robot
UBTECH 1 robot

Specifications Comparison: Robots With LiDAR

Side-by-side comparison of all 17 robots using LiDAR.

Robot Price Status
Agile ONE Development
Atlas (Electric) Active
BellaBot Active
Digit Active
DRC-HUBO+ Prototype
ergoCub Active
Iron $150k Development
Kaleido 9 Prototype
Kuavo 5 $38k Prototype
Luna Prototype
PowerDetect UV Reveal 2-In-1 $1.3k Available
Reachy 2 Active
RoBee R Active
Spot+Scrub Ai $1.2k Available
Stretch Active
W1 $5.0k Available
Walker S Active

LiDAR Across Robot Categories

LiDAR spans 5 robot categories — from consumer to research platforms.

Technologies most often paired with LiDAR across 17 robots.

Browse the full components directory or see the components glossary for detailed explanations of each technology.

Price Context for Robots With LiDAR

5 of 17 robots with LiDAR have public pricing, ranging $1.2k$150k. 12 robots use custom or enterprise pricing.

Lowest

$1.2k

Spot+Scrub Ai

Average

$39.1k

5 robots with pricing

Highest

$150k

Iron

Alternatives to LiDAR

561 other sensor technologies tracked in ui44, ranked by adoption.

Browse all Sensor components or use the robot comparison tool to evaluate how different sensor configurations perform across specific robot models.

LiDAR in the Broader Robotics Industry

The robotics sensor market is one of the fastest-growing segments in the broader sensor industry. As robots move from controlled industrial environments into unstructured home and commercial spaces, the demands on sensor technology increase dramatically.

Key Industry Trends

Multi-modal sensing

Robots combine multiple sensor types (vision, depth, tactile, inertial) to build comprehensive environmental understanding

Miniaturization

Sensors that once occupied entire circuit boards now fit into fingernail-sized packages, making advanced sensing affordable for consumer robots

Edge AI integration

AI processing directly in sensor modules enables faster perception without cloud latency

Industry Adoption Snapshot

LiDAR is adopted by 17 robots from 16 manufacturers in the ui44 database, providing a data-driven view of real-world deployment patterns.

Certifications & Standards

IP67 CE FCC TÜV (sanitization) CE Compliant ISO 27001 EMC-EMI Compliant IPX4 Protection GDPR Compliant

Certifications carried by robots incorporating LiDAR, indicating compliance with safety, EMC, and quality standards.

Integration & Ecosystem Compatibility

Platform compatibility, voice integration, and AI capabilities across robots with LiDAR.

Buyer and operations guidance

The long-form buyer, maintenance, and troubleshooting material kept available without forcing it into the main scan path.

Buyer Considerations for LiDAR

If LiDAR is an important factor in your robot selection, here are key considerations to guide your decision.

What to Look For in Sensor Components

Coverage area

Does the sensor array provide 360° awareness or only forward-facing detection?

Range

How far can the robot sense obstacles or objects?

Resolution

How detailed is the sensor data for recognition tasks?

Redundancy

Are there backup sensors if one fails?

Serviceability

Are sensors user-serviceable or require manufacturer maintenance?

How to Evaluate LiDAR

Integration Quality

A component is only as good as its integration. Check how the manufacturer has incorporated LiDAR into the overall robot design and software stack.

Complementary Components

Review what other sensor technologies are paired with LiDAR in each robot — see the related components section.

Category Fit

Make sure the robot's category matches your use case. LiDAR serves different roles in different robot types.

Manufacturer Track Record

Consider the manufacturer's reputation for software updates, support, and component reliability.

Compare Before You Buy

Use the ui44 comparison tool to evaluate robots with LiDAR side by side.

Maintenance & Longevity: LiDAR

Overview

Sensors are among the most maintenance-sensitive components in a robot. Their performance can degrade over time due to physical wear, environmental exposure, and calibration drift. Understanding the maintenance profile of a robot's sensor suite helps set realistic expectations for long-term ownership and operation.

Durability & Reliability

Sensor durability varies significantly by type. Solid-state sensors like IMUs and accelerometers have no moving parts and typically last the lifetime of the robot.

  • Optical sensors like cameras and LiDAR can accumulate dust, scratches, or condensation on their lenses over time.
  • Mechanical sensors such as bump sensors and encoders may experience wear on moving contacts.
  • Environmental sensors for temperature and humidity are generally robust but can be affected by corrosive environments.
  • Overall, sensor failure rates in modern consumer robots are low, but environmental factors like dust accumulation and UV exposure can gradually degrade performance rather than cause sudden failure.
Ongoing Maintenance

Regular sensor maintenance primarily involves keeping optical surfaces clean. Camera lenses, LiDAR windows, and infrared emitters should be wiped with a soft, lint-free cloth to remove dust and fingerprints.

  • Many modern robots perform automatic sensor self-diagnostics and will alert users when calibration has drifted beyond acceptable limits.
  • Some robots support user-initiated recalibration routines for specific sensors.
  • For robots used in dusty or pet-heavy environments, more frequent cleaning of sensor surfaces may be necessary.
  • Manufacturer documentation typically includes sensor care instructions specific to the robot's sensor configuration.
Future-Proofing Considerations

When evaluating sensor technology for long-term value, consider the manufacturer's track record for software updates that improve sensor utilization. A robot with good sensors and ongoing software development can actually improve its performance over time as algorithms are refined.

  • However, sensor hardware itself cannot be upgraded post-purchase on most consumer robots, making the initial sensor specification an important long-term consideration.
  • Robots with modular sensor designs that allow component replacement offer better long-term maintainability, though this is currently more common in commercial and research platforms than consumer products.

For the 17 robots in the ui44 database using LiDAR, we recommend checking the individual robot pages for manufacturer-specific maintenance guidance and support documentation. Each manufacturer has different support policies, update frequencies, and warranty terms that affect the long-term ownership experience of their sensor technologies.

Troubleshooting & Common Issues: LiDAR

Sensor-related issues are among the most common problems home robot owners encounter. Many sensor issues can be resolved with simple maintenance or environmental adjustments, while others may indicate hardware problems requiring manufacturer support. Understanding common failure modes helps you diagnose and resolve issues quickly, minimizing robot downtime.

Robot bumps into obstacles it should detect

Likely Causes

  • Dirty or obstructed sensor windows are the most frequent cause.
  • Dust, pet hair, fingerprints, or cleaning solution residue on LiDAR, camera, or infrared sensor surfaces significantly reduce detection accuracy.
  • Highly reflective surfaces like mirrors, glass doors, and glossy furniture can also confuse optical and laser-based sensors by creating phantom readings or absorbing signals entirely.

Resolution

  • Clean all sensor windows and lenses with a soft, dry microfiber cloth.
  • Avoid chemical cleaners unless the manufacturer specifically recommends them.
  • If cleaning does not resolve the issue, check for recent firmware updates that may address sensor calibration.
  • For persistent problems with specific surfaces, consider applying anti-reflective film to mirrors or glass surfaces in the robot's operating area.

Robot map becomes inaccurate or corrupted over time

Likely Causes

  • Sensor drift and calibration degradation can cause mapping errors.
  • Significant furniture rearrangement, new obstacles, or changed room layouts may confuse the mapping algorithm.
  • In some cases, electromagnetic interference from nearby electronics can affect sensor readings used for localization.

Resolution

  • Delete and rebuild the map from scratch using the manufacturer's app.
  • Ensure the robot's firmware is up to date, as mapping improvements are frequently included in updates.
  • If the problem recurs, run the robot during periods of minimal household activity to get the cleanest initial map.

Cliff or drop sensors trigger on flat surfaces

Likely Causes

  • Dark-colored flooring, transitions between floor materials, and thick carpet edges can trigger infrared cliff sensors.
  • Direct sunlight hitting the floor near the robot can also interfere with infrared detection by saturating the sensor with ambient infrared light.

Resolution

  • Clean the cliff sensors on the underside of the robot.
  • If the issue occurs at specific locations consistently, check whether the floor has very dark patches, strong color transitions, or high-gloss finishes that might confuse the sensors.
  • Some manufacturers allow cliff sensor sensitivity adjustment through the companion app.

When to Contact the Manufacturer

  • Contact the manufacturer if sensor issues persist after cleaning and firmware updates, if you notice physical damage to any sensor housing, or if the robot reports sensor errors in its diagnostic log.
  • Sensor calibration that cannot be corrected through standard procedures may indicate hardware degradation requiring professional service or component replacement.

For model-specific troubleshooting, visit the individual robot pages for the 17 robots using LiDAR. Each manufacturer provides model-specific support resources and diagnostic tools for their sensor implementations.