The Robot Report

  • Home
  • News
  • Technologies
    • Batteries / Power Supplies
    • Cameras / Imaging / Vision
    • Controllers
    • End Effectors
    • Microprocessors / SoCs
    • Motion Control
    • Sensors
    • Soft Robotics
    • Software / Simulation
  • Development
    • Artificial Intelligence
    • Human Robot Interaction / Haptics
    • Mobility / Navigation
    • Research
  • Robots
    • AGVs
    • AMRs
    • Consumer
    • Collaborative Robots
    • Drones
    • Humanoids
    • Industrial
    • Self-Driving Vehicles
    • Unmanned Maritime Systems
  • Business
    • Financial
      • Investments
      • Mergers & Acquisitions
      • Earnings
    • Markets
      • Agriculture
      • Healthcare
      • Logistics
      • Manufacturing
      • Mining
      • Security
    • RBR50
      • RBR50 Winners 2025
      • RBR50 Winners 2024
      • RBR50 Winners 2023
      • RBR50 Winners 2022
      • RBR50 Winners 2021
  • Resources
    • Automated Warehouse Research Reports
    • Digital Issues
    • eBooks
    • Publications
      • Automated Warehouse
      • Collaborative Robotics Trends
    • Search Robotics Database
    • Videos
    • Webinars / Digital Events
  • Events
    • RoboBusiness
    • Robotics Summit & Expo
    • DeviceTalks
    • R&D 100
    • Robotics Weeks
  • Podcast
    • Episodes
  • Advertise
  • Subscribe

DLRob announces self-training addition to its robot controller

By Brianna Wessling | July 3, 2023

dlrob autostaq

DLRob’s Autostaq allows robots to pack and stack a wide range of objects autonomously. | Source: DLRob

Deep Learning Robotics (DLRob), and AI and robotics technology company, announced a new feature for its vision-based controller that was released earlier this year. This feature, called Autostaq, allows robots to autonomously pack and stack a wide range of objects with little setup time. 

DLRob’s AI controller can enable robots to learn from human demonstrations. Now, with the latest feature, the controller has the ability to self-train using a unique combination of generated synthetic data and real performance data. 

Generating synthetic data and merging that data with the controller’s own real-world performance data allows it to achieve remarkable adaptability and accuracy in handling diverse objects and placing them in optimal locations with little or no setup time. This means there is no user demonstration needed. 

“We are thrilled to introduce this new feature of our vision-based robot controller, which marks a major milestone in the field of AI-powered robots and automation,” Deep Learning Robotics’ CEO Carlos Benaim said. “By leveraging our self-training approach, the controller gains an unprecedented level of proficiency, enabling robots to pack and stack virtually anything by finding optimal locations for each of the objects identified in the scene. This breakthrough has the potential to transform various industries, from logistics and warehousing to manufacturing and beyond.”

The robot controller’s software uses machine learning algorithms to allow robots to learn by observing and mimicking human actions. The software is designed with a user-friendly interface so that anyone with any level of robotic knowledge can teach the robots new tasks. 

The software can handle a wide range of robots and applications, including industrial manufacturing, home automation and more. It uses plug-and-play technology, which DLRob hopes will decrease implementation time. 

DLRob was founded in 2015 and is based in Ashdod, HaDaron, Isreal. It aims to change how robots are programmed and operated in both structured and unstructured environments. 

About The Author

Brianna Wessling

Brianna Wessling is an Associate Editor, Robotics, WTWH Media. She joined WTWH Media in November 2021, after graduating from the University of Kansas with degrees in Journalism and English. She covers a wide range of robotics topics, but specializes in women in robotics, robotics in healthcare, and space robotics.

She can be reached at [email protected]

Comments

  1. Jason Creech says

    July 6, 2023 at 9:21 pm

    This advancement sounds great. Is the vision system using a point cloud modeling to generate the robot off set coordinates for pick positions?

    Reply

Tell Us What You Think! Cancel reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Related Articles Read More >

Different types of robots, such as this cobot arm, rely on different motors for precise motion.
How motor design enables different types of robots
Force- and power-limited robots such as those from UR enable humans to work alongside robots for tasks such as welding, shown here.
The keys to designing safe and shared cobot workspaces
With an accurate robot, the planned path can be executed without manual interaction on the real robot.
High-precision robots: What to do when absolute accuracy is poor?
Gantry systems using ServoBelt technology can provide the automotive industry with flexibility at a fraction of the cost of rack-and-pinion systems, says Bell-Everman.
ServoBelt offers high-end performance for automotive gantry

RBR50 Innovation Awards

“rr
EXPAND YOUR KNOWLEDGE AND STAY CONNECTED
Get the latest info on technologies, tools and strategies for Robotics Professionals.

Latest Episode of The Robot Report Podcast

Automated Warehouse Research Reports

Sponsored Content

  • Supporting the future of medical robotics with smarter motor solutions
  • YUAN Unveils Next-Gen AI Robotics Powered by NVIDIA for Land, Sea & Air
  • ASMPT chooses Renishaw for high-quality motion control
  • Revolutionizing Manufacturing with Smart Factories
  • How to Set Up a Planetary Gear Motion with SOLIDWORKS
The Robot Report
  • Automated Warehouse
  • RoboBusiness Event
  • Robotics Summit & Expo
  • About The Robot Report
  • Subscribe
  • Contact Us

Copyright © 2026 WTWH Media LLC. All Rights Reserved. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of WTWH Media
Privacy Policy | Advertising | About Us

Search The Robot Report

  • Home
  • News
  • Technologies
    • Batteries / Power Supplies
    • Cameras / Imaging / Vision
    • Controllers
    • End Effectors
    • Microprocessors / SoCs
    • Motion Control
    • Sensors
    • Soft Robotics
    • Software / Simulation
  • Development
    • Artificial Intelligence
    • Human Robot Interaction / Haptics
    • Mobility / Navigation
    • Research
  • Robots
    • AGVs
    • AMRs
    • Consumer
    • Collaborative Robots
    • Drones
    • Humanoids
    • Industrial
    • Self-Driving Vehicles
    • Unmanned Maritime Systems
  • Business
    • Financial
      • Investments
      • Mergers & Acquisitions
      • Earnings
    • Markets
      • Agriculture
      • Healthcare
      • Logistics
      • Manufacturing
      • Mining
      • Security
    • RBR50
      • RBR50 Winners 2025
      • RBR50 Winners 2024
      • RBR50 Winners 2023
      • RBR50 Winners 2022
      • RBR50 Winners 2021
  • Resources
    • Automated Warehouse Research Reports
    • Digital Issues
    • eBooks
    • Publications
      • Automated Warehouse
      • Collaborative Robotics Trends
    • Search Robotics Database
    • Videos
    • Webinars / Digital Events
  • Events
    • RoboBusiness
    • Robotics Summit & Expo
    • DeviceTalks
    • R&D 100
    • Robotics Weeks
  • Podcast
    • Episodes
  • Advertise
  • Subscribe