MITB Banner

10 Open-Source Datasets To Learn Robotics

Share

Nowadays, researchers are striving to implement human-level understanding into robots. For instance, understanding the surrounding environment, hand movement, detecting and grasping objects and many more.

In this article, we list down 10 open-source datasets that can be used to learn robotics in an advanced way.

1| Robot Arm Pushing Dataset

The robot arm pushing dataset contains approximately 59,000 examples of robot pushing motions which including one training set (train) and two test sets. The first test set used two different subsets of the objects pushed during training. The second test set involved two subsets of objects. One key application of this dataset is to implement it in a learned model for decision making in vision-based robotic control tasks.

2| Robot Arm Grasping Dataset

The robot arm grasping dataset contains approximately 650,000 examples of robot grasping attempts. The dataset was mainly created to train a large convolutional neural network (CNN) to predict the probability that the task-space motion of the gripper will result in successful grasps.

3| Dataset Of Daily Interactive Manipulation

The dataset of daily interactive manipulation focuses on position, orientation, force, and torque of objects manipulated in daily tasks. It is a collection of 3D position and orientation (PO), force and torque (FT) data of tools/objects being manipulated to fulfill certain tasks. It includes 1,603 trials of 32 types of daily motions and 1,596 trials of pouring alone along with the helper code.

4| Fukuoka Datasets For Place Categorisation

Fukuoka datasets for place categorisation is a collection of several multimodal 3D datasets. It contains several datasets for the task of place categorisation using global information without object segmentation. It can also be used for other segmentation-based methods and can be extended with additional annotations. The datasets include indoor and outdoor scenarios obtained in different locations in Fukuoka city, Japan.

5| MultiDrone Public DataSet

Aerial robots can be used for many robotic tasks, for instance, surveillance, monitoring, filming, among others. The MultiDrone public dataset has been collected using both pre-existing audiovisual material and newly filmed UAV shots. It includes more than 10 hours of UAV footage depicting football, rowing, and cycling (DW). A large subset of this dataset has been annotated for facilitating scientific research, in tasks such as visual detection and tracking of bicycles, football players, human crowds, etc.  

6| Complex Urban Dataset With Multi-Level Sensors From Highly Diverse Urban Environments

The complex urban dataset with multi-level sensors from a highly diverse urban environment captures various complex urban features and addresses the major issues of complex urban areas, such as unreliable and sporadic Global Positioning System (GPS) data, multi-lane roads, complex building structures, and the abundance of highly dynamic objects. In this dataset, typical LiDAR sensor data, 2D and 3D LiDAR, are provided.

7| Natural Language Instructions For Human-Robot Collaborative Manipulation

Natural language instructions for human-robot collaborative manipulation is a dataset of natural language instructions for object reference in manipulation scenarios. It comprises of 1582 individual written instructions which were collected via online crowdsourcing. This dataset is particularly useful for researchers who work in natural language processing, human-robot interaction, and robotic manipulation.

8| Yale-CMU-Berkeley Dataset For Robotic Manipulation Research

Yale-CMU-Berkeley dataset for robotic manipulation research is an image and model dataset of the real-life objects from the Yale-CMU-Berkeley Object Set, which is specifically designed for benchmarking in manipulation research. The dataset includes 600 high-resolution RGB images, 600 RGB-D images and five sets of textured three-dimensional geometric models.

9| Robot-at-Home Dataset

Robot-at-home is a robotic dataset for semantic mapping of home environments. It is a collection of raw and processed sensory data from domestic settings which are used for semantic mapping algorithms through the categorisation of objects or rooms. The dataset contains 87,000+ time-stamped observations gathered by a mobile robot endowed with a rig of four RGB-D cameras and a 2D laser scanner.

10| Robotic 3D Scan Repository

Robotic 3D scan repository is a data repository of standard 3D data sets for the robotics community, log files of robot runs and 3D point clouds from robotic experiments. One can easily use this data for developing SLAM or interpretation algorithms.  

Share
Picture of Ambika Choudhury

Ambika Choudhury

A Technical Journalist who loves writing about Machine Learning and Artificial Intelligence. A lover of music, writing and learning something out of the box.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.