NExT Lab
Maker SpacesFabLabNExT LabRobotics LabPrint Room and Loans
  • NExT Lab
  • Contact Details
  • NExT Lab Access
  • Sustainability
    • 3D Printing
  • Case Studies & Projects
    • |3DS|VR| Voices of Country
    • |3DS| Our Quiet Neighbour
    • |3DS| OFF FORM | OFF MODERN
    • |3DP|AR| Prosthetic Habitats
    • |AR| Studio 40: The Field
    • |VR|3DP| Gravity Sketch: Door Handles
    • |3DS| 3D Scanning Examples
    • |AR|3DP| GRANULAR
  • 3D Printing |3DP|
    • 3D Printing at the NExT Lab
      • Other 3D Printing Options
    • Get Started
    • Design Approaches
    • Modelling Guidelines
    • 3D Print Farm
      • Quick-Start Guide
        • File Naming Conventions
      • Detailed Overview
        • 3D Printing Mesh Preparation
        • Submitting a Print Request
        • Post-Submission: Updating, Paying & Collecting
        • Slicing & Settings
    • Open Access Printers
      • PRUSA Open-Access
        • Workflows
          • Materials Experimentation
          • Experimental Techniques
        • Prusa i3 MK3S Fundamentals
        • Hardware Glossary
          • Extruder
          • Hotend & Nozzle
          • Print Surface, Bed & Y Axis
          • PINDA Inductive Probe
          • X-Axis Gantry
          • Z-Axis Stage
        • Software/Slicer Glossary
          • Plater/Virtual Print Bed
          • Print Settings
          • Filament Settings
          • Printer Settings
        • Troubleshooting
          • Filament Jam/Clog
          • Worn Nozzle
          • Broken/Loose Heatbreak
          • First Layer Issues/Prints Not Sticking to Bed
          • Stringing & Oozing Hotend
    • Use Own Filament
    • Key Techniques
      • Hollowing Models
      • Combating Warping
      • Split Models & Joints
      • Joints and Connections
      • Fillets & Chamfers
      • Accuracy, Precision & Tolerancing
      • Post-Processing & Finishing
        • No Sanding Method
        • Sanding Method
        • Epoxy Method
        • Fillers Method
      • Printing for Transparency
      • Mesh Techniques
        • Meshes 101
        • Working with Meshes
        • Repairing Meshes
        • Other Techniques
          • Thicken a Mesh with Grasshopper
          • Mesh Manipulation with Blender
          • Custom Supports in Meshmixer
      • Topography Models
      • Using the Makerbot Experimental Extruder
      • Troubleshooting
      • Adjusting Print Settings
    • Resources
      • Downloadable Software & Accounts
      • Software Primers
        • Autodesk Meshmixer
        • Blender
    • Mold Making and Casting
  • 3D Scanning |3DS|
    • 3D Scanning at the NExT Lab
    • 3D Scanning Use Cases
    • Guides
      • Principles of 3D Scanning / Digital Reconstruction
      • Photogrammetry
        • Photogrammetry Theory
        • Photogrammetry Benchmark
        • Technical Guides
          • From Photos to 3D Spatial Data
          • Advanced Techniques
          • Taking Measurements + Visualisation
          • From Photogrammetry to 3D Printing
      • BLK360 Terrestrial LiDAR Scanner
        • BLK360 Benchmark
        • Scan
        • Register
          • Export from iPad
        • Process
      • Artec Handheld SLT Scanners
        • Using the Scanners
        • Manual Alignment
        • Fill Holes
        • Smoothing
        • Frame Selection
      • VLX LiDAR SLAM Scanner
        • VLX setup
        • Preparing to Scan
        • Using the Scanner
        • Processing the Scans
      • Working with 3D Scan Data
        • Point Clouds and Rhino
        • Point Clouds and Cloud Compare
        • Point Clouds and Blender
        • Point Clouds to Meshes
    • Troubleshooting
      • General
      • Artec EVA
      • Leica BLK360
      • VLX
  • Augmented Reality |AR|
    • Augmented/Mixed Reality at the NExT Lab
      • Use Case of AR
    • Guides
      • Hololens 2
      • Fologram
        • Fologram Applications
          • Fologram for Hololens
          • Fologram for Mobile
        • Fologram for Rhino
        • Fologram for Grasshopper
        • Shared Experiences / Tracked Models
        • Extended Functionality
          • Preparing Models for AR
          • Interactivity
          • Fabrication
      • Unity and Vuforia
        • Unity Primer
        • 2D Targets (Image Targets)
        • 3D Targets (Object Targets)
        • Vuforia Primer
        • Creating a Simple AR App
          • Unity Next Steps: Interaction
          • Model Recognition
    • Troubleshooting
      • Hololens & Fologram
      • FAQ: Augmented Reality
    • Resources
      • Platforms (Hardware)
        • Microsoft Hololens
        • Mobile
      • Software Packages
      • Student Contact
        • AR: Intro Sessions
        • AR: Workshops and Resources
          • UntYoung Leaders Program Workshopitled
          • Young Leaders Program Workshop
          • Construction as Alchemy
  • Virtual Reality |VR|
    • Virtual Reality at the NExT Lab
    • Guides
      • Virtual Reality Hardware Set Up
        • Meta Quest 3
          • Troubleshooting
        • HTC Vive Headsets
          • HTC Vive
            • Troubleshooting
          • HTC Vive Pro
          • HTC Vive Cosmos
            • Troubleshooting
      • Twinmotion VR
        • Twinmotion VR: Features
        • Twinmotion VR: Troubleshooting
      • Virtual Reality Experiences
        • Unreal Engine
          • Unreal Engine Primer
            • Process: Level Building, Playing & Packaging
            • Actors: Components, Content and Editors
            • Materials & Textures
            • Lighting & Mobility
            • Player: VR and non-VR
            • Interactivity & Blueprints
          • Unreal Engine: Guides
            • Setting up a VR-ready File & Templates
            • Creating a Basic VR Experience
            • Custom Collision and Navigation
            • UV and Lightmaps
            • Outputting Content
            • Unreal Troubleshooting
            • Point Cloud Visualisation
          • VR: Video Tutorial Series
            • Exporting from Rhino
            • Model Preparation in 3DS Max
            • Unreal Engine
      • Designing in Virtual Reality
        • Gravity Sketch
          • Quick Start
        • Masterpiece Creator
    • Student Contact
      • VR: Intro Sessions
  • Sensing
    • Body Tracking
      • Usage
        • Technical Specifications
      • Data Analysis in Grasshopper
        • Analysis Examples
      • Animated Point Clouds(UE)
  • ROBOTICS
    • Robotic Dog
      • Operational Health & Safety
      • Robot Dog Setup
      • Operation Setup
        • Operation Basics
        • Arm Mode
        • Programming Mode
        • Mapping Mode
      • Advanced Operations
      • Expansion Equipment / Attachments
      • Basic Simulation
      • Troubleshooting
Powered by GitBook
On this page
  • Default Specs
  • Dev Cases
  • Navigation and exploration
  • Surveillance and inspection
  • Logistics and Assistance
  • Attachments
  • Lidar kit
  • D1 servo arm
  • GNSS receiver
  • Jetson Orin NX 8GB

Was this helpful?

  1. ROBOTICS
  2. Robotic Dog

Expansion Equipment / Attachments

The Go2 Quadruped robot is a versatile platform suitable for a range of applications. It can navigate and explore, leveraging 4D LIDAR and obstacle avoidance for autonomous movement and terrain mapping. Its capabilities allow for robotics education, programming practise, and AI experimentation, with secondary development support.

The robot is also effective for surveillance and inspection, providing real-time monitoring and autonomous access to hard-to-reach or hazardous areas. In addition, it supports logistics and assistance tasks, such as lightweight object transportation and routine operations, especially when paired with the D1 Servo Arm.

Default Specs

General

Robot Type

Quadruped, consumer-grade interactive robot

Degrees of Freedom

12 (precision joint motors)

Material

Aluminium alloy and high-strength engineering plastics

Speed

Maximum Running Speed

5 m/s (laboratory conditions)

Auto-follow Speeds

Slow: 1.5 m/s; Fast: 3.0 m/s

Remote Control

Distance

Over 30 meters

Modes

Companion remote, handheld remote, app-based

App Features

Image transmission, OTA updates, custom modes, real-time controls

Navigation

Intelligent Avoidance

360°×90° sensing via 4D LIDAR L1, 0.05m minimum detection range

Side-follow System

ISS2.0 with 50% improved positioning accuracy

Obstacle Avoidance

Forward only

Voice Interaction

Voice Control

Supports real-time commands and high-accuracy voice recognition (PRO and EDU versions only)

Build & Mobility

Maximum Climbing Angle

40° (30° for AIR version)

Maximum Stair Step Height

16 cm

Terrain Suitability

Flat surfaces; not for wet, dusty, or soft terrains

Stability Mechanism

Composite force and position control for precision

Battery

Endurance

1–2 hours depending on load and terrain

Battery Placement

Inserted on the side; 3 or more bars recommended for operation

Customization

Custom Modes

Various motions (e.g., dance, moonwalk, stretch) via app or remote

Dimensions

Size and Weight

Not explicitly specified; lightweight design

Accessories

Auto Retractable Strap

Adjustable hand strap (PRO version only)

Consumable Parts

Spare foot-end pads included for rough terrain use

Limitations

Environmental Conditions

Operates best in 5°C–35°C; not waterproof or dust-proof

Friction Requirements

Unsuitable for icy, wet, or low-friction grounds

Dev Cases

Navigation and exploration

Autonomous navigation: Equipped with 4D LIDAR and obstacle avoidance for autonomous movement and terrain mapping.

Terrain exploration: Suitable for navigating and exploring controlled environments.

Surveillance and inspection

Remote monitoring: Real-time image transmission allows for surveillance or inspection in hard-to-reach or unsafe areas.

Inspection tasks: Useful for inspecting industrial or confined areas autonomously.

Logistics and Assistance

Lightweight object transportation: Can carry or move lightweight items, especially when paired with the D1 Servo Arm.

Routine assistance: Assists in controlled tasks such as fetching or organising objects in structured environments.


Attachments

Lidar kit

OS2 Long-Range High-Resolution Imaging Lidar:

The robot dog can be equipped with a Lidar Scanning Kit. We currently have the OS2 Long Range – High resolution Imaging lidar available. This lidar system has been developed to handle the most challenging tasks offering robustness and reliability.

  • Up to a 400m scanning range (depending on lighting and weather conditions)

  • Minimum range 0.8m

  • 22.5 degree Vertical Field of View

  • 360 Degree Horizontal

  • Ip68/69K Rating for a range of weather conditions

  • 128 channels of High-resolution imaging to 0.1cm accuracy

  • 2.62m points per second

  • Near-infrared capture

  • Open-source ROS Drivers

D1 servo arm

Allows lifting light-weight objects with the use of a clamp jaw.

  • The D1 Servo arm is an addition to the Robot Dog which expands the capabilities and interactivity of the unit. The Servo arm allows users to experiment with manual handling, object manipulation and general object interaction.

  • The servo arm offers a max armspan of 550mm (without the jaws) 670mm(with gripper) and can handle up to 500g in weight, suited to education and practical operations.

  • The servo arm can be programmed through the ROS platform for complex tasks. Simple controls are integrated and facilitated through the Go2s control interface.

  • Typically, servo arms are used for research tasks such as data collection, particularly in hard to reach locations.

GNSS receiver

What is a GNSS receiver?

  • The term “GNSS” stands for Global Navigation Satellite System and is used to describe an electronic device that receives and processes digital signals sent by satellites in space. This receiver then usese these signals to calculate its own location, velocity and time. In order for these receivers to operate correctly they should have an unobstructed line of sight to satellites in the sky to ensure accurate information transmission.

  • This is a powerful tool that is used in combination with the lidar kit to assist with georeferencing mapping and scan data.

Emlid Reach RS2+ - GNSS Receiver: The GNSS (Global Navigation Satellite System) is a device that receives satellite signals to calculate location, velocity, and time, and is used with lidar systems for georeferencing mapping and scan data, with the Emlid Reach RS2+ offering features like centimetre-level precision, 22-hour battery life, and IP67 weatherproofing.

  • The Emlid Reach RS2+ features includes:

  • Cm level precision at over 60kms p/h

  • Wifi, Bluetooth and LTE connectivity

  • 22hr battery life

  • Robust construction with IP67 Weatherproof ratings

Jetson Orin NX 8GB

The Jetson Orin NX 8GB is a compact AI computing module by NVIDIA designed for edge computing and robotics. With a processing capability of 40 TOPS (trillion operations per second), it is powered by the NVIDIA Ampere architecture GPU, ARM Cortex CPU, and deep learning accelerators. This enables it to handle multiple neural networks in parallel and process data from high-resolution sensors.

  • Performance: The 40 TOPS allows it to run complex AI models efficiently, including real-time inference for tasks like computer vision, natural language processing, and robotics.

  • Applications :Common use cases include autonomous machines, smart cameras, industrial automation, and robotics operating systems (ROS) development.

  • Connectivity and I/O: Supports high-speed interfaces like PCIe and USB, along with GPIOs for sensor and device integration.

  • Energy Efficiency: Configurable power options (7–15W) make it suitable for power-constrained environments.

  • Software Support: Compatible with the NVIDIA JetPack SDK, offering tools for AI deployment, sensor integration, and robotics programming.

PreviousAdvanced OperationsNextBasic Simulation

Last updated 4 months ago

Was this helpful?