Close Menu
  • Home
  • Aerospace & Defense
    • Automation & Process Control
      • Automotive & Transportation
  • Banking & Finance
    • Chemicals & Materials
    • Consumer Goods & Services
  • Economy
    • Electronics & Semiconductor
  • Energy & Resources
    • Food & Beverage
    • Hospitality & Tourism
    • Information Technology
  • Agriculture
What's Hot

Venezuelans in exile dream of returning home. What’s stopping them? |Nicolas Maduro News

Intel will start manufacturing GPUs, a market dominated by Nvidia

Plans for 2026: What’s next for Startup Battlefield 200?

Facebook X (Twitter) Instagram
USA Business Watch – Insightful News on Economy, Finance, Politics & Industry
  • Home
  • Aerospace & Defense
    • Automation & Process Control
      • Automotive & Transportation
  • Banking & Finance
    • Chemicals & Materials
    • Consumer Goods & Services
  • Economy
    • Electronics & Semiconductor
  • Energy & Resources
    • Food & Beverage
    • Hospitality & Tourism
    • Information Technology
  • Agriculture
  • Home
  • About Us
  • Market Research Reports and Company
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
USA Business Watch – Insightful News on Economy, Finance, Politics & Industry
Home » A new vision-based system teaches machines to understand your body
Electronics & Semiconductor

A new vision-based system teaches machines to understand your body

Bussiness InsightsBy Bussiness InsightsJuly 27, 2025No Comments7 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Robots, Know Yourself: A new vision-based system teaches machines to understand your body

Qualitative results on robustness for variance exclusion scenarios. Credit: Nature (2025). doi:10.1038/s41586-025-09170-0

In the office of MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL), soft robot hands carefully curl their fingers to grab a small object. The interesting part is not the mechanical design or embedded sensors. In fact, there’s nothing in your hand. Instead, the entire system relies on a single camera that monitors the movement of the robot and uses its visual data to control it.

This feature comes from a new system developed by CSAIL Scientists and provides a different perspective on robot control. Rather than using hand-designed models or complex sensor arrays, robots can learn how their body responds to control commands through their vision alone. This approach, called Neural Jacobian Fields (NJF), gives robots physical self-awareness.

Papers about the work were published naturally.

“This work illustrates the transition from programming robots to robot education,” says Sizhe Lester Li, MIT Ph.D. Chief researcher of electrical engineering and computer science students, CSAIL affiliates and researchers.

“Today, many robotic tasks require extensive engineering and coding. In the future, we will be able to learn how to autonomously achieve our goals, assuming that we will show our robots what to do.”

Motivation comes from simple but powerful reconstruction. The main barrier to affordable and flexible robotics is not hardware, but the control of functions that can be achieved in multiple ways.

Traditional robots are built to be stiff and sensor-rich, making it easier to build digital twins, the precise mathematical replica used for control. However, when the robot is soft, deformed or irregularly shaped, those assumptions fall apart. Rather than force the robot to match the model, NJF inverts the script. The robot acquires the ability to learn its internal models from observation.

See and learn

This separation of modeling and hardware design can greatly expand the design space for robotics. In soft, bio-inspired robots, designers often embed sensors and enhance parts of the structure to make modeling feasible.

The NJF lifts that constraint. The system does not require onboard sensors or design adjustments to allow for control. Designers can explore unconventional and unconstrained forms without worrying whether they can model or control them later.

“Think about how to control your fingers. You sway, you observe and adapt,” says Li. “That’s what our system does. It experiments with random actions and knowing which controls move through which parts of the robot.”

The system has proven to be robust across a variety of robot types. The team tested the NJF with pneumatic soft robotic hands that allow pinch and grind pins, rigid Allegro hands, 3D printed robotic arms, and even rotating platforms that do not have built-in sensors.

In all cases, the system learned how it responded to the robot’s shape and control signals with just vision and random movement.

Researchers see possibilities that go far beyond the lab. NJF-equipped robots can one day perform agricultural tasks with centimeter-level localization accuracy, operate on construction sites without elaborate sensor arrays, or navigate dynamic environments where traditional methods break.

The NJF core has neural networks that capture two intertwined aspects of the robotic embodiment: three-dimensional geometry and sensitivity that controls the input. The system is based on Neural Radiation Fields (NERF), a technique that maps spatial coordinates to color and density values to reconstruct a 3D scene from an image.

The NJF extends this approach by learning not only the shape of the robot, but also the Jacobi field. This is a function that predicts how the robot’s body points will move in response to motor commands.

To train the model, the robot performs random movements, with multiple cameras recording the results. No human supervision or prior knowledge of the robot structure is required. The system simply drives the relationship between control signals and movement by watching.

Once the training is complete, the robot only needs one monocular camera for real-time closed-loop control, and runs on approximately 12 HERTZs. This allows one to observe, plan, and act on itself continuously. This speed makes NJF more feasible than many physics-based simulators for soft robots.

In early simulations, even simple 2D fingers and sliders were able to learn this mapping using only a few examples. By modeling how a particular point transforms and shifts according to actions, the NJF constructs a dense map of controllability. This internal model allows generalization of the movement of the entire robot body, even when the data is noisy or incomplete.

“What’s really interesting is that the system knows on its own which part of the robot is controlling,” says Li. “This is unprogrammed. It comes naturally through learning, like people discovering buttons on new devices.”

The future is soft

For decades, robotics have favored hard, easily modeled machines, like industrial weapons found in factories. However, this field is heading towards soft, bio-style robots that can adapt more fluidly to the real world. trade off? These robots are difficult to model.

“Robotics today are often out of reach due to costly sensors and complex programming. The goal at the Neural Jacobian Field is to lower the barriers, making robotics affordable, adaptable and accessible to more people.

“It opens the door to robots that can operate in messy, unstructured environments, from farms to construction sites, without expensive infrastructure.”

“Vision alone can provide the queues needed to elicit the need for GPS, external tracking systems, or complex onboard sensors. This opens the door to robust and adaptive operation underground to underground. RUS, Professor of Electrical Engineering and Computer Science, Director of CSAil.

“By learning from visual feedback, these systems develop unique internal models of movement and dynamics, allowing for flexible, self-monitored operations where traditional localization methods fail.”

Training the NJF now requires multiple cameras and requires redoing for each robot, but researchers already imagine a more accessible version. In the future, enthusiasts will be able to record random movements of robots on their mobile phones, just like they would take a video of a rental car before driving, and will use that footage to create a control model.

The system is not yet generalized between different robots, lacking force and tactile sensing, limiting its effectiveness for contact-rich tasks. However, the team is looking for new ways to address these limitations. It expands the ability of models to improve generalization, handle occlusions, and infer across longer spatial and temporal fields.

“Just as humans have an intuitive understanding of how their bodies move and respond to commands, the NJF gives robots a self-awareness embodied by their vision alone,” says Li.

“This understanding is the foundation of flexible manipulation and control in real-world environments. Essentially, our work reflects the broader trends of robotics: from manually programming detailed models to teaching robots through observation and interaction.”

This paper summarizes Sitzmann Lab’s computer vision and self-monitoring learning tasks, as well as RUS Lab’s soft robot expertise.

More information: Control a variety of robots by inferring Jacobian fields in Sizhe Lester Li et al, Deep Networks, Nature (2025). doi:10.1038/s41586-025-09170-0

Provided by Massachusetts Institute of Technology

This story has been republished courtesy of MIT News (web.mit.edu/newsoffice/), a popular site that covers news about MIT research, innovation and education.

Quote: Robots, Know Yourself: A New Vision-Based System teaches machines to understand the body obtained from https://news/2025-07-thysyself-vision base machines (July 27, 2025)

This document is subject to copyright. Apart from fair transactions for private research or research purposes, there is no part that is reproduced without written permission. Content is provided with information only.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleAs rainy weather stalls progress, mixed bags for British harvest
Next Article Baghdad’s gunfight kills paramilitary groups when killing ministries | Conflict News
Bussiness Insights
  • Website

Related Posts

Dual-mode design improves accuracy of MEMS accelerometers, study finds

November 18, 2025

Researchers complete first real-world validation of maritime IoT communications network

November 18, 2025

Plasma-based method creates efficient, low-cost catalyst for metal-air batteries

November 18, 2025
Leave A Reply Cancel Reply

Latest Posts

McCain launches ‘Farm of the Future’ in UK to test regenerative practices

Grain demand from UK processors falls to 20-year low in 2025/26

Citizens warned against eating eggs produced at home near industrial sites

High-altitude sheep grazing is associated with more stable soil carbon, researchers say

Latest Posts

York Space begins trading at $38 a share, touts ‘Golden Dome’ potential

January 29, 2026

American Airlines flies to Venezuela for the first time since 2019

January 29, 2026

Southwest Airlines (LUV) 2025 Q4 Earnings

January 28, 2026

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • Venezuelans in exile dream of returning home. What’s stopping them? |Nicolas Maduro News
  • Intel will start manufacturing GPUs, a market dominated by Nvidia
  • Plans for 2026: What’s next for Startup Battlefield 200?
  • Former leader’s son Saif al-Islam Gaddafi killed in Libya | Muammar Gaddafi News
  • Homeland Security seeks to force tech companies to hand over data on Trump critics

Recent Comments

  1. Numbersjed on 100% tariffs on Trump’s drugs: What we know | Donald Trump News
  2. JamesPak on Hundreds gather in Barcelona to protest overtourism in southern Europe
  3. vibroanalizador on 100% tariffs on Trump’s drugs: What we know | Donald Trump News
  4. игровой аппарат гейтс оф олимпус on 100% tariffs on Trump’s drugs: What we know | Donald Trump News
  5. online casino games slots on 100% tariffs on Trump’s drugs: What we know | Donald Trump News

Welcome to USA Business Watch – your trusted source for real-time insights, in-depth analysis, and industry trends across the American and global business landscape.

At USABusinessWatch.com, we aim to inform decision-makers, professionals, entrepreneurs, and curious minds with credible news and expert commentary across key sectors that shape the economy and society.

Facebook X (Twitter) Instagram Pinterest YouTube

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Archives

  • February 2026
  • January 2026
  • December 2025
  • November 2025
  • October 2025
  • September 2025
  • August 2025
  • July 2025
  • June 2025
  • March 2022
  • January 2021

Categories

  • Aerospace & Defense
  • Agriculture
  • Automation & Process Control
  • Automotive & Transportation
  • Banking & Finance
  • Chemicals & Materials
  • Consumer Goods & Services
  • Economy
  • Economy
  • Electronics & Semiconductor
  • Energy & Resources
  • Food & Beverage
  • Hospitality & Tourism
  • Information Technology
  • Political
Facebook X (Twitter) Instagram Pinterest
  • Home
  • About Us
  • Market Research Reports and Company
  • Contact us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2026 usabusinesswatch. Designed by usabusinesswatch.

Type above and press Enter to search. Press Esc to cancel.