A new AI model replaces months of simulation with near-instant predictions, changing how spacecraft operations are prepared
Updated
April 24, 2026 10:53 AM

Northrop Grumman Stargaze serves as the mother ship for the Pegasus, an air-launched orbital rocket. PHOTO: UNSPLASH
Flexcompute, a startup that builds software to simulate real-world physics, is working with Northrop Grumman to change how space missions are prepared. Together, they have developed an AI-based system that can predict how spacecraft respond during critical manoeuvres such as docking—when one spacecraft moves in and connects with another in orbit. These steps have traditionally taken months of preparation.
At the centre of this work is a long-standing problem in space operations. When a spacecraft fires its thrusters, the exhaust plume interacts with nearby surfaces. These interactions can affect movement, temperature and stability. Because these effects are difficult to test in real conditions, engineers have relied on large volumes of computer simulations to estimate outcomes before a mission. That process is slow and resource-intensive.
The new system replaces much of that workflow with a trained AI model. Instead of running millions of simulations, the model learns patterns from physics-based data and can make predictions in seconds. It also provides a measure of uncertainty, which helps engineers understand how reliable those predictions are when making decisions.
"At Northrop Grumman, we're pioneering physics AI to accelerate design and solve complex simulation and modelling problems like plume impingement—critical for station keeping, rendezvous and space robotics. Simply put: we're pushing the boundaries of advanced space operations", said Fahad Khan, Director of AI Foundations at Northrop Grumman. "Partnering with Flexcompute and NVIDIA, we're accelerating innovation and mission timelines to deliver superior space capabilities for customers at the speed they need".
The system is built using technology from NVIDIA, which provides the computing framework behind the model. Flexcompute has adapted it to handle the specific challenges of spaceflight, including how gases expand and interact in a vacuum. The result is a tool that can simulate complex scenarios much faster while maintaining the level of accuracy needed for mission planning.
By shortening preparation time, the model changes how engineers approach spacecraft design and operations. Faster predictions mean teams can test more scenarios and adjust plans more quickly. It also helps improve fuel use and extend the lifespan of spacecraft.
"Northrop Grumman's confidence reflects what sets Flexcompute apart", said Vera Yang, President and Co-Founder of Flexcompute. "We are able to take the most accurate and scalable physics foundations and evolve them into highly trained, customized Physics AI solutions that engineers can rely on. This work shows how we are transforming the role of simulation, not just speeding it up, but expanding what engineers can confidently solve and how quickly they can act".
The collaboration points to a broader shift in how engineering problems are being handled. Instead of relying only on detailed simulations that take time to run, companies are beginning to use AI systems that can approximate those results quickly while still reflecting the underlying physics.
"The industry's most ambitious space missions now demand a level of speed and precision that traditional engineering cycles can no longer sustain", said Tim Costa, vice president and general manager of computational engineering at NVIDIA. "By integrating NVIDIA PhysicsNeMo, Northrop Grumman and Flexcompute are transforming complex simulations like plume impingement from days of compute into seconds of insight, drastically accelerating the path from mission concept to orbit".
What emerges from this work is a shift in how missions are prepared. When prediction cycles move from months to seconds, testing and decision-making can happen faster. For space operations, where timing and precision are closely linked, that change could reshape how systems are built and run.
Keep Reading
A smartphone that moves, tracks and responds in real time—but is it real utility or just a marketing gimmick?
Updated
April 15, 2026 6:00 PM

HONOR Robot Phone, with its camera arm extended. PHOTO: HONOR
Smartphones today feel more familiar than new. Each year brings better performance and better cameras, but fewer real surprises. So when a company unveils something called a “Robot Phone”, it’s bound to get attention.
HONOR did exactly that at the Mobile World Congress (MWC) in Barcelona this year. While most smartphone brands are focused on software upgrades, HONOR is trying something different with hardware. Its Robot Phone is built to move and adjust on its own. The camera sits on a motorized system that can tilt, track motion and shift angles automatically. It almost looks like a small robotic head, following whatever is happening in front of it. It can pick up sound, recognize motion and stay visually aware of its surroundings. This result feels less like using a regular phone and more like interacting with something responsive.
So what makes HONOR’s Robot Phone different from the smartphones we already use? Here’s a closer look at its camera system, AI features and design, and whether it is truly something new or simply smart marketing.
At its core, the Robot Phone still works like a regular smartphone. What makes it different is the camera system. It has a 200MP camera that sits on a motorized arm with a three-axis gimbal, which extends when in use and folds back into the phone when not needed. The compact motor gives the camera physical movement, while motion control allows it to sense, track and follow a person or object in real time. That means it can keep a subject in frame without constant manual adjustment.
The camera also adds a more playful side to the experience. It can respond with simple gestures, such as nodding or shaking its head, and it can even move in sync with music.
This setup could be particularly useful for content creators. As CNET tech journalist and YouTuber Andrew Lanxon pointed out, it removes the need to carry a separate gimbal. Since the robotic camera module can easily fold into the body of the phone, it is easier to carry around and more convenient for filming or taking photos on the go.
The Robot Phone also has the practical advantage of a smartphone display. It gives users a bigger screen than a standalone camera for framing, monitoring and reviewing footage. Since it runs on Android, the process of recording, editing and sharing content is also more direct.
The most impressive part of the HONOR Robot Phone design is how it fits a moving camera system into the body of a smartphone without needing external attachments.
To make this possible, HONOR uses a custom micro motor that is 70% smaller than mainstream competitors. The company also says it is the industry’s smallest four-degrees-of-freedom (4DoF) gimbal system. To support the stable movement of the camera module, the internal structure uses high-strength materials such as steel and titanium alloy. These materials help the mechanism stay durable as it shifts and repositions over time.
Battery life is another obvious question. HONOR has not revealed the battery capacity of the Robot Phone itself, but it did showcase its Silicon-Carbon Blade Battery technology at MWC 2026. The company says this battery is designed to increase energy density while keeping devices slim, and that it could support capacities of 7,000 mAh and beyond in future foldable devices.
That is not specific to the Robot Phone, but it does hint at the kind of battery improvements that may be needed for smartphones with moving parts and more advanced camera systems.
The AI features in Honor’s Robot Phone are focused on how the device sees and responds to its surroundings in real time. At the most basic level, the phone can track what is happening in a scene and adjust itself without constant user input.
On the functional side, the system keeps subjects framed and in focus automatically. Its AI Object Tracking ensures subjects stay centred, while AI SpinShot enables controlled 90° and 180° rotations for smoother transitions, even when the phone is used one-handed. It can also detect motion and recognize sound, which lets it respond to activity as it happens instead of reacting frame by frame.
The AI becomes more noticeable in the way the device behaves. When activated, the camera module unfolds and the screen displays a pair of animated eyes that track the user’s face and gaze. Honor calls this “embodied AI”, meaning the assistant expresses itself through movement rather than only voice or text. The camera module can adjust its angle during video calls, which makes it feel a little more physically present.
According to Thomas Bai, AI product expert at Honor, the goal is to move beyond passive assistance. By combining sensing, movement and real-time processing, the device is designed to interact with its environment in a more continuous way. In practice, that could mean interpreting its surroundings and responding as situations change, such as when someone is moving through an unfamiliar space.
The Robot Phone has sparked curiosity, but there is still a lot we do not know. For one thing, it is still a prototype, with a release expected later this year. Early signs also suggest it may be expensive, partly because of rising memory chip costs. Some of its more playful features also feel uncertain. In demos, the phone can move along to music, but with only a handful of pre-set tracks, it is hard to tell whether that feature will be genuinely useful or remain more of a showcase moment.
Then there are the practical questions. A motorized camera system could make the phone heavier and more top-heavy, which may affect comfort during daily use. Running a motor alongside continuous AI tracking will also likely put pressure on battery life. These are not dealbreakers, but they are trade-offs that will matter outside of a demo.
Privacy is another concern that is hard to overlook. Some of the AI features rely on cloud processing, which means certain data is sent to external servers instead of being processed fully on the device. That is common in many AI systems today, but it feels more significant here because the phone is built to actively track movement and reposition its camera in real time. For some people, that level of autonomy may feel intrusive rather than helpful. It also raises bigger questions about what sensors are built into the device and how much data they collect during everyday use.
So, is the HONOR Robot Phone a real step forward, or just a clever idea packaged well?
The answer depends on who it is for.
For content creators, the appeal is obvious. Early indications suggest it could make video capture easier by reducing the need for extra gear. Honor’s collaboration with cinema camera company ARRI also suggests a serious push toward more cinematic smartphone footage.
For everyone else, the value is less clear. Outside of content creation, it is still hard to see how these features would translate into everyday use in a meaningful way.
For now, the Robot Phone sits somewhere between promise and experiment. Whether it turns into a genuinely useful new kind of smartphone or fades away as a novelty will only become clear once it moves beyond controlled demos and into real life.