Skip to content

Simulating embodied sensorimotor control with NeuroMechFly v2

March 2026 Update

We introduced a new FlyGym 2.x.x API in March 2026, with a complete code rewrite and redesigned interface. This version delivers significantly improved performance:

  • ~10x speed-up for CPU-based simulations (~2x real-time throughput)
  • ~300x speed-up for GPU-based simulation via Warp/MJWarp (~60x real-time throughput)

Additional improvements include:

  • Improved scene composition workflow
  • Interactive viewer
  • Simplified dependency stack

This version is not backward compatible, and not all features from FlyGym 1.x.x are available. Feature requests can be submitted via Issues on the Github repository. See more information about the changes here.

Prefer the old version? FlyGym 1.x.x has been migrated to flygym-gymnasium. Its documentation has been migrated to gymnasium.neuromechfly.org.

overview

FlyGym is the Python library for NeuroMechFly, a digital twin of the adult fruit fly Drosophila melanogaster that can see, smell, walk over challenging terrain, and interact with the environment.

For more information, see our NeuroMechFly v2 paper.

NeuroMechFly consists of the following components: - Biomechanical model: The biomechanical model is based on a micro-CT scan of a real adult female fly (see our original NeuroMechFly publication). We have adjusted several body segments (in particular in the antennae) to better reflect the biological reality. - Vision: The fly has compound eyes consisting of individual units called ommatidia arranged on a hexagonal lattice. We have simulated the visual inputs on the fly’s retinas. - Olfaction: The fly has odor receptors in the antennae and the maxillary palps. We have simulated the odor inputs experienced by the fly by computing the odor/chemical intensity at these locations. - Hierarchical control: The fly’s Central Nervous System consists of the brain and the Ventral Nerve Cord (VNC), a hierarchy analogous to our brain-spinal cord organization. The user can build a two-part model — one handling brain-level sensory integration and decision making and one handling VNC-level motor control — with an interface between the two consisting of descending (brain-to-VNC) and ascending (VNC-to-brain) representations. - Leg adhesion: Insects have evolved specialized adhesive structures at the tips of the legs that enable locomotion vertical walls and overhanging ceilings. We have simulated these structures in our model. The mechanism by which the fly lifts the legs during locomotion despite adhesive forces is not well understood; to abstract this, adhesion can be turned on/off during leg stance/swing. - Mechanosensory feedback: The user has access to the joint angles, forces, and contact forces experienced by the fly.

This package is developed at the Neuroengineering Laboratory, EPFL.

Getting Started

For installation, see the documentation page.

To get started, follow tutorials here.