<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
    <channel>
        <title>Flight Dynamics and Control Lab | Taeyoung Lee</title>
        <description>Taeyoung Lee, geometric mechanics, geometric control, quadrotor, George Washington University</description>
        <link>https://fdcl.engineering.gwu.edu/</link>
        <atom:link href="https://fdcl.engineering.gwu.edu/feed.xml" rel="self" type="application/rss+xml"/>
        <pubDate>Thu, 07 May 2026 01:43:55 +0000</pubDate>
        <lastBuildDate>Thu, 07 May 2026 01:43:55 +0000</lastBuildDate>
        <generator>Jekyll v4.4.1</generator>
        
            <item>
                <title>Equivariant Reinforcement Learning Frameworks for Quadrotor Low-Level Control</title>
                <description>&lt;p&gt;Teaching quadrotor unmanned aerial vehicles (UAVs) to fly autonomously is a high-stakes engineering hurdle. 
Because quadrotors are inherently unstable and exhibit complex, non-linear dynamics, training them using standard model-free Reinforcement Learning (RL) is notoriously data-intensive and risky. 
In a typical RL setup, an agent must experience thousands of “crashes” in simulation to learn basic stability, a process that is often computationally expensive and unsafe.&lt;/p&gt;

&lt;p&gt;Standard multilayer perceptron (MLP)-based RL models often treat every new orientation or position as a completely unique configuration, requiring the RL agent to learn all information from scratch. 
To address this, this study introduces &lt;strong&gt;Equivariant Reinforcement Learning&lt;/strong&gt;. 
This framework uses geometric principles to make learning significantly faster and more robust by embedding the invariance and equivariance properties directly into the neural network architecture.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/equiv-rl/framework.jpg&quot; width=&quot;800&quot; /&gt;
&lt;/p&gt;

&lt;h3 id=&quot;geometry-as-a-shortcut-the-power-of-equivariance&quot;&gt;Geometry as a Shortcut: The Power of Equivariance&lt;/h3&gt;

&lt;p&gt;The breakthrough in this research lies in identifying rotational and reflectional symmetries present in the quadrotor dynamics. 
By using &lt;strong&gt;Equivariant Multilayer Perceptrons (EMLPs)&lt;/strong&gt;, the drone can learn an optimal control action in one configuration and automatically apply it to others.&lt;/p&gt;

&lt;div style=&quot;display: flex; justify-content: center; gap: 20px; margin-bottom: 20px;&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/equiv-rl/rot_sym2.jpg&quot; alt=&quot;Rotational Symmetry&quot; style=&quot;width: 50%; height: auto; object-fit: contain;&quot; /&gt;
  &lt;img src=&quot;/images/posts/ben/equiv-rl/ref_sym.jpg&quot; alt=&quot;Reflectional Symmetry&quot; style=&quot;width: 50%; height: auto; object-fit: contain;&quot; /&gt;
&lt;/div&gt;

&lt;h3 id=&quot;two-brains-are-better-than-one-modular-vs-monolithic&quot;&gt;Two Brains are Better Than One: Modular vs. Monolithic&lt;/h3&gt;

&lt;p&gt;A core strategic insight of this work is the move away from “monolithic” (single-agent) brains toward specialized “modular” (multi-agent) architectures.&lt;/p&gt;

&lt;table&gt;
  &lt;tbody&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Feature&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Monolithic (Mono-MLP/EMLP)&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Modular (Mod-MLP/EMLP)&lt;/strong&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Structure&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;Single agent manages all flight aspects.&lt;/td&gt;
      &lt;td&gt;Decoupled: Translation and Yaw modules.&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Geometry&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;Operates on the full $SO(3)$ group.&lt;/td&gt;
      &lt;td&gt;Splits $SO(3)$ into $S^2$ (thrust) and $S^1$ (yaw).&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Efficiency&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;Slower convergence; prone to overfitting.&lt;/td&gt;
      &lt;td&gt;Parallelized learning; superior tracking.&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Sample Requirement&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;Still struggling at $8 \times 10^5$ timesteps.&lt;/td&gt;
      &lt;td&gt;Near-peak rewards within $\approx 2 \times 10^5$ timesteps.&lt;/td&gt;
    &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;

&lt;p&gt;By splitting the three-dimensional orthogonal group $SO(3)$ into a sphere ($S^2$) for translational control and a circle ($S^1$) for yaw control, the modular approach prevents performance degradation during agile maneuvers. 
As evidenced by learning curves, the Mod-EMLP architecture achieves significantly higher returns early in the training phase compared to any other framework.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/equiv-rl/learning_curves.jpg&quot; width=&quot;800&quot; /&gt;
&lt;/p&gt;

&lt;h3 id=&quot;zero-shot-sim-to-real-transfer-validation&quot;&gt;Zero-Shot Sim-to-Real Transfer Validation&lt;/h3&gt;

&lt;p&gt;A major hurdle in robotics is “Sim-to-Real” transfer. 
To bridge this, this work utilized &lt;strong&gt;Domain Randomization&lt;/strong&gt;, where physical parameters–such as mass, arm length, and inertia–were uniformly sampled within $\pm 10\%$ of their nominal values during training. 
This forces the policy to be robust rather than environment-specific.&lt;/p&gt;

&lt;p&gt;The framework was validated through a &lt;strong&gt;Zero-Shot Transfer&lt;/strong&gt; process, and the experimental setup at the Flight Dynamics and Control Lab included:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Vicon Valkyrie VK-8 Motion Capture:&lt;/strong&gt; A 12-camera system running at 200 Hz for precise position and attitude estimation.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;NVIDIA Jetson TX2:&lt;/strong&gt; An onboard flight computer managing the RL controller and EKF state estimation in real-time.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Analysis of figure-eight Lissajous trajectory flight tests shows that &lt;strong&gt;incorporating equivariant learning within a modular design&lt;/strong&gt; is the key:&lt;/p&gt;

&lt;table&gt;
  &lt;tbody&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Metric&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Monolithic MLP (Baseline)&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;Modular EMLP (Winner)&lt;/strong&gt;&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Average Position Error&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;14.63 cm&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;8.44 cm&lt;/strong&gt; (Nearly 42% reduction)&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Average Yaw Error&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;7.15 degrees&lt;/td&gt;
      &lt;td&gt;&lt;strong&gt;4.33 degrees&lt;/strong&gt; (Highly precise)&lt;/td&gt;
    &lt;/tr&gt;
    &lt;tr&gt;
      &lt;td&gt;&lt;strong&gt;Generalization&lt;/strong&gt;&lt;/td&gt;
      &lt;td&gt;Must manually learn every configuration, leading to redundancies in learning.&lt;/td&gt;
      &lt;td&gt;Encodes rotational/reflectional symmetry for 10x more efficient learning.&lt;/td&gt;
    &lt;/tr&gt;
  &lt;/tbody&gt;
&lt;/table&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/equiv-rl/sim-to-real1.jpg&quot; width=&quot;800&quot; /&gt;
&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/equiv-rl/sim-to-real2.jpg&quot; width=&quot;800&quot; /&gt;
&lt;/p&gt;

&lt;h3 id=&quot;the-future-of-geometric-learning-in-robotics&quot;&gt;The Future of Geometric Learning in Robotics&lt;/h3&gt;

&lt;p&gt;While this study utilized quadrotors, these geometric RL methods are a blueprint for any robotics system with underlying symmetries, from bipedal robots to industrial manipulators. 
Now we must ask:&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;If we can eliminate the redundancy of learning through geometric priors, are we finally moving toward an era of “Near-Zero-Data” robotics where physics-informed architectures replace brute-force computation?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p class=&quot;tip&quot;&gt;For a detailed look at the study, check out the &lt;a href=&quot;https://arxiv.org/abs/2502.20500&quot;&gt;research publication&lt;/a&gt; and explore the open-source &lt;a href=&quot;https://github.com/fdcl-gwu/gym-rotor&quot;&gt;training code&lt;/a&gt;.&lt;/p&gt;
</description>
                <pubDate>Tue, 21 Apr 2026 12:50:20 +0000</pubDate>
                <link>https://fdcl.engineering.gwu.edu/equiv-rl</link>
                <guid isPermaLink="true">https://fdcl.engineering.gwu.edu/equiv-rl</guid>
                
                <category>Research</category>
                
                
            </item>
        
            <item>
                <title>Ship-Relative UAV Pose Estimation with 3D LiDAR</title>
                <description>&lt;p&gt;Accurate ship-relative navigation is essential for enabling safe autonomous landing and operations of UAVs at sea. Traditional systems rely heavily on GPS or cameras, but these are often unreliable in maritime environments. For mission critical applications, GPS is vulnerable to to jamming and spoofing, while cameras suffer in low light and harsh weather.&lt;/p&gt;

&lt;p&gt;To address this, we developed a learning-based pose estimation pipeline that uses 3D LiDAR scans to estimate the full six-degree-of-freedom (6DoF) pose of a UAV relative to a ship, trained in simulation and validated with real-world data collected on a US Naval Academy YP689 research vessel.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;challenges-in-shipboard-pose-estimation&quot;&gt;Challenges in Shipboard Pose Estimation&lt;/h2&gt;

&lt;p&gt;LiDAR provides robustness to lighting and visibility compared to cameras, but ship-relative pose estimation introduces several challenges. Scans are often sparse and incomplete, especially at longer ranges. Furthermore, occlusions from ship structures can obscure key features, and when the ship is the only visible object in the scene, the network has limited contextual information to rely on for accurate alignment. These factors make classical registration algorithms like ICP&lt;sup&gt;1&lt;/sup&gt;, as well as learned alignment methods such as PointNetLK&lt;sup&gt;2&lt;/sup&gt;, prone to failure.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/lidar_pipeline.png&quot; loading=&quot;lazy&quot; style=&quot;width: 700px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;!-- TODO: PointNetLK incorrectly aligned example --&gt;
&lt;hr /&gt;

&lt;h2 id=&quot;dataset-generation-and-collection&quot;&gt;Dataset Generation and Collection&lt;/h2&gt;

&lt;p&gt;Developing any deep learning model requires large amounts of training data, which is especially difficult to obtain in real maritime settings. To address this, we built a simulation environment in Gazebo using a CAD model of the YP689 training vessel. Within this environment, a virtual Ouster OS0-32 LiDAR was mounted on a UAV model to generate thousands of scans. These synthetic scans capture the ship geometry as viewed from varied positions and orientations representative of actual flight trajectories.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/gazebo_water_img_long.png&quot; loading=&quot;lazy&quot; style=&quot;width: 500px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;In addition to simulation data, we collected real-world validation data during experiments aboard the US Naval Academy’s YP689 research vessel in the Chesapeake Bay, MD. A UAV equipped with a real Ouster OS0-32 collected LiDAR data during full-length flight trajectories around the ship. High-accuracy ground-truth poses were also obtained using inertial integration&lt;sup&gt;3&lt;/sup&gt;, as well as RTK GPS for benchmarking.&lt;/p&gt;

&lt;div style=&quot;display: flex; justify-content: center; gap: 20px;&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/full_lidar_pod.png&quot; style=&quot;width:45%;&quot; loading=&quot;lazy&quot; /&gt;
  &lt;img src=&quot;/images/posts/karl/drone_yp_flight.png&quot; style=&quot;width:45%;&quot; loading=&quot;lazy&quot; /&gt;
&lt;/div&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;proposed-model-architecture&quot;&gt;Proposed Model Architecture&lt;/h2&gt;

&lt;p&gt;The core of the system is a Point Transformer-based neural network adapted for ship-relative pose estimation.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/model_arch.png&quot; loading=&quot;lazy&quot; style=&quot;width: 700px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;ul&gt;
  &lt;li&gt;&lt;strong&gt;Input&lt;/strong&gt;: A single LiDAR scan (downsampled to 1024 points, normalized).&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Feature extraction&lt;/strong&gt;: Learns local and global scan features using self-attention layers.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Keypoint prediction&lt;/strong&gt;: Network predicts ship keypoints in the LiDAR frame, using fixed CAD keypoint embeddings as decoder queries.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Pose estimation&lt;/strong&gt;: These predicted keypoints are matched with the ground truth ones, obtaining a closed form algorithm&lt;sup&gt;5&lt;/sup&gt;.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Refinement&lt;/strong&gt;: Finally, a lightweight registration further refines the alignment, where the model output serves as the initial guess.&lt;/li&gt;
&lt;/ul&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/kabsch_scrnsht.png&quot; loading=&quot;lazy&quot; style=&quot;width: 700px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;training-procedure&quot;&gt;Training Procedure&lt;/h2&gt;

&lt;p&gt;The model was trained using 20,000 synthetic LiDAR scans. To better reflect real-world conditions, scans were augmented with Gaussian sensor noise and occlusions from flight deck obstructions. For the training, we utilized an NVIDIA A100 GPU and took about 4 hours to complete 100 epochs on our dataset.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/training_curve.png&quot; loading=&quot;lazy&quot; style=&quot;width: 500px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;A key aspect of this work is that the network was trained &lt;em&gt;only&lt;/em&gt; on simulated data, meaning it never saw real LiDAR scans during training. This allowed us to directly assess the sim-to-real gap and whether a transformer-based approach could generalize well from simulation to the real-world. In particular, this ability would allow for easier scalability to new or potentially unknown ship classes.&lt;/p&gt;

&lt;hr /&gt;

&lt;h2 id=&quot;results-and-evaluation&quot;&gt;Results and Evaluation&lt;/h2&gt;
&lt;p&gt;When evaluated on real-world LiDAR scans, the model achieved a mean rotation error of 2.6° and a mean translation error of 0.36 m directly from the network output. After applying a lightweight refinement step with GICP&lt;sup&gt;4&lt;/sup&gt;, the accuracy improved substantially, reducing rotation errors to 0.62° and translation errors to 7 cm. Notably, more than 96% of predictions had less than 20 cm of translation error relative to the ground truth pose.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/real-world-results.png&quot; loading=&quot;lazy&quot; style=&quot;width: 700px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;These results show that a model trained entirely in simulation can generalize effectively to real-world LiDAR scans collected from shipboard UAV operations. Compared to classical methods such as ICP and PointNetLK, our approach proved significantly more robust to sparsity and partial views, non-uniform point distribution, and doesn’t require an initial pose estimate.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/karl/trajs_scrnsht.png&quot; loading=&quot;lazy&quot; style=&quot;width: 700px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;h2 id=&quot;future-directions&quot;&gt;Future Directions&lt;/h2&gt;
&lt;p&gt;Future directions include incorporating environmental disturbances such as sea spray and dynamic occlusions into simulation to better capture real-world operating conditions. Beyond LiDAR alone, more reliable navigation can be achieved through multi-sensor fusion, integrating vision and inertial measurements to complement LiDAR’s robustness. Finally, optimizing the model for embedded hardware such as NVIDIA Jetson platforms would enable efficient, real-time deployment directly on UAV flight computers.&lt;/p&gt;

&lt;hr /&gt;

&lt;p class=&quot;note&quot;&gt;This work is based on my &lt;a href=&quot;https://www.proquest.com/dissertations-theses/ship-relative-uav-pose-estimation-with-3d-lidar/docview/3237086844/se-2&quot;&gt;M.S. thesis&lt;/a&gt; &lt;em&gt;Ship-Relative UAV Pose Estimation with 3D LiDAR&lt;/em&gt; (2025), advised by Prof. Taeyoung Lee. The source code is available at &lt;a href=&quot;https://github.com/fdcl-gwu/point-transformer&quot;&gt;github.com/fdcl-gwu/point-transformer&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;sup&gt;1&lt;/sup&gt; : P.J. Besl and Neil D. McKay. A method for registration of 3-d shapes. IEEE
Transactions on Pattern Analysis and Machine Intelligence, 14(2):239–256, 1992.&lt;/p&gt;

&lt;p&gt;&lt;sup&gt;2&lt;/sup&gt; : Yasuhiro Aoki, Hunter Goforth, Arun Srivatsan Rangaprasad, and Simon Lucey.
Pointnetlk: Robust &amp;amp; efficient point cloud registration using pointnet. pages
7156–7165, 06 2019.&lt;/p&gt;

&lt;p&gt;&lt;sup&gt;3&lt;/sup&gt; : Kenny Chen, Ryan Nemiroff, and Brett T. Lopez. Direct lidar-inertial odom-
etry: Lightweight lio with continuous-time motion correction. In 2023 IEEE
International Conference on Robotics and Automation (ICRA), pages 3983–3989,
2023.&lt;/p&gt;

&lt;p&gt;&lt;sup&gt;4&lt;/sup&gt; : Aleksandr Segal, Dirk Hähnel, and Sebastian Thrun. Generalized-icp. 06 2009.&lt;/p&gt;

&lt;p&gt;&lt;sup&gt;5&lt;/sup&gt; : Jim Lawrence, Javier Bernal, and Christoph Witzgall. A purely algebraic justifi-
cation of the kabsch-umeyama algorithm. Journal of Research of the National
Institute of Standards and Technology, 124, October 2019.&lt;/p&gt;
</description>
                <pubDate>Wed, 27 Aug 2025 17:00:00 +0000</pubDate>
                <link>https://fdcl.engineering.gwu.edu/pointnet</link>
                <guid isPermaLink="true">https://fdcl.engineering.gwu.edu/pointnet</guid>
                
                <category>Research</category>
                
                
            </item>
        
            <item>
                <title>Data-driven Controls of a Flapping Wing UAV</title>
                <description>&lt;!-- With a wingspan of approximately 4 cm, it stands out as one of the larger butterfly species in the region.  --&gt;
&lt;p&gt;The Monarch butterfly, native to North America, is renowned for its distinctive characteristics. 
Each year, millions of Monarchs embark on an incredible migration journey from North America to Mexico, covering a staggering distance of up to 4000 km. 
Flapping wing aerial vehicles can offer significant advantages in energy efficiency and agility compared to conventional fixed or rotary wing types, whose lift-to-drag ratio deteriorates rapidly as their size is reduced. 
&lt;!-- One notable application is the development of Marsbees designed for the low-density atmosphere of Mars, where high efficiency is essential. --&gt;&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;Consequently, the flapping wing mechanism has been envisioned as a critical component for micro autonomous drones of the next generation.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;div class=&quot;gallery-box&quot;&gt;
  &lt;div class=&quot;gallery gallery-column-3&quot;&gt;
    &lt;img src=&quot;/images/posts/tejaswi/fwuav_npr_monarch.png&quot; loading=&quot;lazy&quot; style=&quot;width: 350px; overflow: hidden&quot; /&gt;
    &lt;img src=&quot;/images/posts/tejaswi/fwuav_NASA_marsbee.png&quot; loading=&quot;lazy&quot; style=&quot;width: 350px; overflow: hidden&quot; /&gt;
  &lt;/div&gt;
  &lt;!-- &lt;em&gt;Caption&lt;/em&gt; --&gt;
&lt;/div&gt;

&lt;!-- ### Challenges in control systems for FWUAVs --&gt;

&lt;p&gt;While numerous bioinspired robots have been developed, advancements in control systems for FWUAVs lag behind those for more traditional unmanned aerial vehicles like quadrotors. 
This disparity is due to the complexity of modeling and controlling flapping-wing dynamics. 
Most existing models focus on high-frequency flapping of small wings, neglecting the intricate coupling between wing motion and body dynamics - key features of butterfly flight.&lt;/p&gt;

&lt;h2 id=&quot;geometric-dynamic-model-and-control&quot;&gt;Geometric dynamic model and control&lt;/h2&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/tejaswi/fwuav_flapping_cycle.png&quot; loading=&quot;lazy&quot; style=&quot;width: 700px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Inspired by the Monarch butterfly’s flight, a novel dynamic model has been developed to account for the effects of low-frequency wing flapping and abdomen undulation. 
This provides an elegant, &lt;strong&gt;global formulation&lt;/strong&gt; of the dynamics, avoiding complexities and singularities associated with local coordinates.
Next, an optimal periodic motion that minimizes the energy variations was constructed, and a feedback control system was proposed to asymptotically stabilize it according to the Floquet stability theory &lt;sup&gt;1&lt;/sup&gt;.&lt;/p&gt;

&lt;h2 id=&quot;constrained-imitation-learning-from-optimal-trajectories&quot;&gt;Constrained imitation learning from optimal trajectories&lt;/h2&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/tejaswi/state_error_comp.png&quot; loading=&quot;lazy&quot; style=&quot;width: 500px; overflow: hidden&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;For the global motion of FWUAV which includes both translation and rotation of the body (6DOF), a periodic optimal controller was utilized after identifying specific wing kinematics parameters and their physical relationship to aerodynamics. 
To further improve computational efficiency, imitation learning was uniquely tailored to transform a set of optimal trajectories into &lt;strong&gt;data-driven feedback control&lt;/strong&gt;. 
Compared with conventional methods, constrained imitation learning (&lt;span style=&quot;color: blue;&quot;&gt;COIL&lt;/span&gt;) eliminates the need to generate additional optimal trajectories on-line, while simultaneously improving stability properties &lt;sup&gt;2&lt;/sup&gt;.&lt;/p&gt;

&lt;h2 id=&quot;modular-control-scheme-using-visual-inertial-data&quot;&gt;Modular control scheme using visual-inertial data&lt;/h2&gt;

&lt;p&gt;Finally, a vision-based control scheme was proposed to avoid estimating the state of the flapping wing aerial vehicle in real time. 
A deep neural pose estimator, based on a Siamese network, extracts robust features for better performance compared to traditional keypoint-based methods &lt;sup&gt;3&lt;/sup&gt;.
Instead of training a monolithic network, we proposed a &lt;em&gt;modular construction&lt;/em&gt; where a pose estimation network and a control network are concatenated, which were trained alternatingly to achieve the complex tasks of end-to-end perception and control efficiently. 
To improve convergence when combined with the controller, an alternating learning algorithm (&lt;span style=&quot;color: blue;&quot;&gt;ALICE&lt;/span&gt;) was presented to iteratively refine the individual neural networks so that ultimately the vehicle can be guided to perform given maneuvers.&lt;/p&gt;

&lt;div class=&quot;gallery-box&quot;&gt;
  &lt;div class=&quot;gallery gallery-column-3&quot;&gt;
    &lt;img src=&quot;/images/posts/tejaswi/fwuav_sensorimotor_control.png&quot; loading=&quot;lazy&quot; style=&quot;width: 350px; overflow: hidden&quot; /&gt;
    &lt;img src=&quot;/images/posts/tejaswi/fwuav_best_comparison.png&quot; loading=&quot;lazy&quot; style=&quot;width: 350px; overflow: hidden&quot; /&gt;
  &lt;/div&gt;
  &lt;!-- &lt;em&gt;Caption&lt;/em&gt; --&gt;
&lt;/div&gt;

&lt;p class=&quot;note&quot;&gt;This &lt;a href=&quot;https://scholarspace.library.gwu.edu/etd/gq67js07z&quot;&gt;framework establishes the first nonlinear control system that stabilizes the coupled 6DOF longitudinal and lateral dynamics of FWUAVs&lt;/a&gt; without relying on the common assumptions of averaging or linearization. Possible future directions:&lt;/p&gt;
&lt;ul&gt;
  &lt;li&gt;Including flexibility of wings and fluid-structure interactions for improved modeling&lt;/li&gt;
  &lt;li&gt;Learning-based aerodynamic models that balance accuracy with computational efficiency, further advancing FWUAV control systems.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;sup&gt;1&lt;/sup&gt; : Tejaswi, K.C., Kang, C.K. and Lee, T., 2021, May. &lt;a href=&quot;https://doi.org/10.23919/acc50511.2021.9483293&quot;&gt;Dynamics and control of a flapping wing uav with abdomen undulation inspired by monarch butterfly&lt;/a&gt;. In 2021 American control conference (ACC) (pp. 66-71). IEEE.&lt;/p&gt;

&lt;p&gt;&lt;sup&gt;2&lt;/sup&gt; : Tejaswi, K.C. and Lee, T., 2022. &lt;a href=&quot;https://doi.org/10.1109/lra.2022.3194682&quot;&gt;Constrained Imitation Learning for a Flapping Wing Unmanned Aerial Vehicle&lt;/a&gt;. IEEE Robotics and Automation Letters, 7(4), pp.10534-10541.&lt;/p&gt;

&lt;p&gt;&lt;sup&gt;3&lt;/sup&gt; : Tejaswi, K.C., Lee, T. and Kang, C.K., 2024. &lt;a href=&quot;https://doi.org/10.2514/6.2024-0948&quot;&gt;Deep Neural Pose Estimation for a Flapping Wing Unmanned Aerial Vehicle with Visual-Inertial Sensor Fusion&lt;/a&gt;. In AIAA SCITECH 2024 Forum (p. 0948).&lt;/p&gt;

&lt;!-- &lt;ol&gt;
&lt;li&gt;Kodihalli Chandrappa, Tejaswi, &quot;Data-Driven Controls of a Flapping Wing Unmanned Aerial Vehicle Inspired by Monarch Butterfly,&quot; The George Washington University, 2024.&lt;/li&gt;&lt;br&gt; 
&lt;/ol&gt; --&gt;
</description>
                <pubDate>Mon, 20 Jan 2025 17:00:00 +0000</pubDate>
                <link>https://fdcl.engineering.gwu.edu/flapping-wing-uav-control</link>
                <guid isPermaLink="true">https://fdcl.engineering.gwu.edu/flapping-wing-uav-control</guid>
                
                <category>Research</category>
                
                
            </item>
        
            <item>
                <title>Modular Reinforcement Learning for a Quadrotor UAV</title>
                <description>&lt;p&gt;Imagine a quadrotor drone gliding through complex environments, expertly handling tight corners and sharp turns while maintaining precise control of its &lt;em&gt;heading&lt;/em&gt;. Achieving this level of dexterity is no small feat, particularly because traditional reinforcement learning (RL)-based control methods often struggle to balance the delicate dynamics of translational and yawing motions.&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;What happens when we give drones a modular brain, designed to focus on specific tasks like balance and direction independently?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2 id=&quot;unlocking-the-potential-of-modular-rl&quot;&gt;Unlocking the Potential of Modular RL&lt;/h2&gt;

&lt;p&gt;Traditional RL approaches to controlling quadrotors often rely on an &lt;em&gt;end-to-end monolithic policy&lt;/em&gt;—a single, unified neural network—designed to manage all aspects of flight dynamics simultaneously. While effective in many cases, these models face challenges when dealing with complex interactions of roll, pitch, and yaw motions. Why? Because rapid yawing motions can destabilize the drone’s other motions, as excessive acceleration/deceleration of quadrotor rotors potentially amplifies the imbalances.&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;

  &lt;img src=&quot;/images/posts/ben/frameworks.png&quot; width=&quot;800&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Enter &lt;strong&gt;modular reinforcement learning&lt;/strong&gt;. By dividing the quadrotor’s dynamics into two distinct parts—translational and yaw subsystems—modular RL allows specialized agents to manage each task independently. This decomposition not only simplifies the learning complexity but also enhances system robustness, as one module can continue operating even if the other encounters difficulties.&lt;/p&gt;

&lt;h2 id=&quot;a-look-at-the-learning-curves&quot;&gt;A Look at the Learning Curves&lt;/h2&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/learning_curve.png&quot; width=&quot;650&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;The proposed modular RL policies, namely &lt;span style=&quot;color: red;&quot;&gt;CMP&lt;/span&gt; and &lt;span style=&quot;color: blue;&quot;&gt;DMP&lt;/span&gt;, achieved faster convergence and superior performance compared to their monolithic counterparts, &lt;span style=&quot;color: green;&quot;&gt;NMP&lt;/span&gt;. By focusing on specific subtasks, each agent can optimize its learning for its assigned task, avoiding the complexities and inefficiencies of learning a single, generalized policy.&lt;/p&gt;

&lt;p&gt;Additionally, we explored centralized and decentralized inter-module communication between two agents. Centralized coordination allowed agents to share information during training, resulting in improved synergy and higher performance. Decentralized agents, while effective, occasionally faced coordination challenges in high-demand scenarios.&lt;/p&gt;

&lt;h2 id=&quot;from-simulation-to-sky&quot;&gt;From Simulation to Sky&lt;/h2&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/sim2real.png&quot; width=&quot;700&quot; /&gt;
&lt;/p&gt;

&lt;p&gt;Training RL agents in simulation is just the beginning; the real flight test comes in deploying them in the physical world. By employing sim-to-real techniques such as domain randomization (introducing variability in simulators) and policy regularization (smoothing control signals), trained RL agents transferred seamlessly from simulation to reality.&lt;/p&gt;

&lt;p&gt;In flight tests, our modular framework consistently outperformed traditional monolithic approaches, particularly during aggressive maneuvers requiring high yaw rates. While traditional models struggled to maintain stability at high yaw rates (e.g., 40 deg/s), the modular agents demonstrated remarkable precision and adaptability. Another advantage? If one module needed fine-tuning, it could be adjusted without retraining the entire system, saving time and resources&lt;/p&gt;

&lt;p align=&quot;center&quot;&gt;
  &lt;img src=&quot;/images/posts/ben/flight_traj.png&quot; width=&quot;800&quot; /&gt;
&lt;/p&gt;

&lt;h2 id=&quot;whats-next-for-modular-rl-in-robotics&quot;&gt;What’s Next for Modular RL in Robotics?&lt;/h2&gt;

&lt;p&gt;Our work marks an important step forward in drone autonomy, but the implications go far beyond quadrotors. From self-driving cars to robotic arms in manufacturing, the modular RL approach promises smarter, more efficient, and more adaptable solutions.&lt;/p&gt;

&lt;p&gt;So, next time you see a drone executing a flawless aerial maneuver, remember—it might just have a modular brain guiding its every move.&lt;/p&gt;

&lt;p class=&quot;tip&quot;&gt;For a detailed look at the study, check out the &lt;a href=&quot;https://doi.org/10.1109/LRA.2024.3511412&quot;&gt;research publication&lt;/a&gt; and explore our &lt;a href=&quot;https://github.com/fdcl-gwu/gym-rotor-modularRL&quot;&gt;training code&lt;/a&gt;.&lt;/p&gt;
</description>
                <pubDate>Tue, 14 Jan 2025 12:01:35 +0000</pubDate>
                <link>https://fdcl.engineering.gwu.edu/reinforcement-learning-quadrotor</link>
                <guid isPermaLink="true">https://fdcl.engineering.gwu.edu/reinforcement-learning-quadrotor</guid>
                
                <category>Research</category>
                
                
            </item>
        
    </channel>
</rss>