ece6554:project_invcartpend
Differences
This shows you the differences between two versions of the page.
ece6554:project_invcartpend [2019/03/31 15:18] – [Linearized System: Cart Trajectory Tracking] pvela | ece6554:project_invcartpend [2024/08/20 21:38] (current) – external edit 127.0.0.1 | ||
---|---|---|---|
Line 4: | Line 4: | ||
It is somewhat simple as the project is single input. | It is somewhat simple as the project is single input. | ||
- | ==== Equations of Motion ==== | + | ===== Equations of Motion |
There are many instances or versions of the equations of motion for this classical problem, depending on the nature of the pendulum. | There are many instances or versions of the equations of motion for this classical problem, depending on the nature of the pendulum. | ||
Line 14: | Line 14: | ||
+ \frac 12 m_p l^2 \sin^2(\theta) \dot \theta^2 + \frac 12 J_p \dot \theta^2 | + \frac 12 m_p l^2 \sin^2(\theta) \dot \theta^2 + \frac 12 J_p \dot \theta^2 | ||
+ m_pgl(1 - \cos(\theta)) | + m_pgl(1 - \cos(\theta)) | ||
- | | + | \\ |
& = \frac 12 (m_c + m_p) \dot x^2 + m_p l \cos(\theta) \dot x \dot \theta | & = \frac 12 (m_c + m_p) \dot x^2 + m_p l \cos(\theta) \dot x \dot \theta | ||
+ \frac 12 (J_p + m_p l^2) \dot \theta^2 - m_pgl(1 - \cos(\theta)) | + \frac 12 (J_p + m_p l^2) \dot \theta^2 - m_pgl(1 - \cos(\theta)) | ||
- | \end{split} | + | \end{split} |
\end{equation} | \end{equation} | ||
or the Lagrangian | or the Lagrangian | ||
Line 34: | Line 34: | ||
\begin{bmatrix} u - \delta_\theta \dot x \\ -\delta_\theta \dot \theta \end{bmatrix} | \begin{bmatrix} u - \delta_\theta \dot x \\ -\delta_\theta \dot \theta \end{bmatrix} | ||
\end{equation} | \end{equation} | ||
- | If the pendulum is a ball at the end of a think stick, then the inertia is negligible and $J_p$ can be ignored. | + | If the pendulum is a ball at the end of a thin stick, then the inertia is negligible and $J_p$ can be ignored. |
Given the prevalence of $m_p l$ almost everywhere, it is useful to factor the product out | Given the prevalence of $m_p l$ almost everywhere, it is useful to factor the product out | ||
Line 50: | Line 50: | ||
\begin{bmatrix} \ddot x \\ \ddot \theta \end{bmatrix} | \begin{bmatrix} \ddot x \\ \ddot \theta \end{bmatrix} | ||
= | = | ||
- | \begin{bmatrix} \dot \theta^2 \sin(\theta) - D_x \dot x + u \\ g \sin(\theta) -D_\theta \dot \theta \end{bmatrix} | + | \begin{bmatrix} \dot \theta^2 \sin(\theta) - D_x \dot x \\ g \sin(\theta) -D_\theta \dot \theta \end{bmatrix} |
+ \begin{bmatrix} b u \\ 0 \end{bmatrix} | + \begin{bmatrix} b u \\ 0 \end{bmatrix} | ||
\end{equation} | \end{equation} | ||
Line 106: | Line 106: | ||
- | ==== Parameters and Limits ==== | + | ===== Parameters and Limits |
^ Parameter ^ Value ^ | ^ Parameter ^ Value ^ | ||
Line 119: | Line 119: | ||
My units might be off for the friction, but the values are fine. | My units might be off for the friction, but the values are fine. | ||
+ | ===== Implementation ===== | ||
+ | |||
+ | Functional code stubs for the implementation are provided in the {{ ECE6554: | ||
====== Activities ====== | ====== Activities ====== | ||
---------------------- | ---------------------- | ||
Line 141: | Line 144: | ||
Implement an adaptive system and compare the outcomes. Run the adaptive system twice. | Implement an adaptive system and compare the outcomes. Run the adaptive system twice. | ||
- | ===== Step 2: Nonlinear Controller | + | ===== Step 2: Implement Single-Layer GRBF Neural Network for Function Approximation |
- | TBD | + | |
- | ===== Step 3: What ===== | + | To be fleshed out, but main idea is to first approximate a real function like $\sin(x)$ or $\mathrm{sinc}(x)$ or some polynomial function, over a fixed interval. |
- | TBD | + | |
+ | For completeness, | ||
+ | ===== Step 3: Neuro-Adaptive Controller ===== | ||
+ | |||
+ | The inverted cart-pendulum system is a bit tricky to work with from a nonlinear control perspective because the single control acts through the two states to " | ||
+ | |||
+ | Nevertheless, | ||
+ | |||
+ | Run the system with increasing angular errors until the system loses stability. Note down a few initial conditions that will lead to instability. Run your neuro-adaptive controller on a few stable instances to learn the $\alpha$ parameters (include a few that go near the point of instability). | ||
+ | |||
+ | Turn in plots of the " | ||
+ | |||
+ | How well it works is a function of how many basis elements you have, their bandwidth, etc. Make sure that they cover the domain of operation (including the extended area where you'd like to have stability). Note that the nonlinearities are independent of the cart position $x$, so you should not include it in the function approximator. | ||
+ | |||
+ | ===== Step 4: Matched and Unmatched Uncertainty ===== | ||
+ | |||
+ | The inverted cart pendulum system has a set of nonlinearities that are matched, plus a set that are not. Consider adaptive control for that set: | ||
+ | |||
+ | \begin{equation} | ||
+ | \dot x = A x + f_u(x) + B(\theta) (u + f_m(x)) | ||
+ | \end{equation} | ||
+ | |||
+ | Instead of using a GRBF-NN, consider cancelling only the part that can be cancelled and leaving alone the part that can't be cancelled. | ||
+ | If you are taking the Nonlinear Control class and have covered ISS, then you should be able to come up with a targeted adaptive controller similar to the disturbance rejection one that compensates for the state-dependent unmatched uncertainty with better boundedness or stabilization properties relative to naive adaptive disturbance compensation. | ||
+ | |||
+ | |||
+ | **Note:** The uncancelled parts can actually be recovered using ideas from concurrent learning. Basically store the evolution of the system and use it to recover what that part should look like, assuming that the remainder of the adaptive controller was functioning properly. | ||
+ | |||
+ | |||
+ | ===== Implementation Tips ===== | ||
+ | |||
+ | ==== Tip 1: Neuro-Adaptive Updates ==== | ||
+ | |||
+ | When implementing new things, it is important to start from a position of strength. Do not try to do it all in one step given that there are many smaller steps that you most likely have never done. Break the system down into smaller pieces. From the first step, you should already have functioning adaptive controllers on only the linear system, plus running on the true nonlinear dynamics (but with a linear reference model). Naturally, this means you also have the non-adaptive versions working, as those should have been an initial testing step prior to incorporating adaptation. This is what //Step 1// aims to achieve. | ||
+ | |||
+ | Assuming that you know how to implement a neural network, then the new part is the neuro-adaptive component. | ||
+ | Otherwise you've got two new parts, the neural network and the neuro-adaptive controller. Continuing, rather than try to implement the more complex version of a neuro-adaptive controller for a multi-state system, it is better to implement neuro-adaptive estimation on a simpler system as the only adaptive component. | ||
+ | \begin{equation} | ||
+ | \dot x = f(x) + u(x; \alpha) = f(x) + k x - \alpha^T \Phi(x) | ||
+ | \end{equation} | ||
+ | where $k < 0$. Pick some nonlinear function $f(x)$ and then see if you can estimate it using the neuro-adaptive component. Establish the domain of approximation. Figure out what a decent grid spacing for the Gaussian RBF centers is and how many would fit in the domain based on the grid spacing. Pick a bandwidth consistent with the grid spacing. | ||
+ | the difference and the $\log$ of the difference, $\alpha^T \Phi(x) - f(x)$ and $\log(\alpha^T \Phi(x) - f(x))$, to see that it is learning the function. Usually learning is along the trajectories followed, so evaluate ths function along the trajectory of the previously integrated simulation(s). Run for different time durations. | ||
+ | |||
+ | This is just one tip. Overall, if you are not doing it, you need to learn how to break down a problem into digestable bits. The Step 1 and Step 2 breakdown does that, but you can and should go ever further yourself when resolving this project. | ||
+ | |||
+ | ==== Tip 2: Neural Network Approximation ==== | ||
+ | |||
+ | Suppose that you haven' | ||
+ | |||
+ | Overall, there is an exploration path that you need to follow if you are going to succeed at this. | ||
+ | ==== Tip 3: Implementing the Function Approximator ==== | ||
+ | |||
+ | One difficult thing for many newbies to Matlab is writing compact and efficient code. Efficient usually means avoiding '' | ||
+ | |||
+ | * '' | ||
+ | * '' | ||
+ | * '' | ||
+ | |||
+ | If you do it all properly, the addition of a neuro-adaptive controller shouldn' | ||
+ | |||
+ | |||
+ | ====== Report Considerations ====== | ||
+ | ---------------------- | ||
+ | Even though the neuro-adaptive controller was done first, a sequencing that makes sense goes from no adaptation to matched adaptation and eventually to neuro-adaptation. | ||
- | ===== References ===== | + | ====== References |
---------------------- | ---------------------- | ||
- | TBD | + | No references given. |
------ | ------ |
ece6554/project_invcartpend.1554059904.txt.gz · Last modified: 2024/08/20 21:38 (external edit)