--- a +++ b/Docs/Applications/Daily-activities-and-ergonomics/PosturePredictionModel.md @@ -0,0 +1,101 @@ +--- +gallery_title: "Posture Prediction Model" +gallery_image: "/Applications/images/StandingPosturePredictionModel.webp" +--- + + +(sphx_glr_auto_examples_adls_and_ergonomics_plot_PosturePredictionModel.py)= +(example_posture_prediction)= +# Posture Prediction Model + + +````{sidebar} **Example** +<img src="/Applications/images/StandingPosturePredictionModel.webp" width="70%" align="center"> + +```` + +This is a model which can predict the posture as a consequence of applied loads in hands. +It does this by minimizing joint torques and applying balance drivers which account for external +applied loads. + + +:::{seealso} +**Main file location in AMMR:** + +{menuselection}`Application --> Examples --> StandingPosturePredictionWithLoad --> +StandingPosturePrediction.main.any` +::: + + +The model is driven by a combination of the following drivers: +: - Drivers which minimize the joint moments (arising from gravity and applied loads in hands) in elbow, shoulder, L4L5 and knee + - Driver which tries to keep the CoP inside the foot stance area. + - Feet maintain contact with the ground, but the position can be controlled by widgets + - Hands are linked to an object, of which positioning can be altered using widgets + + +:::{figure} /Applications/images/StandingPosturePredictionModel_UpwardsForce.jpg +:align: center +::: + + +Two type of loads can be applied, either a fixed weight of the object and/or a force vector. + +## Usage + +To run the model +: - Load the model + +:::{figure} /Applications/images/StandingPosturePredictionModel_UponLoading.jpg +:align: center +::: + +The example model has a force vector applied on the object, currently the object has no weight (this is editable) + +```AnyScriptDoc +AnyVar ObjectWeight = 0; //weight of mass applied +``` + +- Try to drag (click and drag) one of the widgets in the ModelView (seen as small coordinate systems). For example the force vector widget: + +:::{figure} /Applications/images/StandingPosturePredictionModel_WidgetMovement.jpg +:align: center +::: + +- When the widget is release the model will run the analysis. The Force vector is shown by the blue arrow, the weight of the object is represented by a red line, and the resulting reaction force is depicted by the green arrow. + +:::{figure} /Applications/images/StandingPosturePredictionModel_PickupBox.jpg +:align: center +::: + +There are three remaining widgets, 2 controlling the starting position of the feet, and 1 controlling object's location. + +:::{figure} /Applications/images/StandingPosturePredictionModel_FootWidgets.jpg +:align: center +::: + +The model can also be adjusted to apply the load to both hands or a single hand. +Three combinations can be made using the define statements + +```AnyScriptDoc +#define LoadInRightHand 1 +#define LoadInLeftHand 1 +``` + +In the following example a force vector is applied to only the right hand. + +:::{figure} /Applications/images/StandingPosturePredictionModel_Onearm.jpg +:align: center +::: + +The following video shows an example of how you can manipulate the multiple widgets to achieve different predictive postures: + +```{raw} html +<video width="100%" style="display:block; margin: 0 auto;" controls autoplay loop> + <source src="../../_static/StandingPosturePrediction.mp4" type="video/mp4"> +Your browser does not support the video tag. +</video> +``` + +**Please note the speed of the video was increased for viewing purposes.** +