GCC Code Coverage Report
Directory: ./ Exec Total Coverage
File: include/crocoddyl/core/solvers/ipopt/ipopt-iface.hpp Lines: 23 42 54.8 %
Date: 2024-02-13 11:12:33 Branches: 35 70 50.0 %

Line Branch Exec Source
1
///////////////////////////////////////////////////////////////////////////////
2
// BSD 3-Clause License
3
//
4
// Copyright (C) 2022-2023, IRI: CSIC-UPC, Heriot-Watt University
5
// Copyright note valid unless otherwise stated in individual files.
6
// All rights reserved.
7
///////////////////////////////////////////////////////////////////////////////
8
9
#ifndef __CROCODDYL_CORE_SOLVERS_IPOPT_IPOPT_IFACE_HPP__
10
#define __CROCODDYL_CORE_SOLVERS_IPOPT_IPOPT_IFACE_HPP__
11
12
#define HAVE_CSTDDEF
13
#include <IpTNLP.hpp>
14
#undef HAVE_CSTDDEF
15
16
#include "crocoddyl/core/mathbase.hpp"
17
#include "crocoddyl/core/optctrl/shooting.hpp"
18
19
namespace crocoddyl {
20
21
struct IpoptInterfaceData;
22
23
/**
24
 * @brief Class for interfacing a crocoddyl::ShootingProblem with IPOPT
25
 *
26
 * This class implements the pure virtual functions from Ipopt::TNLP to solve
27
 * the optimal control problem in `problem_` using a multiple shooting approach.
28
 *
29
 * Ipopt considers its decision variables `x` to belong to the Euclidean space.
30
 * However, Crocoddyl states could lie in a manifold. To ensure that the
31
 * solution of Ipopt lies in the manifold of the state, we perform the
32
 * optimization in the tangent space of a given initial state. Finally we
33
 * retract the Ipopt solution to the manifold. That is:
34
 *  * \f[
35
 * \begin{aligned}
36
 * \mathbf{x}^* = \mathbf{x}^0 \oplus \mathbf{\Delta x}^*
37
 * \end{aligned}
38
 * \f]
39
 *
40
 * where \f$\mathbf{x}^*\f$ is the final solution, \f$\mathbf{x}^0\f$ is the
41
 * initial guess and \f$\mathbf{\Delta x}^*\f$ is the Ipopt solution in the
42
 * tangent space of \f$\mathbf{x}_0\f$. Due to this procedure, the computation
43
 * of the cost function, the dynamic constraint as well as their corresponding
44
 * derivatives should be properly modified.
45
 *
46
 * The Ipopt decision vector is built as follows: \f$x = [ \mathbf{\Delta
47
 * x}_0^\top, \mathbf{u}_0^\top, \mathbf{\Delta x}_1^\top, \mathbf{u}_1^\top,
48
 * \dots, \mathbf{\Delta x}_N^\top ]\f$
49
 *
50
 * Dynamic constraints are posed as: \f$(\mathbf{x}^0_{k+1} \oplus
51
 * \mathbf{\Delta x}_{k+1}) \ominus \mathbf{f}(\mathbf{x}_{k}^0 \oplus
52
 * \mathbf{\Delta x}_{k}, \mathbf{u}_k) = \mathbf{0}\f$
53
 *
54
 * Initial condition: \f$ \mathbf{x}(0) \ominus (\mathbf{x}_{k}^0 \oplus
55
 * \mathbf{\Delta x}_{k}) = \mathbf{0}\f$
56
 *
57
 * Documentation of the methods has been extracted from Ipopt::TNLP.hpp file
58
 *
59
 *  \sa `get_nlp_info()`, `get_bounds_info()`, `eval_f()`, `eval_g()`,
60
 * `eval_grad_f()`, `eval_jac_g()`, `eval_h()`
61
 */
62
63
class IpoptInterface : public Ipopt::TNLP {
64
 public:
65
16
  EIGEN_MAKE_ALIGNED_OPERATOR_NEW
66
67
  /**
68
   * @brief Initialize the Ipopt interface
69
   *
70
   * @param[in] problem  Crocoddyl shooting problem
71
   */
72
  IpoptInterface(const boost::shared_ptr<crocoddyl::ShootingProblem>& problem);
73
74
  virtual ~IpoptInterface();
75
76
  /**
77
   * @brief  Methods to gather information about the NLP
78
   *
79
   * %Ipopt uses this information when allocating the arrays that it will later
80
   * ask you to fill with values. Be careful in this method since incorrect
81
   * values will cause memory bugs which may be very difficult to find.
82
   * @param[out] n            Storage for the number of variables \f$x\f$
83
   * @param[out] m            Storage for the number of constraints \f$g(x)\f$
84
   * @param[out] nnz_jac_g    Storage for the number of nonzero entries in the
85
   * Jacobian
86
   * @param[out] nnz_h_lag    Storage for the number of nonzero entries in the
87
   * Hessian
88
   * @param[out] index_style  Storage for the index style the numbering style
89
   * used for row/col entries in the sparse matrix format
90
   */
91
  virtual bool get_nlp_info(Ipopt::Index& n, Ipopt::Index& m,
92
                            Ipopt::Index& nnz_jac_g, Ipopt::Index& nnz_h_lag,
93
                            IndexStyleEnum& index_style);
94
95
  /**
96
   * @brief Method to request bounds on the variables and constraints.
97
   T
98
   * @param[in]  n    Number of variables \f$x\f$ in the problem
99
   * @param[out] x_l  Lower bounds \f$x^L\f$ for the variables \f$x\f$
100
   * @param[out] x_u  Upper bounds \f$x^U\f$ for the variables \f$x\f$
101
   * @param[in]  m    Number of constraints \f$g(x)\f$ in the problem
102
   * @param[out] g_l  Lower bounds \f$g^L\f$ for the constraints \f$g(x)\f$
103
   * @param[out] g_u  Upper bounds \f$g^U\f$ for the constraints \f$g(x)\f$
104
   *
105
   * @return true if success, false otherwise.
106
   *
107
   * The values of `n` and `m` that were specified in
108
   IpoptInterface::get_nlp_info are passed
109
   * here for debug checking. Setting a lower bound to a value less than or
110
   * equal to the value of the option \ref OPT_nlp_lower_bound_inf
111
   "nlp_lower_bound_inf"
112
   * will cause %Ipopt to assume no lower bound. Likewise, specifying the upper
113
   bound above or
114
   * equal to the value of the option \ref OPT_nlp_upper_bound_inf
115
   "nlp_upper_bound_inf"
116
   * will cause %Ipopt to assume no upper bound. These options are set to
117
   -10<sup>19</sup> and
118
   * 10<sup>19</sup>, respectively, by default, but may be modified by changing
119
   these
120
   * options.
121
   */
122
  virtual bool get_bounds_info(Ipopt::Index n, Ipopt::Number* x_l,
123
                               Ipopt::Number* x_u, Ipopt::Index m,
124
                               Ipopt::Number* g_l, Ipopt::Number* g_u);
125
126
  /**
127
   * \brief Method to request the starting point before iterating.
128
   *
129
   * @param[in]  n            Number of variables \f$x\f$ in the problem; it
130
   * will have the same value that was specified in
131
   * `IpoptInterface::get_nlp_info`
132
   * @param[in]  init_x       If true, this method must provide an initial value
133
   * for \f$x\f$
134
   * @param[out] x            Initial values for the primal variables \f$x\f$
135
   * @param[in]  init_z       If true, this method must provide an initial value
136
   * for the bound multipliers \f$z^L\f$ and \f$z^U\f$
137
   * @param[out] z_L          Initial values for the bound multipliers \f$z^L\f$
138
   * @param[out] z_U          Initial values for the bound multipliers \f$z^U\f$
139
   * @param[in]  m            Number of constraints \f$g(x)\f$ in the problem;
140
   * it will have the same value that was specified in
141
   * `IpoptInterface::get_nlp_info`
142
   * @param[in]  init_lambda  If true, this method must provide an initial value
143
   * for the constraint multipliers \f$\lambda\f$
144
   * @param[out] lambda       Initial values for the constraint multipliers,
145
   * \f$\lambda\f$
146
   *
147
   * @return true if success, false otherwise.
148
   *
149
   * The boolean variables indicate whether the algorithm requires to have x,
150
   * z_L/z_u, and lambda initialized, respectively. If, for some reason, the
151
   * algorithm requires initializations that cannot be provided, false should be
152
   * returned and %Ipopt will stop. The default options only require initial
153
   * values for the primal variables \f$x\f$.
154
   *
155
   * Note, that the initial values for bound multiplier components for absent
156
   * bounds (\f$x^L_i=-\infty\f$ or \f$x^U_i=\infty\f$) are ignored.
157
   */
158
  // [TNLP_get_starting_point]
159
  virtual bool get_starting_point(Ipopt::Index n, bool init_x, Ipopt::Number* x,
160
                                  bool init_z, Ipopt::Number* z_L,
161
                                  Ipopt::Number* z_U, Ipopt::Index m,
162
                                  bool init_lambda, Ipopt::Number* lambda);
163
164
  /**
165
   * @brief Method to request the value of the objective function.
166
   *
167
   * @param[in] n           Number of variables \f$x\f$ in the problem; it will
168
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
169
   * @param[in] x           Values for the primal variables \f$x\f$ at which the
170
   * objective function \f$f(x)\f$ is to be evaluated
171
   * @param[in] new_x       False if any evaluation method (`eval_*`) was
172
   * previously called with the same values in x, true otherwise. This can be
173
   * helpful when users have efficient implementations that calculate multiple
174
   * outputs at once. %Ipopt internally caches results from the TNLP and
175
   *                        generally, this flag can be ignored.
176
   * @param[out] obj_value  Storage for the value of the objective function
177
   * \f$f(x)\f$
178
   *
179
   * @return true if success, false otherwise.
180
   */
181
  virtual bool eval_f(Ipopt::Index n, const Ipopt::Number* x, bool new_x,
182
                      Ipopt::Number& obj_value);
183
184
  /**
185
   * @brief Method to request the gradient of the objective function.
186
   *
187
   * @param[in] n        Number of variables \f$x\f$ in the problem; it will
188
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
189
   * @param[in] x        Values for the primal variables \f$x\f$ at which the
190
   * gradient \f$\nabla f(x)\f$ is to be evaluated
191
   * @param[in] new_x    False if any evaluation method (`eval_*`) was
192
   * previously called with the same values in x, true otherwise; see also
193
   * `IpoptInterface::eval_f`
194
   * @param[out] grad_f  Array to store values of the gradient of the objective
195
   * function \f$\nabla f(x)\f$. The gradient array is in the same order as the
196
   * \f$x\f$ variables (i.e., the gradient of the objective with respect to
197
   * `x[2]` should be put in `grad_f[2]`).
198
   *
199
   * @return true if success, false otherwise.
200
   */
201
  virtual bool eval_grad_f(Ipopt::Index n, const Ipopt::Number* x, bool new_x,
202
                           Ipopt::Number* grad_f);
203
204
  /**
205
   * @brief Method to request the constraint values.
206
   *
207
   * @param[in] n      Number of variables \f$x\f$ in the problem; it will have
208
   * the same value that was specified in `IpoptInterface::get_nlp_info`
209
   * @param[in] x      Values for the primal variables \f$x\f$ at which the
210
   * constraint functions \f$g(x)\f$ are to be evaluated
211
   * @param[in] new_x  False if any evaluation method (`eval_*`) was previously
212
   * called with the same values in x, true otherwise; see also
213
   * `IpoptInterface::eval_f`
214
   * @param[in] m      Number of constraints \f$g(x)\f$ in the problem; it will
215
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
216
   * @param[out] g     Array to store constraint function values \f$g(x)\f$, do
217
   * not add or subtract the bound values \f$g^L\f$ or \f$g^U\f$.
218
   *
219
   * @return true if success, false otherwise.
220
   */
221
  virtual bool eval_g(Ipopt::Index n, const Ipopt::Number* x, bool new_x,
222
                      Ipopt::Index m, Ipopt::Number* g);
223
224
  /**
225
   * @brief Method to request either the sparsity structure or the values of the
226
   * Jacobian of the constraints.
227
   *
228
   * The Jacobian is the matrix of derivatives where the derivative of
229
   * constraint function \f$g_i\f$ with respect to variable \f$x_j\f$ is placed
230
   * in row \f$i\f$ and column \f$j\f$. See \ref TRIPLET for a discussion of the
231
   * sparse matrix format used in this method.
232
   *
233
   * @param[in] n         Number of variables \f$x\f$ in the problem; it will
234
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
235
   * @param[in] x         First call: NULL; later calls: the values for the
236
   * primal variables \f$x\f$ at which the constraint Jacobian \f$\nabla
237
   * g(x)^T\f$ is to be evaluated
238
   * @param[in] new_x     False if any evaluation method (`eval_*`) was
239
   * previously called with the same values in x, true otherwise; see also
240
   * `IpoptInterface::eval_f`
241
   * @param[in] m         Number of constraints \f$g(x)\f$ in the problem; it
242
   * will have the same value that was specified in
243
   * `IpoptInterface::get_nlp_info`
244
   * @param[in] nele_jac  Number of nonzero elements in the Jacobian; it will
245
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
246
   * @param[out] iRow     First call: array of length `nele_jac` to store the
247
   * row indices of entries in the Jacobian f the constraints; later calls: NULL
248
   * @param[out] jCol     First call: array of length `nele_jac` to store the
249
   * column indices of entries in the acobian of the constraints; later calls:
250
   * NULL
251
   * @param[out] values   First call: NULL; later calls: array of length
252
   * nele_jac to store the values of the entries in the Jacobian of the
253
   * constraints
254
   *
255
   * @return true if success, false otherwise.
256
   *
257
   * @note The arrays iRow and jCol only need to be filled once. If the iRow and
258
   * jCol arguments are not NULL (first call to this function), then %Ipopt
259
   * expects that the sparsity structure of the Jacobian (the row and column
260
   * indices only) are written into iRow and jCol. At this call, the arguments
261
   * `x` and `values` will be NULL. If the arguments `x` and `values` are not
262
   * NULL, then %Ipopt expects that the value of the Jacobian as calculated from
263
   * array `x` is stored in array `values` (using the same order as used when
264
   * specifying the sparsity structure). At this call, the arguments `iRow` and
265
   * `jCol` will be NULL.
266
   */
267
  virtual bool eval_jac_g(Ipopt::Index n, const Ipopt::Number* x, bool new_x,
268
                          Ipopt::Index m, Ipopt::Index nele_jac,
269
                          Ipopt::Index* iRow, Ipopt::Index* jCol,
270
                          Ipopt::Number* values);
271
272
  /**
273
   * @brief Method to request either the sparsity structure or the values of the
274
   * Hessian of the Lagrangian.
275
   *
276
   * The Hessian matrix that %Ipopt uses is
277
   * \f[ \sigma_f \nabla^2 f(x_k) + \sum_{i=1}^m\lambda_i\nabla^2 g_i(x_k) \f]
278
   * for the given values for \f$x\f$, \f$\sigma_f\f$, and \f$\lambda\f$.
279
   * See \ref TRIPLET for a discussion of the sparse matrix format used in this
280
   * method.
281
   *
282
   * @param[in] n           Number of variables \f$x\f$ in the problem; it will
283
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
284
   * @param[in] x           First call: NULL; later calls: the values for the
285
   * primal variables \f$x\f$ at which the Hessian is to be evaluated
286
   * @param[in] new_x       False if any evaluation method (`eval_*`) was
287
   * previously called with the same values in x, true otherwise; see also
288
   * IpoptInterface::eval_f
289
   * @param[in] obj_factor  Factor \f$\sigma_f\f$ in front of the objective term
290
   * in the Hessian
291
   * @param[in] m           Number of constraints \f$g(x)\f$ in the problem; it
292
   * will have the same value that was specified in
293
   * `IpoptInterface::get_nlp_info`
294
   * @param[in] lambda      Values for the constraint multipliers \f$\lambda\f$
295
   * at which the Hessian is to be evaluated
296
   * @param[in] new_lambda  False if any evaluation method was previously called
297
   * with the same values in lambda, true otherwise
298
   * @param[in] nele_hess   Number of nonzero elements in the Hessian; it will
299
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
300
   * @param[out] iRow       First call: array of length nele_hess to store the
301
   * row indices of entries in the Hessian; later calls: NULL
302
   * @param[out] jCol       First call: array of length nele_hess to store the
303
   * column indices of entries in the Hessian; later calls: NULL
304
   * @param[out] values     First call: NULL; later calls: array of length
305
   * nele_hess to store the values of the entries in the Hessian
306
   *
307
   * @return true if success, false otherwise.
308
   *
309
   * @note The arrays iRow and jCol only need to be filled once. If the iRow and
310
   * jCol arguments are not NULL (first call to this function), then %Ipopt
311
   * expects that the sparsity structure of the Hessian (the row and column
312
   * indices only) are written into iRow and jCol. At this call, the arguments
313
   * `x`, `lambda`, and `values` will be NULL. If the arguments `x`, `lambda`,
314
   * and `values` are not NULL, then %Ipopt expects that the value of the
315
   * Hessian as calculated from arrays `x` and `lambda` are stored in array
316
   * `values` (using the same order as used when specifying the sparsity
317
   * structure). At this call, the arguments `iRow` and `jCol` will be NULL.
318
   *
319
   * @attention As this matrix is symmetric, %Ipopt expects that only the lower
320
   * diagonal entries are specified.
321
   *
322
   * A default implementation is provided, in case the user wants to set
323
   * quasi-Newton approximations to estimate the second derivatives and doesn't
324
   * not need to implement this method.
325
   */
326
  virtual bool eval_h(Ipopt::Index n, const Ipopt::Number* x, bool new_x,
327
                      Ipopt::Number obj_factor, Ipopt::Index m,
328
                      const Ipopt::Number* lambda, bool new_lambda,
329
                      Ipopt::Index nele_hess, Ipopt::Index* iRow,
330
                      Ipopt::Index* jCol, Ipopt::Number* values);
331
332
  /**
333
   * @brief This method is called when the algorithm has finished (successfully
334
   * or not) so the TNLP can digest the outcome, e.g., store/write the solution,
335
   * if any.
336
   *
337
   * @param[in] status @parblock gives the status of the algorithm
338
   *   - SUCCESS: Algorithm terminated successfully at a locally optimal
339
   *     point, satisfying the convergence tolerances (can be specified
340
   *     by options).
341
   *   - MAXITER_EXCEEDED: Maximum number of iterations exceeded (can be
342
   * specified by an option).
343
   *   - CPUTIME_EXCEEDED: Maximum number of CPU seconds exceeded (can be
344
   * specified by an option).
345
   *   - STOP_AT_TINY_STEP: Algorithm proceeds with very little progress.
346
   *   - STOP_AT_ACCEPTABLE_POINT: Algorithm stopped at a point that was
347
   * converged, not to "desired" tolerances, but to "acceptable" tolerances (see
348
   * the acceptable-... options).
349
   *   - LOCAL_INFEASIBILITY: Algorithm converged to a point of local
350
   * infeasibility. Problem may be infeasible.
351
   *   - USER_REQUESTED_STOP: The user call-back function
352
   * IpoptInterface::intermediate_callback returned false, i.e., the user code
353
   * requested a premature termination of the optimization.
354
   *   - DIVERGING_ITERATES: It seems that the iterates diverge.
355
   *   - RESTORATION_FAILURE: Restoration phase failed, algorithm doesn't know
356
   * how to proceed.
357
   *   - ERROR_IN_STEP_COMPUTATION: An unrecoverable error occurred while %Ipopt
358
   * tried to compute the search direction.
359
   *   - INVALID_NUMBER_DETECTED: Algorithm received an invalid number (such as
360
   * NaN or Inf) from the NLP; see also option check_derivatives_for_nan_inf).
361
   *   - INTERNAL_ERROR: An unknown internal error occurred.
362
   * @endparblock
363
   * @param[in] n           Number of variables \f$x\f$ in the problem; it will
364
   * have the same value that was specified in `IpoptInterface::get_nlp_info`
365
   * @param[in] x           Final values for the primal variables
366
   * @param[in] z_L         Final values for the lower bound multipliers
367
   * @param[in] z_U         Final values for the upper bound multipliers
368
   * @param[in] m           Number of constraints \f$g(x)\f$ in the problem; it
369
   * will have the same value that was specified in
370
   * `IpoptInterface::get_nlp_info`
371
   * @param[in] g           Final values of the constraint functions
372
   * @param[in] lambda      Final values of the constraint multipliers
373
   * @param[in] obj_value   Final value of the objective function
374
   * @param[in] ip_data     Provided for expert users
375
   * @param[in] ip_cq       Provided for expert users
376
   */
377
  virtual void finalize_solution(
378
      Ipopt::SolverReturn status, Ipopt::Index n, const Ipopt::Number* x,
379
      const Ipopt::Number* z_L, const Ipopt::Number* z_U, Ipopt::Index m,
380
      const Ipopt::Number* g, const Ipopt::Number* lambda,
381
      Ipopt::Number obj_value, const Ipopt::IpoptData* ip_data,
382
      Ipopt::IpoptCalculatedQuantities* ip_cq);
383
384
  /**
385
   * @brief Intermediate Callback method for the user.
386
   *
387
   * This method is called once per iteration (during the convergence check),
388
   * and can be used to obtain information about the optimization status while
389
   * %Ipopt solves the problem, and also to request a premature termination.
390
   *
391
   * The information provided by the entities in the argument list correspond to
392
   * what %Ipopt prints in the iteration summary (see also \ref OUTPUT). Further
393
   * information can be obtained from the ip_data and ip_cq objects. The current
394
   * iterate and violations of feasibility and optimality can be accessed via
395
   * the methods IpoptInterface::get_curr_iterate() and
396
   * IpoptInterface::get_curr_violations(). These methods translate values for
397
   * the *internal representation* of the problem from `ip_data` and `ip_cq`
398
   * objects into the TNLP representation.
399
   *
400
   * @return If this method returns false, %Ipopt will terminate with the
401
   * User_Requested_Stop status.
402
   *
403
   * It is not required to implement (overload) this method. The default
404
   * implementation always returns true.
405
   */
406
  bool intermediate_callback(
407
      Ipopt::AlgorithmMode mode, Ipopt::Index iter, Ipopt::Number obj_value,
408
      Ipopt::Number inf_pr, Ipopt::Number inf_du, Ipopt::Number mu,
409
      Ipopt::Number d_norm, Ipopt::Number regularization_size,
410
      Ipopt::Number alpha_du, Ipopt::Number alpha_pr, Ipopt::Index ls_trials,
411
      const Ipopt::IpoptData* ip_data, Ipopt::IpoptCalculatedQuantities* ip_cq);
412
413
  /**
414
   * @brief Create the data structure to store temporary computations
415
   *
416
   * @return the IpoptInterface Data
417
   */
418
  boost::shared_ptr<IpoptInterfaceData> createData(const std::size_t nx,
419
                                                   const std::size_t ndx,
420
                                                   const std::size_t nu);
421
422
  void resizeData();
423
424
  /**
425
   * @brief Return the total number of optimization variables (states and
426
   * controls)
427
   */
428
  std::size_t get_nvar() const;
429
430
  /**
431
   * @brief Return the total number of constraints in the NLP
432
   */
433
  std::size_t get_nconst() const;
434
435
  /**
436
   * @brief Return the state vector
437
   */
438
  const std::vector<Eigen::VectorXd>& get_xs() const;
439
440
  /**
441
   * @brief Return the control vector
442
   */
443
  const std::vector<Eigen::VectorXd>& get_us() const;
444
445
  /**
446
   * @brief Return the crocoddyl::ShootingProblem to be solved
447
   */
448
  const boost::shared_ptr<crocoddyl::ShootingProblem>& get_problem() const;
449
450
  double get_cost() const;
451
452
  /**
453
   * @brief Modify the state vector
454
   */
455
  void set_xs(const std::vector<Eigen::VectorXd>& xs);
456
457
  /**
458
   * @brief Modify the control vector
459
   */
460
  void set_us(const std::vector<Eigen::VectorXd>& us);
461
462
 private:
463
  boost::shared_ptr<crocoddyl::ShootingProblem>
464
      problem_;                      //!< Optimal control problem
465
  std::vector<Eigen::VectorXd> xs_;  //!< Vector of states
466
  std::vector<Eigen::VectorXd> us_;  //!< Vector of controls
467
  std::vector<std::size_t> ixu_;     //!< Index of at node i
468
  std::size_t nvar_;                 //!< Number of NLP variables
469
  std::size_t nconst_;               //!< Number of the NLP constraints
470
  std::vector<boost::shared_ptr<IpoptInterfaceData>>
471
      datas_;    //!< Vector of Datas
472
  double cost_;  //!< Total cost
473
474
  IpoptInterface(const IpoptInterface&);
475
476
  IpoptInterface& operator=(const IpoptInterface&);
477
};
478
479
struct IpoptInterfaceData {
480
  EIGEN_MAKE_ALIGNED_OPERATOR_NEW
481
482
44
  IpoptInterfaceData(const std::size_t nx, const std::size_t ndx,
483
                     const std::size_t nu)
484
44
      : x(nx),
485
        xnext(nx),
486
        dx(ndx),
487
        dxnext(ndx),
488
        x_diff(ndx),
489
        u(nu),
490
        Jint_dx(ndx, ndx),
491
        Jint_dxnext(ndx, ndx),
492
        Jdiff_x(ndx, ndx),
493
        Jdiff_xnext(ndx, ndx),
494
        Jg_dx(ndx, ndx),
495
        Jg_dxnext(ndx, ndx),
496
        Jg_u(ndx, ndx),
497
        Jg_ic(ndx, ndx),
498
        FxJint_dx(ndx, ndx),
499
        Ldx(ndx),
500
        Ldxdx(ndx, ndx),
501








44
        Ldxu(ndx, nu) {
502
44
    x.setZero();
503
44
    xnext.setZero();
504
44
    dx.setZero();
505
44
    dxnext.setZero();
506
44
    x_diff.setZero();
507
44
    u.setZero();
508
44
    Jint_dx.setZero();
509
44
    Jint_dxnext.setZero();
510
44
    Jdiff_x.setZero();
511
44
    Jdiff_xnext.setZero();
512
44
    Jg_dx.setZero();
513
44
    Jg_dxnext.setZero();
514
44
    Jg_u.setZero();
515
44
    Jg_ic.setZero();
516
44
    FxJint_dx.setZero();
517
44
    Ldx.setZero();
518
44
    Ldxdx.setZero();
519
44
    Ldxu.setZero();
520
44
  }
521
522
  void resize(const std::size_t nx, const std::size_t ndx,
523
              const std::size_t nu) {
524
    x.conservativeResize(nx);
525
    xnext.conservativeResize(nx);
526
    dx.conservativeResize(ndx);
527
    dxnext.conservativeResize(ndx);
528
    x_diff.conservativeResize(ndx);
529
    u.conservativeResize(nu);
530
    Jint_dx.conservativeResize(ndx, ndx);
531
    Jint_dxnext.conservativeResize(ndx, ndx);
532
    Jdiff_x.conservativeResize(ndx, ndx);
533
    Jdiff_xnext.conservativeResize(ndx, ndx);
534
    Jg_dx.conservativeResize(ndx, ndx);
535
    Jg_dxnext.conservativeResize(ndx, ndx);
536
    Jg_u.conservativeResize(ndx, ndx);
537
    Jg_ic.conservativeResize(ndx, ndx);
538
    FxJint_dx.conservativeResize(ndx, ndx);
539
    Ldx.conservativeResize(ndx);
540
    Ldxdx.conservativeResize(ndx, ndx);
541
    Ldxu.conservativeResize(ndx, nu);
542
  }
543
544
  Eigen::VectorXd x;        //!< Integrated state
545
  Eigen::VectorXd xnext;    //!< Integrated state at next node
546
  Eigen::VectorXd dx;       //!< Increment in the tangent space
547
  Eigen::VectorXd dxnext;   //!< Increment in the tangent space at next node
548
  Eigen::VectorXd x_diff;   //!< State difference
549
  Eigen::VectorXd u;        //!< Control
550
  Eigen::MatrixXd Jint_dx;  //!< Jacobian of the sum operation w.r.t dx
551
  Eigen::MatrixXd
552
      Jint_dxnext;  //!< Jacobian of the sum operation w.r.t dx at next node
553
  Eigen::MatrixXd
554
      Jdiff_x;  //!< Jacobian of the diff operation w.r.t the first element
555
  Eigen::MatrixXd Jdiff_xnext;  //!< Jacobian of the diff operation w.r.t the
556
                                //!< first element at the next node
557
  Eigen::MatrixXd Jg_dx;        //!< Jacobian of the dynamic constraint w.r.t dx
558
  Eigen::MatrixXd
559
      Jg_dxnext;         //!< Jacobian of the dynamic constraint w.r.t dxnext
560
  Eigen::MatrixXd Jg_u;  //!< Jacobian of the dynamic constraint w.r.t u
561
  Eigen::MatrixXd
562
      Jg_ic;  //!< Jacobian of the initial condition constraint w.r.t dx
563
  Eigen::MatrixXd FxJint_dx;  //!< Intermediate computation needed for Jg_ic
564
  Eigen::VectorXd Ldx;        //!< Jacobian of the cost w.r.t dx
565
  Eigen::MatrixXd Ldxdx;      //!< Hessian of the cost w.r.t dxdx
566
  Eigen::MatrixXd Ldxu;       //!< Hessian of the cost w.r.t dxu
567
};
568
569
}  // namespace crocoddyl
570
571
#endif