@@ -417,48 +417,49 @@ struct general_matrix_matrix_product<Index, stan::math::var, LhsStorageOrder,
417
417
#if EIGEN_VERSION_AT_LEAST(3, 3, 8)
418
418
static void run (Index rows, Index cols, Index depth, const LhsScalar* lhs,
419
419
Index lhsStride, const RhsScalar* rhs, Index rhsStride,
420
- ResScalar* res, Index resIncr, Index resStride, const ResScalar& alpha,
420
+ ResScalar* res, Index resIncr, Index resStride,
421
+ const ResScalar& alpha,
421
422
level3_blocking<LhsScalar, RhsScalar>& /* blocking */ ,
422
- GemmParallelInfo<Index>* /* info = 0 */ ) {
423
+ GemmParallelInfo<Index>* /* info = 0 */ ){
423
424
#else
424
425
static void run (Index rows, Index cols, Index depth, const LhsScalar* lhs,
425
426
Index lhsStride, const RhsScalar* rhs, Index rhsStride,
426
427
ResScalar* res, Index resStride, const ResScalar& alpha,
427
428
level3_blocking<LhsScalar, RhsScalar>& /* blocking */ ,
428
429
GemmParallelInfo<Index>* /* info = 0 */ ) {
429
430
#endif
430
- for (Index i = 0 ; i < cols; i++) {
431
- general_matrix_vector_product<
432
- Index, LhsScalar, LhsMapper, LhsStorageOrder, ConjugateLhs, RhsScalar,
433
- RhsMapper,
434
- ConjugateRhs>::run (rows, depth, lhs, lhsStride,
435
- &rhs[static_cast <int >(RhsStorageOrder)
436
- == static_cast <int >(ColMajor)
437
- ? i * rhsStride
438
- : i],
439
- static_cast <int >(RhsStorageOrder)
440
- == static_cast <int >(ColMajor)
441
- ? 1
442
- : rhsStride,
443
- &res[i * resStride], 1 , alpha);
444
- }
445
- }
446
-
447
- EIGEN_DONT_INLINE
448
- static void run (Index rows, Index cols, Index depth,
449
- const LhsMapper& lhsMapper, const RhsMapper& rhsMapper,
450
- ResScalar* res, Index resStride, const ResScalar& alpha,
451
- level3_blocking<LhsScalar, RhsScalar>& blocking,
452
- GemmParallelInfo<Index>* info = 0 ) {
453
- const LhsScalar* lhs = lhsMapper.data ();
454
- const Index lhsStride = lhsMapper.stride ();
455
- const RhsScalar* rhs = rhsMapper.data ();
456
- const Index rhsStride = rhsMapper.stride ();
431
+ for (Index i = 0 ; i < cols; i++) {
432
+ general_matrix_vector_product<
433
+ Index, LhsScalar, LhsMapper, LhsStorageOrder, ConjugateLhs,
434
+ RhsScalar, RhsMapper,
435
+ ConjugateRhs>::run (rows, depth, lhs, lhsStride,
436
+ &rhs[static_cast <int >(RhsStorageOrder)
437
+ == static_cast <int >(ColMajor)
438
+ ? i * rhsStride
439
+ : i],
440
+ static_cast <int >(RhsStorageOrder)
441
+ == static_cast <int >(ColMajor)
442
+ ? 1
443
+ : rhsStride,
444
+ &res[i * resStride], 1 , alpha);
445
+ }
446
+ } // namespace internal
457
447
458
- run (rows, cols, depth, lhs, lhsStride, rhs, rhsStride, res, resStride,
459
- alpha, blocking, info);
460
- }
461
- };
448
+ EIGEN_DONT_INLINE
449
+ static void run (Index rows, Index cols, Index depth, const LhsMapper& lhsMapper,
450
+ const RhsMapper& rhsMapper, ResScalar* res, Index resStride,
451
+ const ResScalar& alpha,
452
+ level3_blocking<LhsScalar, RhsScalar>& blocking,
453
+ GemmParallelInfo<Index>* info = 0 ) {
454
+ const LhsScalar* lhs = lhsMapper.data ();
455
+ const Index lhsStride = lhsMapper.stride ();
456
+ const RhsScalar* rhs = rhsMapper.data ();
457
+ const Index rhsStride = rhsMapper.stride ();
458
+
459
+ run (rows, cols, depth, lhs, lhsStride, rhs, rhsStride, res, resStride, alpha,
460
+ blocking, info);
461
+ }
462
+ }; // namespace Eigen
462
463
} // namespace internal
463
464
} // namespace Eigen
464
465
#endif
0 commit comments