//| Copyright: (C) 2020-2024 Kevin Larke //| License: GNU GPL version 3.0 or above. See the accompanying LICENSE file. #ifndef cwMtx_h #define cwMtx_h /* Memory Layout: 3x2 mtx: 0 3 1 4 2 5 3x2x2 mtx front back 0 3 | 6 9 1 4 | 7 10 2 5 | 8 11 More about dope and weight vectors. https://stackoverflow.com/questions/30409991/use-a-dope-vector-to-access-arbitrary-axial-slices-of-a-multidimensional-array */ namespace cw { namespace mtx { enum { kAliasReleaseFl = 0x01, // do not allocate memory, use the passed data pointer, and eventually release it kAliasNoReleaseFl = 0x02, // do not allocate memory, use the passed data pointer, and do not ever release it kDimV_NoReleaseFl = 0x04, // do not release the dimV array when the matrix is released. kDuplDataFl = 0x08, // allocate data space and copy the data in kZeroFl = 0x10, // zero the newly allocated data }; template< typename T > struct mtx_str { unsigned flags = 0; unsigned dimN = 0; unsigned* dimV = nullptr; unsigned* mulV = nullptr; T* base = nullptr; unsigned allocEleN = 0; // always 0 if data is aliased }; template< typename T > void release( struct mtx_str& m ) { if( cwIsNotFlag(m.flags,kDimV_NoReleaseFl) ) mem::release(m.dimV); if( cwIsNotFlag(m.flags,kAliasNoReleaseFl) ) mem::release(m.base); } template< typename T > void release( struct mtx_str*& m ) { if( m != nullptr ) { release(*m); mem::release(m); } } // Release data memory when this matrix is released template< typename T > void set_memory_release_flag( struct mtx_str& m, bool linkDimVFl=true ) { m.flags = cwClearFlag(m.flags,kAliasNoReleaseFl); if( linkDimVFl ) m.flags = cwClearFlag(m.flags,kDimV_NoReleaseFl); } // Do NOT release data memory when this matrix is released. template< typename T > void clear_memory_release_flag( struct mtx_str& m, bool linkDimVFl=true ) { m.flags = cwSetFlag(m.flags,kAliasNoReleaseFl); if( linkDimVFl ) m.flags = cwSetFlag(m.flags,kDimV_NoReleaseFl); } // Note that dimV[] is always copied and therefore is the reponsibility of the caller to free. template< typename T > struct mtx_str* _init( struct mtx_str* m, unsigned dimN, const unsigned* dimV, T* base=nullptr, unsigned flags=0 ) { // if a pre-allocated mtx obj was not given then allocate one if( m == nullptr ) m = mem::allocZ>(1); // if the pre-allocd mtx obj has more dim's than the new one if( m->dimN >= dimN ) m->dimN = dimN; else // else expand dimV[] { m->dimV = mem::resize(m->dimV,dimN*2); m->mulV = m->dimV + dimN; m->dimN = dimN; } // update dimV[] with the new extents and calc. the new ele count unsigned eleN = 0; unsigned mul = 1; for(unsigned i=0; idimV[i] = dimV[i]; m->mulV[i] = mul; mul *= dimV[i]; eleN = (i==0 ? 1 : eleN) * dimV[i]; } bool aliasFl = cwIsFlag(flags, kAliasNoReleaseFl | kAliasReleaseFl ); // if the new object data is aliased if( aliasFl ) { // release any memory the pre-allocated obj may own if( cwIsNotFlag(m->flags,kAliasNoReleaseFl) ) mem::release(m->base); m->base = base; m->allocEleN = 0; // always 0 when data is aliased } else // the new object is not aliased { // if the current data space is too small then reallocate it if( eleN > m->allocEleN ) { // don't allow an alias-no-release ptr to be released if( cwIsFlag(m->flags,kAliasNoReleaseFl) ) m->base = nullptr; m->base = mem::resize(m->base, eleN, cwIsFlag(flags,kZeroFl) ? mem::kZeroAllFl : 0 ); m->allocEleN = eleN; } } // if duplication was requested if( cwIsFlag(flags,kDuplDataFl) ) { assert( aliasFl == false ); assert( base != nullptr ); if( base != nullptr ) memcpy(m->base,base, eleN*sizeof(T) ); } m->flags = flags; return m; } template< typename T > struct mtx_str* alloc( const unsigned* dimV, unsigned dimN, T* base, unsigned flags=0 ) { return _init( nullptr, dimN, dimV, base, flags); } template< typename T> struct mtx_str* _alloc_( unsigned flags, unsigned* dimV, unsigned dimN) { return alloc(dimV, dimN, nullptr, flags ); } template< typename T, typename... ARGS> struct mtx_str* _alloc_( unsigned flags, unsigned* dimV, unsigned dimN, unsigned n, ARGS&&... args) { unsigned _dimV[ dimN + 1 ]; vop::copy(_dimV,dimV,dimN); _dimV[dimN] = n; return _alloc_(flags,_dimV, dimN+1, std::forward(args)...); } template< typename T, typename... ARGS> struct mtx_str* alloc( unsigned flags, ARGS&&... args) { return _alloc_(flags,nullptr,0,args...); } // Allocate the matrix w/o zeroing the initial contents template< typename T > struct mtx_str* alloc( const unsigned* dimV, unsigned dimN ) { return _init( nullptr, dimN, dimV, nullptr, 0); } // Allocate the matrix and zero the contents template< typename T > struct mtx_str* allocZ( const unsigned* dimV, unsigned dimN ) { return _init( nullptr, dimN, dimV, nullptr, kZeroFl); } // Allocate the matrix and copy the data from base[] template< typename T > struct mtx_str* allocDupl( unsigned dimN, const unsigned* dimV, const T* base ) { return _init( nullptr, dimN, dimV, const_cast(base), kDuplDataFl); } // Allocate a matrix and use base[] as the data. Release base[] when it is no longer needed. template< typename T > struct mtx_str* allocAlias( unsigned dimN, const unsigned* dimV, T* base ) { return _init( nullptr, dimN, dimV, base, kAliasReleaseFl); } // Allocate a mtrix and use base[] as the data - do NOT release base[]. template< typename T > struct mtx_str* allocAliasNoRelease( unsigned dimN, const unsigned* dimV, const T* base ) { return _init( nullptr, dimN, dimV, const_cast(base), kAliasNoReleaseFl); } unsigned _offsetDimV( const unsigned* dimV, unsigned dimN, unsigned* idxV ); unsigned _offsetMulV( const unsigned* dimV, unsigned dimN, unsigned* idxV ); unsigned _mtx_object_get_degree( const struct object_str* cfg ); rc_t _mtx_object_get_shape( const struct object_str* cfg, unsigned i, unsigned* dimV, unsigned dimN, unsigned& eleN ); // 'i' is the index into 'idxV[]' of the matrix dimension which 'cfg' refers to template< typename T> rc_t _get_mtx_eles_from_cfg( const struct object_str* cfg, struct mtx_str* m, unsigned i, unsigned* idxV ) { rc_t rc = kOkRC; // if cfg is not a list then this must be a value if( !cfg->is_list() ) { // get the value T v; if(cfg->value(v) != kOkRC ) return cwLogError(kSyntaxErrorRC,"Unable to obtain matrix value in dimension index: %i\n",i); // and store it in the current idxV[] location m->base[ _offsetMulV(m->mulV,m->dimN,idxV) ] = v; return kOkRC; } // otherwise this is a list - and the list must contain lists or values for(unsigned j=0; jchild_count(); ++j) { // update idxV[] which the dimension of the ith child elment idxV[i] = j; // recurse! if((rc = _get_mtx_eles_from_cfg(cfg->child_ele(j) ,m, i+1, idxV)) != kOkRC ) break; } return rc; } // Allocate a new matrix by parsing an object_t description. template< typename T > struct mtx_str* allocCfg( const struct object_str* cfg ) { unsigned dimN = 0; // get the degree of the matrix dimN = _mtx_object_get_degree(cfg); // if 'cfg' does not refer to a matrix if( dimN == 0 ) { cwLogError(kSyntaxErrorRC,"The matrix object does not have a list-list syntax."); } else { // allocate the shape vector unsigned dimV[dimN]; unsigned idxV[dimN]; unsigned eleN = 0; struct mtx_str* m = nullptr; // get the shape of the matrix if( _mtx_object_get_shape(cfg,0,dimV,dimN,eleN) != kOkRC ) return nullptr; // allocate the matrix if((m = alloc(dimV,dimN)) == nullptr ) cwLogError(kObjAllocFailRC,"A matrix allocation failed."); else // if(_get_mtx_eles_from_cfg(cfg,m,0,idxV) == kOkRC ) return m; } return nullptr; } template< typename T > void _slice_setup( const struct mtx_str& m, const unsigned* sIdxV, const unsigned* sCntV, unsigned* siV, unsigned *snV ) { // if sIdx is not given then assume it is the origin if( sIdxV != nullptr ) vop::copy(siV,sIdxV,m.dimN); else vop::zero(siV,m.dimN); // calculate the length in each dimension for(unsigned i=0; i struct mtx_str* alloc( const struct mtx_str& src, const unsigned* sIdxV, const unsigned* sCntV ) { struct mtx_str* m; unsigned siV[ src.dimN ]; unsigned snV[ src.dimN ]; unsigned dIdxV[ src.dimN ]; vop::zero(dIdxV,src.dimN); _slice_setup(src,sIdxV,sCntV,siV,snV); if((m = alloc(snV,src.dimN) ) == nullptr ) return nullptr; copy(*m,dIdxV,src,siV,snV); return m; } // Allocate a new matrix by slicing an existing matrix and aliasing the contents into a new matrix. // Set the elements of sCntV to kInvalidCnt to indicate that the entire dimension following the offset index should be copied. // Set sIdxV to nullptr to begin at 0,0,... // Set sCntV to nullptr to take all elements after sIdxV. template< typename T > struct mtx_str* sliceAlias( const struct mtx_str& src, const unsigned* sIdxV, const unsigned* sCntV ) { struct mtx_str* m; unsigned siV[ src.dimN ]; unsigned snV[ src.dimN ]; _slice_setup(src,sIdxV,sCntV,siV,snV); m = allocAliasNoRelease( src.dimN, snV, addr(src,siV) ); // the memory layout in the slice mtx is the same as the matrix // that it aliases and therefore the 'mulV' vector is the same // in both matrices. vop::copy(m->mulV,src.mulV,src.dimN); return m; } template struct mtx_str* _slice( const struct mtx_str& m, unsigned* iV, unsigned iN, unsigned index ) { // the offset index vector must be fully specified if( index < m.dimN ) { cwLogError(kInvalidArgRC,"An invalid number index + count values was given to slice(). %i < %i.",index,m.dimN); return nullptr; } // fill in the end of iV[] with kInvalidCnt to indicate that all values after the offset should be copied for(; index(m,iV,iV+m.dimN); } template< typename T0, typename T1, typename... ARGS> struct mtx_str* _slice( const struct mtx_str& m, unsigned* iV, unsigned iN, unsigned index, unsigned n, ARGS&&... args) { if( index >= iN ) { cwLogError(kInvalidArgRC,"Too many index/count arguments were passed to mtx::slice()."); return nullptr; } iV[index] = n; return _slice(m,iV,iN,index+1,std::forward(args)...); } // This function is a wrapper around alloc(const struct mtx_str& src, const unsigned* sIdxV, const unsigned* sCntV ). // The argument list should specify the values for sIdxV[0:dimN] and sCntV[0:dimN]. // The count of arguments should therefore not exceed src.dimN*2. // The sCntV[] argument list may be truncated, or set to kInvalidCnt, if all values after the offset for a given dimension are to be copied. template< typename T0, typename T1, typename... ARGS> struct mtx_str* slice( const struct mtx_str& m, ARGS&&... args) { unsigned iV[ m.dimN*2 ]; return _slice(m, iV, m.dimN*2, 0, std::forward(args)...); } template struct mtx_str* _sliceAlias( const struct mtx_str& m, unsigned* iV, unsigned iN, unsigned index ) { // the offset index vector must be fully specified if( index < m.dimN ) { cwLogError(kInvalidArgRC,"An invalid number index + count values was given to sliceAlias(). %i < %i.",index,m.dimN); return nullptr; } // fill in the end of iV[] with kInvalidCnt to indicate that all values after the offset should be copied for(; index(m,iV,iV+m.dimN); } template< typename T, typename... ARGS> struct mtx_str* _sliceAlias( const struct mtx_str& m, unsigned* iV, unsigned iN, unsigned index, unsigned n, ARGS&&... args) { if( index >= iN ) { cwLogError(kInvalidArgRC,"Too many index/count arguments were passed to mtx::sliceAlias()."); return nullptr; } iV[index] = n; return _sliceAlias(m,iV,iN,index+1,std::forward(args)...); } template< typename T, typename... ARGS> struct mtx_str* slice_alias( const struct mtx_str& m, ARGS&&... args) { unsigned iV[ m.dimN*2 ]; return _sliceAlias(m,iV, m.dimN*2, 0, args...); } // resize m[] template< typename T > struct mtx_str* resize( struct mtx_str* m, const unsigned* dimV, unsigned dimN, T* base=nullptr, unsigned flags=0 ) { return _init( m, dimN, dimV, base, flags ); } // resize y[] to have the same size as x[] template< typename T > struct mtx_str* resize( struct mtx_str* y, const struct mtx_str& x ) { return resize(y,x->dimV,x->dimN); } // Copy a slice of src[] into dst[] at a particular location. // dst[] is assumed to be allocated with sufficient size to receive src[]. // Set dIdxV to nullptr to copy to the 0,0, ... of the dst matrix // Set sIdxV to nullptr to copy from the 0,0, ... of the src matrix. // Set sCntV to nullptr to cop all of the src matrix. template< typename T0, typename T1 > rc_t copy( struct mtx_str& dst, const unsigned* dIdxV, const struct mtx_str& src, const unsigned* sIdxV, const unsigned* sCntV ) { rc_t rc = kOkRC; unsigned nV[ src.dimN ]; unsigned siV[ src.dimN ]; unsigned diV[ dst.dimN ]; vop::zero(nV,src.dimN); if( sCntV == nullptr ) sCntV = src.dimV; if( sIdxV == nullptr ) vop::zero(siV,src.dimN); else vop::copy(siV,sIdxV,src.dimN); if( dIdxV == nullptr ) vop::zero(diV,dst.dimN); else vop::copy(diV,dIdxV,dst.dimN); #ifndef NDEBUG // verify the starting address assert( is_legal_address(dst, diV) ); assert( is_legal_address(src, siV) ); vop::add(siV,sCntV,src.dimN); vop::add(diV,sCntV,dst.dimN); vop::sub(siV,1,src.dimN); vop::sub(diV,1,dst.dimN); //verify the ending address assert( is_legal_address(dst, diV) ); assert( is_legal_address(src, siV) ); vop::copy(siV,sIdxV,src.dimN); vop::copy(diV,dIdxV,dst.dimN); #endif // copy one element ele(dst, diV ) = ele(src, siV ); for(int j=0; j >= 0; ) { // increment the src and dst addr // from highest to lowest degree for(j=src.dimN-1; j>=0; --j) { // if incrementing the jth dim does not overflow ... if( ++nV[j] < sCntV[j] ) { siV[j] += 1; diV[j] += 1; // copy one element ele(dst, diV ) = ele(src, siV ); break; // .. then incr siV[] and diV[] with the next src/dst address } // otherwise reset the counter and address for this dim and backup by one dim. nV[j] = 0; diV[j] = dIdxV[j]; siV[j] = sIdxV[j]; } } return rc; } template< typename T > rc_t _join_update_dims( unsigned index, unsigned* dimV, unsigned dimN, const struct mtx_str& m ) { rc_t rc = kOkRC; // verify that the degree of all matrices are the same if( m.dimN != dimN ) return cwLogError(kInvalidArgRC,"Join matrix size mismatch. dimN:%i != %i", m.dimN, dimN); // only the dimension specified by 'index' may be different for(unsigned i=0; i void _join_copy( struct mtx_str& dst, const struct mtx_str& src, unsigned index, unsigned ii ) { unsigned dIdxV[ dst.dimN ]; unsigned sIdxV[ src.dimN ]; vop::zero(dIdxV,dst.dimN); vop::zero(sIdxV,src.dimN); dIdxV[ index ] = ii; copy(dst,dIdxV,src,sIdxV,src.dimV); } template< typename T > struct mtx_str* _join( unsigned index, unsigned* dimV, unsigned dimN, unsigned ii ) { // Allocate an empty matrix to copy the joined matrices into. return alloc(dimV,dimN); } template< typename T, typename... ARGS> struct mtx_str* _join( unsigned index, unsigned* dimV, unsigned dimN, unsigned ii, const struct mtx_str& m, ARGS&&... args) { struct mtx_str* y = nullptr; if( _join_update_dims(index,dimV,dimN,m) != kOkRC ) return nullptr; if((y = _join( index, dimV, dimN, ii + m.dimV[index], std::forward(args)...)) != nullptr ) { _join_copy(*y,m,index,ii); } return y; } template< typename T, typename... ARGS> struct mtx_str* join( unsigned index, const struct mtx_str& m, ARGS&&... args) { struct mtx_str* y = nullptr; unsigned dimV[ m.dimN ]; for(unsigned i=0; i(args)...)) != nullptr ) { _join_copy(*y,m,index,0); } return y; } template< typename T > bool is_legal_address( const struct mtx_str& m, const unsigned* idxV ) { for(unsigned i=0; i= m.dimV[i] ) return false; return true; } template< typename T > unsigned offset( const struct mtx_str& m, const unsigned* idxV ) { return vop::mac(idxV,m.mulV,m.dimN); } template< typename T > unsigned _offset( const struct mtx_str& m, int i, unsigned offs ) { return offs; } template< typename T, typename... ARGS> unsigned _offset( const struct mtx_str& m, int i, unsigned offs, unsigned idx, ARGS&&... args) { return _offset(m,i+1, offs + idx*m.mulV[i], std::forward(args)...); } template< typename T, typename... ARGS> unsigned offset( const struct mtx_str& m, unsigned idx, ARGS&&... args) { return _offset(m,0,0,idx,std::forward(args)...); } template< typename T > T* addr( struct mtx_str& m, const unsigned* idxV ) { return m.base + offset(m,idxV); } template< typename T > const T* addr( const struct mtx_str& m, const unsigned* idxV ) { return m.base + offset(m,idxV); } template< typename T, typename... ARGS> T* addr( struct mtx_str& m, unsigned i, ARGS&&... args) { return m.base + offset(m,i,std::forward(args)...); } template< typename T, typename... ARGS> const T* addr( const struct mtx_str& m, unsigned i, ARGS&&... args) { return m.base + offset(m,i,std::forward(args)...); } template< typename T > T& ele( struct mtx_str& m, const unsigned* idxV ) { return *addr(m,idxV); } template< typename T > const T& ele( const struct mtx_str& m, const unsigned* idxV ) { return *addr(m,idxV); } template< typename T, typename... ARGS> T& ele( struct mtx_str& m, unsigned i, ARGS&&... args) { return *addr(m,i,std::forward(args)...); } template< typename T, typename... ARGS> const T& ele( const struct mtx_str& m, unsigned i, ARGS&&... args) { return *addr(m,i,std::forward(args)...); } template< typename T > bool is_col_vector( const struct mtx_str& m ) { return m.dimN==1 || (m.dimN==2 && m.dimV[1]==1); }; template< typename T > bool is_row_vector( const struct mtx_str& m ) { return m.dimN==2 && m.dimV[0]==1; } template< typename T > bool is_vector( const struct mtx_str& m ) { return is_col_vector(m) || is_row_vector(m); } // Return 'true' if the matrices have the same size. template< typename T > bool is_size_equal( const struct mtx_str& x0, const struct mtx_str& x1 ) { if( x0.dimN != x1.dimN ) return false; return vop::is_equal(x0.dimV,x1.dimV,x0.dimN); } template< typename T > bool is_equal( const struct mtx_str& x0, const struct mtx_str& x1 ) { if( !is_size_equal(x0,x1) ) return false; return vop::is_equal(x0.base,x1.base,ele_count(x0)); } // Return the count of elements in the matrix template< typename T > unsigned ele_count( const struct mtx_str& x ) { return vop::prod(x.dimV,x.dimN); } template< typename T > void transpose( struct mtx_str& m ) { for(unsigned i=0; i void _print( const struct mtx_str& m, unsigned* idxV, unsigned i, unsigned decPl, unsigned colWidth ) { if( i == m.dimN ) { double v = ele( m, idxV ); // print the value cwLogPrint("%*.*f ",colWidth,decPl,v); } else { for(unsigned j=0; j=2 && i == m.dimN-2 ) { // print the dimension index for matrices with 3+ dim's if( i > 0 && j == 0 ) cwLogPrint("%i\n",idxV[i-1]); // print the row index for matrices with 2+ dim's if( m.dimN>1 ) cwLogPrint("%i | ",j); } idxV[i] = j; _print(m, idxV, i+1, decPl, colWidth ); } // prevent multiple newlines on last printed line if( m.dimN==1 || (m.dimN>=2 && i > m.dimN-2) ) cwLogPrint("\n"); } } template< typename T > void print( const struct mtx_str& m, unsigned decPl=3, unsigned colWidth=10 ) { unsigned idxV[ m.dimN ]; memset(idxV,0,sizeof(idxV)); if( std::numeric_limits::is_integer ) decPl = 0; _print( m, idxV, 0, decPl, colWidth ); } template< typename T > void report( const struct mtx_str& m, const char* label, unsigned decPl=3, unsigned colWidth=10 ) { cwLogPrint("%s :",label); for(unsigned i=0; i void mult( struct mtx_str& y, const struct mtx_str& x0, const struct mtx_str& x1 ) { assert( is_size_equal(x0,x1) ); resize(&y,x0); // resize y to the same dim's as m unsigned n = ele_count(x0); for(unsigned i=0; i void mult( struct mtx_str& y, const struct mtx_str& x ) { assert( is_size_equal(y,x) ); vop::mul(y.base,x.base,ele_count(x)); } // y = x * scalar (elementwise) template< typename T > void mult( struct mtx_str& y, const struct mtx_str& x, const T& scalar ) { resize(&y,x); // resize y to the same dim's as m vop::mul(y.base,x.base,ele_count(x),scalar); } // y *= scalar (elementwise) template< typename T > void mult( struct mtx_str& y, const T& scalar ) { vop::mul(y.base,scalar,ele_count(y)); } // y = m + x (elementwise) template< typename T > void add( struct mtx_str& y, const struct mtx_str& x0, const struct mtx_str& x1 ) { assert( is_size_equal(x0,x1) ); resize(&y,x0); // resize y to the same dim's as m vop::add(y.base,x0.base,x1.base,ele_count(x0)); } // y += x (elementwise) template< typename T > void add( struct mtx_str& y, const struct mtx_str& x ) { assert( is_size_equal(y,x) ); vop::add(y.base,x.base,ele_count(x)); } // y = x + scalar (elementwise) template< typename T > void add( struct mtx_str& y, const struct mtx_str& x, const T& scalar ) { resize(&y,x); vop::add(y.base,x.base,scalar,ele_count(y)); } // y += scalar (elementwise) template< typename T > void add( struct mtx_str& y, const T& scalar ) { vop::add(y.base,scalar,ele_count(y)); } template const T max( const struct mtx_str& x ) { return vop::max(x.base,ele_count(x)); } template const T min( const struct mtx_str& x ) { return vop::min(x.base,ele_count(x)); } template< typename T> struct mtx_str* alloc_one_hot( const struct mtx_str& mV ) { if( !is_vector(mV) ) { cwLogError(kInvalidArgRC,"Only vectors can be converted to one-hot matrices."); return nullptr; } int min_val = (int)mtx::min(mV); int max_val = (int)mtx::max(mV); unsigned rN = (max_val - min_val) + 1; unsigned cN = ele_count(mV); struct mtx::mtx_str* zM = mtx::alloc(kZeroFl,rN,cN); for(unsigned i=0; i(mV,i) - min_val; ele(*zM, j, i) = 1; } return zM; } template< typename T0, typename T1 > rc_t mtx_mul( struct mtx_str& y, const struct mtx_str& m, const struct mtx_str& x ) { assert( x.dimN >= 1 && m.dimN >= 1 ); unsigned xrn = x.dimN==1 ? ele_count(x) : x.dimV[0]; unsigned xcn = x.dimN==1 ? 1 : x.dimV[1]; unsigned mrn = m.dimN==1 ? 1 : m.dimV[0]; unsigned mcn = m.dimN==1 ? ele_count(m) : m.dimV[1]; unsigned yDimV[] = { mrn, xcn }; if( mcn != xrn ) return cwLogError(kInvalidArgRC, "Mtx mult. failed. Size mismatch: m[%i,%i] x[%i,%i].",mrn,mcn,xrn,xcn); //cwLogPrint("%i %i : %i %i\n",mrn,mcn,xrn,xcn); resize(&y,yDimV, 2 ); // go across the columns of x for(unsigned i=0; i f_t; typedef struct mtx_str d_t; rc_t test( const test::test_args_t& args ); } } #endif