nmatrix 0.0.8 → 0.0.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/.gitignore +3 -8
- data/.rspec +1 -1
- data/.travis.yml +12 -0
- data/CONTRIBUTING.md +27 -12
- data/Gemfile +1 -0
- data/History.txt +38 -0
- data/Manifest.txt +15 -15
- data/README.rdoc +7 -6
- data/Rakefile +40 -5
- data/ext/nmatrix/data/data.cpp +2 -37
- data/ext/nmatrix/data/data.h +19 -121
- data/ext/nmatrix/data/meta.h +70 -0
- data/ext/nmatrix/extconf.rb +40 -12
- data/ext/nmatrix/math/math.h +13 -103
- data/ext/nmatrix/nmatrix.cpp +10 -2018
- data/ext/nmatrix/nmatrix.h +16 -13
- data/ext/nmatrix/ruby_constants.cpp +12 -1
- data/ext/nmatrix/ruby_constants.h +7 -1
- data/ext/nmatrix/ruby_nmatrix.c +2169 -0
- data/ext/nmatrix/storage/dense.cpp +123 -14
- data/ext/nmatrix/storage/dense.h +10 -4
- data/ext/nmatrix/storage/list.cpp +265 -48
- data/ext/nmatrix/storage/list.h +6 -9
- data/ext/nmatrix/storage/storage.cpp +44 -54
- data/ext/nmatrix/storage/storage.h +2 -2
- data/ext/nmatrix/storage/yale/class.h +1070 -0
- data/ext/nmatrix/storage/yale/iterators/base.h +142 -0
- data/ext/nmatrix/storage/yale/iterators/iterator.h +130 -0
- data/ext/nmatrix/storage/yale/iterators/row.h +449 -0
- data/ext/nmatrix/storage/yale/iterators/row_stored.h +139 -0
- data/ext/nmatrix/storage/yale/iterators/row_stored_nd.h +167 -0
- data/ext/nmatrix/storage/yale/iterators/stored_diagonal.h +123 -0
- data/ext/nmatrix/storage/yale/math/transpose.h +110 -0
- data/ext/nmatrix/storage/yale/yale.cpp +1785 -0
- data/ext/nmatrix/storage/{yale.h → yale/yale.h} +23 -55
- data/ext/nmatrix/types.h +2 -0
- data/ext/nmatrix/util/io.cpp +27 -45
- data/ext/nmatrix/util/io.h +0 -2
- data/ext/nmatrix/util/sl_list.cpp +169 -28
- data/ext/nmatrix/util/sl_list.h +9 -3
- data/lib/nmatrix/blas.rb +20 -20
- data/lib/nmatrix/enumerate.rb +1 -1
- data/lib/nmatrix/io/mat5_reader.rb +8 -14
- data/lib/nmatrix/lapack.rb +3 -3
- data/lib/nmatrix/math.rb +3 -3
- data/lib/nmatrix/nmatrix.rb +19 -5
- data/lib/nmatrix/nvector.rb +2 -0
- data/lib/nmatrix/shortcuts.rb +90 -125
- data/lib/nmatrix/version.rb +1 -1
- data/nmatrix.gemspec +7 -8
- data/spec/{nmatrix_spec.rb → 00_nmatrix_spec.rb} +45 -208
- data/spec/01_enum_spec.rb +184 -0
- data/spec/{slice_spec.rb → 02_slice_spec.rb} +55 -39
- data/spec/blas_spec.rb +22 -54
- data/spec/elementwise_spec.rb +9 -8
- data/spec/io_spec.rb +6 -4
- data/spec/lapack_spec.rb +26 -26
- data/spec/math_spec.rb +9 -5
- data/spec/nmatrix_yale_spec.rb +29 -61
- data/spec/shortcuts_spec.rb +34 -22
- data/spec/slice_set_spec.rb +157 -0
- data/spec/spec_helper.rb +42 -2
- data/spec/stat_spec.rb +192 -0
- metadata +52 -55
- data/ext/nmatrix/storage/yale.cpp +0 -2284
- data/spec/nmatrix_list_spec.rb +0 -113
- data/spec/nvector_spec.rb +0 -112
@@ -0,0 +1,142 @@
|
|
1
|
+
/////////////////////////////////////////////////////////////////////
|
2
|
+
// = NMatrix
|
3
|
+
//
|
4
|
+
// A linear algebra library for scientific computation in Ruby.
|
5
|
+
// NMatrix is part of SciRuby.
|
6
|
+
//
|
7
|
+
// NMatrix was originally inspired by and derived from NArray, by
|
8
|
+
// Masahiro Tanaka: http://narray.rubyforge.org
|
9
|
+
//
|
10
|
+
// == Copyright Information
|
11
|
+
//
|
12
|
+
// SciRuby is Copyright (c) 2010 - 2013, Ruby Science Foundation
|
13
|
+
// NMatrix is Copyright (c) 2013, Ruby Science Foundation
|
14
|
+
//
|
15
|
+
// Please see LICENSE.txt for additional copyright notices.
|
16
|
+
//
|
17
|
+
// == Contributing
|
18
|
+
//
|
19
|
+
// By contributing source code to SciRuby, you agree to be bound by
|
20
|
+
// our Contributor Agreement:
|
21
|
+
//
|
22
|
+
// * https://github.com/SciRuby/sciruby/wiki/Contributor-Agreement
|
23
|
+
//
|
24
|
+
// == base.h
|
25
|
+
//
|
26
|
+
// Yale storage pure virtual basic_iterator class.
|
27
|
+
//
|
28
|
+
|
29
|
+
#ifndef YALE_ITERATORS_BASE_H
|
30
|
+
# define YALE_ITERATORS_BASE_H
|
31
|
+
|
32
|
+
#include <type_traits>
|
33
|
+
#include <typeinfo>
|
34
|
+
#include <stdexcept>
|
35
|
+
|
36
|
+
namespace nm {
|
37
|
+
|
38
|
+
template <typename D> class YaleStorage;
|
39
|
+
|
40
|
+
namespace yale_storage {
|
41
|
+
|
42
|
+
template <typename D>
|
43
|
+
VALUE nm_rb_dereference(D const& v) {
|
44
|
+
return nm::RubyObject(v).rval;
|
45
|
+
}
|
46
|
+
|
47
|
+
template <>
|
48
|
+
VALUE nm_rb_dereference<nm::RubyObject>(nm::RubyObject const& v) {
|
49
|
+
return v.rval;
|
50
|
+
}
|
51
|
+
|
52
|
+
/*
|
53
|
+
* Iterator base class (pure virtual).
|
54
|
+
*/
|
55
|
+
template <typename D,
|
56
|
+
typename RefType,
|
57
|
+
typename YaleRef = typename std::conditional<
|
58
|
+
std::is_const<RefType>::value,
|
59
|
+
const nm::YaleStorage<D>,
|
60
|
+
nm::YaleStorage<D>
|
61
|
+
>::type>
|
62
|
+
class basic_iterator_T {
|
63
|
+
|
64
|
+
protected:
|
65
|
+
YaleRef& y;
|
66
|
+
size_t i_;
|
67
|
+
size_t p_;
|
68
|
+
|
69
|
+
public:
|
70
|
+
size_t offset(size_t d) const { return y.offset(d); }
|
71
|
+
size_t shape(size_t d) const { return y.shape(d); }
|
72
|
+
size_t real_shape(size_t d) const { return y.real_shape(d); }
|
73
|
+
|
74
|
+
size_t dense_location() const {
|
75
|
+
return i()*shape(1) + j();
|
76
|
+
}
|
77
|
+
|
78
|
+
template <typename T = typename std::conditional<std::is_const<RefType>::value, const size_t, size_t>::type>
|
79
|
+
T& ija(size_t pp) const { return y.ija(pp); }
|
80
|
+
|
81
|
+
template <typename T = typename std::conditional<std::is_const<RefType>::value, const size_t, size_t>::type>
|
82
|
+
T& ija(size_t pp) { return y.ija(pp); }
|
83
|
+
|
84
|
+
virtual bool diag() const {
|
85
|
+
return p_ < std::min(y.real_shape(0), y.real_shape(1));
|
86
|
+
}
|
87
|
+
virtual bool done_with_diag() const {
|
88
|
+
return p_ == std::min(y.real_shape(0), y.real_shape(1));
|
89
|
+
}
|
90
|
+
virtual bool nondiag() const {
|
91
|
+
return p_ > std::min(y.real_shape(0), y.real_shape(1));
|
92
|
+
}
|
93
|
+
|
94
|
+
basic_iterator_T(YaleRef& obj, size_t ii = 0, size_t pp = 0) : y(obj), i_(ii), p_(pp) { }
|
95
|
+
|
96
|
+
basic_iterator_T<D,RefType,YaleRef>& operator=(const basic_iterator_T<D,RefType,YaleRef>& rhs) {
|
97
|
+
if (&y != &(rhs.y)) throw std::logic_error("can only be used on iterators with the same matrix");
|
98
|
+
i_ = rhs.i_;
|
99
|
+
p_ = rhs.p_;
|
100
|
+
return *this;
|
101
|
+
}
|
102
|
+
|
103
|
+
virtual inline size_t i() const { return i_; }
|
104
|
+
virtual size_t j() const = 0;
|
105
|
+
|
106
|
+
virtual inline VALUE rb_i() const { return LONG2NUM(i()); }
|
107
|
+
virtual inline VALUE rb_j() const { return LONG2NUM(j()); }
|
108
|
+
|
109
|
+
virtual size_t real_i() const { return offset(0) + i(); }
|
110
|
+
virtual size_t real_j() const { return offset(1) + j(); }
|
111
|
+
virtual size_t p() const { return p_; }
|
112
|
+
virtual bool real_ndnz_exists() const { return !y.real_row_empty(real_i()) && ija(p_) == real_j(); }
|
113
|
+
|
114
|
+
virtual RefType& operator*() = 0;
|
115
|
+
virtual RefType& operator*() const = 0;
|
116
|
+
|
117
|
+
|
118
|
+
// Ruby VALUE de-reference
|
119
|
+
inline VALUE operator~() const {
|
120
|
+
return nm_rb_dereference<D>(**this);
|
121
|
+
//virtual VALUE operator~() const {
|
122
|
+
// if (typeid(D) == typeid(RubyObject)) return (**this); // FIXME: return rval instead, faster;
|
123
|
+
// else return RubyObject(*(*this)).rval;
|
124
|
+
}
|
125
|
+
|
126
|
+
virtual bool operator==(const std::pair<size_t,size_t>& ij) {
|
127
|
+
if (p() >= ija(real_shape(0))) return false;
|
128
|
+
else return i() == ij.first && j() == ij.second;
|
129
|
+
}
|
130
|
+
|
131
|
+
virtual bool operator==(const basic_iterator_T<D,RefType,YaleRef>& rhs) const {
|
132
|
+
return i() == rhs.i() && j() == rhs.j();
|
133
|
+
}
|
134
|
+
virtual bool operator!=(const basic_iterator_T<D,RefType,YaleRef>& rhs) const {
|
135
|
+
return i() != rhs.i() || j() != rhs.j();
|
136
|
+
}
|
137
|
+
};
|
138
|
+
|
139
|
+
|
140
|
+
} } // end of namespace nm::yale_storage
|
141
|
+
|
142
|
+
#endif // YALE_ITERATORS_BASE_H
|
@@ -0,0 +1,130 @@
|
|
1
|
+
/////////////////////////////////////////////////////////////////////
|
2
|
+
// = NMatrix
|
3
|
+
//
|
4
|
+
// A linear algebra library for scientific computation in Ruby.
|
5
|
+
// NMatrix is part of SciRuby.
|
6
|
+
//
|
7
|
+
// NMatrix was originally inspired by and derived from NArray, by
|
8
|
+
// Masahiro Tanaka: http://narray.rubyforge.org
|
9
|
+
//
|
10
|
+
// == Copyright Information
|
11
|
+
//
|
12
|
+
// SciRuby is Copyright (c) 2010 - 2013, Ruby Science Foundation
|
13
|
+
// NMatrix is Copyright (c) 2013, Ruby Science Foundation
|
14
|
+
//
|
15
|
+
// Please see LICENSE.txt for additional copyright notices.
|
16
|
+
//
|
17
|
+
// == Contributing
|
18
|
+
//
|
19
|
+
// By contributing source code to SciRuby, you agree to be bound by
|
20
|
+
// our Contributor Agreement:
|
21
|
+
//
|
22
|
+
// * https://github.com/SciRuby/sciruby/wiki/Contributor-Agreement
|
23
|
+
//
|
24
|
+
// == iterator.h
|
25
|
+
//
|
26
|
+
// Iterate over yale as if dense
|
27
|
+
//
|
28
|
+
|
29
|
+
#ifndef YALE_ITERATORS_ITERATOR_H
|
30
|
+
# define YALE_ITERATORS_ITERATOR_H
|
31
|
+
|
32
|
+
#include <type_traits>
|
33
|
+
#include <typeinfo>
|
34
|
+
|
35
|
+
namespace nm { namespace yale_storage {
|
36
|
+
|
37
|
+
/*
|
38
|
+
* Iterator for traversing matrix class as if it were dense (visits each entry in order).
|
39
|
+
*/
|
40
|
+
template <typename D,
|
41
|
+
typename RefType,
|
42
|
+
typename YaleRef = typename std::conditional<
|
43
|
+
std::is_const<RefType>::value,
|
44
|
+
const nm::YaleStorage<D>,
|
45
|
+
nm::YaleStorage<D>
|
46
|
+
>::type>
|
47
|
+
class iterator_T : public basic_iterator_T<D,RefType,YaleRef> {
|
48
|
+
using basic_iterator_T<D,RefType,YaleRef>::i_;
|
49
|
+
using basic_iterator_T<D,RefType,YaleRef>::p_;
|
50
|
+
using basic_iterator_T<D,RefType,YaleRef>::y;
|
51
|
+
using basic_iterator_T<D,RefType,YaleRef>::offset;
|
52
|
+
using basic_iterator_T<D,RefType,YaleRef>::shape;
|
53
|
+
using basic_iterator_T<D,RefType,YaleRef>::ija;
|
54
|
+
|
55
|
+
protected:
|
56
|
+
size_t j_; // These are relative to the slice.
|
57
|
+
|
58
|
+
public:
|
59
|
+
// Create an iterator. May select the row since this is O(1).
|
60
|
+
iterator_T(YaleRef& obj, size_t ii = 0)
|
61
|
+
: basic_iterator_T<D,RefType,YaleRef>(obj, ii, obj.ija(ii + obj.offset(0))), j_(0)
|
62
|
+
{
|
63
|
+
// advance to the beginning of the row
|
64
|
+
if (obj.offset(1) > 0)
|
65
|
+
p_ = y.find_pos_for_insertion(i_,j_);
|
66
|
+
}
|
67
|
+
|
68
|
+
// Prefix ++
|
69
|
+
iterator_T<D,RefType,YaleRef>& operator++() {
|
70
|
+
size_t prev_j = j_++;
|
71
|
+
if (j_ >= shape(1)) {
|
72
|
+
j_ = 0;
|
73
|
+
++i_;
|
74
|
+
|
75
|
+
// Do a binary search to find the beginning of the slice
|
76
|
+
p_ = offset(0) > 0 ? y.find_pos_for_insertion(i_,j_) : ija(i_);
|
77
|
+
} else {
|
78
|
+
// If the last j was actually stored in this row of the matrix, need to advance p.
|
79
|
+
|
80
|
+
if (!y.real_row_empty(i_ + offset(0)) && ija(p_) <= prev_j + offset(1)) ++p_; // this test is the same as real_ndnz_exists
|
81
|
+
}
|
82
|
+
|
83
|
+
return *this;
|
84
|
+
}
|
85
|
+
|
86
|
+
iterator_T<D,RefType,YaleRef> operator++(int dummy) const {
|
87
|
+
iterator_T<D,RefType,YaleRef> iter(*this);
|
88
|
+
return ++iter;
|
89
|
+
}
|
90
|
+
|
91
|
+
virtual bool operator!=(const iterator_T<D,RefType,YaleRef>& rhs) const {
|
92
|
+
return this->dense_location() != rhs.dense_location();
|
93
|
+
}
|
94
|
+
|
95
|
+
virtual bool operator==(const iterator_T<D,RefType,YaleRef>& rhs) const {
|
96
|
+
return this->dense_location() == rhs.dense_location();
|
97
|
+
}
|
98
|
+
|
99
|
+
bool operator<(const iterator_T<D,RefType,YaleRef>& rhs) const {
|
100
|
+
return this->dense_location() < rhs.dense_location();
|
101
|
+
}
|
102
|
+
|
103
|
+
bool operator>(const iterator_T<D,RefType,YaleRef>& rhs) const {
|
104
|
+
return this->dense_location() > rhs.dense_location();
|
105
|
+
}
|
106
|
+
|
107
|
+
virtual bool diag() const { return i_ + offset(0) == j_ + offset(1); }
|
108
|
+
|
109
|
+
// De-reference
|
110
|
+
RefType& operator*() {
|
111
|
+
if (diag()) return y.a( i_ + offset(0) );
|
112
|
+
else if (p_ >= ija(i_+offset(0)+1)) return y.const_default_obj();
|
113
|
+
else if (!y.real_row_empty(i_ + offset(0)) && ija(p_) == j_ + offset(1)) return y.a( p_ );
|
114
|
+
else return y.const_default_obj();
|
115
|
+
}
|
116
|
+
|
117
|
+
RefType& operator*() const {
|
118
|
+
if (diag()) return y.a( i_ + offset(0) );
|
119
|
+
else if (p_ >= ija(i_+offset(0)+1)) return y.const_default_obj();
|
120
|
+
else if (!y.real_row_empty(i_ + offset(0)) && ija(p_) == j_ + offset(1)) return y.a( p_ );
|
121
|
+
else return y.const_default_obj();
|
122
|
+
}
|
123
|
+
|
124
|
+
virtual size_t j() const { return j_; }
|
125
|
+
};
|
126
|
+
|
127
|
+
|
128
|
+
} } // end of namespace nm::yale_storage
|
129
|
+
|
130
|
+
#endif // YALE_ITERATORS_ITERATOR_H
|
@@ -0,0 +1,449 @@
|
|
1
|
+
/////////////////////////////////////////////////////////////////////
|
2
|
+
// = NMatrix
|
3
|
+
//
|
4
|
+
// A linear algebra library for scientific computation in Ruby.
|
5
|
+
// NMatrix is part of SciRuby.
|
6
|
+
//
|
7
|
+
// NMatrix was originally inspired by and derived from NArray, by
|
8
|
+
// Masahiro Tanaka: http://narray.rubyforge.org
|
9
|
+
//
|
10
|
+
// == Copyright Information
|
11
|
+
//
|
12
|
+
// SciRuby is Copyright (c) 2010 - 2013, Ruby Science Foundation
|
13
|
+
// NMatrix is Copyright (c) 2013, Ruby Science Foundation
|
14
|
+
//
|
15
|
+
// Please see LICENSE.txt for additional copyright notices.
|
16
|
+
//
|
17
|
+
// == Contributing
|
18
|
+
//
|
19
|
+
// By contributing source code to SciRuby, you agree to be bound by
|
20
|
+
// our Contributor Agreement:
|
21
|
+
//
|
22
|
+
// * https://github.com/SciRuby/sciruby/wiki/Contributor-Agreement
|
23
|
+
//
|
24
|
+
// == row.h
|
25
|
+
//
|
26
|
+
// Iterator for traversing a matrix row by row. Includes an
|
27
|
+
// orthogonal iterator for visiting each stored entry in a row.
|
28
|
+
// This one cannot be de-referenced; you have to de-reference
|
29
|
+
// the column.
|
30
|
+
|
31
|
+
#ifndef YALE_ITERATORS_ROW_H
|
32
|
+
# define YALE_ITERATORS_ROW_H
|
33
|
+
|
34
|
+
#include <stdexcept>
|
35
|
+
|
36
|
+
namespace nm { namespace yale_storage {
|
37
|
+
|
38
|
+
template <typename D,
|
39
|
+
typename RefType,
|
40
|
+
typename YaleRef = typename std::conditional<
|
41
|
+
std::is_const<RefType>::value,
|
42
|
+
const nm::YaleStorage<D>,
|
43
|
+
nm::YaleStorage<D>
|
44
|
+
>::type>
|
45
|
+
class row_iterator_T {
|
46
|
+
|
47
|
+
protected:
|
48
|
+
YaleRef& y;
|
49
|
+
size_t i_;
|
50
|
+
size_t p_first, p_last; // first and last IJA positions in the row
|
51
|
+
|
52
|
+
|
53
|
+
/*
|
54
|
+
* Update the row positions -- use to ensure a row stays valid after an insert operation. Also
|
55
|
+
* used to initialize a row iterator at a different row index.
|
56
|
+
*/
|
57
|
+
void update() {
|
58
|
+
if (i_ < y.shape(0)) {
|
59
|
+
p_first = p_real_first();
|
60
|
+
p_last = p_real_last();
|
61
|
+
if (!nd_empty()) {
|
62
|
+
// try to find new p_first
|
63
|
+
p_first = y.real_find_left_boundary_pos(p_first, p_last, y.offset(1));
|
64
|
+
if (!nd_empty()) {
|
65
|
+
// also try to find new p_last
|
66
|
+
p_last = y.real_find_left_boundary_pos(p_first, p_last, y.offset(1) + y.shape(1) - 1);
|
67
|
+
if (y.ija(p_last) - y.offset(1) >= shape(1)) --p_last; // searched too far.
|
68
|
+
}
|
69
|
+
}
|
70
|
+
} else { // invalid row -- this is an end iterator.
|
71
|
+
p_first = y.ija(y.real_shape(0));
|
72
|
+
p_last = y.ija(y.real_shape(0))-1; // mark as empty
|
73
|
+
}
|
74
|
+
}
|
75
|
+
|
76
|
+
/*
|
77
|
+
* Indicate to the row iterator that p_first and p_last have moved by some amount. Only
|
78
|
+
* defined for row_iterator, not const_row_iterator. This is a lightweight form of update().
|
79
|
+
*/
|
80
|
+
//template <typename = typename std::enable_if<!std::is_const<RefType>::value>::type>
|
81
|
+
void shift(int amount) {
|
82
|
+
p_first += amount;
|
83
|
+
p_last += amount;
|
84
|
+
}
|
85
|
+
|
86
|
+
|
87
|
+
/*
|
88
|
+
* Enlarge the row by amount by moving p_last over. This is a lightweight form of update().
|
89
|
+
*/
|
90
|
+
//template <typename = typename std::enable_if<!std::is_const<RefType>::value>::type>
|
91
|
+
void adjust_length(int amount) {
|
92
|
+
p_last += amount;
|
93
|
+
}
|
94
|
+
|
95
|
+
public:
|
96
|
+
/* typedef row_stored_iterator_T<D,RefType,YaleRef> row_stored_iterator;
|
97
|
+
typedef row_stored_nd_iterator_T<D,RefType,YaleRef> row_stored_nd_iterator;
|
98
|
+
typedef row_stored_iterator_T<D,const RefType,const YaleRef> const_row_stored_iterator;
|
99
|
+
typedef row_stored_nd_iterator_T<D,const RefType,const YaleRef> const_row_stored_nd_iterator;*/
|
100
|
+
typedef row_stored_iterator_T<D,RefType,YaleRef, row_iterator_T<D,RefType,YaleRef> > row_stored_iterator;
|
101
|
+
typedef row_stored_nd_iterator_T<D,RefType,YaleRef, row_iterator_T<D,RefType,YaleRef> > row_stored_nd_iterator;
|
102
|
+
template <typename E, typename ERefType, typename EYaleRef> friend class row_iterator_T;
|
103
|
+
friend class row_stored_iterator_T<D,RefType,YaleRef, row_iterator_T<D,RefType,YaleRef> >;
|
104
|
+
friend class row_stored_nd_iterator_T<D,RefType,YaleRef, row_iterator_T<D,RefType,YaleRef> >;//row_stored_iterator;
|
105
|
+
friend class row_stored_iterator_T<D,RefType,YaleRef, const row_iterator_T<D,RefType,YaleRef> >;
|
106
|
+
friend class row_stored_nd_iterator_T<D,RefType,YaleRef, const row_iterator_T<D,RefType,YaleRef> >;//row_stored_iterator;
|
107
|
+
friend class nm::YaleStorage<D>;
|
108
|
+
|
109
|
+
//friend row_stored_nd_iterator;
|
110
|
+
|
111
|
+
inline size_t ija(size_t pp) const { return y.ija(pp); }
|
112
|
+
inline size_t& ija(size_t pp) { return y.ija(pp); }
|
113
|
+
inline RefType& a(size_t p) const { return y.a_p()[p]; }
|
114
|
+
inline RefType& a(size_t p) { return y.a_p()[p]; }
|
115
|
+
|
116
|
+
|
117
|
+
|
118
|
+
row_iterator_T(YaleRef& obj, size_t ii = 0)
|
119
|
+
: y(obj), i_(ii)
|
120
|
+
{
|
121
|
+
update();
|
122
|
+
}
|
123
|
+
|
124
|
+
|
125
|
+
template <typename E, typename ERefType = typename std::conditional<std::is_const<RefType>::value, const E, E>::type>
|
126
|
+
bool operator!=(const row_iterator_T<E,ERefType>& rhs) const {
|
127
|
+
return i_ != rhs.i_;
|
128
|
+
}
|
129
|
+
|
130
|
+
template <typename E, typename ERefType = typename std::conditional<std::is_const<RefType>::value, const E, E>::type>
|
131
|
+
bool operator==(const row_iterator_T<E,ERefType>& rhs) const {
|
132
|
+
return i_ == rhs.i_;
|
133
|
+
}
|
134
|
+
|
135
|
+
template <typename E, typename ERefType = typename std::conditional<std::is_const<RefType>::value, const E, E>::type>
|
136
|
+
bool operator<(const row_iterator_T<E,ERefType>& rhs) const {
|
137
|
+
return i_ < rhs.i_;
|
138
|
+
}
|
139
|
+
|
140
|
+
template <typename E, typename ERefType = typename std::conditional<std::is_const<RefType>::value, const E, E>::type>
|
141
|
+
bool operator>(const row_iterator_T<E,ERefType>& rhs) const {
|
142
|
+
return i_ > rhs.i_;
|
143
|
+
}
|
144
|
+
|
145
|
+
row_iterator_T<D,RefType,YaleRef>& operator++() {
|
146
|
+
if (is_end()) throw std::out_of_range("attempted to iterate past end of slice (vertically)");
|
147
|
+
++i_;
|
148
|
+
update();
|
149
|
+
return *this;
|
150
|
+
}
|
151
|
+
|
152
|
+
row_iterator_T<D,RefType,YaleRef> operator++(int dummy) const {
|
153
|
+
row_iterator_T<D,RefType,YaleRef> next(*this);
|
154
|
+
return ++next;
|
155
|
+
}
|
156
|
+
|
157
|
+
bool is_end() const {
|
158
|
+
return i_ == y.shape(0) && p_first == y.ija(y.real_shape(0));
|
159
|
+
}
|
160
|
+
|
161
|
+
size_t real_i() const {
|
162
|
+
return i_ + y.offset(0);
|
163
|
+
}
|
164
|
+
|
165
|
+
size_t i() const {
|
166
|
+
return i_;
|
167
|
+
}
|
168
|
+
|
169
|
+
// last element of the real row
|
170
|
+
size_t p_real_last() const {
|
171
|
+
return y.ija(real_i()+1)-1;
|
172
|
+
}
|
173
|
+
|
174
|
+
// first element of the real row
|
175
|
+
size_t p_real_first() const {
|
176
|
+
return y.ija(real_i());
|
177
|
+
}
|
178
|
+
|
179
|
+
// Is the real row of the original matrix totally empty of NDs?
|
180
|
+
bool real_nd_empty() const {
|
181
|
+
return p_real_last() < p_real_first();
|
182
|
+
}
|
183
|
+
|
184
|
+
bool nd_empty() const {
|
185
|
+
return p_last < p_first;
|
186
|
+
}
|
187
|
+
|
188
|
+
// slice j coord of the diag.
|
189
|
+
size_t diag_j() const {
|
190
|
+
if (!has_diag())
|
191
|
+
throw std::out_of_range("don't call diag_j unless you've checked for one");
|
192
|
+
return real_i() - y.offset(1);
|
193
|
+
}
|
194
|
+
|
195
|
+
// return the actual position of the diagonal element for this real row, regardless of whether
|
196
|
+
// it's in range or not.
|
197
|
+
size_t p_diag() const {
|
198
|
+
return real_i();
|
199
|
+
}
|
200
|
+
|
201
|
+
// Checks to see if there is a diagonal within the slice
|
202
|
+
bool has_diag() const {
|
203
|
+
// real position of diag is real_i == real_j. Is it in range?
|
204
|
+
return (p_diag() >= y.offset(1) && p_diag() - y.offset(1) < y.shape(1));
|
205
|
+
}
|
206
|
+
|
207
|
+
// Checks to see if the diagonal is the first entry in the slice.
|
208
|
+
bool is_diag_first() const {
|
209
|
+
if (!has_diag()) return false;
|
210
|
+
if (nd_empty()) return true;
|
211
|
+
return diag_j() < y.ija(p_first) - y.offset(1);
|
212
|
+
}
|
213
|
+
|
214
|
+
// Checks to see if the diagonal is the last entry in the slice.
|
215
|
+
bool is_diag_last() const {
|
216
|
+
if (!has_diag()) return false;
|
217
|
+
if (nd_empty()) return true;
|
218
|
+
return diag_j() > y.ija(p_last);
|
219
|
+
}
|
220
|
+
|
221
|
+
// Is the row of the slice totally empty of NDs and Ds?
|
222
|
+
// We can only determine that it's empty of Ds if the diagonal
|
223
|
+
// is not a part of the sliced portion of the row.
|
224
|
+
bool empty() const {
|
225
|
+
return nd_empty() && has_diag();
|
226
|
+
}
|
227
|
+
|
228
|
+
|
229
|
+
size_t shape(size_t pp) const {
|
230
|
+
return y.shape(pp);
|
231
|
+
}
|
232
|
+
|
233
|
+
size_t offset(size_t pp) const {
|
234
|
+
return y.offset(pp);
|
235
|
+
}
|
236
|
+
|
237
|
+
inline VALUE rb_i() const { return LONG2NUM(i()); }
|
238
|
+
|
239
|
+
row_stored_iterator_T<D,RefType,YaleRef> begin() { return row_stored_iterator_T<D,RefType,YaleRef>(*this, p_first); }
|
240
|
+
row_stored_nd_iterator_T<D,RefType,YaleRef> ndbegin() { return row_stored_nd_iterator_T<D,RefType,YaleRef>(*this, p_first); }
|
241
|
+
row_stored_iterator_T<D,RefType,YaleRef> end() { return row_stored_iterator_T<D,RefType,YaleRef>(*this, p_last+1, true); }
|
242
|
+
row_stored_nd_iterator_T<D,RefType,YaleRef> ndend() { return row_stored_nd_iterator_T<D,RefType,YaleRef>(*this, p_last+1); }
|
243
|
+
|
244
|
+
row_stored_iterator_T<D,RefType,YaleRef> begin() const { return row_stored_iterator_T<D,RefType,YaleRef>(*this, p_first); }
|
245
|
+
row_stored_nd_iterator_T<D,RefType,YaleRef> ndbegin() const { return row_stored_nd_iterator_T<D,RefType,YaleRef>(*this, p_first); }
|
246
|
+
row_stored_iterator_T<D,RefType,YaleRef> end() const { return row_stored_iterator_T<D,RefType,YaleRef>(*this, p_last+1, true); }
|
247
|
+
row_stored_nd_iterator_T<D,RefType,YaleRef> ndend() const { return row_stored_nd_iterator_T<D,RefType,YaleRef>(*this, p_last+1); }
|
248
|
+
|
249
|
+
|
250
|
+
row_stored_nd_iterator_T<D,RefType,YaleRef> lower_bound(const size_t& j) const {
|
251
|
+
row_stored_nd_iterator_T<D,RefType,YaleRef>(*this, y.real_find_left_boundary_pos(p_first, p_last, y.offset(1)));
|
252
|
+
}
|
253
|
+
|
254
|
+
row_stored_nd_iterator_T<D,RefType,YaleRef> ndfind(size_t j) {
|
255
|
+
if (j == 0) return ndbegin();
|
256
|
+
size_t p = p_first > p_last ? p_first : y.real_find_left_boundary_pos(p_first, p_last, j + y.offset(1));
|
257
|
+
row_stored_nd_iterator iter = row_stored_nd_iterator_T<D,RefType,YaleRef>(*this, p);
|
258
|
+
return iter;
|
259
|
+
}
|
260
|
+
|
261
|
+
row_stored_iterator_T<D,RefType,YaleRef> find(size_t j) {
|
262
|
+
if (j == 0) return begin(); // may or may not be on the diagonal
|
263
|
+
else return row_stored_iterator_T<D,RefType,YaleRef>(*this, ndfind(j).p(), false); // is on the diagonal, definitely
|
264
|
+
}
|
265
|
+
|
266
|
+
/*
|
267
|
+
* Remove an entry from an already found non-diagonal position. Adjust this row appropriately so we can continue to
|
268
|
+
* use it.
|
269
|
+
*/
|
270
|
+
//template <typename = typename std::enable_if<!std::is_const<RefType>::value>::type>
|
271
|
+
row_stored_nd_iterator erase(row_stored_nd_iterator position) {
|
272
|
+
size_t sz = y.size();
|
273
|
+
if (sz - 1 <= y.capacity() / nm::yale_storage::GROWTH_CONSTANT) {
|
274
|
+
y.update_resize_move(position, real_i(), -1);
|
275
|
+
} else {
|
276
|
+
y.move_left(position, 1);
|
277
|
+
y.update_real_row_sizes_from(real_i(), -1);
|
278
|
+
}
|
279
|
+
adjust_length(-1);
|
280
|
+
return row_stored_nd_iterator(*this, position.p()-1);
|
281
|
+
}
|
282
|
+
|
283
|
+
/*
|
284
|
+
* Remove an entry from the matrix at the already-located position. If diagonal, just sets to default; otherwise,
|
285
|
+
* actually removes the entry.
|
286
|
+
*/
|
287
|
+
//template <typename = typename std::enable_if<!std::is_const<RefType>::value>::type>
|
288
|
+
row_stored_nd_iterator erase(const row_stored_iterator& jt) {
|
289
|
+
if (jt.diag()) {
|
290
|
+
*jt = y.const_default_obj(); // diagonal is the easy case -- no movement.
|
291
|
+
return row_stored_nd_iterator(*this, jt.p());
|
292
|
+
} else {
|
293
|
+
return erase(row_stored_nd_iterator(*this, jt.p()));
|
294
|
+
}
|
295
|
+
}
|
296
|
+
|
297
|
+
|
298
|
+
|
299
|
+
//template <typename = typename std::enable_if<!std::is_const<RefType>::value>::type>
|
300
|
+
row_stored_nd_iterator insert(row_stored_nd_iterator position, size_t jj, const D& val) {
|
301
|
+
size_t sz = y.size();
|
302
|
+
while (!position.end() && position.j() < jj) ++position; // position is just a hint. (This loop ideally only has to happen once.)
|
303
|
+
|
304
|
+
if (!position.end() && position.j() == jj) {
|
305
|
+
*position = val; // replace existing
|
306
|
+
} else {
|
307
|
+
|
308
|
+
if (sz + 1 > y.capacity()) {
|
309
|
+
y.update_resize_move(position, real_i(), 1);
|
310
|
+
} else {
|
311
|
+
y.move_right(position, 1);
|
312
|
+
y.update_real_row_sizes_from(real_i(), 1);
|
313
|
+
}
|
314
|
+
ija(position.p()) = jj + y.offset(1); // set column ID
|
315
|
+
a(position.p()) = val;
|
316
|
+
adjust_length(1);
|
317
|
+
}
|
318
|
+
|
319
|
+
return position++;
|
320
|
+
}
|
321
|
+
|
322
|
+
|
323
|
+
/*
|
324
|
+
* This version of insert doesn't return anything. Why, when the others do?
|
325
|
+
*
|
326
|
+
* Well, mainly because j here can be a diagonal entry. Most of the inserters return the *next* element following
|
327
|
+
* the insertion, but to do that, we have to create a row_stored_nd_iterator, which requires at least one binary
|
328
|
+
* search for the location following the diagonal (and as of the writing of this, two binary searches). There's no
|
329
|
+
* reason to do that when we never actually *use* the return value. So instead we just have void.
|
330
|
+
*/
|
331
|
+
//template <typename = typename std::enable_if<!std::is_const<RefType>::value>::type>
|
332
|
+
void insert(size_t j, const D& val) {
|
333
|
+
if (j + y.offset(1) == real_i()) a(real_i()) = val;
|
334
|
+
else {
|
335
|
+
row_stored_nd_iterator jt = ndfind(j);
|
336
|
+
if (!jt.end() && jt.j() == j) {
|
337
|
+
if (val == y.const_default_obj()) erase(jt); // erase
|
338
|
+
else insert(jt, j, val); // replace
|
339
|
+
} else { // only insert if it's not the default
|
340
|
+
if (val != y.const_default_obj()) insert(jt, j, val);
|
341
|
+
}
|
342
|
+
}
|
343
|
+
}
|
344
|
+
|
345
|
+
|
346
|
+
/*
|
347
|
+
* Determines a plan for inserting a single row. Returns an integer giving the amount of the row change.
|
348
|
+
*/
|
349
|
+
int single_row_insertion_plan(row_stored_nd_iterator position, size_t jj, size_t length, D const* v, size_t v_size, size_t& v_offset) {
|
350
|
+
int nd_change = 0;
|
351
|
+
|
352
|
+
for (size_t jc = jj; jc < jj + length; ++jc, ++v_offset) {
|
353
|
+
if (v_offset >= v_size) v_offset %= v_size; // reset v position.
|
354
|
+
|
355
|
+
if (jc + y.offset(1) != real_i()) { // diagonal -- no nd_change here
|
356
|
+
if (position.end()) {
|
357
|
+
if (v[v_offset] != y.const_default_obj()) nd_change++; // insert
|
358
|
+
} else if (position.j() != jc) { // not present -- do we need to add it?
|
359
|
+
if (v[v_offset] != y.const_default_obj()) nd_change++;
|
360
|
+
} else { // position.j() == jc
|
361
|
+
if (v[v_offset] == y.const_default_obj()) nd_change--;
|
362
|
+
++position; // move iterator forward.
|
363
|
+
}
|
364
|
+
}
|
365
|
+
}
|
366
|
+
return nd_change;
|
367
|
+
}
|
368
|
+
|
369
|
+
/*
|
370
|
+
* Determine a plan for inserting a single row -- finds the position first. Returns the position and
|
371
|
+
* the change amount. Don't use this one if you can help it because it requires a binary search of
|
372
|
+
* the row.
|
373
|
+
*/
|
374
|
+
std::pair<int,size_t> single_row_insertion_plan(size_t jj, size_t length, D const* v, size_t v_size, size_t& v_offset) {
|
375
|
+
std::pair<int,size_t> result;
|
376
|
+
row_stored_nd_iterator pos = ndfind(jj);
|
377
|
+
result.first = single_row_insertion_plan(pos, jj, length, v, v_size, v_offset);
|
378
|
+
result.second = pos.p();
|
379
|
+
return result;
|
380
|
+
}
|
381
|
+
|
382
|
+
/*
|
383
|
+
* Insert elements into a single row. Returns an iterator to the end of the insertion range.
|
384
|
+
*/
|
385
|
+
row_stored_nd_iterator insert(row_stored_nd_iterator position, size_t jj, size_t length, D const* v, size_t v_size, size_t& v_offset) {
|
386
|
+
size_t tmp_v_offset = v_offset;
|
387
|
+
int nd_change = single_row_insertion_plan(position, jj, length, v, v_size, tmp_v_offset);
|
388
|
+
|
389
|
+
// First record the position, just in case our iterator becomes invalid.
|
390
|
+
size_t pp = position.p();
|
391
|
+
|
392
|
+
// Resize the array as necessary, or move entries after the insertion point to make room.
|
393
|
+
size_t sz = y.size();
|
394
|
+
if (sz + nd_change > y.capacity() || sz + nd_change <= y.capacity() / nm::yale_storage::GROWTH_CONSTANT)
|
395
|
+
y.update_resize_move(position, real_i(), nd_change);
|
396
|
+
else if (nd_change != 0) {
|
397
|
+
if (nd_change < 0) y.move_left(position, -nd_change);
|
398
|
+
else if (nd_change > 0) y.move_right(position, nd_change);
|
399
|
+
y.update_real_row_sizes_from(real_i(), nd_change);
|
400
|
+
}
|
401
|
+
|
402
|
+
for (size_t jc = jj; jc < jj + length; ++jc, ++v_offset) {
|
403
|
+
if (v_offset >= v_size) v_offset %= v_size; // reset v position.
|
404
|
+
|
405
|
+
if (jc + y.offset(1) == real_i()) {
|
406
|
+
y.a(real_i()) = v[v_offset]; // modify diagonal
|
407
|
+
} else if (v[v_offset] != y.const_default_obj()) {
|
408
|
+
y.ija(pp) = jc; // modify non-diagonal
|
409
|
+
y.a(pp) = v[v_offset];
|
410
|
+
++pp;
|
411
|
+
}
|
412
|
+
}
|
413
|
+
|
414
|
+
// Update this row.
|
415
|
+
adjust_length(nd_change);
|
416
|
+
|
417
|
+
return row_stored_nd_iterator(*this, pp);
|
418
|
+
}
|
419
|
+
|
420
|
+
/*
|
421
|
+
* For when we don't need to worry about the offset, does the same thing as the insert above.
|
422
|
+
*/
|
423
|
+
row_stored_nd_iterator insert(const row_stored_nd_iterator& position, size_t jj, size_t length, D const* v, size_t v_size) {
|
424
|
+
size_t v_offset = 0;
|
425
|
+
return insert(position, jj, length, v, v_size, v_offset);
|
426
|
+
}
|
427
|
+
|
428
|
+
|
429
|
+
/*
|
430
|
+
* Merges elements offered for insertion with existing elements in the row.
|
431
|
+
*/
|
432
|
+
row_stored_nd_iterator insert(size_t jj, size_t length, D const* v, size_t v_size, size_t& v_offset) {
|
433
|
+
return insert(ndfind(jj), jj, length, v, v_size, v_offset);
|
434
|
+
}
|
435
|
+
|
436
|
+
/*
|
437
|
+
* Merges elements offered for insertion with existing elements in the row.
|
438
|
+
*/
|
439
|
+
row_stored_nd_iterator insert(size_t jj, size_t length, D const* v, size_t v_size) {
|
440
|
+
size_t v_offset = 0;
|
441
|
+
return insert(ndfind(jj), jj, length, v, v_size, v_offset);
|
442
|
+
}
|
443
|
+
|
444
|
+
|
445
|
+
};
|
446
|
+
|
447
|
+
} } // end of nm::yale_storage namespace
|
448
|
+
|
449
|
+
#endif // YALE_ITERATORS_ROW_H
|