xnd 0.2.0dev3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +7 -0
- data/CONTRIBUTING.md +42 -0
- data/Gemfile +3 -0
- data/History.md +0 -0
- data/README.md +7 -0
- data/Rakefile +135 -0
- data/ext/ruby_xnd/extconf.rb +70 -0
- data/ext/ruby_xnd/float_pack_unpack.c +277 -0
- data/ext/ruby_xnd/float_pack_unpack.h +39 -0
- data/ext/ruby_xnd/gc_guard.c +36 -0
- data/ext/ruby_xnd/gc_guard.h +12 -0
- data/ext/ruby_xnd/include/xnd.h +449 -0
- data/ext/ruby_xnd/lib/libxnd.a +0 -0
- data/ext/ruby_xnd/lib/libxnd.so +1 -0
- data/ext/ruby_xnd/lib/libxnd.so.0 +1 -0
- data/ext/ruby_xnd/lib/libxnd.so.0.2.0dev3 +0 -0
- data/ext/ruby_xnd/memory_block_object.c +32 -0
- data/ext/ruby_xnd/memory_block_object.h +33 -0
- data/ext/ruby_xnd/ruby_xnd.c +1953 -0
- data/ext/ruby_xnd/ruby_xnd.h +61 -0
- data/ext/ruby_xnd/ruby_xnd_internal.h +85 -0
- data/ext/ruby_xnd/util.h +170 -0
- data/ext/ruby_xnd/xnd/AUTHORS.txt +5 -0
- data/ext/ruby_xnd/xnd/INSTALL.txt +134 -0
- data/ext/ruby_xnd/xnd/LICENSE.txt +29 -0
- data/ext/ruby_xnd/xnd/MANIFEST.in +3 -0
- data/ext/ruby_xnd/xnd/Makefile.in +80 -0
- data/ext/ruby_xnd/xnd/README.rst +44 -0
- data/ext/ruby_xnd/xnd/config.guess +1530 -0
- data/ext/ruby_xnd/xnd/config.h.in +22 -0
- data/ext/ruby_xnd/xnd/config.sub +1782 -0
- data/ext/ruby_xnd/xnd/configure +4867 -0
- data/ext/ruby_xnd/xnd/configure.ac +164 -0
- data/ext/ruby_xnd/xnd/doc/Makefile +14 -0
- data/ext/ruby_xnd/xnd/doc/_static/copybutton.js +66 -0
- data/ext/ruby_xnd/xnd/doc/conf.py +26 -0
- data/ext/ruby_xnd/xnd/doc/index.rst +44 -0
- data/ext/ruby_xnd/xnd/doc/libxnd/data-structures.rst +186 -0
- data/ext/ruby_xnd/xnd/doc/libxnd/functions.rst +148 -0
- data/ext/ruby_xnd/xnd/doc/libxnd/index.rst +25 -0
- data/ext/ruby_xnd/xnd/doc/releases/index.rst +34 -0
- data/ext/ruby_xnd/xnd/doc/xnd/align-pack.rst +96 -0
- data/ext/ruby_xnd/xnd/doc/xnd/buffer-protocol.rst +42 -0
- data/ext/ruby_xnd/xnd/doc/xnd/index.rst +30 -0
- data/ext/ruby_xnd/xnd/doc/xnd/quickstart.rst +62 -0
- data/ext/ruby_xnd/xnd/doc/xnd/types.rst +674 -0
- data/ext/ruby_xnd/xnd/install-sh +527 -0
- data/ext/ruby_xnd/xnd/libxnd/Makefile.in +102 -0
- data/ext/ruby_xnd/xnd/libxnd/Makefile.vc +112 -0
- data/ext/ruby_xnd/xnd/libxnd/bitmaps.c +345 -0
- data/ext/ruby_xnd/xnd/libxnd/contrib.h +313 -0
- data/ext/ruby_xnd/xnd/libxnd/copy.c +944 -0
- data/ext/ruby_xnd/xnd/libxnd/equal.c +1216 -0
- data/ext/ruby_xnd/xnd/libxnd/inline.h +154 -0
- data/ext/ruby_xnd/xnd/libxnd/overflow.h +147 -0
- data/ext/ruby_xnd/xnd/libxnd/split.c +286 -0
- data/ext/ruby_xnd/xnd/libxnd/tests/Makefile.in +39 -0
- data/ext/ruby_xnd/xnd/libxnd/tests/Makefile.vc +44 -0
- data/ext/ruby_xnd/xnd/libxnd/tests/README.txt +2 -0
- data/ext/ruby_xnd/xnd/libxnd/tests/runtest.c +101 -0
- data/ext/ruby_xnd/xnd/libxnd/tests/test.h +48 -0
- data/ext/ruby_xnd/xnd/libxnd/tests/test_fixed.c +108 -0
- data/ext/ruby_xnd/xnd/libxnd/xnd.c +1304 -0
- data/ext/ruby_xnd/xnd/libxnd/xnd.h +449 -0
- data/ext/ruby_xnd/xnd/python/test_xnd.py +3144 -0
- data/ext/ruby_xnd/xnd/python/xnd/__init__.py +290 -0
- data/ext/ruby_xnd/xnd/python/xnd/_xnd.c +2822 -0
- data/ext/ruby_xnd/xnd/python/xnd/contrib/pretty.py +850 -0
- data/ext/ruby_xnd/xnd/python/xnd/docstrings.h +129 -0
- data/ext/ruby_xnd/xnd/python/xnd/pyxnd.h +200 -0
- data/ext/ruby_xnd/xnd/python/xnd/util.h +182 -0
- data/ext/ruby_xnd/xnd/python/xnd_randvalue.py +1121 -0
- data/ext/ruby_xnd/xnd/python/xnd_support.py +106 -0
- data/ext/ruby_xnd/xnd/setup.py +303 -0
- data/ext/ruby_xnd/xnd/vcbuild/INSTALL.txt +42 -0
- data/ext/ruby_xnd/xnd/vcbuild/runtest32.bat +16 -0
- data/ext/ruby_xnd/xnd/vcbuild/runtest64.bat +14 -0
- data/ext/ruby_xnd/xnd/vcbuild/vcbuild32.bat +29 -0
- data/ext/ruby_xnd/xnd/vcbuild/vcbuild64.bat +29 -0
- data/ext/ruby_xnd/xnd/vcbuild/vcclean.bat +13 -0
- data/ext/ruby_xnd/xnd/vcbuild/vcdistclean.bat +14 -0
- data/lib/ruby_xnd.so +0 -0
- data/lib/xnd.rb +306 -0
- data/lib/xnd/monkeys.rb +29 -0
- data/lib/xnd/version.rb +6 -0
- data/spec/debug_spec.rb +9 -0
- data/spec/gc_guard_spec.rb +10 -0
- data/spec/leakcheck.rb +9 -0
- data/spec/spec_helper.rb +877 -0
- data/spec/type_inference_spec.rb +81 -0
- data/spec/xnd_spec.rb +2921 -0
- data/xnd.gemspec +47 -0
- metadata +215 -0
|
@@ -0,0 +1,154 @@
|
|
|
1
|
+
/*
|
|
2
|
+
* BSD 3-Clause License
|
|
3
|
+
*
|
|
4
|
+
* Copyright (c) 2017-2018, plures
|
|
5
|
+
* All rights reserved.
|
|
6
|
+
*
|
|
7
|
+
* Redistribution and use in source and binary forms, with or without
|
|
8
|
+
* modification, are permitted provided that the following conditions are met:
|
|
9
|
+
*
|
|
10
|
+
* 1. Redistributions of source code must retain the above copyright notice,
|
|
11
|
+
* this list of conditions and the following disclaimer.
|
|
12
|
+
*
|
|
13
|
+
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
|
14
|
+
* this list of conditions and the following disclaimer in the documentation
|
|
15
|
+
* and/or other materials provided with the distribution.
|
|
16
|
+
*
|
|
17
|
+
* 3. Neither the name of the copyright holder nor the names of its
|
|
18
|
+
* contributors may be used to endorse or promote products derived from
|
|
19
|
+
* this software without specific prior written permission.
|
|
20
|
+
*
|
|
21
|
+
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
22
|
+
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
23
|
+
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
24
|
+
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
|
25
|
+
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
26
|
+
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
27
|
+
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
28
|
+
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
29
|
+
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
30
|
+
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
31
|
+
*/
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
#ifndef INLINE_H
|
|
35
|
+
#define INLINE_H
|
|
36
|
+
|
|
37
|
+
|
|
38
|
+
#include <stdlib.h>
|
|
39
|
+
#include <stdint.h>
|
|
40
|
+
#include <assert.h>
|
|
41
|
+
#include "ndtypes.h"
|
|
42
|
+
#include "xnd.h"
|
|
43
|
+
|
|
44
|
+
|
|
45
|
+
/*****************************************************************************/
|
|
46
|
+
/* Internal inline functions */
|
|
47
|
+
/*****************************************************************************/
|
|
48
|
+
|
|
49
|
+
/* Counterparts of the functions in xnd.h. These versions ignore the bitmaps. */
|
|
50
|
+
|
|
51
|
+
static inline xnd_t
|
|
52
|
+
_fixed_dim_next(const xnd_t *x, const int64_t i)
|
|
53
|
+
{
|
|
54
|
+
const ndt_t *t = x->type;
|
|
55
|
+
const ndt_t *u = t->FixedDim.type;
|
|
56
|
+
const int64_t step = i * t->Concrete.FixedDim.step;
|
|
57
|
+
xnd_t next;
|
|
58
|
+
|
|
59
|
+
next.bitmap = xnd_bitmap_empty;
|
|
60
|
+
next.index = x->index + step;
|
|
61
|
+
next.type = u;
|
|
62
|
+
next.ptr = u->ndim==0 ? x->ptr + next.index * next.type->datasize : x->ptr;
|
|
63
|
+
|
|
64
|
+
return next;
|
|
65
|
+
}
|
|
66
|
+
|
|
67
|
+
static inline xnd_t
|
|
68
|
+
_var_dim_next(const xnd_t *x, const int64_t start, const int64_t step,
|
|
69
|
+
const int64_t i)
|
|
70
|
+
{
|
|
71
|
+
const ndt_t *t = x->type;
|
|
72
|
+
const ndt_t *u = t->VarDim.type;
|
|
73
|
+
xnd_t next;
|
|
74
|
+
|
|
75
|
+
next.bitmap = xnd_bitmap_empty;
|
|
76
|
+
next.index = start + i * step;
|
|
77
|
+
next.type = u;
|
|
78
|
+
next.ptr = u->ndim==0 ? x->ptr + next.index * next.type->datasize : x->ptr;
|
|
79
|
+
|
|
80
|
+
return next;
|
|
81
|
+
}
|
|
82
|
+
|
|
83
|
+
static inline xnd_t
|
|
84
|
+
_tuple_next(const xnd_t *x, const int64_t i)
|
|
85
|
+
{
|
|
86
|
+
const ndt_t *t = x->type;
|
|
87
|
+
xnd_t next;
|
|
88
|
+
|
|
89
|
+
next.bitmap = xnd_bitmap_empty;
|
|
90
|
+
next.index = 0;
|
|
91
|
+
next.type = t->Tuple.types[i];
|
|
92
|
+
next.ptr = x->ptr + t->Concrete.Tuple.offset[i];
|
|
93
|
+
|
|
94
|
+
return next;
|
|
95
|
+
}
|
|
96
|
+
|
|
97
|
+
static inline xnd_t
|
|
98
|
+
_record_next(const xnd_t *x, const int64_t i)
|
|
99
|
+
{
|
|
100
|
+
const ndt_t *t = x->type;
|
|
101
|
+
xnd_t next;
|
|
102
|
+
|
|
103
|
+
next.bitmap = xnd_bitmap_empty;
|
|
104
|
+
next.index = 0;
|
|
105
|
+
next.type = t->Record.types[i];
|
|
106
|
+
next.ptr = x->ptr + t->Concrete.Record.offset[i];
|
|
107
|
+
|
|
108
|
+
return next;
|
|
109
|
+
}
|
|
110
|
+
|
|
111
|
+
static inline xnd_t
|
|
112
|
+
_ref_next(const xnd_t *x)
|
|
113
|
+
{
|
|
114
|
+
const ndt_t *t = x->type;
|
|
115
|
+
xnd_t next;
|
|
116
|
+
|
|
117
|
+
next.bitmap = xnd_bitmap_empty;
|
|
118
|
+
next.index = 0;
|
|
119
|
+
next.type = t->Ref.type;
|
|
120
|
+
next.ptr = XND_POINTER_DATA(x->ptr);
|
|
121
|
+
|
|
122
|
+
return next;
|
|
123
|
+
}
|
|
124
|
+
|
|
125
|
+
static inline xnd_t
|
|
126
|
+
_constr_next(const xnd_t *x)
|
|
127
|
+
{
|
|
128
|
+
const ndt_t *t = x->type;
|
|
129
|
+
xnd_t next;
|
|
130
|
+
|
|
131
|
+
next.bitmap = xnd_bitmap_empty;
|
|
132
|
+
next.index = 0;
|
|
133
|
+
next.type = t->Constr.type;
|
|
134
|
+
next.ptr = x->ptr;
|
|
135
|
+
|
|
136
|
+
return next;
|
|
137
|
+
}
|
|
138
|
+
|
|
139
|
+
static inline xnd_t
|
|
140
|
+
_nominal_next(const xnd_t *x)
|
|
141
|
+
{
|
|
142
|
+
const ndt_t *t = x->type;
|
|
143
|
+
xnd_t next;
|
|
144
|
+
|
|
145
|
+
next.bitmap = xnd_bitmap_empty;
|
|
146
|
+
next.index = 0;
|
|
147
|
+
next.type = t->Nominal.type;
|
|
148
|
+
next.ptr = x->ptr;
|
|
149
|
+
|
|
150
|
+
return next;
|
|
151
|
+
}
|
|
152
|
+
|
|
153
|
+
|
|
154
|
+
#endif /* INLINE_H */
|
|
@@ -0,0 +1,147 @@
|
|
|
1
|
+
/*
|
|
2
|
+
* BSD 3-Clause License
|
|
3
|
+
*
|
|
4
|
+
* Copyright (c) 2017-2018, plures
|
|
5
|
+
* All rights reserved.
|
|
6
|
+
*
|
|
7
|
+
* Redistribution and use in source and binary forms, with or without
|
|
8
|
+
* modification, are permitted provided that the following conditions are met:
|
|
9
|
+
*
|
|
10
|
+
* 1. Redistributions of source code must retain the above copyright notice,
|
|
11
|
+
* this list of conditions and the following disclaimer.
|
|
12
|
+
*
|
|
13
|
+
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
|
14
|
+
* this list of conditions and the following disclaimer in the documentation
|
|
15
|
+
* and/or other materials provided with the distribution.
|
|
16
|
+
*
|
|
17
|
+
* 3. Neither the name of the copyright holder nor the names of its
|
|
18
|
+
* contributors may be used to endorse or promote products derived from
|
|
19
|
+
* this software without specific prior written permission.
|
|
20
|
+
*
|
|
21
|
+
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
22
|
+
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
23
|
+
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
24
|
+
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
|
25
|
+
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
26
|
+
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
27
|
+
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
28
|
+
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
29
|
+
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
30
|
+
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
31
|
+
*/
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
#include <stdint.h>
|
|
35
|
+
|
|
36
|
+
|
|
37
|
+
/*****************************************************************************/
|
|
38
|
+
/* Arithmetic with overflow checking */
|
|
39
|
+
/*****************************************************************************/
|
|
40
|
+
|
|
41
|
+
#if defined(__GNUC__) && __GNUC__ >= 5 && !defined(__INTEL_COMPILER)
|
|
42
|
+
static inline int64_t
|
|
43
|
+
ADDi64(int64_t a, int64_t b, bool *overflow)
|
|
44
|
+
{
|
|
45
|
+
int64_t c;
|
|
46
|
+
*overflow |= __builtin_add_overflow(a, b, &c);
|
|
47
|
+
return c;
|
|
48
|
+
}
|
|
49
|
+
|
|
50
|
+
static inline int64_t
|
|
51
|
+
SUBi64(int64_t a, int64_t b, bool *overflow)
|
|
52
|
+
{
|
|
53
|
+
int64_t c;
|
|
54
|
+
*overflow |= __builtin_sub_overflow(a, b, &c);
|
|
55
|
+
return c;
|
|
56
|
+
}
|
|
57
|
+
|
|
58
|
+
static inline int64_t
|
|
59
|
+
MULi64(int64_t a, int64_t b, bool *overflow)
|
|
60
|
+
{
|
|
61
|
+
int64_t c;
|
|
62
|
+
*overflow |= __builtin_mul_overflow(a, b, &c);
|
|
63
|
+
return c;
|
|
64
|
+
}
|
|
65
|
+
|
|
66
|
+
static inline size_t
|
|
67
|
+
MULi64_size(int64_t a, int64_t b, bool *overflow)
|
|
68
|
+
{
|
|
69
|
+
int64_t c;
|
|
70
|
+
*overflow |= __builtin_mul_overflow(a, b, &c);
|
|
71
|
+
#if SIZE_MAX < INT64_MAX
|
|
72
|
+
*overflow |= (c > INT32_MAX);
|
|
73
|
+
#endif
|
|
74
|
+
return (size_t)c;
|
|
75
|
+
}
|
|
76
|
+
|
|
77
|
+
static inline int64_t
|
|
78
|
+
ABSi64(int64_t a, bool *overflow)
|
|
79
|
+
{
|
|
80
|
+
if (a == INT64_MIN) {
|
|
81
|
+
*overflow = 1;
|
|
82
|
+
return INT64_MIN;
|
|
83
|
+
}
|
|
84
|
+
return a >= 0 ? a : -a;
|
|
85
|
+
}
|
|
86
|
+
|
|
87
|
+
static inline uint16_t
|
|
88
|
+
ADDu16(uint16_t a, uint16_t b, bool *overflow)
|
|
89
|
+
{
|
|
90
|
+
uint16_t c;
|
|
91
|
+
*overflow |= __builtin_add_overflow(a, b, &c);
|
|
92
|
+
return c;
|
|
93
|
+
}
|
|
94
|
+
#else
|
|
95
|
+
static inline int64_t
|
|
96
|
+
ADDi64(int64_t a, int64_t b, bool *overflow)
|
|
97
|
+
{
|
|
98
|
+
int64_t c = (uint64_t)a + (uint64_t)b;
|
|
99
|
+
*overflow |= ((a < 0 && b < 0 && c >= 0) || (a >= 0 && b >= 0 && c < 0));
|
|
100
|
+
return c;
|
|
101
|
+
}
|
|
102
|
+
|
|
103
|
+
static inline int64_t
|
|
104
|
+
SUBi64(int64_t a, int64_t b, bool *overflow)
|
|
105
|
+
{
|
|
106
|
+
int64_t c = (uint64_t)a - (uint64_t)b;
|
|
107
|
+
*overflow |= ((a < 0 && b >= 0 && c >= 0) || (a >= 0 && b < 0 && c < 0));
|
|
108
|
+
return c;
|
|
109
|
+
}
|
|
110
|
+
|
|
111
|
+
static inline int64_t
|
|
112
|
+
MULi64(int64_t a, int64_t b, bool *overflow)
|
|
113
|
+
{
|
|
114
|
+
int64_t c = (uint64_t)a * (uint64_t)b;
|
|
115
|
+
*overflow |= ((b < 0 && a == INT64_MIN) || (b != 0 && a != c / b));
|
|
116
|
+
return c;
|
|
117
|
+
}
|
|
118
|
+
|
|
119
|
+
static inline size_t
|
|
120
|
+
MULi64_size(int64_t a, int64_t b, bool *overflow)
|
|
121
|
+
{
|
|
122
|
+
int64_t c = (uint64_t)a * (uint64_t)b;
|
|
123
|
+
*overflow |= ((b < 0 && a == INT64_MIN) || (b != 0 && a != c / b));
|
|
124
|
+
#if SIZE_MAX < INT64_MAX
|
|
125
|
+
*overflow |= (c > INT32_MAX);
|
|
126
|
+
#endif
|
|
127
|
+
return (size_t)c;
|
|
128
|
+
}
|
|
129
|
+
|
|
130
|
+
static inline int64_t
|
|
131
|
+
ABSi64(int64_t a, bool *overflow)
|
|
132
|
+
{
|
|
133
|
+
if (a == INT64_MIN) {
|
|
134
|
+
*overflow = 1;
|
|
135
|
+
return INT64_MIN;
|
|
136
|
+
}
|
|
137
|
+
return a >= 0 ? a : -a;
|
|
138
|
+
}
|
|
139
|
+
|
|
140
|
+
static inline uint16_t
|
|
141
|
+
ADDu16(uint16_t a, uint16_t b, bool *overflow)
|
|
142
|
+
{
|
|
143
|
+
uint16_t c = a + b;
|
|
144
|
+
*overflow |= (c < a);
|
|
145
|
+
return c;
|
|
146
|
+
}
|
|
147
|
+
#endif /* OVERFLOW_H */
|
|
@@ -0,0 +1,286 @@
|
|
|
1
|
+
/*
|
|
2
|
+
* BSD 3-Clause License
|
|
3
|
+
*
|
|
4
|
+
* Copyright (c) 2017-2018, plures
|
|
5
|
+
* All rights reserved.
|
|
6
|
+
*
|
|
7
|
+
* Redistribution and use in source and binary forms, with or without
|
|
8
|
+
* modification, are permitted provided that the following conditions are met:
|
|
9
|
+
*
|
|
10
|
+
* 1. Redistributions of source code must retain the above copyright notice,
|
|
11
|
+
* this list of conditions and the following disclaimer.
|
|
12
|
+
*
|
|
13
|
+
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
|
14
|
+
* this list of conditions and the following disclaimer in the documentation
|
|
15
|
+
* and/or other materials provided with the distribution.
|
|
16
|
+
*
|
|
17
|
+
* 3. Neither the name of the copyright holder nor the names of its
|
|
18
|
+
* contributors may be used to endorse or promote products derived from
|
|
19
|
+
* this software without specific prior written permission.
|
|
20
|
+
*
|
|
21
|
+
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
22
|
+
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
23
|
+
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
24
|
+
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
|
25
|
+
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
26
|
+
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
27
|
+
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
28
|
+
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
29
|
+
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
30
|
+
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
31
|
+
*/
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
#include <stdlib.h>
|
|
35
|
+
#include <stdint.h>
|
|
36
|
+
#include <string.h>
|
|
37
|
+
#include <inttypes.h>
|
|
38
|
+
#include "ndtypes.h"
|
|
39
|
+
#include "xnd.h"
|
|
40
|
+
#include "overflow.h"
|
|
41
|
+
|
|
42
|
+
|
|
43
|
+
static const xnd_index_t init_slice =
|
|
44
|
+
{ .tag = Slice,
|
|
45
|
+
.Slice = {.start = -1, .stop = -1, .step = -1}};
|
|
46
|
+
|
|
47
|
+
|
|
48
|
+
static int64_t column(
|
|
49
|
+
int64_t nrows, int64_t ncols,
|
|
50
|
+
xnd_index_t *indices, int *nindices,
|
|
51
|
+
int64_t row, int64_t col, int64_t r, int64_t q,
|
|
52
|
+
int64_t m, const int64_t *ms, int len);
|
|
53
|
+
|
|
54
|
+
|
|
55
|
+
static void
|
|
56
|
+
free_slices(xnd_t *lst, int64_t len)
|
|
57
|
+
{
|
|
58
|
+
for (int64_t i = 0; i < len; i++) {
|
|
59
|
+
ndt_del((ndt_t *)lst[i].type);
|
|
60
|
+
}
|
|
61
|
+
|
|
62
|
+
ndt_free(lst);
|
|
63
|
+
}
|
|
64
|
+
|
|
65
|
+
static inline int64_t
|
|
66
|
+
start(int64_t i, int64_t r, int64_t q)
|
|
67
|
+
{
|
|
68
|
+
return i < r ? i*(q+1) : r+i*q;
|
|
69
|
+
}
|
|
70
|
+
|
|
71
|
+
static inline int64_t
|
|
72
|
+
stop(int64_t i, int64_t r, int64_t q)
|
|
73
|
+
{
|
|
74
|
+
return i < r ? (i+1)*(q+1) : r+(i+1)*q;
|
|
75
|
+
}
|
|
76
|
+
|
|
77
|
+
static inline int64_t
|
|
78
|
+
step(int64_t i, int64_t r, int64_t q)
|
|
79
|
+
{
|
|
80
|
+
return i < r ? q+1 : q;
|
|
81
|
+
}
|
|
82
|
+
|
|
83
|
+
static inline xnd_index_t
|
|
84
|
+
single_step_slice(int64_t i)
|
|
85
|
+
{
|
|
86
|
+
xnd_index_t x;
|
|
87
|
+
|
|
88
|
+
x.tag = Slice;
|
|
89
|
+
x.Slice.start = i;
|
|
90
|
+
x.Slice.stop = i+1;
|
|
91
|
+
x.Slice.step = 1;
|
|
92
|
+
|
|
93
|
+
return x;
|
|
94
|
+
}
|
|
95
|
+
|
|
96
|
+
static inline xnd_index_t
|
|
97
|
+
slice(int64_t i, int64_t r, int64_t q)
|
|
98
|
+
{
|
|
99
|
+
xnd_index_t x;
|
|
100
|
+
|
|
101
|
+
x.tag = Slice;
|
|
102
|
+
x.Slice.start = start(i, r, q);
|
|
103
|
+
x.Slice.stop = stop(i, r, q);
|
|
104
|
+
x.Slice.step = 1;
|
|
105
|
+
|
|
106
|
+
return x;
|
|
107
|
+
}
|
|
108
|
+
|
|
109
|
+
static int64_t
|
|
110
|
+
prepend(int64_t nrows, int64_t ncols,
|
|
111
|
+
xnd_index_t *indices, int *nindices,
|
|
112
|
+
int64_t row, int64_t col, xnd_index_t s, int64_t n)
|
|
113
|
+
{
|
|
114
|
+
for (int64_t i = 0; i < n; i++) {
|
|
115
|
+
assert(row+i < nrows && col < ncols);
|
|
116
|
+
indices[(row+i)*ncols + col] = s;
|
|
117
|
+
nindices[row+i]++;
|
|
118
|
+
}
|
|
119
|
+
|
|
120
|
+
return n;
|
|
121
|
+
}
|
|
122
|
+
|
|
123
|
+
static int64_t
|
|
124
|
+
last_column(int64_t nrows, int64_t ncols,
|
|
125
|
+
xnd_index_t *indices, int *nindices,
|
|
126
|
+
int64_t row, int64_t col, int64_t r, int64_t q, int64_t n)
|
|
127
|
+
{
|
|
128
|
+
for (int64_t i = 0; i < n; i++) {
|
|
129
|
+
assert(row+i < nrows && col < ncols);
|
|
130
|
+
indices[(row+i)*ncols + col] = slice(i, r, q);
|
|
131
|
+
nindices[row+i]++;
|
|
132
|
+
}
|
|
133
|
+
|
|
134
|
+
return n;
|
|
135
|
+
}
|
|
136
|
+
|
|
137
|
+
static int64_t
|
|
138
|
+
schedule(int64_t nrows, int64_t ncols,
|
|
139
|
+
xnd_index_t *indices, int *nindices,
|
|
140
|
+
int64_t row, int64_t col, int64_t n, const int64_t *shape, int len)
|
|
141
|
+
{
|
|
142
|
+
int64_t m;
|
|
143
|
+
int64_t q;
|
|
144
|
+
int64_t r;
|
|
145
|
+
|
|
146
|
+
if (len == 0) {
|
|
147
|
+
return 1;
|
|
148
|
+
}
|
|
149
|
+
|
|
150
|
+
m = shape[0];
|
|
151
|
+
if (n <= m) {
|
|
152
|
+
q = m / n;
|
|
153
|
+
r = m % n;
|
|
154
|
+
return last_column(nrows, ncols, indices, nindices, row, col, r, q, n);
|
|
155
|
+
}
|
|
156
|
+
else {
|
|
157
|
+
q = n / m;
|
|
158
|
+
r = n % m;
|
|
159
|
+
return column(nrows, ncols, indices, nindices, row, col, r, q, m, shape+1, len-1);
|
|
160
|
+
}
|
|
161
|
+
}
|
|
162
|
+
|
|
163
|
+
static int64_t
|
|
164
|
+
column(int64_t nrows, int64_t ncols,
|
|
165
|
+
xnd_index_t *indices, int *nindices,
|
|
166
|
+
int64_t row, int64_t col, int64_t r, int64_t q, int64_t m,
|
|
167
|
+
const int64_t *ms, int len)
|
|
168
|
+
{
|
|
169
|
+
int64_t column_len = 0;
|
|
170
|
+
int64_t n, subtree_len, block_len;
|
|
171
|
+
xnd_index_t s;
|
|
172
|
+
|
|
173
|
+
for (int64_t i = 0; i < m; i++) {
|
|
174
|
+
n = step(i, r, q);
|
|
175
|
+
s = single_step_slice(i);
|
|
176
|
+
subtree_len = schedule(nrows, ncols, indices, nindices, row, col+1, n, ms, len);
|
|
177
|
+
block_len = prepend(nrows, ncols, indices, nindices, row, col, s, subtree_len);
|
|
178
|
+
row += block_len;
|
|
179
|
+
column_len += block_len;
|
|
180
|
+
}
|
|
181
|
+
|
|
182
|
+
return column_len;
|
|
183
|
+
}
|
|
184
|
+
|
|
185
|
+
static int
|
|
186
|
+
get_shape(int64_t *shape, const ndt_t *t, int max_outer, ndt_context_t *ctx)
|
|
187
|
+
{
|
|
188
|
+
int i;
|
|
189
|
+
|
|
190
|
+
if (!ndt_is_ndarray(t)) {
|
|
191
|
+
ndt_err_format(ctx, NDT_ValueError,
|
|
192
|
+
"split function called on non-ndarray");
|
|
193
|
+
return -1;
|
|
194
|
+
}
|
|
195
|
+
|
|
196
|
+
for (i = 0; i < max_outer && t->ndim > 0; i++, t=t->FixedDim.type) {
|
|
197
|
+
shape[i] = t->FixedDim.shape;
|
|
198
|
+
if (shape[i] <= 0) {
|
|
199
|
+
ndt_err_format(ctx, NDT_ValueError,
|
|
200
|
+
"split function called on invalid shape or shape with zeros");
|
|
201
|
+
return -1;
|
|
202
|
+
}
|
|
203
|
+
}
|
|
204
|
+
for (; t->ndim > 0; t=t->FixedDim.type) {
|
|
205
|
+
if (t->FixedDim.shape <= 0) {
|
|
206
|
+
ndt_err_format(ctx, NDT_ValueError,
|
|
207
|
+
"split function called on invalid shape or shape with zeros");
|
|
208
|
+
return -1;
|
|
209
|
+
}
|
|
210
|
+
}
|
|
211
|
+
|
|
212
|
+
return i;
|
|
213
|
+
}
|
|
214
|
+
|
|
215
|
+
xnd_t *
|
|
216
|
+
xnd_split(const xnd_t *x, int64_t *nparts, int max_outer, ndt_context_t *ctx)
|
|
217
|
+
{
|
|
218
|
+
bool overflow = false;
|
|
219
|
+
int64_t shape[NDT_MAX_DIM];
|
|
220
|
+
xnd_index_t *indices;
|
|
221
|
+
int *nindices;
|
|
222
|
+
xnd_t *result;
|
|
223
|
+
int64_t nrows, nmemb;
|
|
224
|
+
int ncols;
|
|
225
|
+
|
|
226
|
+
if (*nparts < 1) {
|
|
227
|
+
ndt_err_format(ctx, NDT_ValueError, "'n' parameter must be >= 1");
|
|
228
|
+
return NULL;
|
|
229
|
+
}
|
|
230
|
+
nrows = *nparts;
|
|
231
|
+
|
|
232
|
+
ncols = get_shape(shape, x->type, max_outer, ctx);
|
|
233
|
+
if (ncols < 0) {
|
|
234
|
+
return NULL;
|
|
235
|
+
}
|
|
236
|
+
|
|
237
|
+
nmemb = MULi64(nrows, ncols, &overflow);
|
|
238
|
+
if (overflow) {
|
|
239
|
+
ndt_err_format(ctx, NDT_ValueError, "'n' parameter is too large");
|
|
240
|
+
return NULL;
|
|
241
|
+
}
|
|
242
|
+
|
|
243
|
+
indices = ndt_alloc(nmemb, sizeof *indices);
|
|
244
|
+
if (indices == NULL) {
|
|
245
|
+
return ndt_memory_error(ctx);
|
|
246
|
+
}
|
|
247
|
+
for (int64_t i = 0; i < nrows; i++) {
|
|
248
|
+
for (int64_t k = 0; k < ncols; k++) {
|
|
249
|
+
indices[i*ncols + k] = init_slice;
|
|
250
|
+
}
|
|
251
|
+
}
|
|
252
|
+
|
|
253
|
+
nindices = ndt_alloc(nrows, sizeof *nindices);
|
|
254
|
+
if (nindices == NULL) {
|
|
255
|
+
ndt_free(indices);
|
|
256
|
+
return ndt_memory_error(ctx);
|
|
257
|
+
}
|
|
258
|
+
for (int64_t i = 0; i < nrows; i++) {
|
|
259
|
+
nindices[i] = 0;
|
|
260
|
+
}
|
|
261
|
+
|
|
262
|
+
nrows = schedule(nrows, ncols, indices, nindices, 0, 0, nrows, shape, ncols);
|
|
263
|
+
|
|
264
|
+
result = ndt_alloc(nrows, sizeof *result);
|
|
265
|
+
if (result == NULL) {
|
|
266
|
+
ndt_free(nindices);
|
|
267
|
+
ndt_free(indices);
|
|
268
|
+
return ndt_memory_error(ctx);
|
|
269
|
+
}
|
|
270
|
+
|
|
271
|
+
for (int64_t i = 0; i < nrows; i++) {
|
|
272
|
+
result[i] = xnd_multikey(x, indices+(i*ncols), nindices[i], ctx);
|
|
273
|
+
if (ndt_err_occurred(ctx)) {
|
|
274
|
+
ndt_free(nindices);
|
|
275
|
+
ndt_free(indices);
|
|
276
|
+
free_slices(result, i);
|
|
277
|
+
return NULL;
|
|
278
|
+
}
|
|
279
|
+
}
|
|
280
|
+
|
|
281
|
+
ndt_free(nindices);
|
|
282
|
+
ndt_free(indices);
|
|
283
|
+
*nparts = nrows;
|
|
284
|
+
|
|
285
|
+
return result;
|
|
286
|
+
}
|